* because it's been added to the accept queue directly.
         */
        req->timeout = tcp_timeout_init(child);
-       inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS,
-                                 req->timeout, TCP_RTO_MAX);
+       tcp_reset_xmit_timer(child, ICSK_TIME_RETRANS,
+                            req->timeout, false);
 
        refcount_set(&req->rsk_refcnt, 2);
 
 
                unsigned long delay = max(usecs_to_jiffies(tp->srtt_us >> 4),
                                          msecs_to_jiffies(10));
 
-               inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                                         delay, TCP_RTO_MAX);
+               tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS, delay, false);
                *ack_flag &= ~FLAG_SET_XMIT_TIMER;
                return true;
        }
                    after(TCP_SKB_CB(skb)->ack_seq, tp->snd_nxt)) {
                        /* Previous FIN/ACK or RST/ACK might be ignored. */
                        if (icsk->icsk_retransmits == 0)
-                               inet_csk_reset_xmit_timer(sk,
-                                               ICSK_TIME_RETRANS,
-                                               TCP_TIMEOUT_MIN, TCP_RTO_MAX);
+                               tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
+                                                    TCP_TIMEOUT_MIN, false);
                        SKB_DR_SET(reason, TCP_INVALID_ACK_SEQUENCE);
                        goto reset_and_undo;
                }
                         */
                        inet_csk_schedule_ack(sk);
                        tcp_enter_quickack_mode(sk, TCP_MAX_QUICKACKS);
-                       inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK,
-                                                 TCP_DELACK_MAX, TCP_RTO_MAX);
+                       tcp_reset_xmit_timer(sk, ICSK_TIME_DACK,
+                                            TCP_DELACK_MAX, false);
                        goto consume;
                }
                tcp_send_ack(sk);
 
        remaining = icsk->icsk_rto - usecs_to_jiffies(delta_us);
 
        if (remaining > 0) {
-               inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                                         remaining, TCP_RTO_MAX);
+               tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS, remaining, false);
        } else {
                /* RTO revert clocked out retransmission.
                 * Will retransmit now.
 
        TCP_INC_STATS(sock_net(sk), TCP_MIB_ACTIVEOPENS);
 
        /* Timer for repeating the SYN until an answer. */
-       inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                                 inet_csk(sk)->icsk_rto, TCP_RTO_MAX);
+       tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
+                            inet_csk(sk)->icsk_rto, false);
        return 0;
 }
 EXPORT_SYMBOL(tcp_connect);
                        icsk->icsk_ack.retry++;
                inet_csk_schedule_ack(sk);
                icsk->icsk_ack.ato = TCP_ATO_MIN;
-               inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK, delay, TCP_RTO_MAX);
+               tcp_reset_xmit_timer(sk, ICSK_TIME_DACK, delay, false);
                return;
        }
 
 
        tcp_update_rto_stats(sk);
        if (!tp->retrans_stamp)
                tp->retrans_stamp = tcp_time_stamp_ts(tp);
-       inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                         req->timeout << req->num_timeout, TCP_RTO_MAX);
+       tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
+                         req->timeout << req->num_timeout, false);
 }
 
 static bool tcp_rtx_probe0_timed_out(const struct sock *sk,
                /* Retransmission failed because of local congestion,
                 * Let senders fight for local resources conservatively.
                 */
-               inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                                         TCP_RESOURCE_PROBE_INTERVAL,
-                                         TCP_RTO_MAX);
+               tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
+                                    TCP_RESOURCE_PROBE_INTERVAL,
+                                    false);
                goto out;
        }
 
                icsk->icsk_backoff++;
                icsk->icsk_rto = min(icsk->icsk_rto << 1, TCP_RTO_MAX);
        }
-       inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
-                                 tcp_clamp_rto_to_user_timeout(sk), TCP_RTO_MAX);
+       tcp_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
+                            tcp_clamp_rto_to_user_timeout(sk), false);
        if (retransmits_timed_out(sk, READ_ONCE(net->ipv4.sysctl_tcp_retries1) + 1, 0))
                __sk_dst_reset(sk);