#define FLAG_FORWARD_PROGRESS  (FLAG_ACKED|FLAG_DATA_SACKED)
 #define FLAG_ANY_PROGRESS      (FLAG_FORWARD_PROGRESS|FLAG_SND_UNA_ADVANCED)
 
-#define IsReno(tp) ((tp)->rx_opt.sack_ok == 0)
-#define IsFack(tp) ((tp)->rx_opt.sack_ok & 2)
-#define IsDSack(tp) ((tp)->rx_opt.sack_ok & 4)
-
 #define IsSackFrto() (sysctl_tcp_frto == 0x2)
 
 #define TCP_REMNANT (TCP_FLAG_FIN|TCP_FLAG_URG|TCP_FLAG_SYN|TCP_FLAG_PSH)
        }
 }
 
+/*
+ * Packet counting of FACK is based on in-order assumptions, therefore TCP
+ * disables it when reordering is detected
+ */
+static void tcp_disable_fack(struct tcp_sock *tp)
+{
+       tp->rx_opt.sack_ok &= ~2;
+}
+
+/* Take a notice that peer is sending DSACKs */
+static void tcp_dsack_seen(struct tcp_sock *tp)
+{
+       tp->rx_opt.sack_ok |= 4;
+}
+
 /* Initialize metrics on socket. */
 
 static void tcp_init_metrics(struct sock *sk)
        }
        if (dst_metric(dst, RTAX_REORDERING) &&
            tp->reordering != dst_metric(dst, RTAX_REORDERING)) {
-               tp->rx_opt.sack_ok &= ~2;
+               tcp_disable_fack(tp);
                tp->reordering = dst_metric(dst, RTAX_REORDERING);
        }
 
                /* This exciting event is worth to be remembered. 8) */
                if (ts)
                        NET_INC_STATS_BH(LINUX_MIB_TCPTSREORDER);
-               else if (IsReno(tp))
+               else if (tcp_is_reno(tp))
                        NET_INC_STATS_BH(LINUX_MIB_TCPRENOREORDER);
-               else if (IsFack(tp))
+               else if (tcp_is_fack(tp))
                        NET_INC_STATS_BH(LINUX_MIB_TCPFACKREORDER);
                else
                        NET_INC_STATS_BH(LINUX_MIB_TCPSACKREORDER);
                       tp->sacked_out,
                       tp->undo_marker ? tp->undo_retrans : 0);
 #endif
-               /* Disable FACK yet. */
-               tp->rx_opt.sack_ok &= ~2;
+               tcp_disable_fack(tp);
        }
 }
 
 
        if (before(start_seq_0, TCP_SKB_CB(ack_skb)->ack_seq)) {
                dup_sack = 1;
-               tp->rx_opt.sack_ok |= 4;
+               tcp_dsack_seen(tp);
                NET_INC_STATS_BH(LINUX_MIB_TCPDSACKRECV);
        } else if (num_sacks > 1) {
                u32 end_seq_1 = ntohl(get_unaligned(&sp[1].end_seq));
                if (!after(end_seq_0, end_seq_1) &&
                    !before(start_seq_0, start_seq_1)) {
                        dup_sack = 1;
-                       tp->rx_opt.sack_ok |= 4;
+                       tcp_dsack_seen(tp);
                        NET_INC_STATS_BH(LINUX_MIB_TCPDSACKOFORECV);
                }
        }
                                continue;
                        if ((TCP_SKB_CB(skb)->sacked&TCPCB_SACKED_RETRANS) &&
                            after(lost_retrans, TCP_SKB_CB(skb)->ack_seq) &&
-                           (IsFack(tp) ||
+                           (tcp_is_fack(tp) ||
                             !before(lost_retrans,
                                     TCP_SKB_CB(skb)->ack_seq + tp->reordering *
                                     tp->mss_cache))) {
 
        tp->lost_out = 0;
        tp->retrans_out = 0;
-       if (IsReno(tp))
+       if (tcp_is_reno(tp))
                tcp_reset_reno_sack(tp);
 
        tcp_for_write_queue(skb, sk) {
 
 static inline int tcp_fackets_out(struct tcp_sock *tp)
 {
-       return IsReno(tp) ? tp->sacked_out+1 : tp->fackets_out;
+       return tcp_is_reno(tp) ? tp->sacked_out+1 : tp->fackets_out;
 }
 
 static inline int tcp_skb_timedout(struct sock *sk, struct sk_buff *skb)
 {
        struct tcp_sock *tp = tcp_sk(sk);
 
-       if (IsFack(tp)) {
+       if (tcp_is_fack(tp)) {
                int lost = tp->fackets_out - tp->reordering;
                if (lost <= 0)
                        lost = 1;
         * Hence, we can detect timed out packets during fast
         * retransmit without falling to slow start.
         */
-       if (!IsReno(tp) && tcp_head_timedout(sk)) {
+       if (!tcp_is_reno(tp) && tcp_head_timedout(sk)) {
                struct sk_buff *skb;
 
                skb = tp->scoreboard_skb_hint ? tp->scoreboard_skb_hint
        int decr = tp->snd_cwnd_cnt + 1;
 
        if ((flag&(FLAG_ANY_PROGRESS|FLAG_DSACKING_ACK)) ||
-           (IsReno(tp) && !(flag&FLAG_NOT_DUP))) {
+           (tcp_is_reno(tp) && !(flag&FLAG_NOT_DUP))) {
                tp->snd_cwnd_cnt = decr&1;
                decr >>= 1;
 
                        NET_INC_STATS_BH(LINUX_MIB_TCPFULLUNDO);
                tp->undo_marker = 0;
        }
-       if (tp->snd_una == tp->high_seq && IsReno(tp)) {
+       if (tp->snd_una == tp->high_seq && tcp_is_reno(tp)) {
                /* Hold old state until something *above* high_seq
                 * is ACKed. For Reno it is MUST to prevent false
                 * fast retransmits (RFC2582). SACK TCP is safe. */
 {
        struct tcp_sock *tp = tcp_sk(sk);
        /* Partial ACK arrived. Force Hoe's retransmit. */
-       int failed = IsReno(tp) || tp->fackets_out>tp->reordering;
+       int failed = tcp_is_reno(tp) || tp->fackets_out>tp->reordering;
 
        if (tcp_may_undo(tp)) {
                /* Plain luck! Hole if filled with delayed
                NET_INC_STATS_BH(LINUX_MIB_TCPLOSSUNDO);
                inet_csk(sk)->icsk_retransmits = 0;
                tp->undo_marker = 0;
-               if (!IsReno(tp))
+               if (tcp_is_sack(tp))
                        tcp_set_ca_state(sk, TCP_CA_Open);
                return 1;
        }
                        if (!tp->undo_marker ||
                            /* For SACK case do not Open to allow to undo
                             * catching for all duplicate ACKs. */
-                           IsReno(tp) || tp->snd_una != tp->high_seq) {
+                           tcp_is_reno(tp) || tp->snd_una != tp->high_seq) {
                                tp->undo_marker = 0;
                                tcp_set_ca_state(sk, TCP_CA_Open);
                        }
                        break;
 
                case TCP_CA_Recovery:
-                       if (IsReno(tp))
+                       if (tcp_is_reno(tp))
                                tcp_reset_reno_sack(tp);
                        if (tcp_try_undo_recovery(sk))
                                return;
        switch (icsk->icsk_ca_state) {
        case TCP_CA_Recovery:
                if (!(flag & FLAG_SND_UNA_ADVANCED)) {
-                       if (IsReno(tp) && is_dupack)
+                       if (tcp_is_reno(tp) && is_dupack)
                                tcp_add_reno_sack(sk);
                } else
                        do_lost = tcp_try_undo_partial(sk, pkts_acked);
                        return;
                /* Loss is undone; fall through to processing in Open state. */
        default:
-               if (IsReno(tp)) {
+               if (tcp_is_reno(tp)) {
                        if (flag & FLAG_SND_UNA_ADVANCED)
                                tcp_reset_reno_sack(tp);
                        if (is_dupack)
 
                /* Otherwise enter Recovery state */
 
-               if (IsReno(tp))
+               if (tcp_is_reno(tp))
                        NET_INC_STATS_BH(LINUX_MIB_TCPRENORECOVERY);
                else
                        NET_INC_STATS_BH(LINUX_MIB_TCPSACKRECOVERY);
                tcp_ack_update_rtt(sk, acked, seq_rtt);
                tcp_ack_packets_out(sk);
 
-               if (IsReno(tp))
+               if (tcp_is_reno(tp))
                        tcp_remove_reno_sacks(sk, pkts_acked);
 
                if (ca_ops->pkts_acked) {
        BUG_TRAP((int)tp->sacked_out >= 0);
        BUG_TRAP((int)tp->lost_out >= 0);
        BUG_TRAP((int)tp->retrans_out >= 0);
-       if (!tp->packets_out && tp->rx_opt.sack_ok) {
+       if (!tp->packets_out && tcp_is_sack(tp)) {
                const struct inet_connection_sock *icsk = inet_csk(sk);
                if (tp->lost_out) {
                        printk(KERN_DEBUG "Leak l=%u %d\n",
                return 1;
        }
 
-       if (!IsSackFrto() || IsReno(tp)) {
+       if (!IsSackFrto() || tcp_is_reno(tp)) {
                /* RFC4138 shortcoming in step 2; should also have case c):
                 * ACK isn't duplicate nor advances window, e.g., opposite dir
                 * data, winupdate
         * Probably, we should reset in this case. For now drop them.
         */
        __skb_queue_purge(&tp->out_of_order_queue);
-       if (tp->rx_opt.sack_ok)
+       if (tcp_is_sack(tp))
                tcp_sack_reset(&tp->rx_opt);
        sk_stream_mem_reclaim(sk);
 
 
 static void tcp_dsack_set(struct tcp_sock *tp, u32 seq, u32 end_seq)
 {
-       if (tp->rx_opt.sack_ok && sysctl_tcp_dsack) {
+       if (tcp_is_sack(tp) && sysctl_tcp_dsack) {
                if (before(seq, tp->rcv_nxt))
                        NET_INC_STATS_BH(LINUX_MIB_TCPDSACKOLDSENT);
                else
                NET_INC_STATS_BH(LINUX_MIB_DELAYEDACKLOST);
                tcp_enter_quickack_mode(sk);
 
-               if (tp->rx_opt.sack_ok && sysctl_tcp_dsack) {
+               if (tcp_is_sack(tp) && sysctl_tcp_dsack) {
                        u32 end_seq = TCP_SKB_CB(skb)->end_seq;
 
                        if (after(TCP_SKB_CB(skb)->end_seq, tp->rcv_nxt))
 
        if (!skb_peek(&tp->out_of_order_queue)) {
                /* Initial out of order segment, build 1 SACK. */
-               if (tp->rx_opt.sack_ok) {
+               if (tcp_is_sack(tp)) {
                        tp->rx_opt.num_sacks = 1;
                        tp->rx_opt.dsack     = 0;
                        tp->rx_opt.eff_sacks = 1;
                }
 
 add_sack:
-               if (tp->rx_opt.sack_ok)
+               if (tcp_is_sack(tp))
                        tcp_sack_new_ofo_skb(sk, seq, end_seq);
        }
 }
                 * is in a sad state like this, we care only about integrity
                 * of the connection not performance.
                 */
-               if (tp->rx_opt.sack_ok)
+               if (tcp_is_sack(tp))
                        tcp_sack_reset(&tp->rx_opt);
                sk_stream_mem_reclaim(sk);
        }
                        tp->tcp_header_len = sizeof(struct tcphdr);
                }
 
-               if (tp->rx_opt.sack_ok && sysctl_tcp_fack)
-                       tp->rx_opt.sack_ok |= 2;
+               if (tcp_is_sack(tp) && sysctl_tcp_fack)
+                       tcp_enable_fack(tp);
 
                tcp_mtup_init(sk);
                tcp_sync_mss(sk, icsk->icsk_pmtu_cookie);