__skb_checksum_complete(skb);
 }
 
-bool tcp_add_backlog(struct sock *sk, struct sk_buff *skb);
+bool tcp_add_backlog(struct sock *sk, struct sk_buff *skb,
+                    enum skb_drop_reason *reason);
 
 #ifdef CONFIG_INET
 void __sk_defer_free_flush(struct sock *sk);
 
        EM(SKB_DROP_REASON_TCP_MD5UNEXPECTED,                   \
           TCP_MD5UNEXPECTED)                                   \
        EM(SKB_DROP_REASON_TCP_MD5FAILURE, TCP_MD5FAILURE)      \
+       EM(SKB_DROP_REASON_SOCKET_BACKLOG, SOCKET_BACKLOG)      \
        EMe(SKB_DROP_REASON_MAX, MAX)
 
 #undef EM
 
        return 0;
 }
 
-bool tcp_add_backlog(struct sock *sk, struct sk_buff *skb)
+bool tcp_add_backlog(struct sock *sk, struct sk_buff *skb,
+                    enum skb_drop_reason *reason)
 {
        u32 limit, tail_gso_size, tail_gso_segs;
        struct skb_shared_info *shinfo;
        if (unlikely(tcp_checksum_complete(skb))) {
                bh_unlock_sock(sk);
                trace_tcp_bad_csum(skb);
+               *reason = SKB_DROP_REASON_TCP_CSUM;
                __TCP_INC_STATS(sock_net(sk), TCP_MIB_CSUMERRORS);
                __TCP_INC_STATS(sock_net(sk), TCP_MIB_INERRS);
                return true;
 
        if (unlikely(sk_add_backlog(sk, skb, limit))) {
                bh_unlock_sock(sk);
+               *reason = SKB_DROP_REASON_SOCKET_BACKLOG;
                __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPBACKLOGDROP);
                return true;
        }
        if (!sock_owned_by_user(sk)) {
                ret = tcp_v4_do_rcv(sk, skb);
        } else {
-               if (tcp_add_backlog(sk, skb))
+               if (tcp_add_backlog(sk, skb, &drop_reason))
                        goto discard_and_relse;
        }
        bh_unlock_sock(sk);
 
        if (!sock_owned_by_user(sk)) {
                ret = tcp_v6_do_rcv(sk, skb);
        } else {
-               if (tcp_add_backlog(sk, skb))
+               if (tcp_add_backlog(sk, skb, &drop_reason))
                        goto discard_and_relse;
        }
        bh_unlock_sock(sk);