static LIST_HEAD(mirred_list);
 static DEFINE_SPINLOCK(mirred_list_lock);
 
-#define MIRRED_RECURSION_LIMIT    4
-static DEFINE_PER_CPU(unsigned int, mirred_rec_level);
+#define MIRRED_NEST_LIMIT    4
+static DEFINE_PER_CPU(unsigned int, mirred_nest_level);
 
 static bool tcf_mirred_is_act_redirect(int action)
 {
        struct sk_buff *skb2 = skb;
        bool m_mac_header_xmit;
        struct net_device *dev;
-       unsigned int rec_level;
+       unsigned int nest_level;
        int retval, err = 0;
        bool use_reinsert;
        bool want_ingress;
        int mac_len;
        bool at_nh;
 
-       rec_level = __this_cpu_inc_return(mirred_rec_level);
-       if (unlikely(rec_level > MIRRED_RECURSION_LIMIT)) {
+       nest_level = __this_cpu_inc_return(mirred_nest_level);
+       if (unlikely(nest_level > MIRRED_NEST_LIMIT)) {
                net_warn_ratelimited("Packet exceeded mirred recursion limit on dev %s\n",
                                     netdev_name(skb->dev));
-               __this_cpu_dec(mirred_rec_level);
+               __this_cpu_dec(mirred_nest_level);
                return TC_ACT_SHOT;
        }
 
                        err = tcf_mirred_forward(want_ingress, skb);
                        if (err)
                                tcf_action_inc_overlimit_qstats(&m->common);
-                       __this_cpu_dec(mirred_rec_level);
+                       __this_cpu_dec(mirred_nest_level);
                        return TC_ACT_CONSUMED;
                }
        }
                if (tcf_mirred_is_act_redirect(m_eaction))
                        retval = TC_ACT_SHOT;
        }
-       __this_cpu_dec(mirred_rec_level);
+       __this_cpu_dec(mirred_nest_level);
 
        return retval;
 }