#include "dev.h"
 #include "sock_destructor.h"
 
-struct kmem_cache *skbuff_head_cache __ro_after_init;
+struct kmem_cache *skbuff_cache __ro_after_init;
 static struct kmem_cache *skbuff_fclone_cache __ro_after_init;
 #ifdef CONFIG_SKB_EXTENSIONS
 static struct kmem_cache *skbuff_ext_cache __ro_after_init;
        struct sk_buff *skb;
 
        if (unlikely(!nc->skb_count)) {
-               nc->skb_count = kmem_cache_alloc_bulk(skbuff_head_cache,
+               nc->skb_count = kmem_cache_alloc_bulk(skbuff_cache,
                                                      GFP_ATOMIC,
                                                      NAPI_SKB_CACHE_BULK,
                                                      nc->skb_cache);
        }
 
        skb = nc->skb_cache[--nc->skb_count];
-       kasan_unpoison_object_data(skbuff_head_cache, skb);
+       kasan_unpoison_object_data(skbuff_cache, skb);
 
        return skb;
 }
        struct sk_buff *skb;
        unsigned int size;
 
-       skb = kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC);
+       skb = kmem_cache_alloc(skbuff_cache, GFP_ATOMIC);
        if (unlikely(!skb))
                return NULL;
 
 {
        struct sk_buff *skb;
 
-       skb = kmem_cache_alloc(skbuff_head_cache, GFP_ATOMIC);
+       skb = kmem_cache_alloc(skbuff_cache, GFP_ATOMIC);
        if (unlikely(!skb))
                return NULL;
 
        u8 *data;
 
        cache = (flags & SKB_ALLOC_FCLONE)
-               ? skbuff_fclone_cache : skbuff_head_cache;
+               ? skbuff_fclone_cache : skbuff_cache;
 
        if (sk_memalloc_socks() && (flags & SKB_ALLOC_RX))
                gfp_mask |= __GFP_MEMALLOC;
 
        switch (skb->fclone) {
        case SKB_FCLONE_UNAVAILABLE:
-               kmem_cache_free(skbuff_head_cache, skb);
+               kmem_cache_free(skbuff_cache, skb);
                return;
 
        case SKB_FCLONE_ORIG:
        sa->skb_array[sa->skb_count++] = skb;
 
        if (unlikely(sa->skb_count == KFREE_SKB_BULK_SIZE)) {
-               kmem_cache_free_bulk(skbuff_head_cache, KFREE_SKB_BULK_SIZE,
+               kmem_cache_free_bulk(skbuff_cache, KFREE_SKB_BULK_SIZE,
                                     sa->skb_array);
                sa->skb_count = 0;
        }
        }
 
        if (sa.skb_count)
-               kmem_cache_free_bulk(skbuff_head_cache, sa.skb_count,
-                                    sa.skb_array);
+               kmem_cache_free_bulk(skbuff_cache, sa.skb_count, sa.skb_array);
 }
 EXPORT_SYMBOL(kfree_skb_list_reason);
 
        struct napi_alloc_cache *nc = this_cpu_ptr(&napi_alloc_cache);
        u32 i;
 
-       kasan_poison_object_data(skbuff_head_cache, skb);
+       kasan_poison_object_data(skbuff_cache, skb);
        nc->skb_cache[nc->skb_count++] = skb;
 
        if (unlikely(nc->skb_count == NAPI_SKB_CACHE_SIZE)) {
                for (i = NAPI_SKB_CACHE_HALF; i < NAPI_SKB_CACHE_SIZE; i++)
-                       kasan_unpoison_object_data(skbuff_head_cache,
+                       kasan_unpoison_object_data(skbuff_cache,
                                                   nc->skb_cache[i]);
 
-               kmem_cache_free_bulk(skbuff_head_cache, NAPI_SKB_CACHE_HALF,
+               kmem_cache_free_bulk(skbuff_cache, NAPI_SKB_CACHE_HALF,
                                     nc->skb_cache + NAPI_SKB_CACHE_HALF);
                nc->skb_count = NAPI_SKB_CACHE_HALF;
        }
                if (skb_pfmemalloc(skb))
                        gfp_mask |= __GFP_MEMALLOC;
 
-               n = kmem_cache_alloc(skbuff_head_cache, gfp_mask);
+               n = kmem_cache_alloc(skbuff_cache, gfp_mask);
                if (!n)
                        return NULL;
 
 
 void __init skb_init(void)
 {
-       skbuff_head_cache = kmem_cache_create_usercopy("skbuff_head_cache",
+       skbuff_cache = kmem_cache_create_usercopy("skbuff_head_cache",
                                              sizeof(struct sk_buff),
                                              0,
                                              SLAB_HWCACHE_ALIGN|SLAB_PANIC,
 {
        if (head_stolen) {
                skb_release_head_state(skb);
-               kmem_cache_free(skbuff_head_cache, skb);
+               kmem_cache_free(skbuff_cache, skb);
        } else {
                __kfree_skb(skb);
        }