bool mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages);
 void mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages);
 #if defined(CONFIG_MEMCG) && defined(CONFIG_INET)
-extern struct static_key memcg_sockets_enabled_key;
-#define mem_cgroup_sockets_enabled static_key_false(&memcg_sockets_enabled_key)
+extern struct static_key_false memcg_sockets_enabled_key;
+#define mem_cgroup_sockets_enabled static_branch_unlikely(&memcg_sockets_enabled_key)
 static inline bool mem_cgroup_under_socket_pressure(struct mem_cgroup *memcg)
 {
 #ifdef CONFIG_MEMCG_KMEM
 #endif
 
 #ifdef CONFIG_MEMCG_KMEM
-extern struct static_key memcg_kmem_enabled_key;
+extern struct static_key_false memcg_kmem_enabled_key;
 
 extern int memcg_nr_cache_ids;
 void memcg_get_cache_ids(void);
 
 static inline bool memcg_kmem_enabled(void)
 {
-       return static_key_false(&memcg_kmem_enabled_key);
+       return static_branch_unlikely(&memcg_kmem_enabled_key);
 }
 
 static inline bool memcg_kmem_is_active(struct mem_cgroup *memcg)
 
  * conditional to this static branch, we'll have to allow modules that does
  * kmem_cache_alloc and the such to see this symbol as well
  */
-struct static_key memcg_kmem_enabled_key;
+DEFINE_STATIC_KEY_FALSE(memcg_kmem_enabled_key);
 EXPORT_SYMBOL(memcg_kmem_enabled_key);
 
 #endif /* CONFIG_MEMCG_KMEM */
        err = page_counter_limit(&memcg->kmem, nr_pages);
        VM_BUG_ON(err);
 
-       static_key_slow_inc(&memcg_kmem_enabled_key);
+       static_branch_inc(&memcg_kmem_enabled_key);
        /*
         * A memory cgroup is considered kmem-active as soon as it gets
         * kmemcg_id. Setting the id after enabling static branching will
 {
        if (memcg->kmem_acct_activated) {
                memcg_destroy_kmem_caches(memcg);
-               static_key_slow_dec(&memcg_kmem_enabled_key);
+               static_branch_dec(&memcg_kmem_enabled_key);
                WARN_ON(page_counter_read(&memcg->kmem));
        }
        tcp_destroy_cgroup(memcg);
 
 #ifdef CONFIG_INET
        if (cgroup_subsys_on_dfl(memory_cgrp_subsys) && !cgroup_memory_nosocket)
-               static_key_slow_inc(&memcg_sockets_enabled_key);
+               static_branch_inc(&memcg_sockets_enabled_key);
 #endif
 
        /*
        memcg_destroy_kmem(memcg);
 #ifdef CONFIG_INET
        if (cgroup_subsys_on_dfl(memory_cgrp_subsys) && !cgroup_memory_nosocket)
-               static_key_slow_dec(&memcg_sockets_enabled_key);
+               static_branch_dec(&memcg_sockets_enabled_key);
 #endif
        __mem_cgroup_free(memcg);
 }
 
 #ifdef CONFIG_INET
 
-struct static_key memcg_sockets_enabled_key;
+DEFINE_STATIC_KEY_FALSE(memcg_sockets_enabled_key);
 EXPORT_SYMBOL(memcg_sockets_enabled_key);
 
 void sock_update_memcg(struct sock *sk)
 
                return;
 
        if (memcg->tcp_mem.active)
-               static_key_slow_dec(&memcg_sockets_enabled_key);
+               static_branch_dec(&memcg_sockets_enabled_key);
 }
 
 static int tcp_update_limit(struct mem_cgroup *memcg, unsigned long nr_pages)
                 * because when this value change, the code to process it is not
                 * patched in yet.
                 */
-               static_key_slow_inc(&memcg_sockets_enabled_key);
+               static_branch_inc(&memcg_sockets_enabled_key);
                memcg->tcp_mem.active = true;
        }