* cgroup root (root_mem_cgroup). So we have to handle
         * dead_memcg from cgroup root separately.
         */
-       if (last != root_mem_cgroup)
+       if (!mem_cgroup_is_root(last))
                __invalidate_reclaim_iterators(root_mem_cgroup,
                                                dead_memcg);
 }
        struct mem_cgroup *iter;
        int ret = 0;
 
-       BUG_ON(memcg == root_mem_cgroup);
+       BUG_ON(mem_cgroup_is_root(memcg));
 
        for_each_mem_cgroup_tree(iter, memcg) {
                struct css_task_iter it;
        memcg = folio_memcg(folio);
 
        if (!memcg)
-               VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != root_mem_cgroup, folio);
+               VM_BUG_ON_FOLIO(!mem_cgroup_is_root(lruvec_memcg(lruvec)), folio);
        else
                VM_BUG_ON_FOLIO(lruvec_memcg(lruvec) != memcg, folio);
 }
        rcu_read_lock();
 
        memcg = mem_cgroup_from_task(victim);
-       if (memcg == root_mem_cgroup)
+       if (mem_cgroup_is_root(memcg))
                goto out;
 
        /*
 {
        struct obj_cgroup *objcg = NULL;
 
-       for (; memcg != root_mem_cgroup; memcg = parent_mem_cgroup(memcg)) {
+       for (; !mem_cgroup_is_root(memcg); memcg = parent_mem_cgroup(memcg)) {
                objcg = rcu_dereference(memcg->objcg);
                if (objcg && obj_cgroup_tryget(objcg))
                        break;
 
        rcu_read_lock();
        memcg = mem_cgroup_from_task(current);
-       if (memcg == root_mem_cgroup)
+       if (mem_cgroup_is_root(memcg))
                goto out;
        if (!cgroup_subsys_on_dfl(memory_cgrp_subsys) && !memcg->tcpmem_active)
                goto out;
                 * The root cgroup cannot be destroyed, so it's refcount must
                 * always be >= 1.
                 */
-               if (WARN_ON_ONCE(memcg == root_mem_cgroup)) {
+               if (WARN_ON_ONCE(mem_cgroup_is_root(memcg))) {
                        VM_BUG_ON(1);
                        break;
                }
 
        if (mem_cgroup_disabled() || do_memsw_account())
                return nr_swap_pages;
-       for (; memcg != root_mem_cgroup; memcg = parent_mem_cgroup(memcg))
+       for (; !mem_cgroup_is_root(memcg); memcg = parent_mem_cgroup(memcg))
                nr_swap_pages = min_t(long, nr_swap_pages,
                                      READ_ONCE(memcg->swap.max) -
                                      page_counter_read(&memcg->swap));
        if (!memcg)
                return false;
 
-       for (; memcg != root_mem_cgroup; memcg = parent_mem_cgroup(memcg)) {
+       for (; !mem_cgroup_is_root(memcg); memcg = parent_mem_cgroup(memcg)) {
                unsigned long usage = page_counter_read(&memcg->swap);
 
                if (usage * 2 >= READ_ONCE(memcg->swap.high) ||
                return true;
 
        original_memcg = get_mem_cgroup_from_objcg(objcg);
-       for (memcg = original_memcg; memcg != root_mem_cgroup;
+       for (memcg = original_memcg; !mem_cgroup_is_root(memcg);
             memcg = parent_mem_cgroup(memcg)) {
                unsigned long max = READ_ONCE(memcg->zswap_max);
                unsigned long pages;