#include <linux/sched/cputime.h>
 
-static DEFINE_MUTEX(cgroup_stat_mutex);
-static DEFINE_PER_CPU(raw_spinlock_t, cgroup_cpu_stat_lock);
+static DEFINE_MUTEX(cgroup_rstat_mutex);
+static DEFINE_PER_CPU(raw_spinlock_t, cgroup_rstat_cpu_lock);
 
-static struct cgroup_cpu_stat *cgroup_cpu_stat(struct cgroup *cgrp, int cpu)
+static struct cgroup_rstat_cpu *cgroup_rstat_cpu(struct cgroup *cgrp, int cpu)
 {
-       return per_cpu_ptr(cgrp->cpu_stat, cpu);
+       return per_cpu_ptr(cgrp->rstat_cpu, cpu);
 }
 
 /**
- * cgroup_cpu_stat_updated - keep track of updated cpu_stat
+ * cgroup_rstat_cpu_updated - keep track of updated rstat_cpu
  * @cgrp: target cgroup
- * @cpu: cpu on which cpu_stat was updated
+ * @cpu: cpu on which rstat_cpu was updated
  *
- * @cgrp's cpu_stat on @cpu was updated.  Put it on the parent's matching
- * cpu_stat->updated_children list.  See the comment on top of
- * cgroup_cpu_stat definition for details.
+ * @cgrp's rstat_cpu on @cpu was updated.  Put it on the parent's matching
+ * rstat_cpu->updated_children list.  See the comment on top of
+ * cgroup_rstat_cpu definition for details.
  */
-static void cgroup_cpu_stat_updated(struct cgroup *cgrp, int cpu)
+static void cgroup_rstat_cpu_updated(struct cgroup *cgrp, int cpu)
 {
-       raw_spinlock_t *cpu_lock = per_cpu_ptr(&cgroup_cpu_stat_lock, cpu);
+       raw_spinlock_t *cpu_lock = per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu);
        struct cgroup *parent;
        unsigned long flags;
 
         * instead of NULL, we can tell whether @cgrp is on the list by
         * testing the next pointer for NULL.
         */
-       if (cgroup_cpu_stat(cgrp, cpu)->updated_next)
+       if (cgroup_rstat_cpu(cgrp, cpu)->updated_next)
                return;
 
        raw_spin_lock_irqsave(cpu_lock, flags);
        /* put @cgrp and all ancestors on the corresponding updated lists */
        for (parent = cgroup_parent(cgrp); parent;
             cgrp = parent, parent = cgroup_parent(cgrp)) {
-               struct cgroup_cpu_stat *cstat = cgroup_cpu_stat(cgrp, cpu);
-               struct cgroup_cpu_stat *pcstat = cgroup_cpu_stat(parent, cpu);
+               struct cgroup_rstat_cpu *rstatc = cgroup_rstat_cpu(cgrp, cpu);
+               struct cgroup_rstat_cpu *prstatc = cgroup_rstat_cpu(parent, cpu);
 
                /*
                 * Both additions and removals are bottom-up.  If a cgroup
                 * is already in the tree, all ancestors are.
                 */
-               if (cstat->updated_next)
+               if (rstatc->updated_next)
                        break;
 
-               cstat->updated_next = pcstat->updated_children;
-               pcstat->updated_children = cgrp;
+               rstatc->updated_next = prstatc->updated_children;
+               prstatc->updated_children = cgrp;
        }
 
        raw_spin_unlock_irqrestore(cpu_lock, flags);
 }
 
 /**
- * cgroup_cpu_stat_pop_updated - iterate and dismantle cpu_stat updated tree
+ * cgroup_rstat_cpu_pop_updated - iterate and dismantle rstat_cpu updated tree
  * @pos: current position
  * @root: root of the tree to traversal
  * @cpu: target cpu
  *
- * Walks the udpated cpu_stat tree on @cpu from @root.  %NULL @pos starts
+ * Walks the udpated rstat_cpu tree on @cpu from @root.  %NULL @pos starts
  * the traversal and %NULL return indicates the end.  During traversal,
  * each returned cgroup is unlinked from the tree.  Must be called with the
- * matching cgroup_cpu_stat_lock held.
+ * matching cgroup_rstat_cpu_lock held.
  *
  * The only ordering guarantee is that, for a parent and a child pair
  * covered by a given traversal, if a child is visited, its parent is
  * guaranteed to be visited afterwards.
  */
-static struct cgroup *cgroup_cpu_stat_pop_updated(struct cgroup *pos,
-                                                 struct cgroup *root, int cpu)
+static struct cgroup *cgroup_rstat_cpu_pop_updated(struct cgroup *pos,
+                                                  struct cgroup *root, int cpu)
 {
-       struct cgroup_cpu_stat *cstat;
+       struct cgroup_rstat_cpu *rstatc;
        struct cgroup *parent;
 
        if (pos == root)
 
        /* walk down to the first leaf */
        while (true) {
-               cstat = cgroup_cpu_stat(pos, cpu);
-               if (cstat->updated_children == pos)
+               rstatc = cgroup_rstat_cpu(pos, cpu);
+               if (rstatc->updated_children == pos)
                        break;
-               pos = cstat->updated_children;
+               pos = rstatc->updated_children;
        }
 
        /*
         * child in most cases. The only exception is @root.
         */
        parent = cgroup_parent(pos);
-       if (parent && cstat->updated_next) {
-               struct cgroup_cpu_stat *pcstat = cgroup_cpu_stat(parent, cpu);
-               struct cgroup_cpu_stat *ncstat;
+       if (parent && rstatc->updated_next) {
+               struct cgroup_rstat_cpu *prstatc = cgroup_rstat_cpu(parent, cpu);
+               struct cgroup_rstat_cpu *nrstatc;
                struct cgroup **nextp;
 
-               nextp = &pcstat->updated_children;
+               nextp = &prstatc->updated_children;
                while (true) {
-                       ncstat = cgroup_cpu_stat(*nextp, cpu);
+                       nrstatc = cgroup_rstat_cpu(*nextp, cpu);
                        if (*nextp == pos)
                                break;
 
                        WARN_ON_ONCE(*nextp == parent);
-                       nextp = &ncstat->updated_next;
+                       nextp = &nrstatc->updated_next;
                }
 
-               *nextp = cstat->updated_next;
-               cstat->updated_next = NULL;
+               *nextp = rstatc->updated_next;
+               rstatc->updated_next = NULL;
        }
 
        return pos;
 static void cgroup_cpu_stat_flush_one(struct cgroup *cgrp, int cpu)
 {
        struct cgroup *parent = cgroup_parent(cgrp);
-       struct cgroup_cpu_stat *cstat = cgroup_cpu_stat(cgrp, cpu);
-       struct task_cputime *last_cputime = &cstat->last_cputime;
+       struct cgroup_rstat_cpu *rstatc = cgroup_rstat_cpu(cgrp, cpu);
+       struct task_cputime *last_cputime = &rstatc->last_cputime;
        struct task_cputime cputime;
        struct cgroup_stat delta;
        unsigned seq;
 
-       lockdep_assert_held(&cgroup_stat_mutex);
+       lockdep_assert_held(&cgroup_rstat_mutex);
 
        /* fetch the current per-cpu values */
        do {
-               seq = __u64_stats_fetch_begin(&cstat->sync);
-               cputime = cstat->cputime;
-       } while (__u64_stats_fetch_retry(&cstat->sync, seq));
+               seq = __u64_stats_fetch_begin(&rstatc->sync);
+               cputime = rstatc->cputime;
+       } while (__u64_stats_fetch_retry(&rstatc->sync, seq));
 
        /* accumulate the deltas to propgate */
        delta.cputime.utime = cputime.utime - last_cputime->utime;
                cgroup_stat_accumulate(&parent->pending_stat, &delta);
 }
 
-/* see cgroup_stat_flush() */
-static void cgroup_stat_flush_locked(struct cgroup *cgrp)
+/* see cgroup_rstat_flush() */
+static void cgroup_rstat_flush_locked(struct cgroup *cgrp)
 {
        int cpu;
 
-       lockdep_assert_held(&cgroup_stat_mutex);
+       lockdep_assert_held(&cgroup_rstat_mutex);
 
        for_each_possible_cpu(cpu) {
-               raw_spinlock_t *cpu_lock = per_cpu_ptr(&cgroup_cpu_stat_lock, cpu);
+               raw_spinlock_t *cpu_lock = per_cpu_ptr(&cgroup_rstat_cpu_lock,
+                                                      cpu);
                struct cgroup *pos = NULL;
 
                raw_spin_lock_irq(cpu_lock);
-               while ((pos = cgroup_cpu_stat_pop_updated(pos, cgrp, cpu)))
+               while ((pos = cgroup_rstat_cpu_pop_updated(pos, cgrp, cpu)))
                        cgroup_cpu_stat_flush_one(pos, cpu);
                raw_spin_unlock_irq(cpu_lock);
        }
 }
 
 /**
- * cgroup_stat_flush - flush stats in @cgrp's subtree
+ * cgroup_rstat_flush - flush stats in @cgrp's subtree
  * @cgrp: target cgroup
  *
  * Collect all per-cpu stats in @cgrp's subtree into the global counters
  * This also gets all cgroups in the subtree including @cgrp off the
  * ->updated_children lists.
  */
-void cgroup_stat_flush(struct cgroup *cgrp)
+void cgroup_rstat_flush(struct cgroup *cgrp)
 {
-       mutex_lock(&cgroup_stat_mutex);
-       cgroup_stat_flush_locked(cgrp);
-       mutex_unlock(&cgroup_stat_mutex);
+       mutex_lock(&cgroup_rstat_mutex);
+       cgroup_rstat_flush_locked(cgrp);
+       mutex_unlock(&cgroup_rstat_mutex);
 }
 
-static struct cgroup_cpu_stat *cgroup_cpu_stat_account_begin(struct cgroup *cgrp)
+static struct cgroup_rstat_cpu *
+cgroup_cpu_stat_account_begin(struct cgroup *cgrp)
 {
-       struct cgroup_cpu_stat *cstat;
+       struct cgroup_rstat_cpu *rstatc;
 
-       cstat = get_cpu_ptr(cgrp->cpu_stat);
-       u64_stats_update_begin(&cstat->sync);
-       return cstat;
+       rstatc = get_cpu_ptr(cgrp->rstat_cpu);
+       u64_stats_update_begin(&rstatc->sync);
+       return rstatc;
 }
 
 static void cgroup_cpu_stat_account_end(struct cgroup *cgrp,
-                                       struct cgroup_cpu_stat *cstat)
+                                       struct cgroup_rstat_cpu *rstatc)
 {
-       u64_stats_update_end(&cstat->sync);
-       cgroup_cpu_stat_updated(cgrp, smp_processor_id());
-       put_cpu_ptr(cstat);
+       u64_stats_update_end(&rstatc->sync);
+       cgroup_rstat_cpu_updated(cgrp, smp_processor_id());
+       put_cpu_ptr(rstatc);
 }
 
 void __cgroup_account_cputime(struct cgroup *cgrp, u64 delta_exec)
 {
-       struct cgroup_cpu_stat *cstat;
+       struct cgroup_rstat_cpu *rstatc;
 
-       cstat = cgroup_cpu_stat_account_begin(cgrp);
-       cstat->cputime.sum_exec_runtime += delta_exec;
-       cgroup_cpu_stat_account_end(cgrp, cstat);
+       rstatc = cgroup_cpu_stat_account_begin(cgrp);
+       rstatc->cputime.sum_exec_runtime += delta_exec;
+       cgroup_cpu_stat_account_end(cgrp, rstatc);
 }
 
 void __cgroup_account_cputime_field(struct cgroup *cgrp,
                                    enum cpu_usage_stat index, u64 delta_exec)
 {
-       struct cgroup_cpu_stat *cstat;
+       struct cgroup_rstat_cpu *rstatc;
 
-       cstat = cgroup_cpu_stat_account_begin(cgrp);
+       rstatc = cgroup_cpu_stat_account_begin(cgrp);
 
        switch (index) {
        case CPUTIME_USER:
        case CPUTIME_NICE:
-               cstat->cputime.utime += delta_exec;
+               rstatc->cputime.utime += delta_exec;
                break;
        case CPUTIME_SYSTEM:
        case CPUTIME_IRQ:
        case CPUTIME_SOFTIRQ:
-               cstat->cputime.stime += delta_exec;
+               rstatc->cputime.stime += delta_exec;
                break;
        default:
                break;
        }
 
-       cgroup_cpu_stat_account_end(cgrp, cstat);
+       cgroup_cpu_stat_account_end(cgrp, rstatc);
 }
 
 void cgroup_stat_show_cputime(struct seq_file *seq)
        if (!cgroup_parent(cgrp))
                return;
 
-       mutex_lock(&cgroup_stat_mutex);
+       mutex_lock(&cgroup_rstat_mutex);
 
-       cgroup_stat_flush_locked(cgrp);
+       cgroup_rstat_flush_locked(cgrp);
 
        usage = cgrp->stat.cputime.sum_exec_runtime;
        cputime_adjust(&cgrp->stat.cputime, &cgrp->stat.prev_cputime,
                       &utime, &stime);
 
-       mutex_unlock(&cgroup_stat_mutex);
+       mutex_unlock(&cgroup_rstat_mutex);
 
        do_div(usage, NSEC_PER_USEC);
        do_div(utime, NSEC_PER_USEC);
                   usage, utime, stime);
 }
 
-int cgroup_stat_init(struct cgroup *cgrp)
+int cgroup_rstat_init(struct cgroup *cgrp)
 {
        int cpu;
 
-       /* the root cgrp has cpu_stat preallocated */
-       if (!cgrp->cpu_stat) {
-               cgrp->cpu_stat = alloc_percpu(struct cgroup_cpu_stat);
-               if (!cgrp->cpu_stat)
+       /* the root cgrp has rstat_cpu preallocated */
+       if (!cgrp->rstat_cpu) {
+               cgrp->rstat_cpu = alloc_percpu(struct cgroup_rstat_cpu);
+               if (!cgrp->rstat_cpu)
                        return -ENOMEM;
        }
 
        /* ->updated_children list is self terminated */
        for_each_possible_cpu(cpu) {
-               struct cgroup_cpu_stat *cstat = cgroup_cpu_stat(cgrp, cpu);
+               struct cgroup_rstat_cpu *rstatc = cgroup_rstat_cpu(cgrp, cpu);
 
-               cstat->updated_children = cgrp;
-               u64_stats_init(&cstat->sync);
+               rstatc->updated_children = cgrp;
+               u64_stats_init(&rstatc->sync);
        }
 
        prev_cputime_init(&cgrp->stat.prev_cputime);
        return 0;
 }
 
-void cgroup_stat_exit(struct cgroup *cgrp)
+void cgroup_rstat_exit(struct cgroup *cgrp)
 {
        int cpu;
 
-       cgroup_stat_flush(cgrp);
+       cgroup_rstat_flush(cgrp);
 
        /* sanity check */
        for_each_possible_cpu(cpu) {
-               struct cgroup_cpu_stat *cstat = cgroup_cpu_stat(cgrp, cpu);
+               struct cgroup_rstat_cpu *rstatc = cgroup_rstat_cpu(cgrp, cpu);
 
-               if (WARN_ON_ONCE(cstat->updated_children != cgrp) ||
-                   WARN_ON_ONCE(cstat->updated_next))
+               if (WARN_ON_ONCE(rstatc->updated_children != cgrp) ||
+                   WARN_ON_ONCE(rstatc->updated_next))
                        return;
        }
 
-       free_percpu(cgrp->cpu_stat);
-       cgrp->cpu_stat = NULL;
+       free_percpu(cgrp->rstat_cpu);
+       cgrp->rstat_cpu = NULL;
 }
 
-void __init cgroup_stat_boot(void)
+void __init cgroup_rstat_boot(void)
 {
        int cpu;
 
        for_each_possible_cpu(cpu)
-               raw_spin_lock_init(per_cpu_ptr(&cgroup_cpu_stat_lock, cpu));
+               raw_spin_lock_init(per_cpu_ptr(&cgroup_rstat_cpu_lock, cpu));
 
-       BUG_ON(cgroup_stat_init(&cgrp_dfl_root.cgrp));
+       BUG_ON(cgroup_rstat_init(&cgrp_dfl_root.cgrp));
 }