#ifdef CONFIG_RCU_FAST_NO_HZ
        struct rcu_head oom_head;
 #endif /* #ifdef CONFIG_RCU_FAST_NO_HZ */
-       atomic_long_t exp_workdone0;    /* # done by workqueue. */
-       atomic_long_t exp_workdone1;    /* # done by others #1. */
-       atomic_long_t exp_workdone2;    /* # done by others #2. */
-       atomic_long_t exp_workdone3;    /* # done by others #3. */
        int exp_dynticks_snap;          /* Double-check need for IPI. */
 
        /* 6) Callback offloading. */
 
 }
 
 /* Common code for synchronize_{rcu,sched}_expedited() work-done checking. */
-static bool sync_exp_work_done(struct rcu_state *rsp, atomic_long_t *stat,
-                              unsigned long s)
+static bool sync_exp_work_done(struct rcu_state *rsp, unsigned long s)
 {
        if (rcu_exp_gp_seq_done(rsp, s)) {
                trace_rcu_exp_grace_period(rsp->name, s, TPS("done"));
                /* Ensure test happens before caller kfree(). */
                smp_mb__before_atomic(); /* ^^^ */
-               atomic_long_inc(stat);
                return true;
        }
        return false;
         * promoting locality and is not strictly needed for correctness.
         */
        for (; rnp != NULL; rnp = rnp->parent) {
-               if (sync_exp_work_done(rsp, &rdp->exp_workdone1, s))
+               if (sync_exp_work_done(rsp, s))
                        return true;
 
                /* Work not done, either wait here or go up. */
                                                  rnp->grplo, rnp->grphi,
                                                  TPS("wait"));
                        wait_event(rnp->exp_wq[rcu_seq_ctr(s) & 0x3],
-                                  sync_exp_work_done(rsp,
-                                                     &rdp->exp_workdone2, s));
+                                  sync_exp_work_done(rsp, s));
                        return true;
                }
                rnp->exp_seq_rq = s; /* Followers can wait on us. */
        }
        mutex_lock(&rsp->exp_mutex);
 fastpath:
-       if (sync_exp_work_done(rsp, &rdp->exp_workdone3, s)) {
+       if (sync_exp_work_done(rsp, s)) {
                mutex_unlock(&rsp->exp_mutex);
                return true;
        }
        rdp = per_cpu_ptr(rsp->rda, raw_smp_processor_id());
        rnp = rcu_get_root(rsp);
        wait_event(rnp->exp_wq[rcu_seq_ctr(s) & 0x3],
-                  sync_exp_work_done(rsp, &rdp->exp_workdone0, s));
+                  sync_exp_work_done(rsp, s));
        smp_mb(); /* Workqueue actions happen before return. */
 
        /* Let the next expedited grace period start. */