goto out;
 
        if (rq->nr_running >= 2) {
-               flags = NOHZ_KICK_MASK;
+               flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
                goto out;
        }
 
                 * on.
                 */
                if (rq->cfs.h_nr_running >= 1 && check_cpu_capacity(rq, sd)) {
-                       flags = NOHZ_KICK_MASK;
+                       flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
                        goto unlock;
                }
        }
                 */
                for_each_cpu_and(i, sched_domain_span(sd), nohz.idle_cpus_mask) {
                        if (sched_asym_prefer(i, cpu)) {
-                               flags = NOHZ_KICK_MASK;
+                               flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
                                goto unlock;
                        }
                }
                 * to run the misfit task on.
                 */
                if (check_misfit_status(rq, sd)) {
-                       flags = NOHZ_KICK_MASK;
+                       flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
                        goto unlock;
                }
 
                 */
                nr_busy = atomic_read(&sds->nr_busy_cpus);
                if (nr_busy > 1) {
-                       flags = NOHZ_KICK_MASK;
+                       flags = NOHZ_STATS_KICK | NOHZ_BALANCE_KICK;
                        goto unlock;
                }
        }
         * setting the flag, we are sure to not clear the state and not
         * check the load of an idle cpu.
         */
-       WRITE_ONCE(nohz.has_blocked, 0);
+       if (flags & NOHZ_STATS_KICK)
+               WRITE_ONCE(nohz.has_blocked, 0);
 
        /*
         * Ensures that if we miss the CPU, we must see the has_blocked
                 * balancing owner will pick it up.
                 */
                if (need_resched()) {
-                       has_blocked_load = true;
+                       if (flags & NOHZ_STATS_KICK)
+                               has_blocked_load = true;
                        goto abort;
                }
 
                rq = cpu_rq(balance_cpu);
 
-               has_blocked_load |= update_nohz_stats(rq);
+               if (flags & NOHZ_STATS_KICK)
+                       has_blocked_load |= update_nohz_stats(rq);
 
                /*
                 * If time for next balance is due,
        if (likely(update_next_balance))
                nohz.next_balance = next_balance;
 
-       WRITE_ONCE(nohz.next_blocked,
-               now + msecs_to_jiffies(LOAD_AVG_PERIOD));
+       if (flags & NOHZ_STATS_KICK)
+               WRITE_ONCE(nohz.next_blocked,
+                          now + msecs_to_jiffies(LOAD_AVG_PERIOD));
 
 abort:
        /* There is still blocked load, enable periodic update */
 
 #define NOHZ_BALANCE_KICK_BIT  0
 #define NOHZ_STATS_KICK_BIT    1
 #define NOHZ_NEWILB_KICK_BIT   2
+#define NOHZ_NEXT_KICK_BIT     3
 
+/* Run rebalance_domains() */
 #define NOHZ_BALANCE_KICK      BIT(NOHZ_BALANCE_KICK_BIT)
+/* Update blocked load */
 #define NOHZ_STATS_KICK                BIT(NOHZ_STATS_KICK_BIT)
+/* Update blocked load when entering idle */
 #define NOHZ_NEWILB_KICK       BIT(NOHZ_NEWILB_KICK_BIT)
+/* Update nohz.next_balance */
+#define NOHZ_NEXT_KICK         BIT(NOHZ_NEXT_KICK_BIT)
 
-#define NOHZ_KICK_MASK (NOHZ_BALANCE_KICK | NOHZ_STATS_KICK)
+#define NOHZ_KICK_MASK (NOHZ_BALANCE_KICK | NOHZ_STATS_KICK | NOHZ_NEXT_KICK)
 
 #define nohz_flags(cpu)        (&cpu_rq(cpu)->nohz_flags)