static int __init init_amu_fie(void)
 {
-       bool invariance_status = topology_scale_freq_invariant();
        cpumask_var_t valid_cpus;
+       bool invariant;
        int ret = 0;
        int cpu;
 
        if (cpumask_equal(valid_cpus, cpu_present_mask))
                cpumask_copy(amu_fie_cpus, cpu_present_mask);
 
-       if (!cpumask_empty(amu_fie_cpus)) {
-               pr_info("CPUs[%*pbl]: counters will be used for FIE.",
-                       cpumask_pr_args(amu_fie_cpus));
-               static_branch_enable(&amu_fie_key);
-       }
+       if (cpumask_empty(amu_fie_cpus))
+               goto free_valid_mask;
 
-       /*
-        * If the system is not fully invariant after AMU init, disable
-        * partial use of counters for frequency invariance.
-        */
-       if (!topology_scale_freq_invariant())
-               static_branch_disable(&amu_fie_key);
+       invariant = topology_scale_freq_invariant();
+
+       /* We aren't fully invariant yet */
+       if (!invariant && !cpumask_equal(amu_fie_cpus, cpu_present_mask))
+               goto free_valid_mask;
+
+       static_branch_enable(&amu_fie_key);
+
+       pr_info("CPUs[%*pbl]: counters will be used for FIE.",
+               cpumask_pr_args(amu_fie_cpus));
 
        /*
         * Task scheduler behavior depends on frequency invariance support,
         * a result of counter initialisation and use, retrigger the build of
         * scheduling domains to ensure the information is propagated properly.
         */
-       if (invariance_status != topology_scale_freq_invariant())
+       if (!invariant)
                rebuild_sched_domains_energy();
 
 free_valid_mask: