&cpu_to_cpu_group,
                                                d->send_covered, d->tmpmask);
                break;
+#endif
+#ifdef CONFIG_SCHED_MC
+       case SD_LV_MC: /* set up multi-core groups */
+               cpumask_and(d->this_core_map, cpu_map, cpu_coregroup_mask(cpu));
+               if (cpu == cpumask_first(d->this_core_map))
+                       init_sched_build_groups(d->this_core_map, cpu_map,
+                                               &cpu_to_core_group,
+                                               d->send_covered, d->tmpmask);
+               break;
 #endif
        default:
                break;
 
        for_each_cpu(i, cpu_map) {
                build_sched_groups(&d, SD_LV_SIBLING, cpu_map, i);
+               build_sched_groups(&d, SD_LV_MC, cpu_map, i);
        }
 
-#ifdef CONFIG_SCHED_MC
-       /* Set up multi-core groups */
-       for_each_cpu(i, cpu_map) {
-               cpumask_and(d.this_core_map, cpu_coregroup_mask(i), cpu_map);
-               if (i != cpumask_first(d.this_core_map))
-                       continue;
-
-               init_sched_build_groups(d.this_core_map, cpu_map,
-                                       &cpu_to_core_group,
-                                       d.send_covered, d.tmpmask);
-       }
-#endif
-
        /* Set up physical groups */
        for (i = 0; i < nr_node_ids; i++) {
                cpumask_and(d.nodemask, cpumask_of_node(i), cpu_map);