return;
wrmsrl(MSR_KVM_STEAL_TIME, (slow_virt_to_phys(st) | KVM_MSR_ENABLED));
- pr_info("stealtime: cpu %d, msr %llx\n", cpu,
+ if (0) pr_info("stealtime: cpu %d, msr %llx\n", cpu,
(unsigned long long) slow_virt_to_phys(st));
}
kvm_sched_clock_offset = kvm_clock_read();
pv_ops.time.sched_clock = kvm_sched_clock_read;
- pr_info("kvm-clock: using sched offset of %llu cycles",
+ if (0) pr_info("kvm-clock: using sched offset of %llu cycles",
kvm_sched_clock_offset);
BUILD_BUG_ON(sizeof(kvm_sched_clock_offset) >
pa = slow_virt_to_phys(&src->pvti) | 0x01ULL;
wrmsrl(msr_kvm_system_time, pa);
- pr_info("kvm-clock: cpu %d, msr %llx, %s", smp_processor_id(), pa, txt);
+ if (0) pr_info("kvm-clock: cpu %d, msr %llx, %s", smp_processor_id(), pa, txt);
}
static void kvm_save_sched_clock_state(void)
goto found;
new = logical_die++;
- if (new != die) {
+ if (0 && new != die) {
pr_info("CPU %u Converting physical %u to logical die %u\n",
cpu, die, new);
}
{
unsigned int cpu;
int n = setup_max_cpus - num_online_cpus();
+ cycles_t t1, t2;
/* ∀ parallel pre-bringup state, bring N CPUs to it */
if (n > 0) {
while (st <= CPUHP_BP_PARALLEL_DYN_END &&
cpuhp_hp_states[st].name) {
int i = n;
-
+ t1 = get_cycles();
for_each_present_cpu(cpu) {
cpu_up(cpu, st);
if (!--i)
break;
}
+ t2 = get_cycles();
+
+ printk("Brought %d CPUs to %s in %lld cycles\n", n - i,
+ cpuhp_hp_states[st].name, t2 - t1);
st++;
}
}
+ t1 = get_cycles();
for_each_present_cpu(cpu) {
if (num_online_cpus() >= setup_max_cpus)
break;
if (!cpu_online(cpu))
cpu_up(cpu, CPUHP_ONLINE);
}
+ t2 = get_cycles();
+ printk("Brought CPUs online in %lld cycles\n", t2 - t1);
}
#ifdef CONFIG_PM_SLEEP_SMP