__read_mostly int xen_have_vector_callback;
 EXPORT_SYMBOL_GPL(xen_have_vector_callback);
 
+static struct notifier_block xen_cpu_notifier;
+
 /*
  * Point at some empty memory to start with. We map the real shared_info
  * page as soon as fixmap is up and running.
        xen_initial_gdt = &per_cpu(gdt_page, 0);
 
        xen_smp_init();
+       register_cpu_notifier(&xen_cpu_notifier);
 
 #ifdef CONFIG_ACPI_NUMA
        /*
        xen_domain_type = XEN_HVM_DOMAIN;
 }
 
-static int xen_hvm_cpu_notify(struct notifier_block *self, unsigned long action,
-                             void *hcpu)
+static int xen_cpu_notify(struct notifier_block *self, unsigned long action,
+                        void *hcpu)
 {
        int cpu = (long)hcpu;
+       int rc;
+
        switch (action) {
        case CPU_UP_PREPARE:
-               if (cpu_acpi_id(cpu) != U32_MAX)
-                       per_cpu(xen_vcpu_id, cpu) = cpu_acpi_id(cpu);
-               else
-                       per_cpu(xen_vcpu_id, cpu) = cpu;
-               xen_vcpu_setup(cpu);
-               if (xen_have_vector_callback) {
-                       if (xen_feature(XENFEAT_hvm_safe_pvclock))
-                               xen_setup_timer(cpu);
+               if (xen_hvm_domain()) {
+                       /*
+                        * This can happen if CPU was offlined earlier and
+                        * offlining timed out in common_cpu_die().
+                        */
+                       if (cpu_report_state(cpu) == CPU_DEAD_FROZEN) {
+                               xen_smp_intr_free(cpu);
+                               xen_uninit_lock_cpu(cpu);
+                       }
+
+                       if (cpu_acpi_id(cpu) != U32_MAX)
+                               per_cpu(xen_vcpu_id, cpu) = cpu_acpi_id(cpu);
+                       else
+                               per_cpu(xen_vcpu_id, cpu) = cpu;
+                       xen_vcpu_setup(cpu);
                }
+
+               if (xen_pv_domain() ||
+                   (xen_have_vector_callback &&
+                    xen_feature(XENFEAT_hvm_safe_pvclock)))
+                       xen_setup_timer(cpu);
+
+               rc = xen_smp_intr_init(cpu);
+               if (rc) {
+                       WARN(1, "xen_smp_intr_init() for CPU %d failed: %d\n",
+                            cpu, rc);
+                       return NOTIFY_BAD;
+               }
+
+               break;
+       case CPU_ONLINE:
+               xen_init_lock_cpu(cpu);
+               break;
+       case CPU_UP_CANCELED:
+               xen_smp_intr_free(cpu);
+               if (xen_pv_domain() ||
+                   (xen_have_vector_callback &&
+                    xen_feature(XENFEAT_hvm_safe_pvclock)))
+                       xen_teardown_timer(cpu);
                break;
        default:
                break;
        return NOTIFY_OK;
 }
 
-static struct notifier_block xen_hvm_cpu_notifier = {
-       .notifier_call  = xen_hvm_cpu_notify,
+static struct notifier_block xen_cpu_notifier = {
+       .notifier_call  = xen_cpu_notify,
 };
 
 #ifdef CONFIG_KEXEC_CORE
        if (xen_feature(XENFEAT_hvm_callback_vector))
                xen_have_vector_callback = 1;
        xen_hvm_smp_init();
-       register_cpu_notifier(&xen_hvm_cpu_notifier);
+       register_cpu_notifier(&xen_cpu_notifier);
        xen_unplug_emulated_devices();
        x86_init.irqs.intr_init = xen_init_IRQ;
        xen_hvm_init_time_ops();
 
        cpu_startup_entry(CPUHP_AP_ONLINE_IDLE);
 }
 
-static void xen_smp_intr_free(unsigned int cpu)
+void xen_smp_intr_free(unsigned int cpu)
 {
        if (per_cpu(xen_resched_irq, cpu).irq >= 0) {
                unbind_from_irqhandler(per_cpu(xen_resched_irq, cpu).irq, NULL);
                per_cpu(xen_pmu_irq, cpu).name = NULL;
        }
 };
-static int xen_smp_intr_init(unsigned int cpu)
+int xen_smp_intr_init(unsigned int cpu)
 {
        int rc;
        char *resched_name, *callfunc_name, *debug_name, *pmu_name;
        common_cpu_up(cpu, idle);
 
        xen_setup_runstate_info(cpu);
-       xen_setup_timer(cpu);
-       xen_init_lock_cpu(cpu);
 
        /*
         * PV VCPUs are always successfully taken down (see 'while' loop
 
        xen_pmu_init(cpu);
 
-       rc = xen_smp_intr_init(cpu);
-       if (rc)
-               return rc;
-
        rc = HYPERVISOR_vcpu_op(VCPUOP_up, xen_vcpu_nr(cpu), NULL);
        BUG_ON(rc);
 
        xen_init_lock_cpu(0);
 }
 
-static int xen_hvm_cpu_up(unsigned int cpu, struct task_struct *tidle)
-{
-       int rc;
-
-       /*
-        * This can happen if CPU was offlined earlier and
-        * offlining timed out in common_cpu_die().
-        */
-       if (cpu_report_state(cpu) == CPU_DEAD_FROZEN) {
-               xen_smp_intr_free(cpu);
-               xen_uninit_lock_cpu(cpu);
-       }
-
-       /*
-        * xen_smp_intr_init() needs to run before native_cpu_up()
-        * so that IPI vectors are set up on the booting CPU before
-        * it is marked online in native_cpu_up().
-       */
-       rc = xen_smp_intr_init(cpu);
-       WARN_ON(rc);
-       if (!rc)
-               rc =  native_cpu_up(cpu, tidle);
-
-       /*
-        * We must initialize the slowpath CPU kicker _after_ the native
-        * path has executed. If we initialized it before none of the
-        * unlocker IPI kicks would reach the booting CPU as the booting
-        * CPU had not set itself 'online' in cpu_online_mask. That mask
-        * is checked when IPIs are sent (on HVM at least).
-        */
-       xen_init_lock_cpu(cpu);
-       return rc;
-}
-
 void __init xen_hvm_smp_init(void)
 {
        if (!xen_have_vector_callback)
                return;
        smp_ops.smp_prepare_cpus = xen_hvm_smp_prepare_cpus;
        smp_ops.smp_send_reschedule = xen_smp_send_reschedule;
-       smp_ops.cpu_up = xen_hvm_cpu_up;
        smp_ops.cpu_die = xen_cpu_die;
        smp_ops.send_call_func_ipi = xen_smp_send_call_function_ipi;
        smp_ops.send_call_func_single_ipi = xen_smp_send_call_function_single_ipi;
 
 #ifndef _XEN_SMP_H
 
+#ifdef CONFIG_SMP
 extern void xen_send_IPI_mask(const struct cpumask *mask,
                              int vector);
 extern void xen_send_IPI_mask_allbutself(const struct cpumask *mask,
 extern void xen_send_IPI_all(int vector);
 extern void xen_send_IPI_self(int vector);
 
+extern int xen_smp_intr_init(unsigned int cpu);
+extern void xen_smp_intr_free(unsigned int cpu);
+
+#else /* CONFIG_SMP */
+
+static inline int xen_smp_intr_init(unsigned int cpu)
+{
+       return 0;
+}
+static inline void xen_smp_intr_free(unsigned int cpu) {}
+#endif /* CONFIG_SMP */
+
 #ifdef CONFIG_XEN_PVH
 extern void xen_pvh_early_cpu_init(int cpu, bool entry);
 #else