@@ -547,16 +547,6 @@ static void kvm_send_ipi_mask_allbutself(const struct cpumask *mask, int vector)
__send_ipi_mask(local_mask, vector);
}
-/*
- * Set the IPI entry points
- */
-static void kvm_setup_pv_ipi(void)
-{
- apic->send_IPI_mask = kvm_send_ipi_mask;
- apic->send_IPI_mask_allbutself = kvm_send_ipi_mask_allbutself;
- pr_info("setup PV IPIs\n");
-}
-
static void kvm_smp_send_call_func_ipi(const struct cpumask *mask)
{
int cpu;
@@ -609,7 +599,24 @@ static int kvm_cpu_down_prepare(unsigned int cpu)
local_irq_enable();
return 0;
}
+#else
+static bool pv_ipi_supported(void)
+{
+ return false;
+}
+#endif
+
+/*
+ * Set the IPI entry points
+ */
+static void kvm_setup_pv_ipi(void)
+{
+#if defined(CONFIG_SMP)
+ apic->send_IPI_mask = kvm_send_ipi_mask;
+ apic->send_IPI_mask_allbutself = kvm_send_ipi_mask_allbutself;
+ pr_info("setup PV IPIs\n");
#endif
+}
static void kvm_flush_tlb_others(const struct cpumask *cpumask,
const struct flush_tlb_info *info)
@@ -619,6 +626,11 @@ static void kvm_flush_tlb_others(const struct cpumask *cpumask,
struct kvm_steal_time *src;
struct cpumask *flushmask = this_cpu_cpumask_var_ptr(__pv_cpu_mask);
+ if (unlikely(!flushmask)) {
+ native_flush_tlb_others(cpumask, info);
+ return;
+ }
+
cpumask_copy(flushmask, cpumask);
/*
* We have to call flush only on online vCPUs. And
@@ -730,18 +742,9 @@ static uint32_t __init kvm_detect(void)
return kvm_cpuid_base();
}
-static void __init kvm_apic_init(void)
-{
-#if defined(CONFIG_SMP)
- if (pv_ipi_supported())
- kvm_setup_pv_ipi();
-#endif
-}
-
static void __init kvm_init_platform(void)
{
kvmclock_init();
- x86_platform.apic_post_init = kvm_apic_init;
}
const __initconst struct hypervisor_x86 x86_hyper_kvm = {
@@ -765,29 +768,39 @@ static __init int activate_jump_labels(void)
}
arch_initcall(activate_jump_labels);
+static void kvm_free_cpumask(void)
+{
+ unsigned int cpu;
+
+ for_each_possible_cpu(cpu)
+ free_cpumask_var(per_cpu(__pv_cpu_mask, cpu));
+}
+
static __init int kvm_alloc_cpumask(void)
{
int cpu;
- bool alloc = false;
if (!kvm_para_available() || nopv)
return 0;
- if (pv_tlb_flush_supported())
- alloc = true;
-
-#if defined(CONFIG_SMP)
- if (pv_ipi_supported())
- alloc = true;
-#endif
-
- if (alloc)
+ if (pv_tlb_flush_supported() || pv_ipi_supported()) {
for_each_possible_cpu(cpu) {
- zalloc_cpumask_var_node(per_cpu_ptr(&__pv_cpu_mask, cpu),
- GFP_KERNEL, cpu_to_node(cpu));
+ if (!zalloc_cpumask_var_node(
+ per_cpu_ptr(&__pv_cpu_mask, cpu),
+ GFP_KERNEL, cpu_to_node(cpu))) {
+ goto zalloc_cpumask_fail;
+ }
}
+ }
+
+ if (pv_ipi_supported())
+ kvm_setup_pv_ipi();
return 0;
+
+zalloc_cpumask_fail:
+ kvm_free_cpumask();
+ return -ENOMEM;
}
arch_initcall(kvm_alloc_cpumask);