Message ID | 20191018093723.102471-2-xiaoyao.li@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | KVM: VMX: Some refactor of VMX vmcs and msr bitmap | expand |
On Fri, Oct 18, 2019 at 05:37:21PM +0800, Xiaoyao Li wrote: > Move vmcs related codes into a new function vmx_vmcs_reset() from > vmx_vcpu_reset(). So that it's more clearer which data is related with > vmcs and can be held in vmcs. > > Suggested-by: Krish Sadhukhan <krish.sadhukhan@oracle.com> > Signed-off-by: Xiaoyao Li <xiaoyao.li@intel.com> > --- > arch/x86/kvm/vmx/vmx.c | 65 ++++++++++++++++++++++++------------------ > 1 file changed, 37 insertions(+), 28 deletions(-) > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > index e660e28e9ae0..ef567df344bf 100644 > --- a/arch/x86/kvm/vmx/vmx.c > +++ b/arch/x86/kvm/vmx/vmx.c > @@ -4271,33 +4271,11 @@ static void vmx_vcpu_setup(struct vcpu_vmx *vmx) > } > } > > -static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > +static void vmx_vmcs_reset(struct kvm_vcpu *vcpu, bool init_event) I'd strongly prefer to keep the existing code. For me, "vmcs_reset" means zeroing out the VMCS, i.e. reset the VMCS to a virgin state. "vcpu_reset" means exactly that, stuff vCPU state to emulate RESET/INIT. And the split is arbitrary and funky, e.g. EFER is integrated into the VMCS on all recent CPUs, but here it's handled in vcpu_reset. > { > struct vcpu_vmx *vmx = to_vmx(vcpu); > - struct msr_data apic_base_msr; > u64 cr0; > > - vmx->rmode.vm86_active = 0; > - vmx->spec_ctrl = 0; > - > - vmx->msr_ia32_umwait_control = 0; > - > - vcpu->arch.microcode_version = 0x100000000ULL; > - vmx->vcpu.arch.regs[VCPU_REGS_RDX] = get_rdx_init_val(); > - vmx->hv_deadline_tsc = -1; > - kvm_set_cr8(vcpu, 0); > - > - if (!init_event) { > - apic_base_msr.data = APIC_DEFAULT_PHYS_BASE | > - MSR_IA32_APICBASE_ENABLE; > - if (kvm_vcpu_is_reset_bsp(vcpu)) > - apic_base_msr.data |= MSR_IA32_APICBASE_BSP; > - apic_base_msr.host_initiated = true; > - kvm_set_apic_base(vcpu, &apic_base_msr); > - } > - > - vmx_segment_cache_clear(vmx); > - > seg_setup(VCPU_SREG_CS); > vmcs_write16(GUEST_CS_SELECTOR, 0xf000); > vmcs_writel(GUEST_CS_BASE, 0xffff0000ul); > @@ -4340,8 +4318,6 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > if (kvm_mpx_supported()) > vmcs_write64(GUEST_BNDCFGS, 0); > > - setup_msrs(vmx); > - > vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, 0); /* 22.2.1 */ > > if (cpu_has_vmx_tpr_shadow() && !init_event) { > @@ -4357,19 +4333,52 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > if (vmx->vpid != 0) > vmcs_write16(VIRTUAL_PROCESSOR_ID, vmx->vpid); > > + vpid_sync_context(vmx->vpid); > + > cr0 = X86_CR0_NW | X86_CR0_CD | X86_CR0_ET; > - vmx->vcpu.arch.cr0 = cr0; > + vcpu->arch.cr0 = cr0; > vmx_set_cr0(vcpu, cr0); /* enter rmode */ > vmx_set_cr4(vcpu, 0); > - vmx_set_efer(vcpu, 0); > > update_exception_bitmap(vcpu); > > - vpid_sync_context(vmx->vpid); > if (init_event) > vmx_clear_hlt(vcpu); > } > > +static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) > +{ > + struct vcpu_vmx *vmx = to_vmx(vcpu); > + struct msr_data apic_base_msr; > + > + vmx->rmode.vm86_active = 0; > + vmx->spec_ctrl = 0; > + > + vmx->msr_ia32_umwait_control = 0; > + > + vcpu->arch.microcode_version = 0x100000000ULL; > + kvm_rdx_write(vcpu, get_rdx_init_val()); > + vmx->hv_deadline_tsc = -1; > + kvm_set_cr8(vcpu, 0); > + > + if (!init_event) { > + apic_base_msr.data = APIC_DEFAULT_PHYS_BASE | > + MSR_IA32_APICBASE_ENABLE; > + if (kvm_vcpu_is_reset_bsp(vcpu)) > + apic_base_msr.data |= MSR_IA32_APICBASE_BSP; > + apic_base_msr.host_initiated = true; > + kvm_set_apic_base(vcpu, &apic_base_msr); > + } > + > + vmx_segment_cache_clear(vmx); > + > + setup_msrs(vmx); > + > + vmx_set_efer(vcpu, 0); Setting EFER before CR0/CR4 is a functional change, and likely wrong, e.g. vmx_set_cr0() queries EFER_LME to trigger exit_lmode() if INIT/RESET is received while the vCPU is in long mode. > + vmx_vmcs_reset(vcpu, init_event); > +} > + > static void enable_irq_window(struct kvm_vcpu *vcpu) > { > exec_controls_setbit(to_vmx(vcpu), CPU_BASED_VIRTUAL_INTR_PENDING); > -- > 2.19.1 >
On 10/19/2019 12:57 AM, Sean Christopherson wrote: > On Fri, Oct 18, 2019 at 05:37:21PM +0800, Xiaoyao Li wrote: >> Move vmcs related codes into a new function vmx_vmcs_reset() from >> vmx_vcpu_reset(). So that it's more clearer which data is related with >> vmcs and can be held in vmcs. >> >> Suggested-by: Krish Sadhukhan <krish.sadhukhan@oracle.com> >> Signed-off-by: Xiaoyao Li <xiaoyao.li@intel.com> >> --- >> arch/x86/kvm/vmx/vmx.c | 65 ++++++++++++++++++++++++------------------ >> 1 file changed, 37 insertions(+), 28 deletions(-) >> >> diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c >> index e660e28e9ae0..ef567df344bf 100644 >> --- a/arch/x86/kvm/vmx/vmx.c >> +++ b/arch/x86/kvm/vmx/vmx.c >> @@ -4271,33 +4271,11 @@ static void vmx_vcpu_setup(struct vcpu_vmx *vmx) >> } >> } >> >> -static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) >> +static void vmx_vmcs_reset(struct kvm_vcpu *vcpu, bool init_event) > > I'd strongly prefer to keep the existing code. For me, "vmcs_reset" means > zeroing out the VMCS, i.e. reset the VMCS to a virgin state. "vcpu_reset" > means exactly that, stuff vCPU state to emulate RESET/INIT. > > And the split is arbitrary and funky, e.g. EFER is integrated into the > VMCS on all recent CPUs, but here it's handled in vcpu_reset. > I left EFER in vcpu_reset() because it doesn't directly lead to a vmcs_write in vmx_set_efer(). OK. I'll drop this patch.
diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c index e660e28e9ae0..ef567df344bf 100644 --- a/arch/x86/kvm/vmx/vmx.c +++ b/arch/x86/kvm/vmx/vmx.c @@ -4271,33 +4271,11 @@ static void vmx_vcpu_setup(struct vcpu_vmx *vmx) } } -static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) +static void vmx_vmcs_reset(struct kvm_vcpu *vcpu, bool init_event) { struct vcpu_vmx *vmx = to_vmx(vcpu); - struct msr_data apic_base_msr; u64 cr0; - vmx->rmode.vm86_active = 0; - vmx->spec_ctrl = 0; - - vmx->msr_ia32_umwait_control = 0; - - vcpu->arch.microcode_version = 0x100000000ULL; - vmx->vcpu.arch.regs[VCPU_REGS_RDX] = get_rdx_init_val(); - vmx->hv_deadline_tsc = -1; - kvm_set_cr8(vcpu, 0); - - if (!init_event) { - apic_base_msr.data = APIC_DEFAULT_PHYS_BASE | - MSR_IA32_APICBASE_ENABLE; - if (kvm_vcpu_is_reset_bsp(vcpu)) - apic_base_msr.data |= MSR_IA32_APICBASE_BSP; - apic_base_msr.host_initiated = true; - kvm_set_apic_base(vcpu, &apic_base_msr); - } - - vmx_segment_cache_clear(vmx); - seg_setup(VCPU_SREG_CS); vmcs_write16(GUEST_CS_SELECTOR, 0xf000); vmcs_writel(GUEST_CS_BASE, 0xffff0000ul); @@ -4340,8 +4318,6 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) if (kvm_mpx_supported()) vmcs_write64(GUEST_BNDCFGS, 0); - setup_msrs(vmx); - vmcs_write32(VM_ENTRY_INTR_INFO_FIELD, 0); /* 22.2.1 */ if (cpu_has_vmx_tpr_shadow() && !init_event) { @@ -4357,19 +4333,52 @@ static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) if (vmx->vpid != 0) vmcs_write16(VIRTUAL_PROCESSOR_ID, vmx->vpid); + vpid_sync_context(vmx->vpid); + cr0 = X86_CR0_NW | X86_CR0_CD | X86_CR0_ET; - vmx->vcpu.arch.cr0 = cr0; + vcpu->arch.cr0 = cr0; vmx_set_cr0(vcpu, cr0); /* enter rmode */ vmx_set_cr4(vcpu, 0); - vmx_set_efer(vcpu, 0); update_exception_bitmap(vcpu); - vpid_sync_context(vmx->vpid); if (init_event) vmx_clear_hlt(vcpu); } +static void vmx_vcpu_reset(struct kvm_vcpu *vcpu, bool init_event) +{ + struct vcpu_vmx *vmx = to_vmx(vcpu); + struct msr_data apic_base_msr; + + vmx->rmode.vm86_active = 0; + vmx->spec_ctrl = 0; + + vmx->msr_ia32_umwait_control = 0; + + vcpu->arch.microcode_version = 0x100000000ULL; + kvm_rdx_write(vcpu, get_rdx_init_val()); + vmx->hv_deadline_tsc = -1; + kvm_set_cr8(vcpu, 0); + + if (!init_event) { + apic_base_msr.data = APIC_DEFAULT_PHYS_BASE | + MSR_IA32_APICBASE_ENABLE; + if (kvm_vcpu_is_reset_bsp(vcpu)) + apic_base_msr.data |= MSR_IA32_APICBASE_BSP; + apic_base_msr.host_initiated = true; + kvm_set_apic_base(vcpu, &apic_base_msr); + } + + vmx_segment_cache_clear(vmx); + + setup_msrs(vmx); + + vmx_set_efer(vcpu, 0); + + vmx_vmcs_reset(vcpu, init_event); +} + static void enable_irq_window(struct kvm_vcpu *vcpu) { exec_controls_setbit(to_vmx(vcpu), CPU_BASED_VIRTUAL_INTR_PENDING);
Move vmcs related codes into a new function vmx_vmcs_reset() from vmx_vcpu_reset(). So that it's more clearer which data is related with vmcs and can be held in vmcs. Suggested-by: Krish Sadhukhan <krish.sadhukhan@oracle.com> Signed-off-by: Xiaoyao Li <xiaoyao.li@intel.com> --- arch/x86/kvm/vmx/vmx.c | 65 ++++++++++++++++++++++++------------------ 1 file changed, 37 insertions(+), 28 deletions(-)