Message ID | 20190927021927.23057-4-weijiang.yang@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Introduce support for Guest CET feature | expand |
On Thu, Sep 26, 2019 at 7:17 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > CET MSRs pass through Guest directly to enhance performance. > CET runtime control settings are stored in MSR_IA32_{U,S}_CET, > Shadow Stack Pointer(SSP) are stored in MSR_IA32_PL{0,1,2,3}_SSP, > SSP table base address is stored in MSR_IA32_INT_SSP_TAB, > these MSRs are defined in kernel and re-used here. All of these new guest MSRs will have to be enumerated by KVM_GET_MSR_INDEX_LIST. > MSR_IA32_U_CET and MSR_IA32_PL3_SSP are used for user mode protection, > the contents could differ from process to process, therefore, > kernel needs to save/restore them during context switch, it makes > sense to pass through them so that the guest kernel can > use xsaves/xrstors to operate them efficiently. Other MSRs are used > for non-user mode protection. See CET spec for detailed info. I assume that XSAVES & XRSTORS bypass the MSR permission bitmap, like other instructions that manipulate MSRs (e.g. SWAPGS, RDTSCP, etc.). Is the guest OS likely to use RDMSR/WRMSR to access these MSRs? > The difference between CET VMCS state fields and xsave components is that, > the former used for CET state storage during VMEnter/VMExit, > whereas the latter used for state retention between Guest task/process > switch. > > Co-developed-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > Signed-off-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > Signed-off-by: Yang Weijiang <weijiang.yang@intel.com> > --- > arch/x86/kvm/cpuid.c | 1 + > arch/x86/kvm/cpuid.h | 2 ++ > arch/x86/kvm/vmx/vmx.c | 39 +++++++++++++++++++++++++++++++++++++++ > 3 files changed, 42 insertions(+) > > diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c > index 1aa86b87b6ab..0a47b9e565be 100644 > --- a/arch/x86/kvm/cpuid.c > +++ b/arch/x86/kvm/cpuid.c > @@ -66,6 +66,7 @@ u64 kvm_supported_xss(void) > { > return KVM_SUPPORTED_XSS & kvm_x86_ops->supported_xss(); > } > +EXPORT_SYMBOL_GPL(kvm_supported_xss); > > #define F(x) bit(X86_FEATURE_##x) > > diff --git a/arch/x86/kvm/cpuid.h b/arch/x86/kvm/cpuid.h > index d78a61408243..1d77b880084d 100644 > --- a/arch/x86/kvm/cpuid.h > +++ b/arch/x86/kvm/cpuid.h > @@ -27,6 +27,8 @@ bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, > > int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); > > +u64 kvm_supported_xss(void); > + > static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) > { > return vcpu->arch.maxphyaddr; > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > index a84198cff397..f720baa7a9ba 100644 > --- a/arch/x86/kvm/vmx/vmx.c > +++ b/arch/x86/kvm/vmx/vmx.c > @@ -7001,6 +7001,43 @@ static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) > vmx->pt_desc.ctl_bitmask &= ~(0xfULL << (32 + i * 4)); > } > > +static void vmx_intercept_cet_msrs(struct kvm_vcpu *vcpu) Nit: It seems like this function adjusts the MSR permission bitmap so as *not* to intercept the CET MSRs. > +{ > + struct vcpu_vmx *vmx = to_vmx(vcpu); > + unsigned long *msr_bitmap; > + u64 kvm_xss; > + bool cet_en; > + > + msr_bitmap = vmx->vmcs01.msr_bitmap; What about nested guests? (i.e. vmcs02). > + kvm_xss = kvm_supported_xss(); > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > + /* > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > + * a bundle for USER CET xsaves. > + */ > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > + } Since this is called from vmx_cpuid_update, what happens if cet_en was previously true and now it's false? > + /* > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > + * for CET KERNEL xsaves. > + */ > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > + > + /* SSP_TAB only available for KERNEL SHSTK.*/ > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > + MSR_TYPE_RW); > + } > +} > + > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > { > struct vcpu_vmx *vmx = to_vmx(vcpu); > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > update_intel_pt_cfg(vcpu); > + > + vmx_intercept_cet_msrs(vcpu); > } > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > -- > 2.17.2 >
On Wed, Oct 02, 2019 at 11:18:32AM -0700, Jim Mattson wrote: > On Thu, Sep 26, 2019 at 7:17 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > > > CET MSRs pass through Guest directly to enhance performance. > > CET runtime control settings are stored in MSR_IA32_{U,S}_CET, > > Shadow Stack Pointer(SSP) are stored in MSR_IA32_PL{0,1,2,3}_SSP, > > SSP table base address is stored in MSR_IA32_INT_SSP_TAB, > > these MSRs are defined in kernel and re-used here. > > All of these new guest MSRs will have to be enumerated by > KVM_GET_MSR_INDEX_LIST. > Since CET feature is Intel platform specific, but looks like KVM_GET_MSR_INDEX_LIST fetchs x86 common MSRs, I have patch in QEMU to support CET MSRs, the patch is here: https://patchwork.ozlabs.org/patch/1058265/ > > MSR_IA32_U_CET and MSR_IA32_PL3_SSP are used for user mode protection, > > the contents could differ from process to process, therefore, > > kernel needs to save/restore them during context switch, it makes > > sense to pass through them so that the guest kernel can > > use xsaves/xrstors to operate them efficiently. Other MSRs are used > > for non-user mode protection. See CET spec for detailed info. > > I assume that XSAVES & XRSTORS bypass the MSR permission bitmap, like > other instructions that manipulate MSRs (e.g. SWAPGS, RDTSCP, etc.). > Is the guest OS likely to use RDMSR/WRMSR to access these MSRs? > Yes, exactly, you may check the CET kernel code. > > The difference between CET VMCS state fields and xsave components is that, > > the former used for CET state storage during VMEnter/VMExit, > > whereas the latter used for state retention between Guest task/process > > switch. > > > > Co-developed-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > Signed-off-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > Signed-off-by: Yang Weijiang <weijiang.yang@intel.com> > > --- > > arch/x86/kvm/cpuid.c | 1 + > > arch/x86/kvm/cpuid.h | 2 ++ > > arch/x86/kvm/vmx/vmx.c | 39 +++++++++++++++++++++++++++++++++++++++ > > 3 files changed, 42 insertions(+) > > > > diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c > > index 1aa86b87b6ab..0a47b9e565be 100644 > > --- a/arch/x86/kvm/cpuid.c > > +++ b/arch/x86/kvm/cpuid.c > > @@ -66,6 +66,7 @@ u64 kvm_supported_xss(void) > > { > > return KVM_SUPPORTED_XSS & kvm_x86_ops->supported_xss(); > > } > > +EXPORT_SYMBOL_GPL(kvm_supported_xss); > > > > #define F(x) bit(X86_FEATURE_##x) > > > > diff --git a/arch/x86/kvm/cpuid.h b/arch/x86/kvm/cpuid.h > > index d78a61408243..1d77b880084d 100644 > > --- a/arch/x86/kvm/cpuid.h > > +++ b/arch/x86/kvm/cpuid.h > > @@ -27,6 +27,8 @@ bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, > > > > int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); > > > > +u64 kvm_supported_xss(void); > > + > > static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) > > { > > return vcpu->arch.maxphyaddr; > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > index a84198cff397..f720baa7a9ba 100644 > > --- a/arch/x86/kvm/vmx/vmx.c > > +++ b/arch/x86/kvm/vmx/vmx.c > > @@ -7001,6 +7001,43 @@ static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) > > vmx->pt_desc.ctl_bitmask &= ~(0xfULL << (32 + i * 4)); > > } > > > > +static void vmx_intercept_cet_msrs(struct kvm_vcpu *vcpu) > > Nit: It seems like this function adjusts the MSR permission bitmap so > as *not* to intercept the CET MSRs. > OK, will rename it. > > +{ > > + struct vcpu_vmx *vmx = to_vmx(vcpu); > > + unsigned long *msr_bitmap; > > + u64 kvm_xss; > > + bool cet_en; > > + > > + msr_bitmap = vmx->vmcs01.msr_bitmap; > > What about nested guests? (i.e. vmcs02). > Hmm, I need to check the nested case, thank you. > > + kvm_xss = kvm_supported_xss(); > > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > > + /* > > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > > + * a bundle for USER CET xsaves. > > + */ > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > > + } > > Since this is called from vmx_cpuid_update, what happens if cet_en was > previously true and now it's false? > Yes, it's likely, but guest CPUID usually is fixed before guest is launched, do you have any suggestion? > > + /* > > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > > + * for CET KERNEL xsaves. > > + */ > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > > + > > + /* SSP_TAB only available for KERNEL SHSTK.*/ > > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > > + MSR_TYPE_RW); > > + } > > +} > > + > > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > { > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > > update_intel_pt_cfg(vcpu); > > + > > + vmx_intercept_cet_msrs(vcpu); > > } > > > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > > -- > > 2.17.2 > >
On Tue, Oct 8, 2019 at 11:13 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > On Wed, Oct 02, 2019 at 11:18:32AM -0700, Jim Mattson wrote: > > On Thu, Sep 26, 2019 at 7:17 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > > > > > CET MSRs pass through Guest directly to enhance performance. > > > CET runtime control settings are stored in MSR_IA32_{U,S}_CET, > > > Shadow Stack Pointer(SSP) are stored in MSR_IA32_PL{0,1,2,3}_SSP, > > > SSP table base address is stored in MSR_IA32_INT_SSP_TAB, > > > these MSRs are defined in kernel and re-used here. > > > > All of these new guest MSRs will have to be enumerated by > > KVM_GET_MSR_INDEX_LIST. > > > Since CET feature is Intel platform specific, but looks like KVM_GET_MSR_INDEX_LIST > fetchs x86 common MSRs, I have patch in QEMU to support CET > MSRs, the patch is here: > https://patchwork.ozlabs.org/patch/1058265/ Qemu is not the only user of kvm. All possible guest MSRs for the platform *must* be enumerated by KVM_GET_MSR_INDEX_LIST. A number of Intel-specific MSRs are already enumerated. > > > MSR_IA32_U_CET and MSR_IA32_PL3_SSP are used for user mode protection, > > > the contents could differ from process to process, therefore, > > > kernel needs to save/restore them during context switch, it makes > > > sense to pass through them so that the guest kernel can > > > use xsaves/xrstors to operate them efficiently. Other MSRs are used > > > for non-user mode protection. See CET spec for detailed info. > > > > I assume that XSAVES & XRSTORS bypass the MSR permission bitmap, like > > other instructions that manipulate MSRs (e.g. SWAPGS, RDTSCP, etc.). > > Is the guest OS likely to use RDMSR/WRMSR to access these MSRs? > > > Yes, exactly, you may check the CET kernel code. > > > > The difference between CET VMCS state fields and xsave components is that, > > > the former used for CET state storage during VMEnter/VMExit, > > > whereas the latter used for state retention between Guest task/process > > > switch. > > > > > > Co-developed-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > > Signed-off-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > > Signed-off-by: Yang Weijiang <weijiang.yang@intel.com> > > > --- > > > arch/x86/kvm/cpuid.c | 1 + > > > arch/x86/kvm/cpuid.h | 2 ++ > > > arch/x86/kvm/vmx/vmx.c | 39 +++++++++++++++++++++++++++++++++++++++ > > > 3 files changed, 42 insertions(+) > > > > > > diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c > > > index 1aa86b87b6ab..0a47b9e565be 100644 > > > --- a/arch/x86/kvm/cpuid.c > > > +++ b/arch/x86/kvm/cpuid.c > > > @@ -66,6 +66,7 @@ u64 kvm_supported_xss(void) > > > { > > > return KVM_SUPPORTED_XSS & kvm_x86_ops->supported_xss(); > > > } > > > +EXPORT_SYMBOL_GPL(kvm_supported_xss); > > > > > > #define F(x) bit(X86_FEATURE_##x) > > > > > > diff --git a/arch/x86/kvm/cpuid.h b/arch/x86/kvm/cpuid.h > > > index d78a61408243..1d77b880084d 100644 > > > --- a/arch/x86/kvm/cpuid.h > > > +++ b/arch/x86/kvm/cpuid.h > > > @@ -27,6 +27,8 @@ bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, > > > > > > int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); > > > > > > +u64 kvm_supported_xss(void); > > > + > > > static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) > > > { > > > return vcpu->arch.maxphyaddr; > > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > > index a84198cff397..f720baa7a9ba 100644 > > > --- a/arch/x86/kvm/vmx/vmx.c > > > +++ b/arch/x86/kvm/vmx/vmx.c > > > @@ -7001,6 +7001,43 @@ static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) > > > vmx->pt_desc.ctl_bitmask &= ~(0xfULL << (32 + i * 4)); > > > } > > > > > > +static void vmx_intercept_cet_msrs(struct kvm_vcpu *vcpu) > > > > Nit: It seems like this function adjusts the MSR permission bitmap so > > as *not* to intercept the CET MSRs. > > > OK, will rename it. > > > +{ > > > + struct vcpu_vmx *vmx = to_vmx(vcpu); > > > + unsigned long *msr_bitmap; > > > + u64 kvm_xss; > > > + bool cet_en; > > > + > > > + msr_bitmap = vmx->vmcs01.msr_bitmap; > > > > What about nested guests? (i.e. vmcs02). > > > Hmm, I need to check the nested case, thank you. > > > > + kvm_xss = kvm_supported_xss(); > > > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > > > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > > > + /* > > > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > > > + * a bundle for USER CET xsaves. > > > + */ > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > > > + } > > > > Since this is called from vmx_cpuid_update, what happens if cet_en was > > previously true and now it's false? > > > Yes, it's likely, but guest CPUID usually is fixed before > guest is launched, do you have any suggestion? How about an else clause? > > > + /* > > > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > > > + * for CET KERNEL xsaves. > > > + */ > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > > > + > > > + /* SSP_TAB only available for KERNEL SHSTK.*/ > > > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > > > + MSR_TYPE_RW); > > > + } > > > +} > > > + > > > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > { > > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > > > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > > > update_intel_pt_cfg(vcpu); > > > + > > > + vmx_intercept_cet_msrs(vcpu); > > > } > > > > > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > > > -- > > > 2.17.2 > > >
On Thu, Oct 10, 2019 at 12:04:40PM -0700, Jim Mattson wrote: > On Tue, Oct 8, 2019 at 11:13 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > > > On Wed, Oct 02, 2019 at 11:18:32AM -0700, Jim Mattson wrote: > > > On Thu, Sep 26, 2019 at 7:17 PM Yang Weijiang <weijiang.yang@intel.com> wrote: > > > > > > > > CET MSRs pass through Guest directly to enhance performance. > > > > CET runtime control settings are stored in MSR_IA32_{U,S}_CET, > > > > Shadow Stack Pointer(SSP) are stored in MSR_IA32_PL{0,1,2,3}_SSP, > > > > SSP table base address is stored in MSR_IA32_INT_SSP_TAB, > > > > these MSRs are defined in kernel and re-used here. > > > > > > All of these new guest MSRs will have to be enumerated by > > > KVM_GET_MSR_INDEX_LIST. > > > > > Since CET feature is Intel platform specific, but looks like KVM_GET_MSR_INDEX_LIST > > fetchs x86 common MSRs, I have patch in QEMU to support CET > > MSRs, the patch is here: > > https://patchwork.ozlabs.org/patch/1058265/ > > Qemu is not the only user of kvm. All possible guest MSRs for the > platform *must* be enumerated by KVM_GET_MSR_INDEX_LIST. A number of > Intel-specific MSRs are already enumerated. > Sure, will do that, thank you! > > > > MSR_IA32_U_CET and MSR_IA32_PL3_SSP are used for user mode protection, > > > > the contents could differ from process to process, therefore, > > > > kernel needs to save/restore them during context switch, it makes > > > > sense to pass through them so that the guest kernel can > > > > use xsaves/xrstors to operate them efficiently. Other MSRs are used > > > > for non-user mode protection. See CET spec for detailed info. > > > > > > I assume that XSAVES & XRSTORS bypass the MSR permission bitmap, like > > > other instructions that manipulate MSRs (e.g. SWAPGS, RDTSCP, etc.). > > > Is the guest OS likely to use RDMSR/WRMSR to access these MSRs? > > > > > Yes, exactly, you may check the CET kernel code. > > > > > > The difference between CET VMCS state fields and xsave components is that, > > > > the former used for CET state storage during VMEnter/VMExit, > > > > whereas the latter used for state retention between Guest task/process > > > > switch. > > > > > > > > Co-developed-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > > > Signed-off-by: Zhang Yi Z <yi.z.zhang@linux.intel.com> > > > > Signed-off-by: Yang Weijiang <weijiang.yang@intel.com> > > > > --- > > > > arch/x86/kvm/cpuid.c | 1 + > > > > arch/x86/kvm/cpuid.h | 2 ++ > > > > arch/x86/kvm/vmx/vmx.c | 39 +++++++++++++++++++++++++++++++++++++++ > > > > 3 files changed, 42 insertions(+) > > > > > > > > diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c > > > > index 1aa86b87b6ab..0a47b9e565be 100644 > > > > --- a/arch/x86/kvm/cpuid.c > > > > +++ b/arch/x86/kvm/cpuid.c > > > > @@ -66,6 +66,7 @@ u64 kvm_supported_xss(void) > > > > { > > > > return KVM_SUPPORTED_XSS & kvm_x86_ops->supported_xss(); > > > > } > > > > +EXPORT_SYMBOL_GPL(kvm_supported_xss); > > > > > > > > #define F(x) bit(X86_FEATURE_##x) > > > > > > > > diff --git a/arch/x86/kvm/cpuid.h b/arch/x86/kvm/cpuid.h > > > > index d78a61408243..1d77b880084d 100644 > > > > --- a/arch/x86/kvm/cpuid.h > > > > +++ b/arch/x86/kvm/cpuid.h > > > > @@ -27,6 +27,8 @@ bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, > > > > > > > > int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); > > > > > > > > +u64 kvm_supported_xss(void); > > > > + > > > > static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) > > > > { > > > > return vcpu->arch.maxphyaddr; > > > > diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c > > > > index a84198cff397..f720baa7a9ba 100644 > > > > --- a/arch/x86/kvm/vmx/vmx.c > > > > +++ b/arch/x86/kvm/vmx/vmx.c > > > > @@ -7001,6 +7001,43 @@ static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) > > > > vmx->pt_desc.ctl_bitmask &= ~(0xfULL << (32 + i * 4)); > > > > } > > > > > > > > +static void vmx_intercept_cet_msrs(struct kvm_vcpu *vcpu) > > > > > > Nit: It seems like this function adjusts the MSR permission bitmap so > > > as *not* to intercept the CET MSRs. > > > > > OK, will rename it. > > > > +{ > > > > + struct vcpu_vmx *vmx = to_vmx(vcpu); > > > > + unsigned long *msr_bitmap; > > > > + u64 kvm_xss; > > > > + bool cet_en; > > > > + > > > > + msr_bitmap = vmx->vmcs01.msr_bitmap; > > > > > > What about nested guests? (i.e. vmcs02). > > > > > Hmm, I need to check the nested case, thank you. > > > > > > + kvm_xss = kvm_supported_xss(); > > > > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > > > > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > > > > + /* > > > > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > > > > + * a bundle for USER CET xsaves. > > > > + */ > > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > > > > + } > > > > > > Since this is called from vmx_cpuid_update, what happens if cet_en was > > > previously true and now it's false? > > > > > Yes, it's likely, but guest CPUID usually is fixed before > > guest is launched, do you have any suggestion? > > How about an else clause? > OK, will add else clauses on the MSRs. thank you. > > > > + /* > > > > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > > > > + * for CET KERNEL xsaves. > > > > + */ > > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > > > > + > > > > + /* SSP_TAB only available for KERNEL SHSTK.*/ > > > > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > > > > + MSR_TYPE_RW); > > > > + } > > > > +} > > > > + > > > > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > > { > > > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > > > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > > > > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > > > > update_intel_pt_cfg(vcpu); > > > > + > > > > + vmx_intercept_cet_msrs(vcpu); > > > > } > > > > > > > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > > > > -- > > > > 2.17.2 > > > >
On Wed, Oct 09, 2019 at 02:15:09PM +0800, Yang Weijiang wrote: > On Wed, Oct 02, 2019 at 11:18:32AM -0700, Jim Mattson wrote: > > > + kvm_xss = kvm_supported_xss(); > > > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > > > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > > > + /* > > > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > > > + * a bundle for USER CET xsaves. > > > + */ > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > > > + } > > > > Since this is called from vmx_cpuid_update, what happens if cet_en was > > previously true and now it's false? > > > Yes, it's likely, but guest CPUID usually is fixed before > guest is launched, do you have any suggestion? Re-enable interception. kvm_x86_ops->cpuid_update() is only called in the ioctl flow, i.e. it's not performance critical. > > > + /* > > > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > > > + * for CET KERNEL xsaves. > > > + */ > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > > > + > > > + /* SSP_TAB only available for KERNEL SHSTK.*/ > > > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > > > + MSR_TYPE_RW); > > > + } > > > +} > > > + > > > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > { > > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > > > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > > > update_intel_pt_cfg(vcpu); > > > + > > > + vmx_intercept_cet_msrs(vcpu); > > > } > > > > > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > > > -- > > > 2.17.2 > > >
On Thu, Oct 17, 2019 at 01:04:11PM -0700, Sean Christopherson wrote: > On Wed, Oct 09, 2019 at 02:15:09PM +0800, Yang Weijiang wrote: > > On Wed, Oct 02, 2019 at 11:18:32AM -0700, Jim Mattson wrote: > > > > + kvm_xss = kvm_supported_xss(); > > > > + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || > > > > + guest_cpuid_has(vcpu, X86_FEATURE_IBT); > > > > + /* > > > > + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are > > > > + * a bundle for USER CET xsaves. > > > > + */ > > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); > > > > + } > > > > > > Since this is called from vmx_cpuid_update, what happens if cet_en was > > > previously true and now it's false? > > > > > Yes, it's likely, but guest CPUID usually is fixed before > > guest is launched, do you have any suggestion? > > Re-enable interception. kvm_x86_ops->cpuid_update() is only called in > the ioctl flow, i.e. it's not performance critical. > OK, will add a *else* clause to re-intercept the MSRs as Jim suggested. > > > > + /* > > > > + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle > > > > + * for CET KERNEL xsaves. > > > > + */ > > > > + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); > > > > + > > > > + /* SSP_TAB only available for KERNEL SHSTK.*/ > > > > + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) > > > > + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, > > > > + MSR_TYPE_RW); > > > > + } > > > > +} > > > > + > > > > static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > > { > > > > struct vcpu_vmx *vmx = to_vmx(vcpu); > > > > @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) > > > > if (boot_cpu_has(X86_FEATURE_INTEL_PT) && > > > > guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) > > > > update_intel_pt_cfg(vcpu); > > > > + > > > > + vmx_intercept_cet_msrs(vcpu); > > > > } > > > > > > > > static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry) > > > > -- > > > > 2.17.2 > > > >
diff --git a/arch/x86/kvm/cpuid.c b/arch/x86/kvm/cpuid.c index 1aa86b87b6ab..0a47b9e565be 100644 --- a/arch/x86/kvm/cpuid.c +++ b/arch/x86/kvm/cpuid.c @@ -66,6 +66,7 @@ u64 kvm_supported_xss(void) { return KVM_SUPPORTED_XSS & kvm_x86_ops->supported_xss(); } +EXPORT_SYMBOL_GPL(kvm_supported_xss); #define F(x) bit(X86_FEATURE_##x) diff --git a/arch/x86/kvm/cpuid.h b/arch/x86/kvm/cpuid.h index d78a61408243..1d77b880084d 100644 --- a/arch/x86/kvm/cpuid.h +++ b/arch/x86/kvm/cpuid.h @@ -27,6 +27,8 @@ bool kvm_cpuid(struct kvm_vcpu *vcpu, u32 *eax, u32 *ebx, int cpuid_query_maxphyaddr(struct kvm_vcpu *vcpu); +u64 kvm_supported_xss(void); + static inline int cpuid_maxphyaddr(struct kvm_vcpu *vcpu) { return vcpu->arch.maxphyaddr; diff --git a/arch/x86/kvm/vmx/vmx.c b/arch/x86/kvm/vmx/vmx.c index a84198cff397..f720baa7a9ba 100644 --- a/arch/x86/kvm/vmx/vmx.c +++ b/arch/x86/kvm/vmx/vmx.c @@ -7001,6 +7001,43 @@ static void update_intel_pt_cfg(struct kvm_vcpu *vcpu) vmx->pt_desc.ctl_bitmask &= ~(0xfULL << (32 + i * 4)); } +static void vmx_intercept_cet_msrs(struct kvm_vcpu *vcpu) +{ + struct vcpu_vmx *vmx = to_vmx(vcpu); + unsigned long *msr_bitmap; + u64 kvm_xss; + bool cet_en; + + msr_bitmap = vmx->vmcs01.msr_bitmap; + kvm_xss = kvm_supported_xss(); + cet_en = guest_cpuid_has(vcpu, X86_FEATURE_SHSTK) || + guest_cpuid_has(vcpu, X86_FEATURE_IBT); + + /* + * U_CET is a must for USER CET, per CET spec., U_CET and PL3_SPP are + * a bundle for USER CET xsaves. + */ + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_USER)) { + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_U_CET, MSR_TYPE_RW); + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL3_SSP, MSR_TYPE_RW); + } + /* + * S_CET is a must for KERNEL CET, PL0_SSP ... PL2_SSP are a bundle + * for CET KERNEL xsaves. + */ + if (cet_en && (kvm_xss & XFEATURE_MASK_CET_KERNEL)) { + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_S_CET, MSR_TYPE_RW); + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL0_SSP, MSR_TYPE_RW); + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL1_SSP, MSR_TYPE_RW); + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_PL2_SSP, MSR_TYPE_RW); + + /* SSP_TAB only available for KERNEL SHSTK.*/ + if (guest_cpuid_has(vcpu, X86_FEATURE_SHSTK)) + vmx_disable_intercept_for_msr(msr_bitmap, MSR_IA32_INT_SSP_TAB, + MSR_TYPE_RW); + } +} + static void vmx_cpuid_update(struct kvm_vcpu *vcpu) { struct vcpu_vmx *vmx = to_vmx(vcpu); @@ -7025,6 +7062,8 @@ static void vmx_cpuid_update(struct kvm_vcpu *vcpu) if (boot_cpu_has(X86_FEATURE_INTEL_PT) && guest_cpuid_has(vcpu, X86_FEATURE_INTEL_PT)) update_intel_pt_cfg(vcpu); + + vmx_intercept_cet_msrs(vcpu); } static void vmx_set_supported_cpuid(u32 func, struct kvm_cpuid_entry2 *entry)