Message ID | a71c9dbd-ec32-8bb6-9d36-c138d9b4f001@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On 25/08/2017 18:53, Brijesh Singh wrote: >> > > Thanks for the tip, I have expanded the patch to cover tdp cases and > have verified > that it works fine with SME enabled KVM. If you are okay with this then > I can > send patch. > > diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c > index ccb70b8..7a8edc0 100644 > --- a/arch/x86/kvm/mmu.c > +++ b/arch/x86/kvm/mmu.c > @@ -4109,16 +4109,30 @@ void > reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu > *context) > { > bool uses_nx = context->nx || context->base_role.smep_andnot_wp; > + struct rsvd_bits_validate *shadow_zero_check; > + int i; > > /* > diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c > index ccb70b8..7a8edc0 100644 > --- a/arch/x86/kvm/mmu.c > +++ b/arch/x86/kvm/mmu.c > @@ -4109,16 +4109,30 @@ void > reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu > *context) > { > bool uses_nx = context->nx || context->base_role.smep_andnot_wp; > + struct rsvd_bits_validate *shadow_zero_check; > + int i; > > /* > * Passing "true" to the last argument is okay; it adds a check > * on bit 8 of the SPTEs which KVM doesn't use anyway. > */ > - __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, > + shadow_zero_check = &context->shadow_zero_check; > + __reset_rsvds_bits_mask(vcpu, shadow_zero_check, > boot_cpu_data.x86_phys_bits, > context->shadow_root_level, uses_nx, > guest_cpuid_has_gbpages(vcpu), > is_pse(vcpu), > true); > + > + if (!shadow_me_mask) > + return; > + > + for (i = context->shadow_root_level; --i >= 0;) { > + shadow_zero_check->rsvd_bits_mask[i][0] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][1] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][2] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][3] &= ~shadow_me_mask; Neither my version nor yours is correct. :) The right one has [0][i] and [1][i] (I inverted the indices by mistake). With that change, you can include my Acked-by: Paolo Bonzini <pbonzini@redhat.com> > + } > + > } > EXPORT_SYMBOL_GPL(reset_shadow_zero_bits_mask); > > @@ -4136,8 +4150,13 @@ static void > reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, > struct kvm_mmu *context) > { > + struct rsvd_bits_validate *shadow_zero_check; > + int i; > + > + shadow_zero_check = &context->shadow_zero_check; > + > if (boot_cpu_is_amd()) > - __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, > + __reset_rsvds_bits_mask(vcpu, shadow_zero_check, > boot_cpu_data.x86_phys_bits, > context->shadow_root_level, false, > boot_cpu_has(X86_FEATURE_GBPAGES), Please use shadow_zero_check here too: __reset_rsvds_bits_mask_ept(&context->shadow_zero_check, Thanks, Paolo > @@ -4147,6 +4166,15 @@ reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu > *vcpu, > boot_cpu_data.x86_phys_bits, > false); > > + if (!shadow_me_mask) > + return; > + > + for (i = context->shadow_root_level; --i >= 0;) { > + shadow_zero_check->rsvd_bits_mask[i][0] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][1] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][2] &= ~shadow_me_mask; > + shadow_zero_check->rsvd_bits_mask[i][3] &= ~shadow_me_mask; > + } > } > > /* > diff --git a/arch/x86/kvm/mmu.h b/arch/x86/kvm/mmu.h > index 3cc7255..d7d248a 100644 > --- a/arch/x86/kvm/mmu.h > +++ b/arch/x86/kvm/mmu.h > @@ -48,7 +48,7 @@ > > static inline u64 rsvd_bits(int s, int e) > { > - return __sme_clr(((1ULL << (e - s + 1)) - 1) << s); > + return ((1ULL << (e - s + 1)) - 1) << s; > } > > void kvm_mmu_set_mmio_spte_mask(u64 mmio_mask, u64 mmio_value); > >
On 08/25/2017 03:05 PM, Paolo Bonzini wrote: > On 25/08/2017 18:53, Brijesh Singh wrote: >>> > Neither my version nor yours is correct. :) The right one has [0][i] > and [1][i] (I inverted the indices by mistake). > > With that change, you can include my > > Acked-by: Paolo Bonzini <pbonzini@redhat.com> > Ingo, I assuming that this patch should be sent through the tip since SME support came from tip. I will be submitting the patch very soon. -Brijesh
On 25/08/2017 22:41, Brijesh Singh wrote: >>>> > >> Neither my version nor yours is correct. :) The right one has [0][i] >> and [1][i] (I inverted the indices by mistake). >> >> With that change, you can include my >> >> Acked-by: Paolo Bonzini <pbonzini@redhat.com> >> > > Ingo, > > I assuming that this patch should be sent through the tip since SME support > came from tip. I will be submitting the patch very soon. Yes, that is correct. I cannot apply it directly to the KVM tree. Paolo
* Paolo Bonzini <pbonzini@redhat.com> wrote: > On 25/08/2017 22:41, Brijesh Singh wrote: > >>>> > > > >> Neither my version nor yours is correct. :) The right one has [0][i] > >> and [1][i] (I inverted the indices by mistake). > >> > >> With that change, you can include my > >> > >> Acked-by: Paolo Bonzini <pbonzini@redhat.com> > >> > > > > Ingo, > > > > I assuming that this patch should be sent through the tip since SME support > > came from tip. I will be submitting the patch very soon. > > Yes, that is correct. I cannot apply it directly to the KVM tree. I've merged it to tip:x86/mm where the SME bits live and will propagate it to linux-next ASAP, once it's gone through my local testing. Thanks, Ingo
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index ccb70b8..7a8edc0 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c @@ -4109,16 +4109,30 @@ void reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) { bool uses_nx = context->nx || context->base_role.smep_andnot_wp; + struct rsvd_bits_validate *shadow_zero_check; + int i; /* diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index ccb70b8..7a8edc0 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c @@ -4109,16 +4109,30 @@ void reset_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) { bool uses_nx = context->nx || context->base_role.smep_andnot_wp; + struct rsvd_bits_validate *shadow_zero_check; + int i; /* * Passing "true" to the last argument is okay; it adds a check * on bit 8 of the SPTEs which KVM doesn't use anyway. */ - __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, + shadow_zero_check = &context->shadow_zero_check; + __reset_rsvds_bits_mask(vcpu, shadow_zero_check, boot_cpu_data.x86_phys_bits, context->shadow_root_level, uses_nx, guest_cpuid_has_gbpages(vcpu), is_pse(vcpu), true); + + if (!shadow_me_mask) + return; + + for (i = context->shadow_root_level; --i >= 0;) { + shadow_zero_check->rsvd_bits_mask[i][0] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][1] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][2] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][3] &= ~shadow_me_mask; + } + } EXPORT_SYMBOL_GPL(reset_shadow_zero_bits_mask); @@ -4136,8 +4150,13 @@ static void reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) { + struct rsvd_bits_validate *shadow_zero_check; + int i; + + shadow_zero_check = &context->shadow_zero_check; + if (boot_cpu_is_amd()) - __reset_rsvds_bits_mask(vcpu, &context->shadow_zero_check, + __reset_rsvds_bits_mask(vcpu, shadow_zero_check, boot_cpu_data.x86_phys_bits, context->shadow_root_level, false, boot_cpu_has(X86_FEATURE_GBPAGES), @@ -4147,6 +4166,15 @@ reset_tdp_shadow_zero_bits_mask(struct kvm_vcpu *vcpu, boot_cpu_data.x86_phys_bits, false); + if (!shadow_me_mask) + return; + + for (i = context->shadow_root_level; --i >= 0;) { + shadow_zero_check->rsvd_bits_mask[i][0] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][1] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][2] &= ~shadow_me_mask; + shadow_zero_check->rsvd_bits_mask[i][3] &= ~shadow_me_mask; + } } /* diff --git a/arch/x86/kvm/mmu.h b/arch/x86/kvm/mmu.h index 3cc7255..d7d248a 100644 --- a/arch/x86/kvm/mmu.h +++ b/arch/x86/kvm/mmu.h @@ -48,7 +48,7 @@ static inline u64 rsvd_bits(int s, int e) { - return __sme_clr(((1ULL << (e - s + 1)) - 1) << s); + return ((1ULL << (e - s + 1)) - 1) << s; } void kvm_mmu_set_mmio_spte_mask(u64 mmio_mask, u64 mmio_value);