diff mbox series

[RFC,v5,083/104] KVM: x86: Split core of hypercall emulation to helper function

Message ID f3293bd872a916bf33165a2ec0d6fc50533b817f.1646422845.git.isaku.yamahata@intel.com (mailing list archive)
State New, archived
Headers show
Series KVM TDX basic feature support | expand

Commit Message

Isaku Yamahata March 4, 2022, 7:49 p.m. UTC
From: Sean Christopherson <sean.j.christopherson@intel.com>

By necessity, TDX will use a different register ABI for hypercalls.
Break out the core functionality so that it may be reused for TDX.

Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
Signed-off-by: Isaku Yamahata <isaku.yamahata@intel.com>
---
 arch/x86/include/asm/kvm_host.h |  4 +++
 arch/x86/kvm/x86.c              | 54 ++++++++++++++++++++-------------
 2 files changed, 37 insertions(+), 21 deletions(-)

Comments

Sagi Shahar March 21, 2022, 6:32 p.m. UTC | #1
On Fri, Mar 4, 2022 at 12:00 PM <isaku.yamahata@intel.com> wrote:
>
> From: Sean Christopherson <sean.j.christopherson@intel.com>
>
> By necessity, TDX will use a different register ABI for hypercalls.
> Break out the core functionality so that it may be reused for TDX.
>
> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
> Signed-off-by: Isaku Yamahata <isaku.yamahata@intel.com>
> ---
>  arch/x86/include/asm/kvm_host.h |  4 +++
>  arch/x86/kvm/x86.c              | 54 ++++++++++++++++++++-------------
>  2 files changed, 37 insertions(+), 21 deletions(-)
>
> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
> index 8dab9f16f559..33b75b0e3de1 100644
> --- a/arch/x86/include/asm/kvm_host.h
> +++ b/arch/x86/include/asm/kvm_host.h
> @@ -1818,6 +1818,10 @@ void kvm_request_apicv_update(struct kvm *kvm, bool activate,
>  void __kvm_request_apicv_update(struct kvm *kvm, bool activate,
>                                 unsigned long bit);
>
> +unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
> +                                     unsigned long a0, unsigned long a1,
> +                                     unsigned long a2, unsigned long a3,
> +                                     int op_64_bit);
>  int kvm_emulate_hypercall(struct kvm_vcpu *vcpu);
>
>  int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> index 314ae43e07bf..9acb33a17445 100644
> --- a/arch/x86/kvm/x86.c
> +++ b/arch/x86/kvm/x86.c
> @@ -9090,26 +9090,15 @@ static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
>         return kvm_skip_emulated_instruction(vcpu);
>  }
>
> -int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> +unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
> +                                     unsigned long a0, unsigned long a1,
> +                                     unsigned long a2, unsigned long a3,
> +                                     int op_64_bit)
>  {
> -       unsigned long nr, a0, a1, a2, a3, ret;
> -       int op_64_bit;
> -
> -       if (kvm_xen_hypercall_enabled(vcpu->kvm))
> -               return kvm_xen_hypercall(vcpu);
> -
> -       if (kvm_hv_hypercall_enabled(vcpu))
> -               return kvm_hv_hypercall(vcpu);
> -
> -       nr = kvm_rax_read(vcpu);
> -       a0 = kvm_rbx_read(vcpu);
> -       a1 = kvm_rcx_read(vcpu);
> -       a2 = kvm_rdx_read(vcpu);
> -       a3 = kvm_rsi_read(vcpu);
> +       unsigned long ret;
>
>         trace_kvm_hypercall(nr, a0, a1, a2, a3);
>
> -       op_64_bit = is_64_bit_hypercall(vcpu);
>         if (!op_64_bit) {
>                 nr &= 0xFFFFFFFF;
>                 a0 &= 0xFFFFFFFF;
> @@ -9118,11 +9107,6 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
>                 a3 &= 0xFFFFFFFF;
>         }
>
> -       if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
> -               ret = -KVM_EPERM;
> -               goto out;
> -       }
> -
>         ret = -KVM_ENOSYS;
>
>         switch (nr) {
> @@ -9181,6 +9165,34 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
>                 ret = -KVM_ENOSYS;
>                 break;
>         }
> +       return ret;
> +}
> +EXPORT_SYMBOL_GPL(__kvm_emulate_hypercall);
> +
> +int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> +{
> +       unsigned long nr, a0, a1, a2, a3, ret;
> +       int op_64_bit;
> +
> +       if (kvm_xen_hypercall_enabled(vcpu->kvm))
> +               return kvm_xen_hypercall(vcpu);
> +
> +       if (kvm_hv_hypercall_enabled(vcpu))
> +               return kvm_hv_hypercall(vcpu);
> +
> +       nr = kvm_rax_read(vcpu);
> +       a0 = kvm_rbx_read(vcpu);
> +       a1 = kvm_rcx_read(vcpu);
> +       a2 = kvm_rdx_read(vcpu);
> +       a3 = kvm_rsi_read(vcpu);
> +       op_64_bit = is_64_bit_mode(vcpu);

I think this should be "op_64_bit = is_64_bit_hypercall(vcpu);"
is_64_bit_mode was replaced with is_64_bit_hypercall to support
protected guests here:
https://lore.kernel.org/all/87cztf8h43.fsf@vitty.brq.redhat.com/T/

Without it, op_64_bit will be set to 0 for TD VMs which will cause the
upper 32 bit of the registers to be cleared in __kvm_emulate_hypercall

> +
> +       if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
> +               ret = -KVM_EPERM;
> +               goto out;
> +       }
> +
> +       ret = __kvm_emulate_hypercall(vcpu, nr, a0, a1, a2, a3, op_64_bit);
>  out:
>         if (!op_64_bit)
>                 ret = (u32)ret;
> --
> 2.25.1
>

Sagi
Isaku Yamahata March 23, 2022, 5:53 p.m. UTC | #2
On Mon, Mar 21, 2022 at 11:32:21AM -0700,
Sagi Shahar <sagis@google.com> wrote:

> > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> > index 314ae43e07bf..9acb33a17445 100644
> > --- a/arch/x86/kvm/x86.c
> > +++ b/arch/x86/kvm/x86.c
> > @@ -9090,26 +9090,15 @@ static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
> >         return kvm_skip_emulated_instruction(vcpu);
> >  }
> >
> > -int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> > +unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
> > +                                     unsigned long a0, unsigned long a1,
> > +                                     unsigned long a2, unsigned long a3,
> > +                                     int op_64_bit)
> >  {
> > -       unsigned long nr, a0, a1, a2, a3, ret;
> > -       int op_64_bit;
> > -
> > -       if (kvm_xen_hypercall_enabled(vcpu->kvm))
> > -               return kvm_xen_hypercall(vcpu);
> > -
> > -       if (kvm_hv_hypercall_enabled(vcpu))
> > -               return kvm_hv_hypercall(vcpu);
> > -
> > -       nr = kvm_rax_read(vcpu);
> > -       a0 = kvm_rbx_read(vcpu);
> > -       a1 = kvm_rcx_read(vcpu);
> > -       a2 = kvm_rdx_read(vcpu);
> > -       a3 = kvm_rsi_read(vcpu);
> > +       unsigned long ret;
> >
> >         trace_kvm_hypercall(nr, a0, a1, a2, a3);
> >
> > -       op_64_bit = is_64_bit_hypercall(vcpu);
> >         if (!op_64_bit) {
> >                 nr &= 0xFFFFFFFF;
> >                 a0 &= 0xFFFFFFFF;
> > @@ -9118,11 +9107,6 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> >                 a3 &= 0xFFFFFFFF;
> >         }
> >
> > -       if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
> > -               ret = -KVM_EPERM;
> > -               goto out;
> > -       }
> > -
> >         ret = -KVM_ENOSYS;
> >
> >         switch (nr) {
> > @@ -9181,6 +9165,34 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> >                 ret = -KVM_ENOSYS;
> >                 break;
> >         }
> > +       return ret;
> > +}
> > +EXPORT_SYMBOL_GPL(__kvm_emulate_hypercall);
> > +
> > +int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
> > +{
> > +       unsigned long nr, a0, a1, a2, a3, ret;
> > +       int op_64_bit;
> > +
> > +       if (kvm_xen_hypercall_enabled(vcpu->kvm))
> > +               return kvm_xen_hypercall(vcpu);
> > +
> > +       if (kvm_hv_hypercall_enabled(vcpu))
> > +               return kvm_hv_hypercall(vcpu);
> > +
> > +       nr = kvm_rax_read(vcpu);
> > +       a0 = kvm_rbx_read(vcpu);
> > +       a1 = kvm_rcx_read(vcpu);
> > +       a2 = kvm_rdx_read(vcpu);
> > +       a3 = kvm_rsi_read(vcpu);
> > +       op_64_bit = is_64_bit_mode(vcpu);
> 
> I think this should be "op_64_bit = is_64_bit_hypercall(vcpu);"
> is_64_bit_mode was replaced with is_64_bit_hypercall to support
> protected guests here:
> https://lore.kernel.org/all/87cztf8h43.fsf@vitty.brq.redhat.com/T/
> 
> Without it, op_64_bit will be set to 0 for TD VMs which will cause the
> upper 32 bit of the registers to be cleared in __kvm_emulate_hypercall

Oops, thanks for pointing it out.  I'll fix it up with next respin.
Paolo Bonzini April 7, 2022, 1:12 p.m. UTC | #3
On 3/21/22 19:32, Sagi Shahar wrote:
> On Fri, Mar 4, 2022 at 12:00 PM <isaku.yamahata@intel.com> wrote:
>>
>> From: Sean Christopherson <sean.j.christopherson@intel.com>
>>
>> By necessity, TDX will use a different register ABI for hypercalls.
>> Break out the core functionality so that it may be reused for TDX.
>>
>> Signed-off-by: Sean Christopherson <sean.j.christopherson@intel.com>
>> Signed-off-by: Isaku Yamahata <isaku.yamahata@intel.com>
>> ---
>>   arch/x86/include/asm/kvm_host.h |  4 +++
>>   arch/x86/kvm/x86.c              | 54 ++++++++++++++++++++-------------
>>   2 files changed, 37 insertions(+), 21 deletions(-)
>>
>> diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
>> index 8dab9f16f559..33b75b0e3de1 100644
>> --- a/arch/x86/include/asm/kvm_host.h
>> +++ b/arch/x86/include/asm/kvm_host.h
>> @@ -1818,6 +1818,10 @@ void kvm_request_apicv_update(struct kvm *kvm, bool activate,
>>   void __kvm_request_apicv_update(struct kvm *kvm, bool activate,
>>                                  unsigned long bit);
>>
>> +unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
>> +                                     unsigned long a0, unsigned long a1,
>> +                                     unsigned long a2, unsigned long a3,
>> +                                     int op_64_bit);
>>   int kvm_emulate_hypercall(struct kvm_vcpu *vcpu);
>>
>>   int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
>> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
>> index 314ae43e07bf..9acb33a17445 100644
>> --- a/arch/x86/kvm/x86.c
>> +++ b/arch/x86/kvm/x86.c
>> @@ -9090,26 +9090,15 @@ static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
>>          return kvm_skip_emulated_instruction(vcpu);
>>   }
>>
>> -int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
>> +unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
>> +                                     unsigned long a0, unsigned long a1,
>> +                                     unsigned long a2, unsigned long a3,
>> +                                     int op_64_bit)
>>   {
>> -       unsigned long nr, a0, a1, a2, a3, ret;
>> -       int op_64_bit;
>> -
>> -       if (kvm_xen_hypercall_enabled(vcpu->kvm))
>> -               return kvm_xen_hypercall(vcpu);
>> -
>> -       if (kvm_hv_hypercall_enabled(vcpu))
>> -               return kvm_hv_hypercall(vcpu);

Please keep Xen and Hyper-V hypercalls to kvm_emulate_hypercall (more on 
this in the reply to patch 89).  __kvm_emulate_hypercall should only 
handle KVM hypercalls.

>> +       if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
>> +               ret = -KVM_EPERM;
>> +               goto out;
>> +       }

Is this guaranteed by TDG.VP.VMCALL?

Paolo

>> +       ret = __kvm_emulate_hypercall(vcpu, nr, a0, a1, a2, a3, op_64_bit);
>>   out:
>>          if (!op_64_bit)
>>                  ret = (u32)ret;
>> --
>> 2.25.1
>>
> 
> Sagi
>
Isaku Yamahata April 8, 2022, 5:34 a.m. UTC | #4
On Thu, Apr 07, 2022 at 03:12:57PM +0200,
Paolo Bonzini <pbonzini@redhat.com> wrote:

> > > +       if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
> > > +               ret = -KVM_EPERM;
> > > +               goto out;
> > > +       }
> 
> Is this guaranteed by TDG.VP.VMCALL?

Yes. TDCALL instruction in TD results in #GP(0) if CPL > 0.
It's documented in trust domain CPU architectural extensions spec.
https://www.intel.com/content/dam/develop/external/us/en/documents-tps/intel-tdx-cpu-architectural-specification.pdf

Anyway VMM can't know TD guest CPL (or other CPU state).
diff mbox series

Patch

diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h
index 8dab9f16f559..33b75b0e3de1 100644
--- a/arch/x86/include/asm/kvm_host.h
+++ b/arch/x86/include/asm/kvm_host.h
@@ -1818,6 +1818,10 @@  void kvm_request_apicv_update(struct kvm *kvm, bool activate,
 void __kvm_request_apicv_update(struct kvm *kvm, bool activate,
 				unsigned long bit);
 
+unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
+				      unsigned long a0, unsigned long a1,
+				      unsigned long a2, unsigned long a3,
+				      int op_64_bit);
 int kvm_emulate_hypercall(struct kvm_vcpu *vcpu);
 
 int kvm_mmu_page_fault(struct kvm_vcpu *vcpu, gpa_t cr2_or_gpa, u64 error_code,
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 314ae43e07bf..9acb33a17445 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -9090,26 +9090,15 @@  static int complete_hypercall_exit(struct kvm_vcpu *vcpu)
 	return kvm_skip_emulated_instruction(vcpu);
 }
 
-int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
+unsigned long __kvm_emulate_hypercall(struct kvm_vcpu *vcpu, unsigned long nr,
+				      unsigned long a0, unsigned long a1,
+				      unsigned long a2, unsigned long a3,
+				      int op_64_bit)
 {
-	unsigned long nr, a0, a1, a2, a3, ret;
-	int op_64_bit;
-
-	if (kvm_xen_hypercall_enabled(vcpu->kvm))
-		return kvm_xen_hypercall(vcpu);
-
-	if (kvm_hv_hypercall_enabled(vcpu))
-		return kvm_hv_hypercall(vcpu);
-
-	nr = kvm_rax_read(vcpu);
-	a0 = kvm_rbx_read(vcpu);
-	a1 = kvm_rcx_read(vcpu);
-	a2 = kvm_rdx_read(vcpu);
-	a3 = kvm_rsi_read(vcpu);
+	unsigned long ret;
 
 	trace_kvm_hypercall(nr, a0, a1, a2, a3);
 
-	op_64_bit = is_64_bit_hypercall(vcpu);
 	if (!op_64_bit) {
 		nr &= 0xFFFFFFFF;
 		a0 &= 0xFFFFFFFF;
@@ -9118,11 +9107,6 @@  int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
 		a3 &= 0xFFFFFFFF;
 	}
 
-	if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
-		ret = -KVM_EPERM;
-		goto out;
-	}
-
 	ret = -KVM_ENOSYS;
 
 	switch (nr) {
@@ -9181,6 +9165,34 @@  int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
 		ret = -KVM_ENOSYS;
 		break;
 	}
+	return ret;
+}
+EXPORT_SYMBOL_GPL(__kvm_emulate_hypercall);
+
+int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
+{
+	unsigned long nr, a0, a1, a2, a3, ret;
+	int op_64_bit;
+
+	if (kvm_xen_hypercall_enabled(vcpu->kvm))
+		return kvm_xen_hypercall(vcpu);
+
+	if (kvm_hv_hypercall_enabled(vcpu))
+		return kvm_hv_hypercall(vcpu);
+
+	nr = kvm_rax_read(vcpu);
+	a0 = kvm_rbx_read(vcpu);
+	a1 = kvm_rcx_read(vcpu);
+	a2 = kvm_rdx_read(vcpu);
+	a3 = kvm_rsi_read(vcpu);
+	op_64_bit = is_64_bit_mode(vcpu);
+
+	if (static_call(kvm_x86_get_cpl)(vcpu) != 0) {
+		ret = -KVM_EPERM;
+		goto out;
+	}
+
+	ret = __kvm_emulate_hypercall(vcpu, nr, a0, a1, a2, a3, op_64_bit);
 out:
 	if (!op_64_bit)
 		ret = (u32)ret;