@@ -197,11 +197,9 @@ struct vmx_inst_decoded {
#define VMX_INST_MEMREG_TYPE_REG 1
struct vmx_inst_op {
int type;
+ unsigned int bytes;
union {
- struct {
- unsigned long mem;
- unsigned int len;
- };
+ unsigned long mem;
unsigned int reg_idx;
};
} op[2];
@@ -464,6 +462,8 @@ static int decode_vmx_inst(struct cpu_user_regs *regs,
unsigned long base, index, seg_base, disp, offset;
int scale, size;
+ unsigned int bytes = vmx_guest_x86_mode(v);
+
if ( vmx_inst_check_privilege(regs, vmxon_check) != X86EMUL_OKAY )
return X86EMUL_EXCEPTION;
@@ -473,10 +473,11 @@ static int decode_vmx_inst(struct cpu_user_regs *regs,
if ( info.fields.memreg ) {
decode->op[0].type = VMX_INST_MEMREG_TYPE_REG;
decode->op[0].reg_idx = info.fields.reg1;
+ decode->op[0].bytes = bytes;
}
else
{
- bool mode_64bit = (vmx_guest_x86_mode(v) == 8);
+ bool mode_64bit = (bytes == 8);
decode->op[0].type = VMX_INST_MEMREG_TYPE_MEMORY;
@@ -508,11 +509,12 @@ static int decode_vmx_inst(struct cpu_user_regs *regs,
goto gp_fault;
decode->op[0].mem = base;
- decode->op[0].len = size;
+ decode->op[0].bytes = bytes;
}
decode->op[1].type = VMX_INST_MEMREG_TYPE_REG;
decode->op[1].reg_idx = info.fields.reg2;
+ decode->op[1].bytes = bytes;
return X86EMUL_OKAY;
@@ -1494,7 +1496,7 @@ int nvmx_handle_vmxon(struct cpu_user_regs *regs)
struct nestedvmx *nvmx = &vcpu_2_nvmx(v);
struct nestedvcpu *nvcpu = &vcpu_nestedhvm(v);
struct vmx_inst_decoded decode;
- unsigned long gpa = 0;
+ uint64_t gpa;
uint32_t nvmcs_revid;
int rc;
@@ -1502,7 +1504,7 @@ int nvmx_handle_vmxon(struct cpu_user_regs *regs)
if ( rc != X86EMUL_OKAY )
return rc;
- rc = operand_read(&gpa, &decode.op[0], regs, decode.op[0].len);
+ rc = operand_read(&gpa, &decode.op[0], regs, sizeof(gpa));
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1715,14 +1717,14 @@ int nvmx_handle_vmptrld(struct cpu_user_regs *regs)
struct vcpu *v = current;
struct vmx_inst_decoded decode;
struct nestedvcpu *nvcpu = &vcpu_nestedhvm(v);
- unsigned long gpa = 0;
+ uint64_t gpa;
int rc;
rc = decode_vmx_inst(regs, &decode, 0);
if ( rc != X86EMUL_OKAY )
return rc;
- rc = operand_read(&gpa, &decode.op[0], regs, decode.op[0].len);
+ rc = operand_read(&gpa, &decode.op[0], regs, sizeof(gpa));
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1801,7 +1803,7 @@ int nvmx_handle_vmptrst(struct cpu_user_regs *regs)
gpa = nvcpu->nv_vvmcxaddr;
rc = hvm_copy_to_guest_linear(decode.op[0].mem, &gpa,
- decode.op[0].len, 0, &pfinfo);
+ decode.op[0].bytes, 0, &pfinfo);
if ( rc == HVMTRANS_bad_linear_to_gfn )
hvm_inject_page_fault(pfinfo.ec, pfinfo.linear);
if ( rc != HVMTRANS_okay )
@@ -1817,7 +1819,7 @@ int nvmx_handle_vmclear(struct cpu_user_regs *regs)
struct vmx_inst_decoded decode;
struct nestedvcpu *nvcpu = &vcpu_nestedhvm(v);
struct nestedvmx *nvmx = &vcpu_2_nvmx(v);
- unsigned long gpa = 0;
+ uint64_t gpa;
void *vvmcs;
int rc;
@@ -1825,7 +1827,7 @@ int nvmx_handle_vmclear(struct cpu_user_regs *regs)
if ( rc != X86EMUL_OKAY )
return rc;
- rc = operand_read(&gpa, &decode.op[0], regs, decode.op[0].len);
+ rc = operand_read(&gpa, &decode.op[0], regs, sizeof(gpa));
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1886,7 +1888,7 @@ int nvmx_handle_vmread(struct cpu_user_regs *regs)
return X86EMUL_OKAY;
}
- rc = operand_read(&vmcs_encoding, &decode.op[1], regs, decode.op[1].len);
+ rc = operand_read(&vmcs_encoding, &decode.op[1], regs, decode.op[1].bytes);
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1900,7 +1902,7 @@ int nvmx_handle_vmread(struct cpu_user_regs *regs)
switch ( decode.op[0].type ) {
case VMX_INST_MEMREG_TYPE_MEMORY:
rc = hvm_copy_to_guest_linear(decode.op[0].mem, &value,
- decode.op[0].len, 0, &pfinfo);
+ decode.op[0].bytes, 0, &pfinfo);
if ( rc == HVMTRANS_bad_linear_to_gfn )
hvm_inject_page_fault(pfinfo.ec, pfinfo.linear);
if ( rc != HVMTRANS_okay )
@@ -1928,7 +1930,7 @@ int nvmx_handle_vmwrite(struct cpu_user_regs *regs)
if ( decode_vmx_inst(regs, &decode, 0) != X86EMUL_OKAY )
return X86EMUL_EXCEPTION;
- rc = operand_read(&operand, &decode.op[0], regs, decode.op[0].len);
+ rc = operand_read(&operand, &decode.op[0], regs, decode.op[0].bytes);
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1938,7 +1940,7 @@ int nvmx_handle_vmwrite(struct cpu_user_regs *regs)
return X86EMUL_OKAY;
}
- rc = operand_read(&vmcs_encoding, &decode.op[1], regs, decode.op[1].len);
+ rc = operand_read(&vmcs_encoding, &decode.op[1], regs, decode.op[1].bytes);
if ( rc != X86EMUL_OKAY )
return rc;
@@ -1973,13 +1975,13 @@ int nvmx_handle_vmwrite(struct cpu_user_regs *regs)
int nvmx_handle_invept(struct cpu_user_regs *regs)
{
struct vmx_inst_decoded decode;
- unsigned long invept_type = 0;
+ uint64_t invept_type;
int ret;
if ( (ret = decode_vmx_inst(regs, &decode, 0)) != X86EMUL_OKAY )
return ret;
- ret = operand_read(&invept_type, &decode.op[1], regs, decode.op[1].len);
+ ret = operand_read(&invept_type, &decode.op[1], regs, decode.op[1].bytes);
if ( ret != X86EMUL_OKAY )
return ret;
@@ -1987,9 +1989,9 @@ int nvmx_handle_invept(struct cpu_user_regs *regs)
{
case INVEPT_SINGLE_CONTEXT:
{
- unsigned long eptp;
+ uint64_t eptp;
- ret = operand_read(&eptp, &decode.op[0], regs, decode.op[0].len);
+ ret = operand_read(&eptp, &decode.op[0], regs, sizeof(eptp));
if ( ret )
return ret;
@@ -2011,13 +2013,13 @@ int nvmx_handle_invept(struct cpu_user_regs *regs)
int nvmx_handle_invvpid(struct cpu_user_regs *regs)
{
struct vmx_inst_decoded decode;
- unsigned long invvpid_type = 0;
+ uint64_t invvpid_type;
int ret;
if ( (ret = decode_vmx_inst(regs, &decode, 0)) != X86EMUL_OKAY )
return ret;
- ret = operand_read(&invvpid_type, &decode.op[1], regs, decode.op[1].len);
+ ret = operand_read(&invvpid_type, &decode.op[1], regs, decode.op[1].bytes);
if ( ret != X86EMUL_OKAY )
return ret;
The sizes of VMX operands are defined in the Intel SDM and have nothing to do with the addr_size field of struct vmx_inst_info: invept: r32/r64, m128 invvpid: r32/r64, m128 vmclear: m64 vmptrld: m64 vmptrst: m64 vmread: r32/64 or m32/64, r32/64 vmwrite: r32/r64, r32/64 or m32/64 vmon: m64 * Register operands are 32-bit or 64-bit depending on the guest mode. * Memory operands are almost always of fixed size, usually 64-bit, but for vmread and vmwrite their size depends on the guest mode. * invept has a 128-bit memory operand but the upper 64 bits are reserved and therefore need not be read. * invvpid has a 128-bit memory operand but we only require the VPID value which lies in the lower 64 bits. When reading variable-size operands, we pass the operand size calculated by decode_vmx_inst() and stored in strcut vmx_inst_op. When reading fixed-size operands, we pass the size of the variable into which the operand is to be read. Signed-off-by: Euan Harris <euan.harris@citrix.com> --- xen/arch/x86/hvm/vmx/vvmx.c | 48 +++++++++++++++++++++++---------------------- 1 file changed, 25 insertions(+), 23 deletions(-)