Message ID | 20200508032919.52147-6-gshan@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | kvm/arm64: Support Async Page Fault | expand |
On Fri, May 08, 2020 at 01:29:15PM +1000, Gavin Shan wrote: > This replace the variable names to make them self-explaining. The > tracepoint isn't changed accordingly because they're part of ABI: > > * @hsr to @esr > * @hsr_ec to @ec > * Use kvm_vcpu_trap_get_class() helper if possible > > Signed-off-by: Gavin Shan <gshan@redhat.com> As with patch 3, I think this cleanup makes sense independent from the rest of the series, and I think it'd make sense to bundle all the patches renaming hsr -> esr, and send those as a preparatory series. Thanks, Mark. > --- > arch/arm64/kvm/handle_exit.c | 28 ++++++++++++++-------------- > arch/arm64/kvm/hyp/switch.c | 9 ++++----- > arch/arm64/kvm/sys_regs.c | 30 +++++++++++++++--------------- > 3 files changed, 33 insertions(+), 34 deletions(-) > > diff --git a/arch/arm64/kvm/handle_exit.c b/arch/arm64/kvm/handle_exit.c > index 00858db82a64..e3b3dcd5b811 100644 > --- a/arch/arm64/kvm/handle_exit.c > +++ b/arch/arm64/kvm/handle_exit.c > @@ -123,13 +123,13 @@ static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) > */ > static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) > { > - u32 hsr = kvm_vcpu_get_esr(vcpu); > + u32 esr = kvm_vcpu_get_esr(vcpu); > int ret = 0; > > run->exit_reason = KVM_EXIT_DEBUG; > - run->debug.arch.hsr = hsr; > + run->debug.arch.hsr = esr; > > - switch (ESR_ELx_EC(hsr)) { > + switch (kvm_vcpu_trap_get_class(esr)) { > case ESR_ELx_EC_WATCHPT_LOW: > run->debug.arch.far = vcpu->arch.fault.far_el2; > /* fall through */ > @@ -139,8 +139,8 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) > case ESR_ELx_EC_BRK64: > break; > default: > - kvm_err("%s: un-handled case hsr: %#08x\n", > - __func__, (unsigned int) hsr); > + kvm_err("%s: un-handled case esr: %#08x\n", > + __func__, (unsigned int)esr); > ret = -1; > break; > } > @@ -150,10 +150,10 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) > > static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) > { > - u32 hsr = kvm_vcpu_get_esr(vcpu); > + u32 esr = kvm_vcpu_get_esr(vcpu); > > - kvm_pr_unimpl("Unknown exception class: hsr: %#08x -- %s\n", > - hsr, esr_get_class_string(hsr)); > + kvm_pr_unimpl("Unknown exception class: esr: %#08x -- %s\n", > + esr, esr_get_class_string(esr)); > > kvm_inject_undefined(vcpu); > return 1; > @@ -230,10 +230,10 @@ static exit_handle_fn arm_exit_handlers[] = { > > static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu) > { > - u32 hsr = kvm_vcpu_get_esr(vcpu); > - u8 hsr_ec = ESR_ELx_EC(hsr); > + u32 esr = kvm_vcpu_get_esr(vcpu); > + u8 ec = kvm_vcpu_trap_get_class(esr); > > - return arm_exit_handlers[hsr_ec]; > + return arm_exit_handlers[ec]; > } > > /* > @@ -273,15 +273,15 @@ int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, > { > if (ARM_SERROR_PENDING(exception_index)) { > u32 esr = kvm_vcpu_get_esr(vcpu); > - u8 hsr_ec = ESR_ELx_EC(esr); > + u8 ec = kvm_vcpu_trap_get_class(esr); > > /* > * HVC/SMC already have an adjusted PC, which we need > * to correct in order to return to after having > * injected the SError. > */ > - if (hsr_ec == ESR_ELx_EC_HVC32 || hsr_ec == ESR_ELx_EC_HVC64 || > - hsr_ec == ESR_ELx_EC_SMC32 || hsr_ec == ESR_ELx_EC_SMC64) { > + if (ec == ESR_ELx_EC_HVC32 || ec == ESR_ELx_EC_HVC64 || > + ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64) { > u32 adj = kvm_vcpu_trap_il_is32bit(esr) ? 4 : 2; > *vcpu_pc(vcpu) -= adj; > } > diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c > index 369f22f49f3d..7bf4840bf90e 100644 > --- a/arch/arm64/kvm/hyp/switch.c > +++ b/arch/arm64/kvm/hyp/switch.c > @@ -356,8 +356,8 @@ static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) > static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) > { > u32 esr = kvm_vcpu_get_esr(vcpu); > + u8 ec = kvm_vcpu_trap_get_class(esr); > bool vhe, sve_guest, sve_host; > - u8 hsr_ec; > > if (!system_supports_fpsimd()) > return false; > @@ -372,14 +372,13 @@ static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) > vhe = has_vhe(); > } > > - hsr_ec = kvm_vcpu_trap_get_class(esr); > - if (hsr_ec != ESR_ELx_EC_FP_ASIMD && > - hsr_ec != ESR_ELx_EC_SVE) > + if (ec != ESR_ELx_EC_FP_ASIMD && > + ec != ESR_ELx_EC_SVE) > return false; > > /* Don't handle SVE traps for non-SVE vcpus here: */ > if (!sve_guest) > - if (hsr_ec != ESR_ELx_EC_FP_ASIMD) > + if (ec != ESR_ELx_EC_FP_ASIMD) > return false; > > /* Valid trap. Switch the context: */ > diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c > index 012fff834a4b..58f81ab519af 100644 > --- a/arch/arm64/kvm/sys_regs.c > +++ b/arch/arm64/kvm/sys_regs.c > @@ -2182,10 +2182,10 @@ static void unhandled_cp_access(struct kvm_vcpu *vcpu, > struct sys_reg_params *params) > { > u32 esr = kvm_vcpu_get_esr(vcpu); > - u8 hsr_ec = kvm_vcpu_trap_get_class(esr); > + u8 ec = kvm_vcpu_trap_get_class(esr); > int cp = -1; > > - switch(hsr_ec) { > + switch (ec) { > case ESR_ELx_EC_CP15_32: > case ESR_ELx_EC_CP15_64: > cp = 15; > @@ -2216,17 +2216,17 @@ static int kvm_handle_cp_64(struct kvm_vcpu *vcpu, > size_t nr_specific) > { > struct sys_reg_params params; > - u32 hsr = kvm_vcpu_get_esr(vcpu); > - int Rt = kvm_vcpu_sys_get_rt(hsr); > - int Rt2 = (hsr >> 10) & 0x1f; > + u32 esr = kvm_vcpu_get_esr(vcpu); > + int Rt = kvm_vcpu_sys_get_rt(esr); > + int Rt2 = (esr >> 10) & 0x1f; > > params.is_aarch32 = true; > params.is_32bit = false; > - params.CRm = (hsr >> 1) & 0xf; > - params.is_write = ((hsr & 1) == 0); > + params.CRm = (esr >> 1) & 0xf; > + params.is_write = ((esr & 1) == 0); > > params.Op0 = 0; > - params.Op1 = (hsr >> 16) & 0xf; > + params.Op1 = (esr >> 16) & 0xf; > params.Op2 = 0; > params.CRn = 0; > > @@ -2273,18 +2273,18 @@ static int kvm_handle_cp_32(struct kvm_vcpu *vcpu, > size_t nr_specific) > { > struct sys_reg_params params; > - u32 hsr = kvm_vcpu_get_esr(vcpu); > - int Rt = kvm_vcpu_sys_get_rt(hsr); > + u32 esr = kvm_vcpu_get_esr(vcpu); > + int Rt = kvm_vcpu_sys_get_rt(esr); > > params.is_aarch32 = true; > params.is_32bit = true; > - params.CRm = (hsr >> 1) & 0xf; > + params.CRm = (esr >> 1) & 0xf; > params.regval = vcpu_get_reg(vcpu, Rt); > - params.is_write = ((hsr & 1) == 0); > - params.CRn = (hsr >> 10) & 0xf; > + params.is_write = ((esr & 1) == 0); > + params.CRn = (esr >> 10) & 0xf; > params.Op0 = 0; > - params.Op1 = (hsr >> 14) & 0x7; > - params.Op2 = (hsr >> 17) & 0x7; > + params.Op1 = (esr >> 14) & 0x7; > + params.Op2 = (esr >> 17) & 0x7; > > if (!emulate_cp(vcpu, ¶ms, target_specific, nr_specific) || > !emulate_cp(vcpu, ¶ms, global, nr_global)) { > -- > 2.23.0 >
Hi Mark, On 5/26/20 8:45 PM, Mark Rutland wrote: > On Fri, May 08, 2020 at 01:29:15PM +1000, Gavin Shan wrote: >> This replace the variable names to make them self-explaining. The >> tracepoint isn't changed accordingly because they're part of ABI: >> >> * @hsr to @esr >> * @hsr_ec to @ec >> * Use kvm_vcpu_trap_get_class() helper if possible >> >> Signed-off-by: Gavin Shan <gshan@redhat.com> > > As with patch 3, I think this cleanup makes sense independent from the > rest of the series, and I think it'd make sense to bundle all the > patches renaming hsr -> esr, and send those as a preparatory series. > Yes, PATCH[3/4/5] will be posted independently, as part of the preparatory work, as you suggested. Thanks, Gavin > Thanks, > Mark. > >> --- >> arch/arm64/kvm/handle_exit.c | 28 ++++++++++++++-------------- >> arch/arm64/kvm/hyp/switch.c | 9 ++++----- >> arch/arm64/kvm/sys_regs.c | 30 +++++++++++++++--------------- >> 3 files changed, 33 insertions(+), 34 deletions(-) >> >> diff --git a/arch/arm64/kvm/handle_exit.c b/arch/arm64/kvm/handle_exit.c >> index 00858db82a64..e3b3dcd5b811 100644 >> --- a/arch/arm64/kvm/handle_exit.c >> +++ b/arch/arm64/kvm/handle_exit.c >> @@ -123,13 +123,13 @@ static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) >> */ >> static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) >> { >> - u32 hsr = kvm_vcpu_get_esr(vcpu); >> + u32 esr = kvm_vcpu_get_esr(vcpu); >> int ret = 0; >> >> run->exit_reason = KVM_EXIT_DEBUG; >> - run->debug.arch.hsr = hsr; >> + run->debug.arch.hsr = esr; >> >> - switch (ESR_ELx_EC(hsr)) { >> + switch (kvm_vcpu_trap_get_class(esr)) { >> case ESR_ELx_EC_WATCHPT_LOW: >> run->debug.arch.far = vcpu->arch.fault.far_el2; >> /* fall through */ >> @@ -139,8 +139,8 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) >> case ESR_ELx_EC_BRK64: >> break; >> default: >> - kvm_err("%s: un-handled case hsr: %#08x\n", >> - __func__, (unsigned int) hsr); >> + kvm_err("%s: un-handled case esr: %#08x\n", >> + __func__, (unsigned int)esr); >> ret = -1; >> break; >> } >> @@ -150,10 +150,10 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) >> >> static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) >> { >> - u32 hsr = kvm_vcpu_get_esr(vcpu); >> + u32 esr = kvm_vcpu_get_esr(vcpu); >> >> - kvm_pr_unimpl("Unknown exception class: hsr: %#08x -- %s\n", >> - hsr, esr_get_class_string(hsr)); >> + kvm_pr_unimpl("Unknown exception class: esr: %#08x -- %s\n", >> + esr, esr_get_class_string(esr)); >> >> kvm_inject_undefined(vcpu); >> return 1; >> @@ -230,10 +230,10 @@ static exit_handle_fn arm_exit_handlers[] = { >> >> static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu) >> { >> - u32 hsr = kvm_vcpu_get_esr(vcpu); >> - u8 hsr_ec = ESR_ELx_EC(hsr); >> + u32 esr = kvm_vcpu_get_esr(vcpu); >> + u8 ec = kvm_vcpu_trap_get_class(esr); >> >> - return arm_exit_handlers[hsr_ec]; >> + return arm_exit_handlers[ec]; >> } >> >> /* >> @@ -273,15 +273,15 @@ int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, >> { >> if (ARM_SERROR_PENDING(exception_index)) { >> u32 esr = kvm_vcpu_get_esr(vcpu); >> - u8 hsr_ec = ESR_ELx_EC(esr); >> + u8 ec = kvm_vcpu_trap_get_class(esr); >> >> /* >> * HVC/SMC already have an adjusted PC, which we need >> * to correct in order to return to after having >> * injected the SError. >> */ >> - if (hsr_ec == ESR_ELx_EC_HVC32 || hsr_ec == ESR_ELx_EC_HVC64 || >> - hsr_ec == ESR_ELx_EC_SMC32 || hsr_ec == ESR_ELx_EC_SMC64) { >> + if (ec == ESR_ELx_EC_HVC32 || ec == ESR_ELx_EC_HVC64 || >> + ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64) { >> u32 adj = kvm_vcpu_trap_il_is32bit(esr) ? 4 : 2; >> *vcpu_pc(vcpu) -= adj; >> } >> diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c >> index 369f22f49f3d..7bf4840bf90e 100644 >> --- a/arch/arm64/kvm/hyp/switch.c >> +++ b/arch/arm64/kvm/hyp/switch.c >> @@ -356,8 +356,8 @@ static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) >> static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) >> { >> u32 esr = kvm_vcpu_get_esr(vcpu); >> + u8 ec = kvm_vcpu_trap_get_class(esr); >> bool vhe, sve_guest, sve_host; >> - u8 hsr_ec; >> >> if (!system_supports_fpsimd()) >> return false; >> @@ -372,14 +372,13 @@ static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) >> vhe = has_vhe(); >> } >> >> - hsr_ec = kvm_vcpu_trap_get_class(esr); >> - if (hsr_ec != ESR_ELx_EC_FP_ASIMD && >> - hsr_ec != ESR_ELx_EC_SVE) >> + if (ec != ESR_ELx_EC_FP_ASIMD && >> + ec != ESR_ELx_EC_SVE) >> return false; >> >> /* Don't handle SVE traps for non-SVE vcpus here: */ >> if (!sve_guest) >> - if (hsr_ec != ESR_ELx_EC_FP_ASIMD) >> + if (ec != ESR_ELx_EC_FP_ASIMD) >> return false; >> >> /* Valid trap. Switch the context: */ >> diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c >> index 012fff834a4b..58f81ab519af 100644 >> --- a/arch/arm64/kvm/sys_regs.c >> +++ b/arch/arm64/kvm/sys_regs.c >> @@ -2182,10 +2182,10 @@ static void unhandled_cp_access(struct kvm_vcpu *vcpu, >> struct sys_reg_params *params) >> { >> u32 esr = kvm_vcpu_get_esr(vcpu); >> - u8 hsr_ec = kvm_vcpu_trap_get_class(esr); >> + u8 ec = kvm_vcpu_trap_get_class(esr); >> int cp = -1; >> >> - switch(hsr_ec) { >> + switch (ec) { >> case ESR_ELx_EC_CP15_32: >> case ESR_ELx_EC_CP15_64: >> cp = 15; >> @@ -2216,17 +2216,17 @@ static int kvm_handle_cp_64(struct kvm_vcpu *vcpu, >> size_t nr_specific) >> { >> struct sys_reg_params params; >> - u32 hsr = kvm_vcpu_get_esr(vcpu); >> - int Rt = kvm_vcpu_sys_get_rt(hsr); >> - int Rt2 = (hsr >> 10) & 0x1f; >> + u32 esr = kvm_vcpu_get_esr(vcpu); >> + int Rt = kvm_vcpu_sys_get_rt(esr); >> + int Rt2 = (esr >> 10) & 0x1f; >> >> params.is_aarch32 = true; >> params.is_32bit = false; >> - params.CRm = (hsr >> 1) & 0xf; >> - params.is_write = ((hsr & 1) == 0); >> + params.CRm = (esr >> 1) & 0xf; >> + params.is_write = ((esr & 1) == 0); >> >> params.Op0 = 0; >> - params.Op1 = (hsr >> 16) & 0xf; >> + params.Op1 = (esr >> 16) & 0xf; >> params.Op2 = 0; >> params.CRn = 0; >> >> @@ -2273,18 +2273,18 @@ static int kvm_handle_cp_32(struct kvm_vcpu *vcpu, >> size_t nr_specific) >> { >> struct sys_reg_params params; >> - u32 hsr = kvm_vcpu_get_esr(vcpu); >> - int Rt = kvm_vcpu_sys_get_rt(hsr); >> + u32 esr = kvm_vcpu_get_esr(vcpu); >> + int Rt = kvm_vcpu_sys_get_rt(esr); >> >> params.is_aarch32 = true; >> params.is_32bit = true; >> - params.CRm = (hsr >> 1) & 0xf; >> + params.CRm = (esr >> 1) & 0xf; >> params.regval = vcpu_get_reg(vcpu, Rt); >> - params.is_write = ((hsr & 1) == 0); >> - params.CRn = (hsr >> 10) & 0xf; >> + params.is_write = ((esr & 1) == 0); >> + params.CRn = (esr >> 10) & 0xf; >> params.Op0 = 0; >> - params.Op1 = (hsr >> 14) & 0x7; >> - params.Op2 = (hsr >> 17) & 0x7; >> + params.Op1 = (esr >> 14) & 0x7; >> + params.Op2 = (esr >> 17) & 0x7; >> >> if (!emulate_cp(vcpu, ¶ms, target_specific, nr_specific) || >> !emulate_cp(vcpu, ¶ms, global, nr_global)) { >> -- >> 2.23.0 >> >
diff --git a/arch/arm64/kvm/handle_exit.c b/arch/arm64/kvm/handle_exit.c index 00858db82a64..e3b3dcd5b811 100644 --- a/arch/arm64/kvm/handle_exit.c +++ b/arch/arm64/kvm/handle_exit.c @@ -123,13 +123,13 @@ static int kvm_handle_wfx(struct kvm_vcpu *vcpu, struct kvm_run *run) */ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) { - u32 hsr = kvm_vcpu_get_esr(vcpu); + u32 esr = kvm_vcpu_get_esr(vcpu); int ret = 0; run->exit_reason = KVM_EXIT_DEBUG; - run->debug.arch.hsr = hsr; + run->debug.arch.hsr = esr; - switch (ESR_ELx_EC(hsr)) { + switch (kvm_vcpu_trap_get_class(esr)) { case ESR_ELx_EC_WATCHPT_LOW: run->debug.arch.far = vcpu->arch.fault.far_el2; /* fall through */ @@ -139,8 +139,8 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) case ESR_ELx_EC_BRK64: break; default: - kvm_err("%s: un-handled case hsr: %#08x\n", - __func__, (unsigned int) hsr); + kvm_err("%s: un-handled case esr: %#08x\n", + __func__, (unsigned int)esr); ret = -1; break; } @@ -150,10 +150,10 @@ static int kvm_handle_guest_debug(struct kvm_vcpu *vcpu, struct kvm_run *run) static int kvm_handle_unknown_ec(struct kvm_vcpu *vcpu, struct kvm_run *run) { - u32 hsr = kvm_vcpu_get_esr(vcpu); + u32 esr = kvm_vcpu_get_esr(vcpu); - kvm_pr_unimpl("Unknown exception class: hsr: %#08x -- %s\n", - hsr, esr_get_class_string(hsr)); + kvm_pr_unimpl("Unknown exception class: esr: %#08x -- %s\n", + esr, esr_get_class_string(esr)); kvm_inject_undefined(vcpu); return 1; @@ -230,10 +230,10 @@ static exit_handle_fn arm_exit_handlers[] = { static exit_handle_fn kvm_get_exit_handler(struct kvm_vcpu *vcpu) { - u32 hsr = kvm_vcpu_get_esr(vcpu); - u8 hsr_ec = ESR_ELx_EC(hsr); + u32 esr = kvm_vcpu_get_esr(vcpu); + u8 ec = kvm_vcpu_trap_get_class(esr); - return arm_exit_handlers[hsr_ec]; + return arm_exit_handlers[ec]; } /* @@ -273,15 +273,15 @@ int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, { if (ARM_SERROR_PENDING(exception_index)) { u32 esr = kvm_vcpu_get_esr(vcpu); - u8 hsr_ec = ESR_ELx_EC(esr); + u8 ec = kvm_vcpu_trap_get_class(esr); /* * HVC/SMC already have an adjusted PC, which we need * to correct in order to return to after having * injected the SError. */ - if (hsr_ec == ESR_ELx_EC_HVC32 || hsr_ec == ESR_ELx_EC_HVC64 || - hsr_ec == ESR_ELx_EC_SMC32 || hsr_ec == ESR_ELx_EC_SMC64) { + if (ec == ESR_ELx_EC_HVC32 || ec == ESR_ELx_EC_HVC64 || + ec == ESR_ELx_EC_SMC32 || ec == ESR_ELx_EC_SMC64) { u32 adj = kvm_vcpu_trap_il_is32bit(esr) ? 4 : 2; *vcpu_pc(vcpu) -= adj; } diff --git a/arch/arm64/kvm/hyp/switch.c b/arch/arm64/kvm/hyp/switch.c index 369f22f49f3d..7bf4840bf90e 100644 --- a/arch/arm64/kvm/hyp/switch.c +++ b/arch/arm64/kvm/hyp/switch.c @@ -356,8 +356,8 @@ static bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu) static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) { u32 esr = kvm_vcpu_get_esr(vcpu); + u8 ec = kvm_vcpu_trap_get_class(esr); bool vhe, sve_guest, sve_host; - u8 hsr_ec; if (!system_supports_fpsimd()) return false; @@ -372,14 +372,13 @@ static bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu) vhe = has_vhe(); } - hsr_ec = kvm_vcpu_trap_get_class(esr); - if (hsr_ec != ESR_ELx_EC_FP_ASIMD && - hsr_ec != ESR_ELx_EC_SVE) + if (ec != ESR_ELx_EC_FP_ASIMD && + ec != ESR_ELx_EC_SVE) return false; /* Don't handle SVE traps for non-SVE vcpus here: */ if (!sve_guest) - if (hsr_ec != ESR_ELx_EC_FP_ASIMD) + if (ec != ESR_ELx_EC_FP_ASIMD) return false; /* Valid trap. Switch the context: */ diff --git a/arch/arm64/kvm/sys_regs.c b/arch/arm64/kvm/sys_regs.c index 012fff834a4b..58f81ab519af 100644 --- a/arch/arm64/kvm/sys_regs.c +++ b/arch/arm64/kvm/sys_regs.c @@ -2182,10 +2182,10 @@ static void unhandled_cp_access(struct kvm_vcpu *vcpu, struct sys_reg_params *params) { u32 esr = kvm_vcpu_get_esr(vcpu); - u8 hsr_ec = kvm_vcpu_trap_get_class(esr); + u8 ec = kvm_vcpu_trap_get_class(esr); int cp = -1; - switch(hsr_ec) { + switch (ec) { case ESR_ELx_EC_CP15_32: case ESR_ELx_EC_CP15_64: cp = 15; @@ -2216,17 +2216,17 @@ static int kvm_handle_cp_64(struct kvm_vcpu *vcpu, size_t nr_specific) { struct sys_reg_params params; - u32 hsr = kvm_vcpu_get_esr(vcpu); - int Rt = kvm_vcpu_sys_get_rt(hsr); - int Rt2 = (hsr >> 10) & 0x1f; + u32 esr = kvm_vcpu_get_esr(vcpu); + int Rt = kvm_vcpu_sys_get_rt(esr); + int Rt2 = (esr >> 10) & 0x1f; params.is_aarch32 = true; params.is_32bit = false; - params.CRm = (hsr >> 1) & 0xf; - params.is_write = ((hsr & 1) == 0); + params.CRm = (esr >> 1) & 0xf; + params.is_write = ((esr & 1) == 0); params.Op0 = 0; - params.Op1 = (hsr >> 16) & 0xf; + params.Op1 = (esr >> 16) & 0xf; params.Op2 = 0; params.CRn = 0; @@ -2273,18 +2273,18 @@ static int kvm_handle_cp_32(struct kvm_vcpu *vcpu, size_t nr_specific) { struct sys_reg_params params; - u32 hsr = kvm_vcpu_get_esr(vcpu); - int Rt = kvm_vcpu_sys_get_rt(hsr); + u32 esr = kvm_vcpu_get_esr(vcpu); + int Rt = kvm_vcpu_sys_get_rt(esr); params.is_aarch32 = true; params.is_32bit = true; - params.CRm = (hsr >> 1) & 0xf; + params.CRm = (esr >> 1) & 0xf; params.regval = vcpu_get_reg(vcpu, Rt); - params.is_write = ((hsr & 1) == 0); - params.CRn = (hsr >> 10) & 0xf; + params.is_write = ((esr & 1) == 0); + params.CRn = (esr >> 10) & 0xf; params.Op0 = 0; - params.Op1 = (hsr >> 14) & 0x7; - params.Op2 = (hsr >> 17) & 0x7; + params.Op1 = (esr >> 14) & 0x7; + params.Op2 = (esr >> 17) & 0x7; if (!emulate_cp(vcpu, ¶ms, target_specific, nr_specific) || !emulate_cp(vcpu, ¶ms, global, nr_global)) {
This replace the variable names to make them self-explaining. The tracepoint isn't changed accordingly because they're part of ABI: * @hsr to @esr * @hsr_ec to @ec * Use kvm_vcpu_trap_get_class() helper if possible Signed-off-by: Gavin Shan <gshan@redhat.com> --- arch/arm64/kvm/handle_exit.c | 28 ++++++++++++++-------------- arch/arm64/kvm/hyp/switch.c | 9 ++++----- arch/arm64/kvm/sys_regs.c | 30 +++++++++++++++--------------- 3 files changed, 33 insertions(+), 34 deletions(-)