Message ID | 20240321130812.2983113-7-ruanjinjie@huawei.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | target/arm: Implement FEAT_NMI and FEAT_GICv3_NMI | expand |
On Thu, 21 Mar 2024 at 13:10, Jinjie Ruan <ruanjinjie@huawei.com> wrote: > > This only implements the external delivery method via the GICv3. > > Signed-off-by: Jinjie Ruan <ruanjinjie@huawei.com> > Reviewed-by: Richard Henderson <richard.henderson@linaro.org> > --- > v9: > - Update the GPIOs passed in the arm_cpu_kvm_set_irq, and update the comment. > - Definitely not merge VINMI and VFNMI into EXCP_VNMI. > - Update VINMI and VFNMI when writing HCR_EL2 or HCRX_EL2. > v8: > - Fix the rcu stall after sending a VNMI in qemu VM. > v7: > - Add Reviewed-by. > v6: > - env->cp15.hcr_el2 -> arm_hcr_el2_eff(). > - env->cp15.hcrx_el2 -> arm_hcrx_el2_eff(). > - Not include VF && VFNMI in CPU_INTERRUPT_VNMI. > v4: > - Accept NMI unconditionally for arm_cpu_has_work() but add comment. > - Change from & to && for EXCP_IRQ or EXCP_FIQ. > - Refator nmi mask in arm_excp_unmasked(). > - Also handle VNMI in arm_cpu_exec_interrupt() and arm_cpu_set_irq(). > - Rename virtual to Virtual. > v3: > - Not include CPU_INTERRUPT_NMI when FEAT_NMI not enabled > - Add ARM_CPU_VNMI. > - Refator nmi mask in arm_excp_unmasked(). > - Test SCTLR_ELx.NMI for ALLINT mask for NMI. > --- > target/arm/cpu-qom.h | 5 +- > target/arm/cpu.c | 124 ++++++++++++++++++++++++++++++++++++++--- > target/arm/cpu.h | 6 ++ > target/arm/helper.c | 33 +++++++++-- > target/arm/internals.h | 18 ++++++ > 5 files changed, 172 insertions(+), 14 deletions(-) > > diff --git a/target/arm/cpu-qom.h b/target/arm/cpu-qom.h > index 8e032691db..b497667d61 100644 > --- a/target/arm/cpu-qom.h > +++ b/target/arm/cpu-qom.h > @@ -36,11 +36,14 @@ DECLARE_CLASS_CHECKERS(AArch64CPUClass, AARCH64_CPU, > #define ARM_CPU_TYPE_SUFFIX "-" TYPE_ARM_CPU > #define ARM_CPU_TYPE_NAME(name) (name ARM_CPU_TYPE_SUFFIX) > > -/* Meanings of the ARMCPU object's four inbound GPIO lines */ > +/* Meanings of the ARMCPU object's seven inbound GPIO lines */ > #define ARM_CPU_IRQ 0 > #define ARM_CPU_FIQ 1 > #define ARM_CPU_VIRQ 2 > #define ARM_CPU_VFIQ 3 > +#define ARM_CPU_NMI 4 > +#define ARM_CPU_VINMI 5 > +#define ARM_CPU_VFNMI 6 > > /* For M profile, some registers are banked secure vs non-secure; > * these are represented as a 2-element array where the first element > diff --git a/target/arm/cpu.c b/target/arm/cpu.c > index ab8d007a86..f1e7ae0975 100644 > --- a/target/arm/cpu.c > +++ b/target/arm/cpu.c > @@ -122,6 +122,13 @@ void arm_restore_state_to_opc(CPUState *cs, > } > #endif /* CONFIG_TCG */ > > +/* > + * With SCTLR_ELx.NMI == 0, IRQ with Superpriority is masked identically with > + * IRQ without Superpriority. Moreover, if the GIC is configured so that > + * FEAT_GICv3_NMI is only set if FEAT_NMI is set, then we won't ever see > + * CPU_INTERRUPT_*NMI anyway. So we might as well accept NMI here > + * unconditionally. > + */ > static bool arm_cpu_has_work(CPUState *cs) > { > ARMCPU *cpu = ARM_CPU(cs); > @@ -129,6 +136,7 @@ static bool arm_cpu_has_work(CPUState *cs) > return (cpu->power_state != PSCI_OFF) > && cs->interrupt_request & > (CPU_INTERRUPT_FIQ | CPU_INTERRUPT_HARD > + | CPU_INTERRUPT_NMI | CPU_INTERRUPT_VINMI | CPU_INTERRUPT_VFNMI > | CPU_INTERRUPT_VFIQ | CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VSERR > | CPU_INTERRUPT_EXITTB); > } > @@ -668,6 +676,7 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, > CPUARMState *env = cpu_env(cs); > bool pstate_unmasked; > bool unmasked = false; > + bool allIntMask = false; > > /* > * Don't take exceptions if they target a lower EL. > @@ -678,13 +687,36 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, > return false; > } > > + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && > + env->cp15.sctlr_el[target_el] & SCTLR_NMI && cur_el == target_el) { > + allIntMask = env->pstate & PSTATE_ALLINT || > + ((env->cp15.sctlr_el[target_el] & SCTLR_SPINTMASK) && > + (env->pstate & PSTATE_SP)); > + } > + > switch (excp_idx) { > + case EXCP_NMI: > + pstate_unmasked = !allIntMask; > + break; > + > + case EXCP_VINMI: > + if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { > + /* VINMIs are only taken when hypervized. */ > + return false; > + } > + return !allIntMask; > + case EXCP_VFNMI: > + if (!(hcr_el2 & HCR_FMO) || (hcr_el2 & HCR_TGE)) { > + /* VFNMIs are only taken when hypervized. */ > + return false; > + } > + return !allIntMask; > case EXCP_FIQ: > - pstate_unmasked = !(env->daif & PSTATE_F); > + pstate_unmasked = (!(env->daif & PSTATE_F)) && (!allIntMask); > break; > > case EXCP_IRQ: > - pstate_unmasked = !(env->daif & PSTATE_I); > + pstate_unmasked = (!(env->daif & PSTATE_I)) && (!allIntMask); > break; > > case EXCP_VFIQ: > @@ -692,13 +724,13 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, > /* VFIQs are only taken when hypervized. */ > return false; > } > - return !(env->daif & PSTATE_F); > + return !(env->daif & PSTATE_F) && (!allIntMask); > case EXCP_VIRQ: > if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { > /* VIRQs are only taken when hypervized. */ > return false; > } > - return !(env->daif & PSTATE_I); > + return !(env->daif & PSTATE_I) && (!allIntMask); > case EXCP_VSERR: > if (!(hcr_el2 & HCR_AMO) || (hcr_el2 & HCR_TGE)) { > /* VIRQs are only taken when hypervized. */ > @@ -804,6 +836,32 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request) > > /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ > > + if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { > + if (interrupt_request & CPU_INTERRUPT_NMI) { > + excp_idx = EXCP_NMI; > + target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > + if (arm_excp_unmasked(cs, excp_idx, target_el, > + cur_el, secure, hcr_el2)) { > + goto found; > + } > + } > + if (interrupt_request & CPU_INTERRUPT_VINMI) { > + excp_idx = EXCP_VINMI; > + target_el = 1; > + if (arm_excp_unmasked(cs, excp_idx, target_el, > + cur_el, secure, hcr_el2)) { > + goto found; > + } > + } > + if (interrupt_request & CPU_INTERRUPT_VFNMI) { > + excp_idx = EXCP_VFNMI; > + target_el = 1; > + if (arm_excp_unmasked(cs, excp_idx, target_el, > + cur_el, secure, hcr_el2)) { > + goto found; > + } > + } > + } Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". At the moment nothing does that: * arm_cpu_update_vinmi() doesn't look at the NMI bit before deciding whether to set CPU_INTERRUPT_VINMI * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its default value of false and so arm_excp_unmasked() returns true, so VINMI is not masked * arm_cpu_exec_interrupt() doesn't look at the NMI bit before deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI if it's set up in the HCR_EL2 bits. However we do this the required behaviour is that if NMI is 0 then it is as if the interrupt doesn't have superpriority and it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. I think the best place to do this is probably here in arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and the NMI bit like IRQ. > if (interrupt_request & CPU_INTERRUPT_FIQ) { > excp_idx = EXCP_FIQ; > target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > @@ -900,6 +958,48 @@ void arm_cpu_update_vfiq(ARMCPU *cpu) > } > } > > +void arm_cpu_update_vinmi(ARMCPU *cpu) > +{ > + /* > + * Update the interrupt level for VINMI, which is the logical OR of > + * the HCRX_EL2.VINMI bit and the input line level from the GIC. > + */ > + CPUARMState *env = &cpu->env; > + CPUState *cs = CPU(cpu); > + > + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VI) && > + (arm_hcrx_el2_eff(env) & HCRX_VINMI)) || > + (env->irq_line_state & CPU_INTERRUPT_VINMI); > + > + if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VINMI) != 0)) { > + if (new_state) { > + cpu_interrupt(cs, CPU_INTERRUPT_VINMI); > + } else { > + cpu_reset_interrupt(cs, CPU_INTERRUPT_VINMI); > + } > + } > +} What semantics do we intend for the VINMI/VFNMI bits in interrupt_request and for the incoming IRQ, FIQ, NMI lines? The GIC spec suggests (but doesn't mandate) that NMI could be signalled by asserting both NMI and IRQ, and plain IRQ by asserting just IRQ (table 4-6 in the GIC spec). I think the GIC changes in this patchset assert only the NMI line for an IRQNMI, and not both NMI and IRQ. That's OK and I think makes more sense for QEMU than signalling both lines, but it's not the same as what we wind up doing with the handling of the HCR_EL2 bits in these functions, because you don't change the existing arm_cpu_update_virq() so that it only sets the CPU_INTERRUPT_VIRQ bit if this is a VIRQ and not a VIRQNMI. So if the guest sets HCR_EL2.VI and HCRX_EL2.VINMI then arm_cpu_update_virq() will say "this is a VIRQ" and also arm_cpu_update_vinmi() will say "This is a VINMI" and so both bits get set in the interrupt_request field. I think the fix for this is probably to have arm_cpu_update_virq() and arm_cpu_update_vfiq() check that this is not a VINMI/VFNMI, so we only set 1 bit in interrupt_request, not 2. thanks -- PMM
On Thu, 21 Mar 2024 at 15:46, Peter Maydell <peter.maydell@linaro.org> wrote: > Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then > we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". > At the moment nothing does that: > * arm_cpu_update_vinmi() doesn't look at the NMI bit before > deciding whether to set CPU_INTERRUPT_VINMI > * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its > default value of false and so arm_excp_unmasked() returns true, > so VINMI is not masked > * arm_cpu_exec_interrupt() doesn't look at the NMI bit before > deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request > > So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI > if it's set up in the HCR_EL2 bits. > > However we do this the required behaviour is that if NMI is 0 > then it is as if the interrupt doesn't have superpriority and > it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. > I think the best place to do this is probably here in > arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then > treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and > the NMI bit like IRQ. Folding in something like this I think will work: diff --git a/target/arm/cpu.c b/target/arm/cpu.c index 91c2896de0f..797ae3eb805 100644 --- a/target/arm/cpu.c +++ b/target/arm/cpu.c @@ -837,7 +837,8 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request) /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ - if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && + (arm_sctlr(env, cur_el) & SCTLR_NMI)) { if (interrupt_request & CPU_INTERRUPT_NMI) { excp_idx = EXCP_NMI; target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); @@ -862,7 +863,22 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request) goto found; } } + } else { + /* + * NMI disabled: interrupts with superpriority are handled + * as if they didn't have it + */ + if (interrupt_request & CPU_INTERRUPT_NMI) { + interrupt_request |= CPU_INTERRUPT_HARD; + } + if (interrupt_request & CPU_INTERRUPT_VINMI) { + interrupt_request |= CPU_INTERRUPT_VIRQ; + } + if (interrupt_request & CPU_INTERRUPT_VFNMI) { + interrupt_request |= CPU_INTERRUPT_VFIQ; + } } + if (interrupt_request & CPU_INTERRUPT_FIQ) { excp_idx = EXCP_FIQ; target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > What semantics do we intend for the VINMI/VFNMI bits in interrupt_request > and for the incoming IRQ, FIQ, NMI lines? The GIC spec suggests > (but doesn't mandate) that NMI could be signalled by asserting > both NMI and IRQ, and plain IRQ by asserting just IRQ (table 4-6 > in the GIC spec). I think the GIC changes in this patchset assert > only the NMI line for an IRQNMI, and not both NMI and IRQ. That's OK > and I think makes more sense for QEMU than signalling both lines, > but it's not the same as what we wind up doing with the handling > of the HCR_EL2 bits in these functions, because you don't change > the existing arm_cpu_update_virq() so that it only sets the > CPU_INTERRUPT_VIRQ bit if this is a VIRQ and not a VIRQNMI. > So if the guest sets HCR_EL2.VI and HCRX_EL2.VINMI then > arm_cpu_update_virq() will say "this is a VIRQ" and also > arm_cpu_update_vinmi() will say "This is a VINMI" and so both bits > get set in the interrupt_request field. > > I think the fix for this is probably to have arm_cpu_update_virq() > and arm_cpu_update_vfiq() check that this is not a VINMI/VFNMI, > so we only set 1 bit in interrupt_request, not 2. And for this a change like: diff --git a/target/arm/cpu.c b/target/arm/cpu.c index 03a48a41366..91c2896de0f 100644 --- a/target/arm/cpu.c +++ b/target/arm/cpu.c @@ -926,7 +926,8 @@ void arm_cpu_update_virq(ARMCPU *cpu) CPUARMState *env = &cpu->env; CPUState *cs = CPU(cpu); - bool new_state = (env->cp15.hcr_el2 & HCR_VI) || + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VI) && + !(arm_hcrx_el2_eff(env) & HCRX_VINMI)) || (env->irq_line_state & CPU_INTERRUPT_VIRQ); if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VIRQ) != 0)) { @@ -947,7 +948,8 @@ void arm_cpu_update_vfiq(ARMCPU *cpu) CPUARMState *env = &cpu->env; CPUState *cs = CPU(cpu); - bool new_state = (env->cp15.hcr_el2 & HCR_VF) || + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VF) && + !(arm_hcrx_el2_eff(env) & HCRX_VFNMI)) || (env->irq_line_state & CPU_INTERRUPT_VFIQ); if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VFIQ) != 0)) { thanks -- PMM
On 2024/3/21 23:46, Peter Maydell wrote: > On Thu, 21 Mar 2024 at 13:10, Jinjie Ruan <ruanjinjie@huawei.com> wrote: >> >> This only implements the external delivery method via the GICv3. >> >> Signed-off-by: Jinjie Ruan <ruanjinjie@huawei.com> >> Reviewed-by: Richard Henderson <richard.henderson@linaro.org> >> --- >> v9: >> - Update the GPIOs passed in the arm_cpu_kvm_set_irq, and update the comment. >> - Definitely not merge VINMI and VFNMI into EXCP_VNMI. >> - Update VINMI and VFNMI when writing HCR_EL2 or HCRX_EL2. >> v8: >> - Fix the rcu stall after sending a VNMI in qemu VM. >> v7: >> - Add Reviewed-by. >> v6: >> - env->cp15.hcr_el2 -> arm_hcr_el2_eff(). >> - env->cp15.hcrx_el2 -> arm_hcrx_el2_eff(). >> - Not include VF && VFNMI in CPU_INTERRUPT_VNMI. >> v4: >> - Accept NMI unconditionally for arm_cpu_has_work() but add comment. >> - Change from & to && for EXCP_IRQ or EXCP_FIQ. >> - Refator nmi mask in arm_excp_unmasked(). >> - Also handle VNMI in arm_cpu_exec_interrupt() and arm_cpu_set_irq(). >> - Rename virtual to Virtual. >> v3: >> - Not include CPU_INTERRUPT_NMI when FEAT_NMI not enabled >> - Add ARM_CPU_VNMI. >> - Refator nmi mask in arm_excp_unmasked(). >> - Test SCTLR_ELx.NMI for ALLINT mask for NMI. >> --- >> target/arm/cpu-qom.h | 5 +- >> target/arm/cpu.c | 124 ++++++++++++++++++++++++++++++++++++++--- >> target/arm/cpu.h | 6 ++ >> target/arm/helper.c | 33 +++++++++-- >> target/arm/internals.h | 18 ++++++ >> 5 files changed, 172 insertions(+), 14 deletions(-) >> >> diff --git a/target/arm/cpu-qom.h b/target/arm/cpu-qom.h >> index 8e032691db..b497667d61 100644 >> --- a/target/arm/cpu-qom.h >> +++ b/target/arm/cpu-qom.h >> @@ -36,11 +36,14 @@ DECLARE_CLASS_CHECKERS(AArch64CPUClass, AARCH64_CPU, >> #define ARM_CPU_TYPE_SUFFIX "-" TYPE_ARM_CPU >> #define ARM_CPU_TYPE_NAME(name) (name ARM_CPU_TYPE_SUFFIX) >> >> -/* Meanings of the ARMCPU object's four inbound GPIO lines */ >> +/* Meanings of the ARMCPU object's seven inbound GPIO lines */ >> #define ARM_CPU_IRQ 0 >> #define ARM_CPU_FIQ 1 >> #define ARM_CPU_VIRQ 2 >> #define ARM_CPU_VFIQ 3 >> +#define ARM_CPU_NMI 4 >> +#define ARM_CPU_VINMI 5 >> +#define ARM_CPU_VFNMI 6 >> >> /* For M profile, some registers are banked secure vs non-secure; >> * these are represented as a 2-element array where the first element >> diff --git a/target/arm/cpu.c b/target/arm/cpu.c >> index ab8d007a86..f1e7ae0975 100644 >> --- a/target/arm/cpu.c >> +++ b/target/arm/cpu.c >> @@ -122,6 +122,13 @@ void arm_restore_state_to_opc(CPUState *cs, >> } >> #endif /* CONFIG_TCG */ >> >> +/* >> + * With SCTLR_ELx.NMI == 0, IRQ with Superpriority is masked identically with >> + * IRQ without Superpriority. Moreover, if the GIC is configured so that >> + * FEAT_GICv3_NMI is only set if FEAT_NMI is set, then we won't ever see >> + * CPU_INTERRUPT_*NMI anyway. So we might as well accept NMI here >> + * unconditionally. >> + */ >> static bool arm_cpu_has_work(CPUState *cs) >> { >> ARMCPU *cpu = ARM_CPU(cs); >> @@ -129,6 +136,7 @@ static bool arm_cpu_has_work(CPUState *cs) >> return (cpu->power_state != PSCI_OFF) >> && cs->interrupt_request & >> (CPU_INTERRUPT_FIQ | CPU_INTERRUPT_HARD >> + | CPU_INTERRUPT_NMI | CPU_INTERRUPT_VINMI | CPU_INTERRUPT_VFNMI >> | CPU_INTERRUPT_VFIQ | CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VSERR >> | CPU_INTERRUPT_EXITTB); >> } >> @@ -668,6 +676,7 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, >> CPUARMState *env = cpu_env(cs); >> bool pstate_unmasked; >> bool unmasked = false; >> + bool allIntMask = false; >> >> /* >> * Don't take exceptions if they target a lower EL. >> @@ -678,13 +687,36 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, >> return false; >> } >> >> + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && >> + env->cp15.sctlr_el[target_el] & SCTLR_NMI && cur_el == target_el) { >> + allIntMask = env->pstate & PSTATE_ALLINT || >> + ((env->cp15.sctlr_el[target_el] & SCTLR_SPINTMASK) && >> + (env->pstate & PSTATE_SP)); >> + } >> + >> switch (excp_idx) { >> + case EXCP_NMI: >> + pstate_unmasked = !allIntMask; >> + break; >> + >> + case EXCP_VINMI: >> + if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { >> + /* VINMIs are only taken when hypervized. */ >> + return false; >> + } >> + return !allIntMask; >> + case EXCP_VFNMI: >> + if (!(hcr_el2 & HCR_FMO) || (hcr_el2 & HCR_TGE)) { >> + /* VFNMIs are only taken when hypervized. */ >> + return false; >> + } >> + return !allIntMask; >> case EXCP_FIQ: >> - pstate_unmasked = !(env->daif & PSTATE_F); >> + pstate_unmasked = (!(env->daif & PSTATE_F)) && (!allIntMask); >> break; >> >> case EXCP_IRQ: >> - pstate_unmasked = !(env->daif & PSTATE_I); >> + pstate_unmasked = (!(env->daif & PSTATE_I)) && (!allIntMask); >> break; >> >> case EXCP_VFIQ: >> @@ -692,13 +724,13 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, >> /* VFIQs are only taken when hypervized. */ >> return false; >> } >> - return !(env->daif & PSTATE_F); >> + return !(env->daif & PSTATE_F) && (!allIntMask); >> case EXCP_VIRQ: >> if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { >> /* VIRQs are only taken when hypervized. */ >> return false; >> } >> - return !(env->daif & PSTATE_I); >> + return !(env->daif & PSTATE_I) && (!allIntMask); >> case EXCP_VSERR: >> if (!(hcr_el2 & HCR_AMO) || (hcr_el2 & HCR_TGE)) { >> /* VIRQs are only taken when hypervized. */ >> @@ -804,6 +836,32 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request) >> >> /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ >> >> + if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { >> + if (interrupt_request & CPU_INTERRUPT_NMI) { >> + excp_idx = EXCP_NMI; >> + target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); >> + if (arm_excp_unmasked(cs, excp_idx, target_el, >> + cur_el, secure, hcr_el2)) { >> + goto found; >> + } >> + } >> + if (interrupt_request & CPU_INTERRUPT_VINMI) { >> + excp_idx = EXCP_VINMI; >> + target_el = 1; >> + if (arm_excp_unmasked(cs, excp_idx, target_el, >> + cur_el, secure, hcr_el2)) { >> + goto found; >> + } >> + } >> + if (interrupt_request & CPU_INTERRUPT_VFNMI) { >> + excp_idx = EXCP_VFNMI; >> + target_el = 1; >> + if (arm_excp_unmasked(cs, excp_idx, target_el, >> + cur_el, secure, hcr_el2)) { >> + goto found; >> + } >> + } >> + } > > Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then > we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". > At the moment nothing does that: > * arm_cpu_update_vinmi() doesn't look at the NMI bit before > deciding whether to set CPU_INTERRUPT_VINMI > * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its > default value of false and so arm_excp_unmasked() returns true, > so VINMI is not masked > * arm_cpu_exec_interrupt() doesn't look at the NMI bit before > deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request > > So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI > if it's set up in the HCR_EL2 bits. > > However we do this the required behaviour is that if NMI is 0 > then it is as if the interrupt doesn't have superpriority and > it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. > I think the best place to do this is probably here in > arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then > treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and > the NMI bit like IRQ. A PE that implements FEAT_NMI and FEAT_GICv3 also implements FEAT_GICv3_NMI. A PE that does not implement FEAT_NMI, does not implement FEAT_GICv3_NMI. As the GIC side has checked the FEAT_GICv3_NMI is implemented or Not. So if cpu_isar_feature(aa64_nmi, env_archcpu(env)) is false, the FEAT_GICv3_NMI will also not implemented,the CPU_INTERRUPT_NMI/VINMI can not come from GIC, so we only need to check cpu_isar_feature(aa64_nmi, env_archcpu(env)) and SCTLR_ELx.NMI is set in hcr_write() and hcrx_write() for CPU side. > >> if (interrupt_request & CPU_INTERRUPT_FIQ) { >> excp_idx = EXCP_FIQ; >> target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); >> @@ -900,6 +958,48 @@ void arm_cpu_update_vfiq(ARMCPU *cpu) >> } >> } >> >> +void arm_cpu_update_vinmi(ARMCPU *cpu) >> +{ >> + /* >> + * Update the interrupt level for VINMI, which is the logical OR of >> + * the HCRX_EL2.VINMI bit and the input line level from the GIC. >> + */ >> + CPUARMState *env = &cpu->env; >> + CPUState *cs = CPU(cpu); >> + >> + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VI) && >> + (arm_hcrx_el2_eff(env) & HCRX_VINMI)) || >> + (env->irq_line_state & CPU_INTERRUPT_VINMI); >> + >> + if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VINMI) != 0)) { >> + if (new_state) { >> + cpu_interrupt(cs, CPU_INTERRUPT_VINMI); >> + } else { >> + cpu_reset_interrupt(cs, CPU_INTERRUPT_VINMI); >> + } >> + } >> +} > > What semantics do we intend for the VINMI/VFNMI bits in interrupt_request > and for the incoming IRQ, FIQ, NMI lines? The GIC spec suggests > (but doesn't mandate) that NMI could be signalled by asserting > both NMI and IRQ, and plain IRQ by asserting just IRQ (table 4-6 > in the GIC spec). I think the GIC changes in this patchset assert > only the NMI line for an IRQNMI, and not both NMI and IRQ. That's OK > and I think makes more sense for QEMU than signalling both lines, > but it's not the same as what we wind up doing with the handling > of the HCR_EL2 bits in these functions, because you don't change > the existing arm_cpu_update_virq() so that it only sets the > CPU_INTERRUPT_VIRQ bit if this is a VIRQ and not a VIRQNMI. > So if the guest sets HCR_EL2.VI and HCRX_EL2.VINMI then > arm_cpu_update_virq() will say "this is a VIRQ" and also > arm_cpu_update_vinmi() will say "This is a VINMI" and so both bits > get set in the interrupt_request field. > > I think the fix for this is probably to have arm_cpu_update_virq() > and arm_cpu_update_vfiq() check that this is not a VINMI/VFNMI, > so we only set 1 bit in interrupt_request, not 2. > > thanks > -- PMM
On 2024/3/22 2:28, Peter Maydell wrote: > On Thu, 21 Mar 2024 at 15:46, Peter Maydell <peter.maydell@linaro.org> wrote: >> Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then >> we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". >> At the moment nothing does that: >> * arm_cpu_update_vinmi() doesn't look at the NMI bit before >> deciding whether to set CPU_INTERRUPT_VINMI >> * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its >> default value of false and so arm_excp_unmasked() returns true, >> so VINMI is not masked >> * arm_cpu_exec_interrupt() doesn't look at the NMI bit before >> deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request >> >> So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI >> if it's set up in the HCR_EL2 bits. >> >> However we do this the required behaviour is that if NMI is 0 >> then it is as if the interrupt doesn't have superpriority and >> it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. >> I think the best place to do this is probably here in >> arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then >> treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and >> the NMI bit like IRQ. > > Folding in something like this I think will work: > > diff --git a/target/arm/cpu.c b/target/arm/cpu.c > index 91c2896de0f..797ae3eb805 100644 > --- a/target/arm/cpu.c > +++ b/target/arm/cpu.c > @@ -837,7 +837,8 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, > int interrupt_request) > > /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ > > - if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { > + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && > + (arm_sctlr(env, cur_el) & SCTLR_NMI)) { > if (interrupt_request & CPU_INTERRUPT_NMI) { > excp_idx = EXCP_NMI; > target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > @@ -862,7 +863,22 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, > int interrupt_request) > goto found; > } > } > + } else { > + /* > + * NMI disabled: interrupts with superpriority are handled > + * as if they didn't have it > + */ > + if (interrupt_request & CPU_INTERRUPT_NMI) { > + interrupt_request |= CPU_INTERRUPT_HARD; The CPU_INTERRUPT_NMI and CPU_INTERRUPT_HARD are set simultaneously, should the CPU_INTERRUPT_NMI be cleared? > + } > + if (interrupt_request & CPU_INTERRUPT_VINMI) { > + interrupt_request |= CPU_INTERRUPT_VIRQ; > + } > + if (interrupt_request & CPU_INTERRUPT_VFNMI) { > + interrupt_request |= CPU_INTERRUPT_VFIQ; > + } > }> + > if (interrupt_request & CPU_INTERRUPT_FIQ) { > excp_idx = EXCP_FIQ; > target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > > >> What semantics do we intend for the VINMI/VFNMI bits in interrupt_request >> and for the incoming IRQ, FIQ, NMI lines? The GIC spec suggests >> (but doesn't mandate) that NMI could be signalled by asserting >> both NMI and IRQ, and plain IRQ by asserting just IRQ (table 4-6 >> in the GIC spec). I think the GIC changes in this patchset assert >> only the NMI line for an IRQNMI, and not both NMI and IRQ. That's OK >> and I think makes more sense for QEMU than signalling both lines, >> but it's not the same as what we wind up doing with the handling >> of the HCR_EL2 bits in these functions, because you don't change >> the existing arm_cpu_update_virq() so that it only sets the >> CPU_INTERRUPT_VIRQ bit if this is a VIRQ and not a VIRQNMI. >> So if the guest sets HCR_EL2.VI and HCRX_EL2.VINMI then >> arm_cpu_update_virq() will say "this is a VIRQ" and also >> arm_cpu_update_vinmi() will say "This is a VINMI" and so both bits >> get set in the interrupt_request field. >> >> I think the fix for this is probably to have arm_cpu_update_virq() >> and arm_cpu_update_vfiq() check that this is not a VINMI/VFNMI, >> so we only set 1 bit in interrupt_request, not 2. > > And for this a change like: > > diff --git a/target/arm/cpu.c b/target/arm/cpu.c > index 03a48a41366..91c2896de0f 100644 > --- a/target/arm/cpu.c > +++ b/target/arm/cpu.c > @@ -926,7 +926,8 @@ void arm_cpu_update_virq(ARMCPU *cpu) > CPUARMState *env = &cpu->env; > CPUState *cs = CPU(cpu); > > - bool new_state = (env->cp15.hcr_el2 & HCR_VI) || > + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VI) && > + !(arm_hcrx_el2_eff(env) & HCRX_VINMI)) || > (env->irq_line_state & CPU_INTERRUPT_VIRQ); > > if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VIRQ) != 0)) { > @@ -947,7 +948,8 @@ void arm_cpu_update_vfiq(ARMCPU *cpu) > CPUARMState *env = &cpu->env; > CPUState *cs = CPU(cpu); > > - bool new_state = (env->cp15.hcr_el2 & HCR_VF) || > + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VF) && > + !(arm_hcrx_el2_eff(env) & HCRX_VFNMI)) || > (env->irq_line_state & CPU_INTERRUPT_VFIQ); > > if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VFIQ) != 0)) { > > > thanks > -- PMM
On Fri, 22 Mar 2024 at 05:05, Jinjie Ruan <ruanjinjie@huawei.com> wrote: > > > > On 2024/3/22 2:28, Peter Maydell wrote: > > On Thu, 21 Mar 2024 at 15:46, Peter Maydell <peter.maydell@linaro.org> wrote: > >> Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then > >> we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". > >> At the moment nothing does that: > >> * arm_cpu_update_vinmi() doesn't look at the NMI bit before > >> deciding whether to set CPU_INTERRUPT_VINMI > >> * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its > >> default value of false and so arm_excp_unmasked() returns true, > >> so VINMI is not masked > >> * arm_cpu_exec_interrupt() doesn't look at the NMI bit before > >> deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request > >> > >> So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI > >> if it's set up in the HCR_EL2 bits. > >> > >> However we do this the required behaviour is that if NMI is 0 > >> then it is as if the interrupt doesn't have superpriority and > >> it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. > >> I think the best place to do this is probably here in > >> arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then > >> treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and > >> the NMI bit like IRQ. > > > > Folding in something like this I think will work: > > > > diff --git a/target/arm/cpu.c b/target/arm/cpu.c > > index 91c2896de0f..797ae3eb805 100644 > > --- a/target/arm/cpu.c > > +++ b/target/arm/cpu.c > > @@ -837,7 +837,8 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, > > int interrupt_request) > > > > /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ > > > > - if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { > > + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && > > + (arm_sctlr(env, cur_el) & SCTLR_NMI)) { > > if (interrupt_request & CPU_INTERRUPT_NMI) { > > excp_idx = EXCP_NMI; > > target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); > > @@ -862,7 +863,22 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, > > int interrupt_request) > > goto found; > > } > > } > > + } else { > > + /* > > + * NMI disabled: interrupts with superpriority are handled > > + * as if they didn't have it > > + */ > > + if (interrupt_request & CPU_INTERRUPT_NMI) { > > + interrupt_request |= CPU_INTERRUPT_HARD; > > The CPU_INTERRUPT_NMI and CPU_INTERRUPT_HARD are set simultaneously, > should the CPU_INTERRUPT_NMI be cleared? interrupt_request here is a local variable, and we will never check the NMI bit in it beyond this point in the function, so there's no need to explicitly clear the NMI bit. thanks -- PMM
On Fri, 22 Mar 2024 at 03:56, Jinjie Ruan <ruanjinjie@huawei.com> wrote: > > > > On 2024/3/21 23:46, Peter Maydell wrote: > > Something somewhere needs to implement "if SCTLR_ELx.NMI is 0 then > > we don't take EXCP_VINMI etc but instead (maybe) EXCP_VIRQ etc". > > At the moment nothing does that: > > * arm_cpu_update_vinmi() doesn't look at the NMI bit before > > deciding whether to set CPU_INTERRUPT_VINMI > > * in arm_excp_unmasked() if NMI is 0 then allIntMask takes its > > default value of false and so arm_excp_unmasked() returns true, > > so VINMI is not masked > > * arm_cpu_exec_interrupt() doesn't look at the NMI bit before > > deciding whether to check the CPU_INTERRUPT_VINMI bit in interrupt_request > > > > So even if SCTLR_ELx.NMI is 0 we'll still try to take a VINMI > > if it's set up in the HCR_EL2 bits. > > > > However we do this the required behaviour is that if NMI is 0 > > then it is as if the interrupt doesn't have superpriority and > > it falls back to being handled as an ordinary IRQ, VIRQ, VFIQ etc. > > I think the best place to do this is probably here in > > arm_cpu_exec_interrupt() -- if SCTLR_ELx.NMI isn't set then > > treat the VFNMI bit like VFIQ, the VINMI bit like VIRQ, and > > the NMI bit like IRQ. > > A PE that implements FEAT_NMI and FEAT_GICv3 also implements > FEAT_GICv3_NMI. A PE that does not implement FEAT_NMI, does not > implement FEAT_GICv3_NMI. > > As the GIC side has checked the FEAT_GICv3_NMI is implemented or Not. So > if cpu_isar_feature(aa64_nmi, env_archcpu(env)) is false, the > FEAT_GICv3_NMI will also not implemented,the CPU_INTERRUPT_NMI/VINMI can > not come from GIC, so we only need to check cpu_isar_feature(aa64_nmi, > env_archcpu(env)) and SCTLR_ELx.NMI is set in hcr_write() and > hcrx_write() for CPU side. If the CPU implements FEAT_NMI then the guest is allowed to write to HCRX_EL2.VINMI even if SCTLR_ELx.NMI is 0. (It might for example choose to set up HCRX_EL2.VINMI first and then set NMI = 1 afterwards.) So you can't fix this by checking for NMI in hcr_write/hcrx_write. If you wanted you could not set the VINMI etc bits in arm_cpu_update_vinmi() and friends if SCTLR_ELx.NMI is 0, as long as you ensure that the update functions all get called again when the NMI bit is written and on exception handling change. -- PMM
diff --git a/target/arm/cpu-qom.h b/target/arm/cpu-qom.h index 8e032691db..b497667d61 100644 --- a/target/arm/cpu-qom.h +++ b/target/arm/cpu-qom.h @@ -36,11 +36,14 @@ DECLARE_CLASS_CHECKERS(AArch64CPUClass, AARCH64_CPU, #define ARM_CPU_TYPE_SUFFIX "-" TYPE_ARM_CPU #define ARM_CPU_TYPE_NAME(name) (name ARM_CPU_TYPE_SUFFIX) -/* Meanings of the ARMCPU object's four inbound GPIO lines */ +/* Meanings of the ARMCPU object's seven inbound GPIO lines */ #define ARM_CPU_IRQ 0 #define ARM_CPU_FIQ 1 #define ARM_CPU_VIRQ 2 #define ARM_CPU_VFIQ 3 +#define ARM_CPU_NMI 4 +#define ARM_CPU_VINMI 5 +#define ARM_CPU_VFNMI 6 /* For M profile, some registers are banked secure vs non-secure; * these are represented as a 2-element array where the first element diff --git a/target/arm/cpu.c b/target/arm/cpu.c index ab8d007a86..f1e7ae0975 100644 --- a/target/arm/cpu.c +++ b/target/arm/cpu.c @@ -122,6 +122,13 @@ void arm_restore_state_to_opc(CPUState *cs, } #endif /* CONFIG_TCG */ +/* + * With SCTLR_ELx.NMI == 0, IRQ with Superpriority is masked identically with + * IRQ without Superpriority. Moreover, if the GIC is configured so that + * FEAT_GICv3_NMI is only set if FEAT_NMI is set, then we won't ever see + * CPU_INTERRUPT_*NMI anyway. So we might as well accept NMI here + * unconditionally. + */ static bool arm_cpu_has_work(CPUState *cs) { ARMCPU *cpu = ARM_CPU(cs); @@ -129,6 +136,7 @@ static bool arm_cpu_has_work(CPUState *cs) return (cpu->power_state != PSCI_OFF) && cs->interrupt_request & (CPU_INTERRUPT_FIQ | CPU_INTERRUPT_HARD + | CPU_INTERRUPT_NMI | CPU_INTERRUPT_VINMI | CPU_INTERRUPT_VFNMI | CPU_INTERRUPT_VFIQ | CPU_INTERRUPT_VIRQ | CPU_INTERRUPT_VSERR | CPU_INTERRUPT_EXITTB); } @@ -668,6 +676,7 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, CPUARMState *env = cpu_env(cs); bool pstate_unmasked; bool unmasked = false; + bool allIntMask = false; /* * Don't take exceptions if they target a lower EL. @@ -678,13 +687,36 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, return false; } + if (cpu_isar_feature(aa64_nmi, env_archcpu(env)) && + env->cp15.sctlr_el[target_el] & SCTLR_NMI && cur_el == target_el) { + allIntMask = env->pstate & PSTATE_ALLINT || + ((env->cp15.sctlr_el[target_el] & SCTLR_SPINTMASK) && + (env->pstate & PSTATE_SP)); + } + switch (excp_idx) { + case EXCP_NMI: + pstate_unmasked = !allIntMask; + break; + + case EXCP_VINMI: + if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { + /* VINMIs are only taken when hypervized. */ + return false; + } + return !allIntMask; + case EXCP_VFNMI: + if (!(hcr_el2 & HCR_FMO) || (hcr_el2 & HCR_TGE)) { + /* VFNMIs are only taken when hypervized. */ + return false; + } + return !allIntMask; case EXCP_FIQ: - pstate_unmasked = !(env->daif & PSTATE_F); + pstate_unmasked = (!(env->daif & PSTATE_F)) && (!allIntMask); break; case EXCP_IRQ: - pstate_unmasked = !(env->daif & PSTATE_I); + pstate_unmasked = (!(env->daif & PSTATE_I)) && (!allIntMask); break; case EXCP_VFIQ: @@ -692,13 +724,13 @@ static inline bool arm_excp_unmasked(CPUState *cs, unsigned int excp_idx, /* VFIQs are only taken when hypervized. */ return false; } - return !(env->daif & PSTATE_F); + return !(env->daif & PSTATE_F) && (!allIntMask); case EXCP_VIRQ: if (!(hcr_el2 & HCR_IMO) || (hcr_el2 & HCR_TGE)) { /* VIRQs are only taken when hypervized. */ return false; } - return !(env->daif & PSTATE_I); + return !(env->daif & PSTATE_I) && (!allIntMask); case EXCP_VSERR: if (!(hcr_el2 & HCR_AMO) || (hcr_el2 & HCR_TGE)) { /* VIRQs are only taken when hypervized. */ @@ -804,6 +836,32 @@ static bool arm_cpu_exec_interrupt(CPUState *cs, int interrupt_request) /* The prioritization of interrupts is IMPLEMENTATION DEFINED. */ + if (cpu_isar_feature(aa64_nmi, env_archcpu(env))) { + if (interrupt_request & CPU_INTERRUPT_NMI) { + excp_idx = EXCP_NMI; + target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); + if (arm_excp_unmasked(cs, excp_idx, target_el, + cur_el, secure, hcr_el2)) { + goto found; + } + } + if (interrupt_request & CPU_INTERRUPT_VINMI) { + excp_idx = EXCP_VINMI; + target_el = 1; + if (arm_excp_unmasked(cs, excp_idx, target_el, + cur_el, secure, hcr_el2)) { + goto found; + } + } + if (interrupt_request & CPU_INTERRUPT_VFNMI) { + excp_idx = EXCP_VFNMI; + target_el = 1; + if (arm_excp_unmasked(cs, excp_idx, target_el, + cur_el, secure, hcr_el2)) { + goto found; + } + } + } if (interrupt_request & CPU_INTERRUPT_FIQ) { excp_idx = EXCP_FIQ; target_el = arm_phys_excp_target_el(cs, excp_idx, cur_el, secure); @@ -900,6 +958,48 @@ void arm_cpu_update_vfiq(ARMCPU *cpu) } } +void arm_cpu_update_vinmi(ARMCPU *cpu) +{ + /* + * Update the interrupt level for VINMI, which is the logical OR of + * the HCRX_EL2.VINMI bit and the input line level from the GIC. + */ + CPUARMState *env = &cpu->env; + CPUState *cs = CPU(cpu); + + bool new_state = ((arm_hcr_el2_eff(env) & HCR_VI) && + (arm_hcrx_el2_eff(env) & HCRX_VINMI)) || + (env->irq_line_state & CPU_INTERRUPT_VINMI); + + if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VINMI) != 0)) { + if (new_state) { + cpu_interrupt(cs, CPU_INTERRUPT_VINMI); + } else { + cpu_reset_interrupt(cs, CPU_INTERRUPT_VINMI); + } + } +} + +void arm_cpu_update_vfnmi(ARMCPU *cpu) +{ + /* + * Update the interrupt level for VFNMI, which is the HCRX_EL2.VFNMI bit. + */ + CPUARMState *env = &cpu->env; + CPUState *cs = CPU(cpu); + + bool new_state = (arm_hcr_el2_eff(env) & HCR_VF) && + (arm_hcrx_el2_eff(env) & HCRX_VFNMI); + + if (new_state != ((cs->interrupt_request & CPU_INTERRUPT_VFNMI) != 0)) { + if (new_state) { + cpu_interrupt(cs, CPU_INTERRUPT_VFNMI); + } else { + cpu_reset_interrupt(cs, CPU_INTERRUPT_VFNMI); + } + } +} + void arm_cpu_update_vserr(ARMCPU *cpu) { /* @@ -929,7 +1029,9 @@ static void arm_cpu_set_irq(void *opaque, int irq, int level) [ARM_CPU_IRQ] = CPU_INTERRUPT_HARD, [ARM_CPU_FIQ] = CPU_INTERRUPT_FIQ, [ARM_CPU_VIRQ] = CPU_INTERRUPT_VIRQ, - [ARM_CPU_VFIQ] = CPU_INTERRUPT_VFIQ + [ARM_CPU_VFIQ] = CPU_INTERRUPT_VFIQ, + [ARM_CPU_NMI] = CPU_INTERRUPT_NMI, + [ARM_CPU_VINMI] = CPU_INTERRUPT_VINMI, }; if (!arm_feature(env, ARM_FEATURE_EL2) && @@ -955,8 +1057,12 @@ static void arm_cpu_set_irq(void *opaque, int irq, int level) case ARM_CPU_VFIQ: arm_cpu_update_vfiq(cpu); break; + case ARM_CPU_VINMI: + arm_cpu_update_vinmi(cpu); + break; case ARM_CPU_IRQ: case ARM_CPU_FIQ: + case ARM_CPU_NMI: if (level) { cpu_interrupt(cs, mask[irq]); } else { @@ -1350,12 +1456,12 @@ static void arm_cpu_initfn(Object *obj) #else /* Our inbound IRQ and FIQ lines */ if (kvm_enabled()) { - /* VIRQ and VFIQ are unused with KVM but we add them to maintain - * the same interface as non-KVM CPUs. + /* VIRQ, VFIQ, NMI, VINMI are unused with KVM but we add + * them to maintain the same interface as non-KVM CPUs. */ - qdev_init_gpio_in(DEVICE(cpu), arm_cpu_kvm_set_irq, 4); + qdev_init_gpio_in(DEVICE(cpu), arm_cpu_kvm_set_irq, 6); } else { - qdev_init_gpio_in(DEVICE(cpu), arm_cpu_set_irq, 4); + qdev_init_gpio_in(DEVICE(cpu), arm_cpu_set_irq, 6); } qdev_init_gpio_out(DEVICE(cpu), cpu->gt_timer_outputs, diff --git a/target/arm/cpu.h b/target/arm/cpu.h index de740d223f..08a6bc50de 100644 --- a/target/arm/cpu.h +++ b/target/arm/cpu.h @@ -61,6 +61,9 @@ #define EXCP_DIVBYZERO 23 /* v7M DIVBYZERO UsageFault */ #define EXCP_VSERR 24 #define EXCP_GPC 25 /* v9 Granule Protection Check Fault */ +#define EXCP_NMI 26 +#define EXCP_VINMI 27 +#define EXCP_VFNMI 28 /* NB: add new EXCP_ defines to the array in arm_log_exception() too */ #define ARMV7M_EXCP_RESET 1 @@ -80,6 +83,9 @@ #define CPU_INTERRUPT_VIRQ CPU_INTERRUPT_TGT_EXT_2 #define CPU_INTERRUPT_VFIQ CPU_INTERRUPT_TGT_EXT_3 #define CPU_INTERRUPT_VSERR CPU_INTERRUPT_TGT_INT_0 +#define CPU_INTERRUPT_NMI CPU_INTERRUPT_TGT_EXT_4 +#define CPU_INTERRUPT_VINMI CPU_INTERRUPT_TGT_EXT_0 +#define CPU_INTERRUPT_VFNMI CPU_INTERRUPT_TGT_INT_1 /* The usual mapping for an AArch64 system register to its AArch32 * counterpart is for the 32 bit world to have access to the lower diff --git a/target/arm/helper.c b/target/arm/helper.c index a65729af66..1868235499 100644 --- a/target/arm/helper.c +++ b/target/arm/helper.c @@ -6042,15 +6042,19 @@ static void do_hcr_write(CPUARMState *env, uint64_t value, uint64_t valid_mask) * and the state of the input lines from the GIC. (This requires * that we have the BQL, which is done by marking the * reginfo structs as ARM_CP_IO.) - * Note that if a write to HCR pends a VIRQ or VFIQ it is never - * possible for it to be taken immediately, because VIRQ and - * VFIQ are masked unless running at EL0 or EL1, and HCR - * can only be written at EL2. + * Note that if a write to HCR pends a VIRQ or VFIQ or VINMI or + * VFNMI, it is never possible for it to be taken immediately + * because VIRQ, VFIQ, VINMI and VFNMI are masked unless running + * at EL0 or EL1, and HCR can only be written at EL2. */ g_assert(bql_locked()); arm_cpu_update_virq(cpu); arm_cpu_update_vfiq(cpu); arm_cpu_update_vserr(cpu); + if (cpu_isar_feature(aa64_nmi, cpu)) { + arm_cpu_update_vinmi(cpu); + arm_cpu_update_vfnmi(cpu); + } } static void hcr_write(CPUARMState *env, const ARMCPRegInfo *ri, uint64_t value) @@ -6199,6 +6203,23 @@ static void hcrx_write(CPUARMState *env, const ARMCPRegInfo *ri, /* Clear RES0 bits. */ env->cp15.hcrx_el2 = value & valid_mask; + + /* + * Updates to VINMI and VFNMI require us to update the status of + * virtual NMI, which are the logical OR of these bits + * and the state of the input lines from the GIC. (This requires + * that we have the BQL, which is done by marking the + * reginfo structs as ARM_CP_IO.) + * Note that if a write to HCRX pends a VINMI or VFNMI it is never + * possible for it to be taken immediately, because VINMI and + * VFNMI are masked unless running at EL0 or EL1, and HCRX + * can only be written at EL2. + */ + if (cpu_isar_feature(aa64_nmi, cpu)) { + g_assert(bql_locked()); + arm_cpu_update_vinmi(cpu); + arm_cpu_update_vfnmi(cpu); + } } static CPAccessResult access_hxen(CPUARMState *env, const ARMCPRegInfo *ri, @@ -6214,6 +6235,7 @@ static CPAccessResult access_hxen(CPUARMState *env, const ARMCPRegInfo *ri, static const ARMCPRegInfo hcrx_el2_reginfo = { .name = "HCRX_EL2", .state = ARM_CP_STATE_AA64, + .type = ARM_CP_IO, .opc0 = 3, .opc1 = 4, .crn = 1, .crm = 2, .opc2 = 2, .access = PL2_RW, .writefn = hcrx_write, .accessfn = access_hxen, .nv2_redirect_offset = 0xa0, @@ -10796,6 +10818,9 @@ void arm_log_exception(CPUState *cs) [EXCP_DIVBYZERO] = "v7M DIVBYZERO UsageFault", [EXCP_VSERR] = "Virtual SERR", [EXCP_GPC] = "Granule Protection Check", + [EXCP_NMI] = "NMI", + [EXCP_VINMI] = "Virtual IRQ NMI", + [EXCP_VFNMI] = "Virtual FIQ NMI", }; if (idx >= 0 && idx < ARRAY_SIZE(excnames)) { diff --git a/target/arm/internals.h b/target/arm/internals.h index 516e0584bf..b53f5e8ff2 100644 --- a/target/arm/internals.h +++ b/target/arm/internals.h @@ -1109,6 +1109,24 @@ void arm_cpu_update_virq(ARMCPU *cpu); */ void arm_cpu_update_vfiq(ARMCPU *cpu); +/** + * arm_cpu_update_vinmi: Update CPU_INTERRUPT_VINMI bit in cs->interrupt_request + * + * Update the CPU_INTERRUPT_VINMI bit in cs->interrupt_request, following + * a change to either the input VNMI line from the GIC or the HCRX_EL2.VINMI. + * Must be called with the BQL held. + */ +void arm_cpu_update_vinmi(ARMCPU *cpu); + +/** + * arm_cpu_update_vfnmi: Update CPU_INTERRUPT_VFNMI bit in cs->interrupt_request + * + * Update the CPU_INTERRUPT_VFNMI bit in cs->interrupt_request, following + * a change to the HCRX_EL2.VFNMI. + * Must be called with the BQL held. + */ +void arm_cpu_update_vfnmi(ARMCPU *cpu); + /** * arm_cpu_update_vserr: Update CPU_INTERRUPT_VSERR bit *