Message ID | 1410435078-28462-3-git-send-email-daniel.thompson@linaro.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: > - .macro svc_entry, stack_hole=0 > + .macro svc_entry, stack_hole=0, call_trace=1 > UNWIND(.fnstart ) > UNWIND(.save {r0 - pc} ) > sub sp, sp, #(S_FRAME_SIZE + \stack_hole - 4) > @@ -183,7 +183,9 @@ ENDPROC(__und_invalid) > stmia r7, {r2 - r6} > > #ifdef CONFIG_TRACE_IRQFLAGS > + .if \call_trace > bl trace_hardirqs_off > + .endif > #endif Good, you picked this up from my patch. But what about the call into lockdep from usr_entry? Yes, it should be safe if we're entering from user mode, because by definition, the kernel can't be holding any locks at that point. However, I'd much prefer to keep to a set of simple rules here: avoid lockdep in FIQ code altogether. That's much easier to understand than "we can call into lockdep provided we've been entered from user mode". The other thing you miss is that /potentially/ call into the scheduler as well from a FIQ. Do we /really/ want to do that kind of work here? Not happy.
On Fri, Sep 12, 2014 at 06:03:07PM +0100, Russell King - ARM Linux wrote: > On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: > > - .macro svc_entry, stack_hole=0 > > + .macro svc_entry, stack_hole=0, call_trace=1 > > UNWIND(.fnstart ) > > UNWIND(.save {r0 - pc} ) > > sub sp, sp, #(S_FRAME_SIZE + \stack_hole - 4) > > @@ -183,7 +183,9 @@ ENDPROC(__und_invalid) > > stmia r7, {r2 - r6} > > > > #ifdef CONFIG_TRACE_IRQFLAGS > > + .if \call_trace > > bl trace_hardirqs_off > > + .endif > > #endif > > Good, you picked this up from my patch. But what about the call into > lockdep from usr_entry? > > Yes, it should be safe if we're entering from user mode, because by > definition, the kernel can't be holding any locks at that point. > However, I'd much prefer to keep to a set of simple rules here: avoid > lockdep in FIQ code altogether. > > That's much easier to understand than "we can call into lockdep provided > we've been entered from user mode". > > The other thing you miss is that /potentially/ call into the scheduler > as well from a FIQ. Do we /really/ want to do that kind of work here? > > Not happy. And you're also missing a .cantunwind for __fiq_usr, which means that the Dwarf doesn't contain an explicit point to stop unwinding. Lastly, don't thread your new patches to the old ones. I utterly hate that behaviour. It makes subject lines totally pointless because all I end up seeing is "[PAT" on the very right hand of the screen. Far from useful.
And another thing... On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: > + .align 5 > +__fiq_abt: > + svc_entry 0, 0 Right, so this function is called __fiq_abt. ... > + svc_exit_via_fiq > + UNWIND(.fnend ) > +ENDPROC(__fiq_svc) Or is it called __fiq_svc ?
On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: > This patch introduces a new default FIQ handler that is structured in a > similar way to the existing ARM exception handler and result in the FIQ > being handled by C code running on the SVC stack (despite this code run > in the FIQ handler is subject to severe limitations with respect to > locking making normal interaction with the kernel impossible). > > This default handler allows concepts that on x86 would be handled using > NMIs to be realized on ARM. Okay, lastly... I sent you my version of this change, which contained the changes I've detailed in the previous three emails, which are absent from your version. However, you've taken on board the "trace" thing to svc_entry, and renamed it to "call_trace". Now if I add this to usr_entry, "call_trace" doesn't make any sense, nor does the .if/.endif placement because it's not just trace_hardirqs_off which needs to be disabled there, but also ct_user_exit as well. I'm beginning to wonder why I tried to be nice here... it would've been a lot faster for me to take your patch, make my own changes and commit that instead. Frustrated.
On Fri, Sep 12, 2014 at 06:14:04PM +0100, Russell King - ARM Linux wrote: > On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: > > This patch introduces a new default FIQ handler that is structured in a > > similar way to the existing ARM exception handler and result in the FIQ > > being handled by C code running on the SVC stack (despite this code run > > in the FIQ handler is subject to severe limitations with respect to > > locking making normal interaction with the kernel impossible). > > > > This default handler allows concepts that on x86 would be handled using > > NMIs to be realized on ARM. > > Okay, lastly... I sent you my version of this change, which contained > the changes I've detailed in the previous three emails, which are absent > from your version. > > However, you've taken on board the "trace" thing to svc_entry, and > renamed it to "call_trace". > > Now if I add this to usr_entry, "call_trace" doesn't make any sense, > nor does the .if/.endif placement because it's not just trace_hardirqs_off > which needs to be disabled there, but also ct_user_exit as well. > > I'm beginning to wonder why I tried to be nice here... it would've been > a lot faster for me to take your patch, make my own changes and commit > that instead. > > Frustrated. And another thing you're missing are the updates to arch/arm/kernel/fiq.c to ensure that the FIQ registers are preserved when we restore this new default FIQ handler.
On 12/09/14 18:03, Russell King - ARM Linux wrote: > On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: >> - .macro svc_entry, stack_hole=0 >> + .macro svc_entry, stack_hole=0, call_trace=1 >> UNWIND(.fnstart ) >> UNWIND(.save {r0 - pc} ) >> sub sp, sp, #(S_FRAME_SIZE + \stack_hole - 4) >> @@ -183,7 +183,9 @@ ENDPROC(__und_invalid) >> stmia r7, {r2 - r6} >> >> #ifdef CONFIG_TRACE_IRQFLAGS >> + .if \call_trace >> bl trace_hardirqs_off >> + .endif >> #endif > > Good, you picked this up from my patch. But what about the call into > lockdep from usr_entry? That was writen from your review comment rather than taken from your patch. > Yes, it should be safe if we're entering from user mode, because by > definition, the kernel can't be holding any locks at that point. > However, I'd much prefer to keep to a set of simple rules here: avoid > lockdep in FIQ code altogether. Ok. You're right that I followed the "can't be holding any locks" logic when I didn't update usr_entry in reaction to the original review comment. I'm also happy with the "avoid lockdep in FIQ code altogether" approach. I'll do this. > That's much easier to understand than "we can call into lockdep provided > we've been entered from user mode". > > The other thing you miss is that /potentially/ call into the scheduler > as well from a FIQ. Do we /really/ want to do that kind of work here? > > Not happy. Sorry. I will fix these. Daniel.
On 12/09/14 18:14, Russell King - ARM Linux wrote: > On Thu, Sep 11, 2014 at 12:31:14PM +0100, Daniel Thompson wrote: >> This patch introduces a new default FIQ handler that is structured in a >> similar way to the existing ARM exception handler and result in the FIQ >> being handled by C code running on the SVC stack (despite this code run >> in the FIQ handler is subject to severe limitations with respect to >> locking making normal interaction with the kernel impossible). >> >> This default handler allows concepts that on x86 would be handled using >> NMIs to be realized on ARM. > > Okay, lastly... I sent you my version of this change, which contained > the changes I've detailed in the previous three emails, which are absent > from your version. > > However, you've taken on board the "trace" thing to svc_entry, and > renamed it to "call_trace". > > Now if I add this to usr_entry, "call_trace" doesn't make any sense, > nor does the .if/.endif placement because it's not just trace_hardirqs_off > which needs to be disabled there, but also ct_user_exit as well. > > I'm beginning to wonder why I tried to be nice here... it would've been > a lot faster for me to take your patch, make my own changes and commit > that instead. I did not do a side by side diff of your FYI patch with my current code and hence overlooked all these changes. Sorry. I should have done that.
diff --git a/arch/arm/kernel/entry-armv.S b/arch/arm/kernel/entry-armv.S index 36276cd..0c70fee 100644 --- a/arch/arm/kernel/entry-armv.S +++ b/arch/arm/kernel/entry-armv.S @@ -146,7 +146,7 @@ ENDPROC(__und_invalid) #define SPFIX(code...) #endif - .macro svc_entry, stack_hole=0 + .macro svc_entry, stack_hole=0, call_trace=1 UNWIND(.fnstart ) UNWIND(.save {r0 - pc} ) sub sp, sp, #(S_FRAME_SIZE + \stack_hole - 4) @@ -183,7 +183,9 @@ ENDPROC(__und_invalid) stmia r7, {r2 - r6} #ifdef CONFIG_TRACE_IRQFLAGS + .if \call_trace bl trace_hardirqs_off + .endif #endif .endm @@ -295,6 +297,15 @@ __pabt_svc: ENDPROC(__pabt_svc) .align 5 +__fiq_svc: + svc_entry 0, 0 + mov r0, sp @ struct pt_regs *regs + bl handle_fiq_as_nmi + svc_exit_via_fiq + UNWIND(.fnend ) +ENDPROC(__fiq_svc) + + .align 5 .LCcralign: .word cr_alignment #ifdef MULTI_DABORT @@ -305,6 +316,46 @@ ENDPROC(__pabt_svc) .word fp_enter /* + * Abort mode handlers + */ + +@ +@ Taking a FIQ in abort mode is similar to taking a FIQ in SVC mode +@ and reuses the same macros. However in abort mode we must also +@ save/restore lr_abt and spsr_abt to make nested aborts safe. +@ + .align 5 +__fiq_abt: + svc_entry 0, 0 + + ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( mov r0, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( msr cpsr_c, r0 ) + mov r1, lr @ Save lr_abt + mrs r2, spsr @ Save spsr_abt, abort is now safe + ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( mov r0, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( msr cpsr_c, r0 ) + stmfd sp!, {r1 - r2} + + add r0, sp, #8 @ struct pt_regs *regs + bl handle_fiq_as_nmi + + ldmfd sp!, {r1 - r2} + ARM( msr cpsr_c, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( mov r0, #ABT_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( msr cpsr_c, r0 ) + mov lr, r1 @ Restore lr_abt, abort is unsafe + msr spsr_cxsf, r2 @ Restore spsr_abt + ARM( msr cpsr_c, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( mov r0, #SVC_MODE | PSR_I_BIT | PSR_F_BIT ) + THUMB( msr cpsr_c, r0 ) + + svc_exit_via_fiq + UNWIND(.fnend ) +ENDPROC(__fiq_svc) + +/* * User mode handlers * * EABI note: sp_svc is always 64-bit aligned here, so should S_FRAME_SIZE @@ -683,6 +734,18 @@ ENTRY(ret_from_exception) ENDPROC(__pabt_usr) ENDPROC(ret_from_exception) + .align 5 +__fiq_usr: + usr_entry + kuser_cmpxchg_check + mov r0, sp @ struct pt_regs *regs + bl handle_fiq_as_nmi + get_thread_info tsk + mov why, #0 + b ret_to_user_from_irq + UNWIND(.fnend ) +ENDPROC(__fiq_usr) + /* * Register switch for ARMv3 and ARMv4 processors * r0 = previous task_struct, r1 = previous thread_info, r2 = next thread_info @@ -1118,17 +1181,29 @@ vector_addrexcptn: b vector_addrexcptn /*============================================================================= - * Undefined FIQs + * FIQ "NMI" handler *----------------------------------------------------------------------------- - * Enter in FIQ mode, spsr = ANY CPSR, lr = ANY PC - * MUST PRESERVE SVC SPSR, but need to switch to SVC mode to show our msg. - * Basically to switch modes, we *HAVE* to clobber one register... brain - * damage alert! I don't think that we can execute any code in here in any - * other mode than FIQ... Ok you can switch to another mode, but you can't - * get out of that mode without clobbering one register. + * Handle a FIQ using the SVC stack allowing FIQ act like NMI on x86 + * systems. */ -vector_fiq: - subs pc, lr, #4 + vector_stub fiq, FIQ_MODE, 4 + + .long __fiq_usr @ 0 (USR_26 / USR_32) + .long __fiq_svc @ 1 (FIQ_26 / FIQ_32) + .long __fiq_svc @ 2 (IRQ_26 / IRQ_32) + .long __fiq_svc @ 3 (SVC_26 / SVC_32) + .long __fiq_svc @ 4 + .long __fiq_svc @ 5 + .long __fiq_svc @ 6 + .long __fiq_abt @ 7 + .long __fiq_svc @ 8 + .long __fiq_svc @ 9 + .long __fiq_svc @ a + .long __fiq_svc @ b + .long __fiq_svc @ c + .long __fiq_svc @ d + .long __fiq_svc @ e + .long __fiq_svc @ f .globl vector_fiq_offset .equ vector_fiq_offset, vector_fiq diff --git a/arch/arm/kernel/entry-header.S b/arch/arm/kernel/entry-header.S index 2fdf867..0d91ca0 100644 --- a/arch/arm/kernel/entry-header.S +++ b/arch/arm/kernel/entry-header.S @@ -216,6 +216,34 @@ ldmia sp, {r0 - pc}^ @ load r0 - pc, cpsr .endm + @ + @ svc_exit_via_fiq - like svc_exit but switches to FIQ mode before exit + @ + @ This macro acts in a similar manner to svc_exit but switches to FIQ + @ mode to restore the final part of the register state. + @ + @ We cannot use the normal svc_exit procedure because that would + @ clobber spsr_svc (FIQ could be delivered during the first few + @ instructions of vector_swi meaning its contents have not been + @ saved anywhere). + @ + @ Note that, unlike svc_exit, this macro also does not allow a caller + @ supplied rpsr. This is because the FIQ exceptions are not re-entrant + @ and the handlers cannot call into the scheduler (meaning the value + @ on the stack remains correct). + @ + .macro svc_exit_via_fiq + mov r0, sp + ldmib r0, {r1 - r14} @ abort is deadly from here onward (it will + @ clobber state restored below) + msr cpsr_c, #FIQ_MODE | PSR_I_BIT | PSR_F_BIT + add r8, r0, #S_PC + ldr r9, [r0, #S_PSR] + msr spsr_cxsf, r9 + ldr r0, [r0, #S_R0] + ldmia r8, {pc}^ + .endm + .macro restore_user_regs, fast = 0, offset = 0 ldr r1, [sp, #\offset + S_PSR] @ get calling cpsr ldr lr, [sp, #\offset + S_PC]! @ get pc @@ -267,6 +295,25 @@ rfeia sp! .endm + @ + @ svc_exit_via_fiq - like svc_exit but switches to FIQ mode before exit + @ + @ For full details see non-Thumb implementation above. + @ + .macro svc_exit_via_fiq + add r0, sp, #S_R2 + ldr lr, [sp, #S_LR] + ldr sp, [sp, #S_SP] @ abort is deadly from here onward (it will + @ clobber state restored below) + ldmia r0, {r2 - r12} + mov r1, #FIQ_MODE | PSR_I_BIT | PSR_F_BIT + msr cpsr_c, r1 + sub r0, #S_R2 + add r8, r0, #S_PC + ldmia r0, {r0 - r1} + rfeia r8 + .endm + #ifdef CONFIG_CPU_V7M /* * Note we don't need to do clrex here as clearing the local monitor is diff --git a/arch/arm/kernel/setup.c b/arch/arm/kernel/setup.c index 84db893d..c031063 100644 --- a/arch/arm/kernel/setup.c +++ b/arch/arm/kernel/setup.c @@ -133,6 +133,7 @@ struct stack { u32 irq[3]; u32 abt[3]; u32 und[3]; + u32 fiq[3]; } ____cacheline_aligned; #ifndef CONFIG_CPU_V7M @@ -470,7 +471,10 @@ void notrace cpu_init(void) "msr cpsr_c, %5\n\t" "add r14, %0, %6\n\t" "mov sp, r14\n\t" - "msr cpsr_c, %7" + "msr cpsr_c, %7\n\t" + "add r14, %0, %8\n\t" + "mov sp, r14\n\t" + "msr cpsr_c, %9" : : "r" (stk), PLC (PSR_F_BIT | PSR_I_BIT | IRQ_MODE), @@ -479,6 +483,8 @@ void notrace cpu_init(void) "I" (offsetof(struct stack, abt[0])), PLC (PSR_F_BIT | PSR_I_BIT | UND_MODE), "I" (offsetof(struct stack, und[0])), + PLC (PSR_F_BIT | PSR_I_BIT | FIQ_MODE), + "I" (offsetof(struct stack, fiq[0])), PLC (PSR_F_BIT | PSR_I_BIT | SVC_MODE) : "r14"); #endif diff --git a/arch/arm/kernel/traps.c b/arch/arm/kernel/traps.c index a447dcc..439138d 100644 --- a/arch/arm/kernel/traps.c +++ b/arch/arm/kernel/traps.c @@ -25,6 +25,7 @@ #include <linux/delay.h> #include <linux/init.h> #include <linux/sched.h> +#include <linux/irq.h> #include <linux/atomic.h> #include <asm/cacheflush.h> @@ -461,6 +462,31 @@ die_sig: } /* + * Handle FIQ similarly to NMI on x86 systems. + * + * The runtime environment for NMIs is extremely restrictive + * (NMIs can pre-empt critical sections meaning almost all locking is + * forbidden) meaning this default FIQ handling must only be used in + * circumstances where non-maskability improves robustness, such as + * watchdog or debug logic. + * + * This handler is not appropriate for general purpose use in drivers + * platform code and can be overrideen using set_fiq_handler. + */ +asmlinkage void __exception_irq_entry handle_fiq_as_nmi(struct pt_regs *regs) +{ + struct pt_regs *old_regs = set_irq_regs(regs); + + nmi_enter(); + + /* nop. FIQ handlers for special arch/arm features can be added here. */ + + nmi_exit(); + + set_irq_regs(old_regs); +} + +/* * bad_mode handles the impossible case in the vectors. If you see one of * these, then it's extremely serious, and could mean you have buggy hardware. * It never returns, and never tries to sync. We hope that we can at least