Message ID | 92a00580828a1bdf96e7e36545f6d229809af04f.1722154575.git.calvin@wbinvd.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | ARM: vfp: Use asm volatile in fmrx/fmxr macros | expand |
(cc Arnd, Nathan, zhuqiuer) On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote: > > Floating point instructions in userspace can crash some arm kernels > built with clang/LLD 17.0.6: > > BUG: unsupported FP instruction in kernel mode > FPEXC == 0xc0000780 > Internal error: Oops - undefined instruction: 0 [#1] ARM > CPU: 0 PID: 196 Comm: vfp-reproducer Not tainted 6.10.0 #1 > Hardware name: BCM2835 > PC is at vfp_support_entry+0xc8/0x2cc > LR is at do_undefinstr+0xa8/0x250 > pc : [<c0101d50>] lr : [<c010a80c>] psr: a0000013 > sp : dc8d1f68 ip : 60000013 fp : bedea19c > r10: ec532b17 r9 : 00000010 r8 : 0044766c > r7 : c0000780 r6 : ec532b17 r5 : c1c13800 r4 : dc8d1fb0 > r3 : c10072c4 r2 : c0101c88 r1 : ec532b17 r0 : 0044766c > Flags: NzCv IRQs on FIQs on Mode SVC_32 ISA ARM Segment none > Control: 00c5387d Table: 0251c008 DAC: 00000051 > Register r0 information: non-paged memory > Register r1 information: vmalloc memory > Register r2 information: non-slab/vmalloc memory > Register r3 information: non-slab/vmalloc memory > Register r4 information: 2-page vmalloc region > Register r5 information: slab kmalloc-cg-2k > Register r6 information: vmalloc memory > Register r7 information: non-slab/vmalloc memory > Register r8 information: non-paged memory > Register r9 information: zero-size pointer > Register r10 information: vmalloc memory > Register r11 information: non-paged memory > Register r12 information: non-paged memory > Process vfp-reproducer (pid: 196, stack limit = 0x61aaaf8b) > Stack: (0xdc8d1f68 to 0xdc8d2000) > 1f60: 0000081f b6f69300 0000000f c10073f4 c10072c4 dc8d1fb0 > 1f80: ec532b17 0c532b17 0044766c b6f9ccd8 00000000 c010a80c 00447670 60000010 > 1fa0: ffffffff c1c13800 00c5387d c0100f10 b6f68af8 00448fc0 00000000 bedea188 > 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c > 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff 00000000 00000000 > Call trace: > [<c0101d50>] (vfp_support_entry) from [<c010a80c>] (do_undefinstr+0xa8/0x250) > [<c010a80c>] (do_undefinstr) from [<c0100f10>] (__und_usr+0x70/0x80) > Exception stack(0xdc8d1fb0 to 0xdc8d1ff8) > 1fa0: b6f68af8 00448fc0 00000000 bedea188 > 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c > 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff > Code: 0a000061 e3877202 e594003c e3a09010 (eef16a10) > ---[ end trace 0000000000000000 ]--- > Kernel panic - not syncing: Fatal exception in interrupt > ---[ end Kernel panic - not syncing: Fatal exception in interrupt ]--- > > This is a minimal userspace reproducer on a Raspberry Pi Zero W: > > #include <stdio.h> > #include <math.h> > > int main(void) > { > double v = 1.0; > printf("%f\n", NAN + *(volatile double *)&v); > return 0; > } > > Another way to consistently trigger the oops is: > > calvin@raspberry-pi-zero-w ~$ python -c "import json" > > The bug reproduces only when the kernel is built with DYNAMIC_DEBUG=n, > because the pr_debug() calls act as barriers even when not activated. > > This is the output from the same kernel source built with the same > compiler and DYNAMIC_DEBUG=y, where the userspace reproducer works as > expected: > > VFP: bounce: trigger ec532b17 fpexc c0000780 > VFP: emulate: INST=0xee377b06 SCR=0x00000000 > VFP: bounce: trigger eef1fa10 fpexc c0000780 > VFP: emulate: INST=0xeeb40b40 SCR=0x00000000 > VFP: raising exceptions 30000000 > > calvin@raspberry-pi-zero-w ~$ ./vfp-reproducer > nan > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly > idential text for vfp_support_entry() makes the problem obvious: > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0 > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7 > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0 > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2 > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0 > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7 > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM! > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0 > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2 > > I think LLVM's reordering is valid as the code is currently written: the > compiler doesn't know the instructions have side effects in hardware. > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be > reordered with respect to each other. The original compiler now produces > working kernels on my hardware with DYNAMIC_DEBUG=n. > > This is the relevant piece of the diff of the vfp_support_entry() text, > from the original oopsing kernel to a working kernel with this patch: > > vmrs r0, fpscr > tst r0, #4096 > bne 0xc0101d48 > tst r0, #458752 > beq 0xc0101ecc > orr r7, r7, #536870912 > ldr r0, [r4, #0x3c] > mov r9, #16 > -vmrs r6, fpscr > orr r9, r9, #251658240 > add r0, r0, #4 > str r0, [r4, #0x3c] > mvn r0, #159 > sub r0, r0, #-1207959552 > and r0, r7, r0 > vmsr fpexc, r0 > vmrs r0, fpsid > +vmrs r6, fpscr > and r0, r0, #983040 > cmp r0, #65536 > bne 0xc0101d88 > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code") > Signed-off-by: Calvin Owens <calvin@wbinvd.org> > --- > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- > 1 file changed, 26 insertions(+), 22 deletions(-) > Thanks for the patch, and for the excellent analysis. Note that this fix has been proposed in the past, as well as another one addressing the same issue, but we've been incredibly sloppy getting it merged. https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/ https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/ What both of us appear to have missed is that there are two versions of these routines, which should either be dropped (as they are obsolete now that the minimum binutils version is 2.25) or fixed up as well, as you do below. Anyone have any thoughts on using a memory clobber as opposed to volatile? I think volatile means that the access cannot be elided, but it is unclear to me whether that implies any ordering. A 'memory' clobber implies that globally visible state is updated, which seems like a stronger guarantee to me. In any case, let's work together to get /some/ version of this fix merged asap. > diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h > index 3c7938fd40aa..32090b0fb250 100644 > --- a/arch/arm/vfp/vfpinstr.h > +++ b/arch/arm/vfp/vfpinstr.h > @@ -64,33 +64,37 @@ > > #ifdef CONFIG_AS_VFP_VMRS_FPINST > > -#define fmrx(_vfp_) ({ \ > - u32 __v; \ > - asm(".fpu vfpv2\n" \ > - "vmrs %0, " #_vfp_ \ > - : "=r" (__v) : : "cc"); \ > - __v; \ > - }) > - > -#define fmxr(_vfp_,_var_) \ > - asm(".fpu vfpv2\n" \ > - "vmsr " #_vfp_ ", %0" \ > - : : "r" (_var_) : "cc") > +#define fmrx(_vfp_) ({ \ > + u32 __v; \ > + asm volatile (".fpu vfpv2\n" \ > + "vmrs %0, " #_vfp_ \ > + : "=r" (__v) : : "cc"); \ > + __v; \ > +}) > + > +#define fmxr(_vfp_, _var_) ({ \ > + asm volatile (".fpu vfpv2\n" \ > + "vmsr " #_vfp_ ", %0" \ > + : : "r" (_var_) : "cc"); \ > +}) > > #else > > #define vfpreg(_vfp_) #_vfp_ > > -#define fmrx(_vfp_) ({ \ > - u32 __v; \ > - asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \ > - : "=r" (__v) : : "cc"); \ > - __v; \ > - }) > - > -#define fmxr(_vfp_,_var_) \ > - asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \ > - : : "r" (_var_) : "cc") > +#define fmrx(_vfp_) ({ \ > + u32 __v; \ > + asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \ > + "cr0, 0 @ fmrx %0, " #_vfp_ \ > + : "=r" (__v) : : "cc"); \ > + __v; \ > +}) > + > +#define fmxr(_vfp_, _var_) ({ \ > + asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \ > + "cr0, 0 @ fmxr " #_vfp_ ", %0" \ > + : : "r" (_var_) : "cc"); \ > +}) > > #endif > > -- > 2.39.2 >
On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote: > (cc Arnd, Nathan, zhuqiuer) > > On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote: > > <snip> > > > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly > > idential text for vfp_support_entry() makes the problem obvious: > > > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc > > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0 > > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7 > > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc > > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr > > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0 > > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid > > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr > > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst > > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2 > > > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc > > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0 > > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7 > > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr > > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM! > > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0 > > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid > > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst > > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc > > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2 > > > > I think LLVM's reordering is valid as the code is currently written: the > > compiler doesn't know the instructions have side effects in hardware. > > > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be > > reordered with respect to each other. The original compiler now produces > > working kernels on my hardware with DYNAMIC_DEBUG=n. > > > > This is the relevant piece of the diff of the vfp_support_entry() text, > > from the original oopsing kernel to a working kernel with this patch: > > > > vmrs r0, fpscr > > tst r0, #4096 > > bne 0xc0101d48 > > tst r0, #458752 > > beq 0xc0101ecc > > orr r7, r7, #536870912 > > ldr r0, [r4, #0x3c] > > mov r9, #16 > > -vmrs r6, fpscr > > orr r9, r9, #251658240 > > add r0, r0, #4 > > str r0, [r4, #0x3c] > > mvn r0, #159 > > sub r0, r0, #-1207959552 > > and r0, r7, r0 > > vmsr fpexc, r0 > > vmrs r0, fpsid > > +vmrs r6, fpscr > > and r0, r0, #983040 > > cmp r0, #65536 > > bne 0xc0101d88 > > > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code") > > Signed-off-by: Calvin Owens <calvin@wbinvd.org> > > --- > > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- > > 1 file changed, 26 insertions(+), 22 deletions(-) > > > > Thanks for the patch, and for the excellent analysis. > > Note that this fix has been proposed in the past, as well as another > one addressing the same issue, but we've been incredibly sloppy > getting it merged. > > https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/ > https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/ Ah sorry for missing that, I searched for the symptom not the cure. > What both of us appear to have missed is that there are two versions > of these routines, which should either be dropped (as they are > obsolete now that the minimum binutils version is 2.25) or fixed up as > well, as you do below. > > Anyone have any thoughts on using a memory clobber as opposed to > volatile? I think volatile means that the access cannot be elided, but > it is unclear to me whether that implies any ordering. A 'memory' > clobber implies that globally visible state is updated, which seems > like a stronger guarantee to me. My thinking was that if 'asm volatile' is sufficient, it will suppress less optimization than the clobber, since the clobber might force the compiler to assume unrelated memory could have been modified when it really never is. But I'm not sure about that. Out of curiousity, I tried it both ways with the same compiler just now, the only tiny difference in the emitted vfp_support_entry() is here: --- /volatile 2024-07-28 13:28:59.890505404 -0700 +++ /memclobber 2024-07-28 13:28:59.890505404 -0700 str r0, [r5, #0x4] vmrs r7, fpexc tst r7, #1073741824 bne 0xc0101d28 orr r7, r7, #1073741824 bic r0, r7, #-2147483648 vmsr fpexc, r0 +ldr r8, [pc, #0x26c] ldr r0, [r5, #0x8] -ldr r8, [pc, #0x268] add r6, r5, #224 ldr r0, [r8, r0, lsl #2] cmp r0, r6 beq 0xc0101d18 Thanks, Calvin > In any case, let's work together to get /some/ version of this fix merged asap. > > > diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h > > index 3c7938fd40aa..32090b0fb250 100644 > > --- a/arch/arm/vfp/vfpinstr.h > > +++ b/arch/arm/vfp/vfpinstr.h > > @@ -64,33 +64,37 @@ > > > > #ifdef CONFIG_AS_VFP_VMRS_FPINST > > > > -#define fmrx(_vfp_) ({ \ > > - u32 __v; \ > > - asm(".fpu vfpv2\n" \ > > - "vmrs %0, " #_vfp_ \ > > - : "=r" (__v) : : "cc"); \ > > - __v; \ > > - }) > > - > > -#define fmxr(_vfp_,_var_) \ > > - asm(".fpu vfpv2\n" \ > > - "vmsr " #_vfp_ ", %0" \ > > - : : "r" (_var_) : "cc") > > +#define fmrx(_vfp_) ({ \ > > + u32 __v; \ > > + asm volatile (".fpu vfpv2\n" \ > > + "vmrs %0, " #_vfp_ \ > > + : "=r" (__v) : : "cc"); \ > > + __v; \ > > +}) > > + > > +#define fmxr(_vfp_, _var_) ({ \ > > + asm volatile (".fpu vfpv2\n" \ > > + "vmsr " #_vfp_ ", %0" \ > > + : : "r" (_var_) : "cc"); \ > > +}) > > > > #else > > > > #define vfpreg(_vfp_) #_vfp_ > > > > -#define fmrx(_vfp_) ({ \ > > - u32 __v; \ > > - asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \ > > - : "=r" (__v) : : "cc"); \ > > - __v; \ > > - }) > > - > > -#define fmxr(_vfp_,_var_) \ > > - asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \ > > - : : "r" (_var_) : "cc") > > +#define fmrx(_vfp_) ({ \ > > + u32 __v; \ > > + asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \ > > + "cr0, 0 @ fmrx %0, " #_vfp_ \ > > + : "=r" (__v) : : "cc"); \ > > + __v; \ > > +}) > > + > > +#define fmxr(_vfp_, _var_) ({ \ > > + asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \ > > + "cr0, 0 @ fmxr " #_vfp_ ", %0" \ > > + : : "r" (_var_) : "cc"); \ > > +}) > > > > #endif > > > > -- > > 2.39.2 > >
On Sun, 28 Jul 2024 at 23:29, Calvin Owens <calvin@wbinvd.org> wrote: > > On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote: > > (cc Arnd, Nathan, zhuqiuer) > > > > On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote: > > > <snip> > > > > > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly > > > idential text for vfp_support_entry() makes the problem obvious: > > > > > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc > > > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0 > > > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7 > > > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc > > > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr > > > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0 > > > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid > > > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr > > > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst > > > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2 > > > > > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc > > > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0 > > > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7 > > > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr > > > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM! > > > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0 > > > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid > > > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst > > > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc > > > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2 > > > > > > I think LLVM's reordering is valid as the code is currently written: the > > > compiler doesn't know the instructions have side effects in hardware. > > > > > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be > > > reordered with respect to each other. The original compiler now produces > > > working kernels on my hardware with DYNAMIC_DEBUG=n. > > > > > > This is the relevant piece of the diff of the vfp_support_entry() text, > > > from the original oopsing kernel to a working kernel with this patch: > > > > > > vmrs r0, fpscr > > > tst r0, #4096 > > > bne 0xc0101d48 > > > tst r0, #458752 > > > beq 0xc0101ecc > > > orr r7, r7, #536870912 > > > ldr r0, [r4, #0x3c] > > > mov r9, #16 > > > -vmrs r6, fpscr > > > orr r9, r9, #251658240 > > > add r0, r0, #4 > > > str r0, [r4, #0x3c] > > > mvn r0, #159 > > > sub r0, r0, #-1207959552 > > > and r0, r7, r0 > > > vmsr fpexc, r0 > > > vmrs r0, fpsid > > > +vmrs r6, fpscr > > > and r0, r0, #983040 > > > cmp r0, #65536 > > > bne 0xc0101d88 > > > > > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code") > > > Signed-off-by: Calvin Owens <calvin@wbinvd.org> > > > --- > > > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- > > > 1 file changed, 26 insertions(+), 22 deletions(-) > > > > > > > Thanks for the patch, and for the excellent analysis. > > > > Note that this fix has been proposed in the past, as well as another > > one addressing the same issue, but we've been incredibly sloppy > > getting it merged. > > > > https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/ > > https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/ > > Ah sorry for missing that, I searched for the symptom not the cure. > > > What both of us appear to have missed is that there are two versions > > of these routines, which should either be dropped (as they are > > obsolete now that the minimum binutils version is 2.25) or fixed up as > > well, as you do below. > > > > Anyone have any thoughts on using a memory clobber as opposed to > > volatile? I think volatile means that the access cannot be elided, but > > it is unclear to me whether that implies any ordering. A 'memory' > > clobber implies that globally visible state is updated, which seems > > like a stronger guarantee to me. > > My thinking was that if 'asm volatile' is sufficient, it will suppress > less optimization than the clobber, since the clobber might force the > compiler to assume unrelated memory could have been modified when it > really never is. But I'm not sure about that. > > Out of curiousity, I tried it both ways with the same compiler just now, > the only tiny difference in the emitted vfp_support_entry() is here: > > --- /volatile 2024-07-28 13:28:59.890505404 -0700 > +++ /memclobber 2024-07-28 13:28:59.890505404 -0700 > str r0, [r5, #0x4] > vmrs r7, fpexc > tst r7, #1073741824 > bne 0xc0101d28 > orr r7, r7, #1073741824 > bic r0, r7, #-2147483648 > vmsr fpexc, r0 > +ldr r8, [pc, #0x26c] > ldr r0, [r5, #0x8] > -ldr r8, [pc, #0x268] > add r6, r5, #224 > ldr r0, [r8, r0, lsl #2] > cmp r0, r6 > beq 0xc0101d18 > Right. So it doesn't matter in practice - good to know. So in the 'volatile' case, I guess we are relying on the compiler not reordering those with respect to each other, which should be sufficient to ensure that we do not access VFP status register before enabling the VFP unit via the control register, as all are accessed using inline asm. In any case, this should go into the patch system. https://www.armlinux.org.uk/developer/patches/section.php?section=0 Note to self and other: follow-up with a patch that removes CONFIG_AS_VFP_VMRS_FPINST, which is no longer needed.
On Monday 07/29 at 09:12 +0200, Ard Biesheuvel wrote: > On Sun, 28 Jul 2024 at 23:29, Calvin Owens <calvin@wbinvd.org> wrote: > > > > On Sunday 07/28 at 19:09 +0200, Ard Biesheuvel wrote: > > > (cc Arnd, Nathan, zhuqiuer) > > > > > > On Sun, 28 Jul 2024 at 10:21, Calvin Owens <calvin@wbinvd.org> wrote: > > > > <snip> > > > > > > > > Crudely grepping for vmsr/vmrs instructions in the otherwise nearly > > > > idential text for vfp_support_entry() makes the problem obvious: > > > > > > > > vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc > > > > vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0 > > > > vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7 > > > > vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc > > > > vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr > > > > vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0 > > > > vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid > > > > vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr > > > > vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst > > > > vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2 > > > > > > > > vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc > > > > vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0 > > > > vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7 > > > > vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr > > > > vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM! > > > > vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0 > > > > vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid > > > > vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst > > > > vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc > > > > vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2 > > > > > > > > I think LLVM's reordering is valid as the code is currently written: the > > > > compiler doesn't know the instructions have side effects in hardware. > > > > > > > > Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be > > > > reordered with respect to each other. The original compiler now produces > > > > working kernels on my hardware with DYNAMIC_DEBUG=n. > > > > > > > > This is the relevant piece of the diff of the vfp_support_entry() text, > > > > from the original oopsing kernel to a working kernel with this patch: > > > > > > > > vmrs r0, fpscr > > > > tst r0, #4096 > > > > bne 0xc0101d48 > > > > tst r0, #458752 > > > > beq 0xc0101ecc > > > > orr r7, r7, #536870912 > > > > ldr r0, [r4, #0x3c] > > > > mov r9, #16 > > > > -vmrs r6, fpscr > > > > orr r9, r9, #251658240 > > > > add r0, r0, #4 > > > > str r0, [r4, #0x3c] > > > > mvn r0, #159 > > > > sub r0, r0, #-1207959552 > > > > and r0, r7, r0 > > > > vmsr fpexc, r0 > > > > vmrs r0, fpsid > > > > +vmrs r6, fpscr > > > > and r0, r0, #983040 > > > > cmp r0, #65536 > > > > bne 0xc0101d88 > > > > > > > > Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code") > > > > Signed-off-by: Calvin Owens <calvin@wbinvd.org> > > > > --- > > > > arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- > > > > 1 file changed, 26 insertions(+), 22 deletions(-) > > > > > > > > > > Thanks for the patch, and for the excellent analysis. > > > > > > Note that this fix has been proposed in the past, as well as another > > > one addressing the same issue, but we've been incredibly sloppy > > > getting it merged. > > > > > > https://lore.kernel.org/linux-arm-kernel/20240410024126.21589-2-zhuqiuer1@huawei.com/ > > > https://lore.kernel.org/linux-arm-kernel/20240318093004.117153-2-ardb+git@google.com/ > > > > Ah sorry for missing that, I searched for the symptom not the cure. > > > > > What both of us appear to have missed is that there are two versions > > > of these routines, which should either be dropped (as they are > > > obsolete now that the minimum binutils version is 2.25) or fixed up as > > > well, as you do below. > > > > > > Anyone have any thoughts on using a memory clobber as opposed to > > > volatile? I think volatile means that the access cannot be elided, but > > > it is unclear to me whether that implies any ordering. A 'memory' > > > clobber implies that globally visible state is updated, which seems > > > like a stronger guarantee to me. > > > > My thinking was that if 'asm volatile' is sufficient, it will suppress > > less optimization than the clobber, since the clobber might force the > > compiler to assume unrelated memory could have been modified when it > > really never is. But I'm not sure about that. > > > > Out of curiousity, I tried it both ways with the same compiler just now, > > the only tiny difference in the emitted vfp_support_entry() is here: > > > > --- /volatile 2024-07-28 13:28:59.890505404 -0700 > > +++ /memclobber 2024-07-28 13:28:59.890505404 -0700 > > str r0, [r5, #0x4] > > vmrs r7, fpexc > > tst r7, #1073741824 > > bne 0xc0101d28 > > orr r7, r7, #1073741824 > > bic r0, r7, #-2147483648 > > vmsr fpexc, r0 > > +ldr r8, [pc, #0x26c] > > ldr r0, [r5, #0x8] > > -ldr r8, [pc, #0x268] > > add r6, r5, #224 > > ldr r0, [r8, r0, lsl #2] > > cmp r0, r6 > > beq 0xc0101d18 > > > > Right. So it doesn't matter in practice - good to know. > > So in the 'volatile' case, I guess we are relying on the compiler not > reordering those with respect to each other, which should be > sufficient to ensure that we do not access VFP status register before > enabling the VFP unit via the control register, as all are accessed > using inline asm. > > In any case, this should go into the patch system. > https://www.armlinux.org.uk/developer/patches/section.php?section=0 Done: https://www.armlinux.org.uk/developer/patches/viewpatch.php?id=9410/1 Thanks, Calvin > Note to self and other: follow-up with a patch that removes > CONFIG_AS_VFP_VMRS_FPINST, which is no longer needed.
diff --git a/arch/arm/vfp/vfpinstr.h b/arch/arm/vfp/vfpinstr.h index 3c7938fd40aa..32090b0fb250 100644 --- a/arch/arm/vfp/vfpinstr.h +++ b/arch/arm/vfp/vfpinstr.h @@ -64,33 +64,37 @@ #ifdef CONFIG_AS_VFP_VMRS_FPINST -#define fmrx(_vfp_) ({ \ - u32 __v; \ - asm(".fpu vfpv2\n" \ - "vmrs %0, " #_vfp_ \ - : "=r" (__v) : : "cc"); \ - __v; \ - }) - -#define fmxr(_vfp_,_var_) \ - asm(".fpu vfpv2\n" \ - "vmsr " #_vfp_ ", %0" \ - : : "r" (_var_) : "cc") +#define fmrx(_vfp_) ({ \ + u32 __v; \ + asm volatile (".fpu vfpv2\n" \ + "vmrs %0, " #_vfp_ \ + : "=r" (__v) : : "cc"); \ + __v; \ +}) + +#define fmxr(_vfp_, _var_) ({ \ + asm volatile (".fpu vfpv2\n" \ + "vmsr " #_vfp_ ", %0" \ + : : "r" (_var_) : "cc"); \ +}) #else #define vfpreg(_vfp_) #_vfp_ -#define fmrx(_vfp_) ({ \ - u32 __v; \ - asm("mrc p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmrx %0, " #_vfp_ \ - : "=r" (__v) : : "cc"); \ - __v; \ - }) - -#define fmxr(_vfp_,_var_) \ - asm("mcr p10, 7, %0, " vfpreg(_vfp_) ", cr0, 0 @ fmxr " #_vfp_ ", %0" \ - : : "r" (_var_) : "cc") +#define fmrx(_vfp_) ({ \ + u32 __v; \ + asm volatile ("mrc p10, 7, %0, " vfpreg(_vfp_) "," \ + "cr0, 0 @ fmrx %0, " #_vfp_ \ + : "=r" (__v) : : "cc"); \ + __v; \ +}) + +#define fmxr(_vfp_, _var_) ({ \ + asm volatile ("mcr p10, 7, %0, " vfpreg(_vfp_) "," \ + "cr0, 0 @ fmxr " #_vfp_ ", %0" \ + : : "r" (_var_) : "cc"); \ +}) #endif
Floating point instructions in userspace can crash some arm kernels built with clang/LLD 17.0.6: BUG: unsupported FP instruction in kernel mode FPEXC == 0xc0000780 Internal error: Oops - undefined instruction: 0 [#1] ARM CPU: 0 PID: 196 Comm: vfp-reproducer Not tainted 6.10.0 #1 Hardware name: BCM2835 PC is at vfp_support_entry+0xc8/0x2cc LR is at do_undefinstr+0xa8/0x250 pc : [<c0101d50>] lr : [<c010a80c>] psr: a0000013 sp : dc8d1f68 ip : 60000013 fp : bedea19c r10: ec532b17 r9 : 00000010 r8 : 0044766c r7 : c0000780 r6 : ec532b17 r5 : c1c13800 r4 : dc8d1fb0 r3 : c10072c4 r2 : c0101c88 r1 : ec532b17 r0 : 0044766c Flags: NzCv IRQs on FIQs on Mode SVC_32 ISA ARM Segment none Control: 00c5387d Table: 0251c008 DAC: 00000051 Register r0 information: non-paged memory Register r1 information: vmalloc memory Register r2 information: non-slab/vmalloc memory Register r3 information: non-slab/vmalloc memory Register r4 information: 2-page vmalloc region Register r5 information: slab kmalloc-cg-2k Register r6 information: vmalloc memory Register r7 information: non-slab/vmalloc memory Register r8 information: non-paged memory Register r9 information: zero-size pointer Register r10 information: vmalloc memory Register r11 information: non-paged memory Register r12 information: non-paged memory Process vfp-reproducer (pid: 196, stack limit = 0x61aaaf8b) Stack: (0xdc8d1f68 to 0xdc8d2000) 1f60: 0000081f b6f69300 0000000f c10073f4 c10072c4 dc8d1fb0 1f80: ec532b17 0c532b17 0044766c b6f9ccd8 00000000 c010a80c 00447670 60000010 1fa0: ffffffff c1c13800 00c5387d c0100f10 b6f68af8 00448fc0 00000000 bedea188 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff 00000000 00000000 Call trace: [<c0101d50>] (vfp_support_entry) from [<c010a80c>] (do_undefinstr+0xa8/0x250) [<c010a80c>] (do_undefinstr) from [<c0100f10>] (__und_usr+0x70/0x80) Exception stack(0xdc8d1fb0 to 0xdc8d1ff8) 1fa0: b6f68af8 00448fc0 00000000 bedea188 1fc0: bedea314 00000001 00448ebc b6f9d000 00447608 b6f9ccd8 00000000 bedea19c 1fe0: bede9198 bedea188 b6e1061c 0044766c 60000010 ffffffff Code: 0a000061 e3877202 e594003c e3a09010 (eef16a10) ---[ end trace 0000000000000000 ]--- Kernel panic - not syncing: Fatal exception in interrupt ---[ end Kernel panic - not syncing: Fatal exception in interrupt ]--- This is a minimal userspace reproducer on a Raspberry Pi Zero W: #include <stdio.h> #include <math.h> int main(void) { double v = 1.0; printf("%f\n", NAN + *(volatile double *)&v); return 0; } Another way to consistently trigger the oops is: calvin@raspberry-pi-zero-w ~$ python -c "import json" The bug reproduces only when the kernel is built with DYNAMIC_DEBUG=n, because the pr_debug() calls act as barriers even when not activated. This is the output from the same kernel source built with the same compiler and DYNAMIC_DEBUG=y, where the userspace reproducer works as expected: VFP: bounce: trigger ec532b17 fpexc c0000780 VFP: emulate: INST=0xee377b06 SCR=0x00000000 VFP: bounce: trigger eef1fa10 fpexc c0000780 VFP: emulate: INST=0xeeb40b40 SCR=0x00000000 VFP: raising exceptions 30000000 calvin@raspberry-pi-zero-w ~$ ./vfp-reproducer nan Crudely grepping for vmsr/vmrs instructions in the otherwise nearly idential text for vfp_support_entry() makes the problem obvious: vmlinux.llvm.good [0xc0101cb8] <+48>: vmrs r7, fpexc vmlinux.llvm.good [0xc0101cd8] <+80>: vmsr fpexc, r0 vmlinux.llvm.good [0xc0101d20] <+152>: vmsr fpexc, r7 vmlinux.llvm.good [0xc0101d38] <+176>: vmrs r4, fpexc vmlinux.llvm.good [0xc0101d6c] <+228>: vmrs r0, fpscr vmlinux.llvm.good [0xc0101dc4] <+316>: vmsr fpexc, r0 vmlinux.llvm.good [0xc0101dc8] <+320>: vmrs r0, fpsid vmlinux.llvm.good [0xc0101dcc] <+324>: vmrs r6, fpscr vmlinux.llvm.good [0xc0101e10] <+392>: vmrs r10, fpinst vmlinux.llvm.good [0xc0101eb8] <+560>: vmrs r10, fpinst2 vmlinux.llvm.bad [0xc0101cb8] <+48>: vmrs r7, fpexc vmlinux.llvm.bad [0xc0101cd8] <+80>: vmsr fpexc, r0 vmlinux.llvm.bad [0xc0101d20] <+152>: vmsr fpexc, r7 vmlinux.llvm.bad [0xc0101d30] <+168>: vmrs r0, fpscr vmlinux.llvm.bad [0xc0101d50] <+200>: vmrs r6, fpscr <== BOOM! vmlinux.llvm.bad [0xc0101d6c] <+228>: vmsr fpexc, r0 vmlinux.llvm.bad [0xc0101d70] <+232>: vmrs r0, fpsid vmlinux.llvm.bad [0xc0101da4] <+284>: vmrs r10, fpinst vmlinux.llvm.bad [0xc0101df8] <+368>: vmrs r4, fpexc vmlinux.llvm.bad [0xc0101e5c] <+468>: vmrs r10, fpinst2 I think LLVM's reordering is valid as the code is currently written: the compiler doesn't know the instructions have side effects in hardware. Fix by using "asm volatile" in fmxr() and fmrx(), so they cannot be reordered with respect to each other. The original compiler now produces working kernels on my hardware with DYNAMIC_DEBUG=n. This is the relevant piece of the diff of the vfp_support_entry() text, from the original oopsing kernel to a working kernel with this patch: vmrs r0, fpscr tst r0, #4096 bne 0xc0101d48 tst r0, #458752 beq 0xc0101ecc orr r7, r7, #536870912 ldr r0, [r4, #0x3c] mov r9, #16 -vmrs r6, fpscr orr r9, r9, #251658240 add r0, r0, #4 str r0, [r4, #0x3c] mvn r0, #159 sub r0, r0, #-1207959552 and r0, r7, r0 vmsr fpexc, r0 vmrs r0, fpsid +vmrs r6, fpscr and r0, r0, #983040 cmp r0, #65536 bne 0xc0101d88 Fixes: 4708fb041346 ("ARM: vfp: Reimplement VFP exception entry in C code") Signed-off-by: Calvin Owens <calvin@wbinvd.org> --- arch/arm/vfp/vfpinstr.h | 48 ++++++++++++++++++++++------------------- 1 file changed, 26 insertions(+), 22 deletions(-)