Message ID | 20240825130943.7738-2-leon.hwang@linux.dev (mailing list archive) |
---|---|
State | Superseded |
Delegated to: | BPF |
Headers | show |
Series | bpf: Fix tailcall infinite loop caused by freplace | expand |
On Sun, 2024-08-25 at 21:09 +0800, Leon Hwang wrote: > This patch fixes a tailcall infinite loop issue caused by freplace. > > Since commit 1c123c567fb1 ("bpf: Resolve fext program type when checking map compatibility"), > freplace prog is allowed to tail call its target prog. Then, when a > freplace prog attaches to its target prog's subprog and tail calls its > target prog, kernel will panic. > > For example: > > tc_bpf2bpf.c: > > // SPDX-License-Identifier: GPL-2.0 > \#include <linux/bpf.h> > \#include <bpf/bpf_helpers.h> > > __noinline > int subprog_tc(struct __sk_buff *skb) > { > return skb->len * 2; > } > > SEC("tc") > int entry_tc(struct __sk_buff *skb) > { > return subprog_tc(skb); > } > > char __license[] SEC("license") = "GPL"; > > tailcall_bpf2bpf_hierarchy_freplace.c: > > // SPDX-License-Identifier: GPL-2.0 > \#include <linux/bpf.h> > \#include <bpf/bpf_helpers.h> > > struct { > __uint(type, BPF_MAP_TYPE_PROG_ARRAY); > __uint(max_entries, 1); > __uint(key_size, sizeof(__u32)); > __uint(value_size, sizeof(__u32)); > } jmp_table SEC(".maps"); > > int count = 0; > > static __noinline > int subprog_tail(struct __sk_buff *skb) > { > bpf_tail_call_static(skb, &jmp_table, 0); > return 0; > } > > SEC("freplace") > int entry_freplace(struct __sk_buff *skb) > { > count++; > subprog_tail(skb); > subprog_tail(skb); > return count; > } > > char __license[] SEC("license") = "GPL"; > > The attach target of entry_freplace is subprog_tc, and the tail callee > in subprog_tail is entry_tc. > > Then, the infinite loop will be entry_tc -> subprog_tc -> entry_freplace > -> subprog_tail --tailcall-> entry_tc, because tail_call_cnt in > entry_freplace will count from zero for every time of entry_freplace > execution. Kernel will panic: > > [ 15.310490] BUG: TASK stack guard page was hit at (____ptrval____) > (stack is (____ptrval____)..(____ptrval____)) > [ 15.310490] Oops: stack guard page: 0000 [#1] PREEMPT SMP NOPTI > [ 15.310490] CPU: 1 PID: 89 Comm: test_progs Tainted: G OE > 6.10.0-rc6-g026dcdae8d3e-dirty #72 > [ 15.310490] Hardware name: QEMU Ubuntu 24.04 PC (i440FX + PIIX, > 1996), BIOS 1.16.3-debian-1.16.3-2 04/01/2014 > [ 15.310490] RIP: 0010:bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 > [ 15.310490] Code: cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc > cc cc cc cc cc f3 0f 1e fa 0f 1f 44 00 00 0f 1f 00 55 48 89 e5 f3 0f 1e > fa <50> 50 53 41 55 48 89 fb 49 bd 00 2a 46 82 98 9c ff ff 48 89 df 4c > [ 15.310490] RSP: 0018:ffffb500c0aa0000 EFLAGS: 00000202 > [ 15.310490] RAX: ffffb500c0aa0028 RBX: ffff9c98808b7e00 RCX: > 0000000000008cb5 > [ 15.310490] RDX: 0000000000000000 RSI: ffff9c9882462a00 RDI: > ffff9c98808b7e00 > [ 15.310490] RBP: ffffb500c0aa0000 R08: 0000000000000000 R09: > 0000000000000000 > [ 15.310490] R10: 0000000000000001 R11: 0000000000000000 R12: > ffffb500c01af000 > [ 15.310490] R13: ffffb500c01cd000 R14: 0000000000000000 R15: > 0000000000000000 > [ 15.310490] FS: 00007f133b665140(0000) GS:ffff9c98bbd00000(0000) > knlGS:0000000000000000 > [ 15.310490] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > [ 15.310490] CR2: ffffb500c0a9fff8 CR3: 0000000102478000 CR4: > 00000000000006f0 > [ 15.310490] Call Trace: > [ 15.310490] <#DF> > [ 15.310490] ? die+0x36/0x90 > [ 15.310490] ? handle_stack_overflow+0x4d/0x60 > [ 15.310490] ? exc_double_fault+0x117/0x1a0 > [ 15.310490] ? asm_exc_double_fault+0x23/0x30 > [ 15.310490] ? bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 > [ 15.310490] </#DF> > [ 15.310490] <TASK> > [ 15.310490] bpf_prog_85781a698094722f_entry+0x4c/0x64 > [ 15.310490] bpf_prog_1c515f389a9059b4_entry2+0x19/0x1b > [ 15.310490] ... > [ 15.310490] bpf_prog_85781a698094722f_entry+0x4c/0x64 > [ 15.310490] bpf_prog_1c515f389a9059b4_entry2+0x19/0x1b > [ 15.310490] bpf_test_run+0x210/0x370 > [ 15.310490] ? bpf_test_run+0x128/0x370 > [ 15.310490] bpf_prog_test_run_skb+0x388/0x7a0 > [ 15.310490] __sys_bpf+0xdbf/0x2c40 > [ 15.310490] ? clockevents_program_event+0x52/0xf0 > [ 15.310490] ? lock_release+0xbf/0x290 > [ 15.310490] __x64_sys_bpf+0x1e/0x30 > [ 15.310490] do_syscall_64+0x68/0x140 > [ 15.310490] entry_SYSCALL_64_after_hwframe+0x76/0x7e > [ 15.310490] RIP: 0033:0x7f133b52725d > [ 15.310490] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa > 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f > 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 8b bb 0d 00 f7 d8 64 89 01 48 > [ 15.310490] RSP: 002b:00007ffddbc10258 EFLAGS: 00000206 ORIG_RAX: > 0000000000000141 > [ 15.310490] RAX: ffffffffffffffda RBX: 00007ffddbc10828 RCX: > 00007f133b52725d > [ 15.310490] RDX: 0000000000000050 RSI: 00007ffddbc102a0 RDI: > 000000000000000a > [ 15.310490] RBP: 00007ffddbc10270 R08: 0000000000000000 R09: > 00007ffddbc102a0 > [ 15.310490] R10: 0000000000000064 R11: 0000000000000206 R12: > 0000000000000004 > [ 15.310490] R13: 0000000000000000 R14: 0000558ec4c24890 R15: > 00007f133b6ed000 > [ 15.310490] </TASK> > [ 15.310490] Modules linked in: bpf_testmod(OE) > [ 15.310490] ---[ end trace 0000000000000000 ]--- > [ 15.310490] RIP: 0010:bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 > [ 15.310490] Code: cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc > cc cc cc cc cc f3 0f 1e fa 0f 1f 44 00 00 0f 1f 00 55 48 89 e5 f3 0f 1e > fa <50> 50 53 41 55 48 89 fb 49 bd 00 2a 46 82 98 9c ff ff 48 89 df 4c > [ 15.310490] RSP: 0018:ffffb500c0aa0000 EFLAGS: 00000202 > [ 15.310490] RAX: ffffb500c0aa0028 RBX: ffff9c98808b7e00 RCX: > 0000000000008cb5 > [ 15.310490] RDX: 0000000000000000 RSI: ffff9c9882462a00 RDI: > ffff9c98808b7e00 > [ 15.310490] RBP: ffffb500c0aa0000 R08: 0000000000000000 R09: > 0000000000000000 > [ 15.310490] R10: 0000000000000001 R11: 0000000000000000 R12: > ffffb500c01af000 > [ 15.310490] R13: ffffb500c01cd000 R14: 0000000000000000 R15: > 0000000000000000 > [ 15.310490] FS: 00007f133b665140(0000) GS:ffff9c98bbd00000(0000) > knlGS:0000000000000000 > [ 15.310490] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 > [ 15.310490] CR2: ffffb500c0a9fff8 CR3: 0000000102478000 CR4: > 00000000000006f0 > [ 15.310490] Kernel panic - not syncing: Fatal exception in interrupt > [ 15.310490] Kernel Offset: 0x30000000 from 0xffffffff81000000 > (relocation range: 0xffffffff80000000-0xffffffffbfffffff) > > This patch fixes the issue by initializing tail_call_cnt at the prologue > of entry_tc. > > Next, when call subprog_tc, the tail_call_cnt_ptr is propagated to > subprog_tc by rax. > Next, when jump to entry_freplace, the tail_call_cnt_ptr will be reused to > count tailcall in freplace prog. > Next, when call subprog_tail, the tail_call_cnt_ptr is propagated to > subprog_tail by rax. > Next, while tail calling to entry_tc, the tail_call_cnt on the stack of > entry_tc increments via the tail_call_cnt_ptr. > > The whole procedure shows as the following JITed prog dumping. > > bpftool p d j n entry_tc: > > int entry_tc(struct __sk_buff * skb): > bpf_prog_1c515f389a9059b4_entry_tc: > ; return subprog_tc(skb); > 0: endbr64 > 4: xorq %rax, %rax ;; rax = 0 (tail_call_cnt) > 7: nopl (%rax,%rax) > c: pushq %rbp > d: movq %rsp, %rbp > 10: endbr64 > 14: cmpq $33, %rax ;; if rax > 33, rax = tcc_ptr > 18: ja 0x20 ;; if rax > 33 goto 0x20 ---+ > 1a: pushq %rax ;; [rbp - 8] = rax = 0 | > 1b: movq %rsp, %rax ;; rax = rbp - 8 | > 1e: jmp 0x21 ;; ---------+ | > 20: pushq %rax ;; <--------|---------------+ > 21: pushq %rax ;; <--------+ [rbp - 16] = rax > 22: movq -16(%rbp), %rax ;; rax = tcc_ptr > 29: callq 0x70 ;; call subprog_tc() > ; return subprog_tc(skb); > 2e: leave > 2f: retq > > int subprog_tc(struct __sk_buff * skb): > bpf_prog_1e8f76e2374a0607_subprog_tc: > ; return skb->len * 2; > 0: endbr64 > 4: nopl (%rax) ;; do not touch tail_call_cnt > 7: jmp 0x108 ;; jump to entry_freplace() > c: pushq %rbp > d: movq %rsp, %rbp > 10: endbr64 > 14: pushq %rax > 15: pushq %rax > 16: movl 112(%rdi), %eax > ; return skb->len * 2; > 19: shll %eax > ; return skb->len * 2; > 1b: leave > 1c: retq > > bpftool p d j n entry_freplace: > > int entry_freplace(struct __sk_buff * skb): > bpf_prog_85781a698094722f_entry_freplace: > ; int entry_freplace(struct __sk_buff *skb) > 0: endbr64 > 4: nopl (%rax) ;; do not touch tail_call_cnt > 7: nopl (%rax,%rax) > c: pushq %rbp > d: movq %rsp, %rbp > 10: endbr64 > 14: cmpq $33, %rax ;; if rax > 33, rax = tcc_ptr > 18: ja 0x20 ;; if rax > 33 goto 0x20 ---+ > 1a: pushq %rax ;; [rbp - 8] = rax = 0 | > 1b: movq %rsp, %rax ;; rax = rbp - 8 | > 1e: jmp 0x21 ;; ---------+ | > 20: pushq %rax ;; <--------|---------------+ > 21: pushq %rax ;; <--------+ [rbp - 16] = rax > 22: pushq %rbx ;; callee saved > 23: pushq %r13 ;; callee saved > 25: movq %rdi, %rbx ;; rbx = skb (callee saved) > ; count++; > 28: movabsq $-123406219759616, %r13 > 32: movl (%r13), %edi > 36: addl $1, %edi > 39: movl %edi, (%r13) > ; subprog_tail(skb); > 3d: movq %rbx, %rdi ;; rdi = skb > 40: movq -16(%rbp), %rax ;; rax = tcc_ptr > 47: callq 0x80 ;; call subprog_tail() > ; subprog_tail(skb); > 4c: movq %rbx, %rdi ;; rdi = skb > 4f: movq -16(%rbp), %rax ;; rax = tcc_ptr > 56: callq 0x80 ;; call subprog_tail() > ; return count; > 5b: movl (%r13), %eax > ; return count; > 5f: popq %r13 > 61: popq %rbx > 62: leave > 63: retq > > int subprog_tail(struct __sk_buff * skb): > bpf_prog_3a140cef239a4b4f_subprog_tail: > ; int subprog_tail(struct __sk_buff *skb) > 0: endbr64 > 4: nopl (%rax) ;; do not touch tail_call_cnt > 7: nopl (%rax,%rax) > c: pushq %rbp > d: movq %rsp, %rbp > 10: endbr64 > 14: pushq %rax ;; [rbp - 8] = rax (tcc_ptr) > 15: pushq %rax ;; [rbp - 16] = rax (tcc_ptr) > 16: pushq %rbx ;; callee saved > 17: pushq %r13 ;; callee saved > 19: movq %rdi, %rbx ;; rbx = skb > ; asm volatile("r1 = %[ctx]\n\t" > 1c: movabsq $-128494642337280, %r13 ;; r13 = jmp_table > 26: movq %rbx, %rdi ;; 1st arg, skb > 29: movq %r13, %rsi ;; 2nd arg, jmp_table > 2c: xorl %edx, %edx ;; 3rd arg, index = 0 > 2e: movq -16(%rbp), %rax ;; rax = [rbp - 16] (tcc_ptr) > 35: cmpq $33, (%rax) > 39: jae 0x4e ;; if *tcc_ptr >= 33 goto 0x4e --------+ > 3b: nopl (%rax,%rax) ;; jmp bypass, toggled by poking | > 40: addq $1, (%rax) ;; (*tcc_ptr)++ | > 44: popq %r13 ;; callee saved | > 46: popq %rbx ;; callee saved | > 47: popq %rax ;; undo rbp-16 push | > 48: popq %rax ;; undo rbp-8 push | > 49: jmp 0xfffffffffffffe18 ;; tail call target, toggled by poking | > ; return 0; ;; | > 4e: popq %r13 ;; restore callee saved <--------------+ > 50: popq %rbx ;; restore callee saved > 51: leave > 52: retq > > As a result, the tail_call_cnt is stored on the stack of entry_tc. And > the tail_call_cnt_ptr is propagated between subprog_tc, entry_freplace, > subprog_tail and entry_tc. > > Furthermore, trampoline is required to propagate > tail_call_cnt/tail_call_cnt_ptr always, no matter whether there is > tailcall at run time. > So, it reuses trampoline flag "BIT(7)" to tell trampoline to propagate > the tail_call_cnt/tail_call_cnt_ptr, as BPF_TRAMP_F_TAIL_CALL_CTX is not > used by any other arch BPF JIT. This change seem to be correct. Could you please add an explanation somewhere why nop3/xor and nop5 are swapped in the prologue? As far as I understand, this is done so that freplace program would reuse xor generated for replaced program, is that right? E.g. for tailcall_bpf2bpf_freplace test case disasm looks as follows: --------------- entry_tc -------------- func #0: 0: f3 0f 1e fa endbr64 4: 48 31 c0 xorq %rax, %rax 7: 0f 1f 44 00 00 nopl (%rax,%rax) c: 55 pushq %rbp d: 48 89 e5 movq %rsp, %rbp 10: f3 0f 1e fa endbr64 ... ------------ entry_freplace ----------- func #0: 0: f3 0f 1e fa endbr64 4: 0f 1f 00 nopl (%rax) 7: 0f 1f 44 00 00 nopl (%rax,%rax) c: 55 pushq %rbp d: 48 89 e5 movq %rsp, %rbp ... So, if entry_freplace would be used to replace entry_tc instead of subprog_tc, the disasm would change to: --------------- entry_tc -------------- func #0: 0: f3 0f 1e fa endbr64 4: 48 31 c0 xorq %rax, %rax 7: 0f 1f 44 00 00 jmp <entry_freplace> Thus reusing %rax initialization from entry_tc. > However, the bad effect is that it requires initializing tail_call_cnt at > prologue always no matter whether the prog is tail_call_reachable, because > it is unable to confirm itself or its subprog[s] whether to be attached by > freplace prog. > And, when call subprog, tail_call_cnt_ptr is required to be propagated > to subprog always. This seems unfortunate. I wonder if disallowing to freplace programs when replacement.tail_call_reachable != replaced.tail_call_reachable would be a better option? [...]
On 2024/8/27 18:37, Eduard Zingerman wrote: > On Sun, 2024-08-25 at 21:09 +0800, Leon Hwang wrote: >> This patch fixes a tailcall infinite loop issue caused by freplace. >> >> Since commit 1c123c567fb1 ("bpf: Resolve fext program type when checking map compatibility"), >> freplace prog is allowed to tail call its target prog. Then, when a >> freplace prog attaches to its target prog's subprog and tail calls its >> target prog, kernel will panic. >> [...] >> >> As a result, the tail_call_cnt is stored on the stack of entry_tc. And >> the tail_call_cnt_ptr is propagated between subprog_tc, entry_freplace, >> subprog_tail and entry_tc. >> >> Furthermore, trampoline is required to propagate >> tail_call_cnt/tail_call_cnt_ptr always, no matter whether there is >> tailcall at run time. >> So, it reuses trampoline flag "BIT(7)" to tell trampoline to propagate >> the tail_call_cnt/tail_call_cnt_ptr, as BPF_TRAMP_F_TAIL_CALL_CTX is not >> used by any other arch BPF JIT. > > This change seem to be correct. > Could you please add an explanation somewhere why nop3/xor and nop5 > are swapped in the prologue? OK, I'll explain it in message of patch v2. > > As far as I understand, this is done so that freplace program > would reuse xor generated for replaced program, is that right? > E.g. for tailcall_bpf2bpf_freplace test case disasm looks as follows: > > --------------- entry_tc -------------- > func #0: > 0: f3 0f 1e fa endbr64 > 4: 48 31 c0 xorq %rax, %rax > 7: 0f 1f 44 00 00 nopl (%rax,%rax) > c: 55 pushq %rbp > d: 48 89 e5 movq %rsp, %rbp > 10: f3 0f 1e fa endbr64 > ... > > ------------ entry_freplace ----------- > func #0: > 0: f3 0f 1e fa endbr64 > 4: 0f 1f 00 nopl (%rax) > 7: 0f 1f 44 00 00 nopl (%rax,%rax) > c: 55 pushq %rbp > d: 48 89 e5 movq %rsp, %rbp > ... > > So, if entry_freplace would be used to replace entry_tc instead > of subprog_tc, the disasm would change to: > > --------------- entry_tc -------------- > func #0: > 0: f3 0f 1e fa endbr64 > 4: 48 31 c0 xorq %rax, %rax > 7: 0f 1f 44 00 00 jmp <entry_freplace> > > Thus reusing %rax initialization from entry_tc. > Great. You understand it correctly. >> However, the bad effect is that it requires initializing tail_call_cnt at >> prologue always no matter whether the prog is tail_call_reachable, because >> it is unable to confirm itself or its subprog[s] whether to be attached by >> freplace prog. >> And, when call subprog, tail_call_cnt_ptr is required to be propagated >> to subprog always. > > This seems unfortunate. > I wonder if disallowing to freplace programs when > replacement.tail_call_reachable != replaced.tail_call_reachable > would be a better option? > This idea is wonderful. We can disallow attaching tail_call_reachable freplace prog to not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. 2. attach not-tail_call_reachable freplace prog to tail_call_reachable bpf prog. 3. attach not-tail_call_reachable freplace prog to not-tail_call_reachable bpf prog. I would like to add this limit in patch v2, that tail_call_reachable freplace is disallowed to attach to not-tail_call_reachable bpf prog. Thereafter, tail_call_cnt_ptr is required to be propagated to subprog only when subprog is tail_call_reachable. Thanks, Leon
On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: > > > I wonder if disallowing to freplace programs when > > replacement.tail_call_reachable != replaced.tail_call_reachable > > would be a better option? > > > > This idea is wonderful. > > We can disallow attaching tail_call_reachable freplace prog to > not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. > > 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. > 2. attach not-tail_call_reachable freplace prog to tail_call_reachable > bpf prog. > 3. attach not-tail_call_reachable freplace prog to > not-tail_call_reachable bpf prog. I think it's fine to disable freplace and tail_call combination altogether. And speaking of the patch. The following: - if (tail_call_reachable) { - LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); - ip += 7; - } + LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); + ip += 7; Is too high of a penalty for every call for freplace+tail_call combo. So disable it in the verifier. pw-bot: cr
On 28/8/24 04:50, Alexei Starovoitov wrote: > On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: >> >>> I wonder if disallowing to freplace programs when >>> replacement.tail_call_reachable != replaced.tail_call_reachable >>> would be a better option? >>> >> >> This idea is wonderful. >> >> We can disallow attaching tail_call_reachable freplace prog to >> not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. >> >> 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. >> 2. attach not-tail_call_reachable freplace prog to tail_call_reachable >> bpf prog. >> 3. attach not-tail_call_reachable freplace prog to >> not-tail_call_reachable bpf prog. > > I think it's fine to disable freplace and tail_call combination altogether. I don't think so. My XDP project heavily relies on freplace and tailcall combination. > > And speaking of the patch. The following: > - if (tail_call_reachable) { > - > LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); > - ip += 7; > - } > + LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); > + ip += 7; > > Is too high of a penalty for every call for freplace+tail_call combo. > > So disable it in the verifier. > I think, it's enough to disallow attaching tail_call_reachable freplace prog to not-tail_call_reachable prog in verifier. As for this code snippet in x64 JIT: func = (u8 *) __bpf_call_base + imm32; if (tail_call_reachable) { LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); ip += 7; } if (!imm32) return -EINVAL; ip += x86_call_depth_emit_accounting(&prog, func, ip); if (emit_call(&prog, func, ip)) return -EINVAL; when a subprog is tail_call_reachable, its caller has to propagate tail_call_cnt_ptr by rax. It's fine to attach tail_call_reachable freplace prog to this subprog as for this case. When the subprog is not tail_call_reachable, its caller is unnecessary to propagate tail_call_cnt_ptr by rax. Then it's disallowed to attach tail_call_reachable freplace prog to the subprog. However, it's fine to attach not-tail_call_reachable freplace prog to the subprog. In conclusion, if disallow attaching tail_call_reachable freplace prog to not-tail_call_reachable prog in verifier, the above code snippet won't be changed. Thanks, Leon
On Tue, Aug 27, 2024 at 7:36 PM Leon Hwang <leon.hwang@linux.dev> wrote: > > > > On 28/8/24 04:50, Alexei Starovoitov wrote: > > On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: > >> > >>> I wonder if disallowing to freplace programs when > >>> replacement.tail_call_reachable != replaced.tail_call_reachable > >>> would be a better option? > >>> > >> > >> This idea is wonderful. > >> > >> We can disallow attaching tail_call_reachable freplace prog to > >> not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. > >> > >> 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. > >> 2. attach not-tail_call_reachable freplace prog to tail_call_reachable > >> bpf prog. > >> 3. attach not-tail_call_reachable freplace prog to > >> not-tail_call_reachable bpf prog. > > > > I think it's fine to disable freplace and tail_call combination altogether. > > I don't think so. > > My XDP project heavily relies on freplace and tailcall combination. Pls share the link to the code. > > > > And speaking of the patch. The following: > > - if (tail_call_reachable) { > > - > > LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); > > - ip += 7; > > - } > > + LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); > > + ip += 7; > > > > Is too high of a penalty for every call for freplace+tail_call combo. > > > > So disable it in the verifier. > > > > I think, it's enough to disallow attaching tail_call_reachable freplace > prog to not-tail_call_reachable prog in verifier. > > As for this code snippet in x64 JIT: > > func = (u8 *) __bpf_call_base + imm32; > if (tail_call_reachable) { > LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); > ip += 7; > } > if (!imm32) > return -EINVAL; > ip += x86_call_depth_emit_accounting(&prog, func, ip); > if (emit_call(&prog, func, ip)) > return -EINVAL; > > when a subprog is tail_call_reachable, its caller has to propagate > tail_call_cnt_ptr by rax. It's fine to attach tail_call_reachable > freplace prog to this subprog as for this case. > > When the subprog is not tail_call_reachable, its caller is unnecessary > to propagate tail_call_cnt_ptr by rax. Then it's disallowed to attach > tail_call_reachable freplace prog to the subprog. However, it's fine to > attach not-tail_call_reachable freplace prog to the subprog. > > In conclusion, if disallow attaching tail_call_reachable freplace prog > to not-tail_call_reachable prog in verifier, the above code snippet > won't be changed. As long as there are no more JIT changes it's ok to go with this partial verifier restriction, but if more issues are found we'll have to restrict it further.
On 29/8/24 00:01, Alexei Starovoitov wrote: > On Tue, Aug 27, 2024 at 7:36 PM Leon Hwang <leon.hwang@linux.dev> wrote: >> >> >> >> On 28/8/24 04:50, Alexei Starovoitov wrote: >>> On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: >>>> >>>>> I wonder if disallowing to freplace programs when >>>>> replacement.tail_call_reachable != replaced.tail_call_reachable >>>>> would be a better option? >>>>> >>>> >>>> This idea is wonderful. >>>> >>>> We can disallow attaching tail_call_reachable freplace prog to >>>> not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. >>>> >>>> 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. >>>> 2. attach not-tail_call_reachable freplace prog to tail_call_reachable >>>> bpf prog. >>>> 3. attach not-tail_call_reachable freplace prog to >>>> not-tail_call_reachable bpf prog. >>> >>> I think it's fine to disable freplace and tail_call combination altogether. >> >> I don't think so. >> >> My XDP project heavily relies on freplace and tailcall combination. > > Pls share the link to the code. > I'm willing to share it with you. But it's an in-house project of my company. Sorry. >>> >>> And speaking of the patch. The following: >>> - if (tail_call_reachable) { >>> - >>> LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); >>> - ip += 7; >>> - } >>> + LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); >>> + ip += 7; >>> >>> Is too high of a penalty for every call for freplace+tail_call combo. >>> >>> So disable it in the verifier. >>> >> >> I think, it's enough to disallow attaching tail_call_reachable freplace >> prog to not-tail_call_reachable prog in verifier. >> >> As for this code snippet in x64 JIT: >> >> func = (u8 *) __bpf_call_base + imm32; >> if (tail_call_reachable) { >> LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); >> ip += 7; >> } >> if (!imm32) >> return -EINVAL; >> ip += x86_call_depth_emit_accounting(&prog, func, ip); >> if (emit_call(&prog, func, ip)) >> return -EINVAL; >> >> when a subprog is tail_call_reachable, its caller has to propagate >> tail_call_cnt_ptr by rax. It's fine to attach tail_call_reachable >> freplace prog to this subprog as for this case. >> >> When the subprog is not tail_call_reachable, its caller is unnecessary >> to propagate tail_call_cnt_ptr by rax. Then it's disallowed to attach >> tail_call_reachable freplace prog to the subprog. However, it's fine to >> attach not-tail_call_reachable freplace prog to the subprog. >> >> In conclusion, if disallow attaching tail_call_reachable freplace prog >> to not-tail_call_reachable prog in verifier, the above code snippet >> won't be changed. > > As long as there are no more JIT changes it's ok to go > with this partial verifier restriction, > but if more issues are found we'll have to restrict it further. OK. I'll do the restriction in verifier. Thanks, Leon
Alexei Starovoitov <alexei.starovoitov@gmail.com> writes: > On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: >> >> > I wonder if disallowing to freplace programs when >> > replacement.tail_call_reachable != replaced.tail_call_reachable >> > would be a better option? >> > >> >> This idea is wonderful. >> >> We can disallow attaching tail_call_reachable freplace prog to >> not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. >> >> 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. >> 2. attach not-tail_call_reachable freplace prog to tail_call_reachable >> bpf prog. >> 3. attach not-tail_call_reachable freplace prog to >> not-tail_call_reachable bpf prog. > > I think it's fine to disable freplace and tail_call combination > altogether. In the libxdp dispatcher we rely on the fact that an freplace program is equivalent to a directly attached XDP program. And we've definitely seen people using tail calls along with the libxdp dispatcher (e.g., https://github.com/xdp-project/xdp-tools/issues/377), so I don't think it's a good idea to disable it entirely. I think restricting the combinations should be fine, though - the libxdp dispatcher will not end up in a tail call map unless someone is going out of their way to do weird things :) -Toke
On Mon, Sep 2, 2024 at 3:20 AM Toke Høiland-Jørgensen <toke@redhat.com> wrote: > > Alexei Starovoitov <alexei.starovoitov@gmail.com> writes: > > > On Tue, Aug 27, 2024 at 5:48 AM Leon Hwang <leon.hwang@linux.dev> wrote: > >> > >> > I wonder if disallowing to freplace programs when > >> > replacement.tail_call_reachable != replaced.tail_call_reachable > >> > would be a better option? > >> > > >> > >> This idea is wonderful. > >> > >> We can disallow attaching tail_call_reachable freplace prog to > >> not-tail_call_reachable bpf prog. So, the following 3 cases are allowed. > >> > >> 1. attach tail_call_reachable freplace prog to tail_call_reachable bpf prog. > >> 2. attach not-tail_call_reachable freplace prog to tail_call_reachable > >> bpf prog. > >> 3. attach not-tail_call_reachable freplace prog to > >> not-tail_call_reachable bpf prog. > > > > I think it's fine to disable freplace and tail_call combination > > altogether. > > In the libxdp dispatcher we rely on the fact that an freplace program is > equivalent to a directly attached XDP program. And we've definitely seen > people using tail calls along with the libxdp dispatcher (e.g., > https://github.com/xdp-project/xdp-tools/issues/377), so I don't think > it's a good idea to disable it entirely. > Thanks Toke to mention this use case, I have xdp-loader to load DNS XDP program with tail calls to do DNS ratelimit and DNS cookie verification see here https://github.com/vincentmli/xdp-tools/blob/vli-xdp-synproxy/xdp-dnsrrl/xdp_dnsrrl.bpf.c#L635-L644 and I have it as part of XDP DDoS in an open source firewall project https://github.com/vincentmli/BPFire. I hope this is continued to be supported in future :) > I think restricting the combinations should be fine, though - the libxdp > dispatcher will not end up in a tail call map unless someone is going > out of their way to do weird things :) > > -Toke > >
diff --git a/arch/x86/net/bpf_jit_comp.c b/arch/x86/net/bpf_jit_comp.c index 074b41fafbe3f..85d9af13f329f 100644 --- a/arch/x86/net/bpf_jit_comp.c +++ b/arch/x86/net/bpf_jit_comp.c @@ -274,6 +274,8 @@ struct jit_context { #define X86_PATCH_SIZE 5 /* Number of bytes that will be skipped on tailcall */ #define X86_TAIL_CALL_OFFSET (12 + ENDBR_INSN_SIZE) +/* Number of extra bytes that will be skipped on poke */ +#define X86_POKE_EXTRA 3 static void push_r12(u8 **pprog) { @@ -440,18 +442,14 @@ static void emit_prologue_tail_call(u8 **pprog, bool is_subprog) * while jumping to another program */ static void emit_prologue(u8 **pprog, u32 stack_depth, bool ebpf_from_cbpf, - bool tail_call_reachable, bool is_subprog, + bool is_extension, bool is_subprog, bool is_exception_cb) { u8 *prog = *pprog; emit_cfi(&prog, is_subprog ? cfi_bpf_subprog_hash : cfi_bpf_hash); - /* BPF trampoline can be made to work without these nops, - * but let's waste 5 bytes for now and optimize later - */ - emit_nops(&prog, X86_PATCH_SIZE); if (!ebpf_from_cbpf) { - if (tail_call_reachable && !is_subprog) + if (!is_extension && !is_subprog) /* When it's the entry of the whole tailcall context, * zeroing rax means initialising tail_call_cnt. */ @@ -460,6 +458,10 @@ static void emit_prologue(u8 **pprog, u32 stack_depth, bool ebpf_from_cbpf, /* Keep the same instruction layout. */ emit_nops(&prog, 3); /* nop3 */ } + /* BPF trampoline can be made to work without these nops, + * but let's waste 5 bytes for now and optimize later + */ + emit_nops(&prog, X86_PATCH_SIZE); /* Exception callback receives FP as third parameter */ if (is_exception_cb) { EMIT3(0x48, 0x89, 0xF4); /* mov rsp, rsi */ @@ -483,7 +485,7 @@ static void emit_prologue(u8 **pprog, u32 stack_depth, bool ebpf_from_cbpf, /* sub rsp, rounded_stack_depth */ if (stack_depth) EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); - if (tail_call_reachable) + if (!ebpf_from_cbpf) emit_prologue_tail_call(&prog, is_subprog); *pprog = prog; } @@ -573,10 +575,13 @@ int bpf_arch_text_poke(void *ip, enum bpf_text_poke_type t, /* * See emit_prologue(), for IBT builds the trampoline hook is preceded - * with an ENDBR instruction. + * with an ENDBR instruction and 3 bytes tail_call_cnt initialization + * instruction. */ if (is_endbr(*(u32 *)ip)) ip += ENDBR_INSN_SIZE; + if (is_bpf_text_address((long)ip)) + ip += X86_POKE_EXTRA; return __bpf_arch_text_poke(ip, t, old_addr, new_addr); } @@ -1365,7 +1370,7 @@ static void emit_shiftx(u8 **pprog, u32 dst_reg, u8 src_reg, bool is64, u8 op) static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image, u8 *rw_image, int oldproglen, struct jit_context *ctx, bool jmp_padding) { - bool tail_call_reachable = bpf_prog->aux->tail_call_reachable; + bool is_extension = bpf_prog->type == BPF_PROG_TYPE_EXT; struct bpf_insn *insn = bpf_prog->insnsi; bool callee_regs_used[4] = {}; int insn_cnt = bpf_prog->len; @@ -1383,7 +1388,7 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image, u8 *rw_image detect_reg_usage(insn, insn_cnt, callee_regs_used); emit_prologue(&prog, bpf_prog->aux->stack_depth, - bpf_prog_was_classic(bpf_prog), tail_call_reachable, + bpf_prog_was_classic(bpf_prog), is_extension, bpf_is_subprog(bpf_prog), bpf_prog->aux->exception_cb); /* Exception callback will clobber callee regs for its own use, and * restore the original callee regs from main prog's stack frame. @@ -2077,10 +2082,8 @@ st: if (is_imm8(insn->off)) u8 *ip = image + addrs[i - 1]; func = (u8 *) __bpf_call_base + imm32; - if (tail_call_reachable) { - LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); - ip += 7; - } + LOAD_TAIL_CALL_CNT_PTR(bpf_prog->aux->stack_depth); + ip += 7; if (!imm32) return -EINVAL; ip += x86_call_depth_emit_accounting(&prog, func, ip); @@ -2877,7 +2880,7 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im * [ ... ] * [ stack_arg2 ] * RBP - arg_stack_off [ stack_arg1 ] - * RSP [ tail_call_cnt_ptr ] BPF_TRAMP_F_TAIL_CALL_CTX + * RSP [ tail_call_cnt_ptr ] BPF_TRAMP_F_TAIL_CALL */ /* room for return value of orig_call or fentry prog */ @@ -2923,6 +2926,8 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im */ if (is_endbr(*(u32 *)orig_call)) orig_call += ENDBR_INSN_SIZE; + if (is_bpf_text_address((long)orig_call)) + orig_call += X86_POKE_EXTRA; orig_call += X86_PATCH_SIZE; } @@ -2949,8 +2954,8 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im /* sub rsp, stack_size */ EMIT4(0x48, 0x83, 0xEC, stack_size); } - if (flags & BPF_TRAMP_F_TAIL_CALL_CTX) - EMIT1(0x50); /* push rax */ + if (flags & BPF_TRAMP_F_TAIL_CALL) + EMIT1(0x50); /* push rax */ /* mov QWORD PTR [rbp - rbx_off], rbx */ emit_stx(&prog, BPF_DW, BPF_REG_FP, BPF_REG_6, -rbx_off); @@ -3005,7 +3010,7 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im restore_regs(m, &prog, regs_off); save_args(m, &prog, arg_stack_off, true); - if (flags & BPF_TRAMP_F_TAIL_CALL_CTX) { + if (flags & BPF_TRAMP_F_TAIL_CALL) { /* Before calling the original function, load the * tail_call_cnt_ptr from stack to rax. */ @@ -3025,6 +3030,7 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im /* remember return value in a stack for bpf prog to access */ emit_stx(&prog, BPF_DW, BPF_REG_FP, BPF_REG_0, -8); im->ip_after_call = image + (prog - (u8 *)rw_image); + emit_nops(&prog, X86_POKE_EXTRA); emit_nops(&prog, X86_PATCH_SIZE); } @@ -3067,7 +3073,7 @@ static int __arch_prepare_bpf_trampoline(struct bpf_tramp_image *im, void *rw_im ret = -EINVAL; goto cleanup; } - } else if (flags & BPF_TRAMP_F_TAIL_CALL_CTX) { + } else if (flags & BPF_TRAMP_F_TAIL_CALL) { /* Before running the original function, load the * tail_call_cnt_ptr from stack to rax. */ diff --git a/include/linux/bpf.h b/include/linux/bpf.h index 00dc4dd28cbd9..aa4a2799eaed1 100644 --- a/include/linux/bpf.h +++ b/include/linux/bpf.h @@ -1074,10 +1074,10 @@ struct btf_func_model { */ #define BPF_TRAMP_F_SHARE_IPMODIFY BIT(6) -/* Indicate that current trampoline is in a tail call context. Then, it has to - * cache and restore tail_call_cnt to avoid infinite tail call loop. +/* Indicate that current trampoline is required to cache and restore + * tail_call_cnt or tail_call_cnt_ptr to avoid infinite tail call loop. */ -#define BPF_TRAMP_F_TAIL_CALL_CTX BIT(7) +#define BPF_TRAMP_F_TAIL_CALL BIT(7) /* * Indicate the trampoline should be suitable to receive indirect calls; diff --git a/kernel/bpf/trampoline.c b/kernel/bpf/trampoline.c index f8302a5ca400d..257adbc83bae3 100644 --- a/kernel/bpf/trampoline.c +++ b/kernel/bpf/trampoline.c @@ -409,8 +409,8 @@ static int bpf_trampoline_update(struct bpf_trampoline *tr, bool lock_direct_mut goto out; } - /* clear all bits except SHARE_IPMODIFY and TAIL_CALL_CTX */ - tr->flags &= (BPF_TRAMP_F_SHARE_IPMODIFY | BPF_TRAMP_F_TAIL_CALL_CTX); + /* clear all bits except SHARE_IPMODIFY and TAIL_CALL */ + tr->flags &= (BPF_TRAMP_F_SHARE_IPMODIFY | BPF_TRAMP_F_TAIL_CALL); if (tlinks[BPF_TRAMP_FEXIT].nr_links || tlinks[BPF_TRAMP_MODIFY_RETURN].nr_links) { diff --git a/kernel/bpf/verifier.c b/kernel/bpf/verifier.c index 1485364464576..aecd4d7cdf205 100644 --- a/kernel/bpf/verifier.c +++ b/kernel/bpf/verifier.c @@ -22096,8 +22096,8 @@ static int check_attach_btf_id(struct bpf_verifier_env *env) if (!tr) return -ENOMEM; - if (tgt_prog && tgt_prog->aux->tail_call_reachable) - tr->flags = BPF_TRAMP_F_TAIL_CALL_CTX; + if (tgt_prog) + tr->flags = BPF_TRAMP_F_TAIL_CALL; prog->aux->dst_trampoline = tr; return 0;
This patch fixes a tailcall infinite loop issue caused by freplace. Since commit 1c123c567fb1 ("bpf: Resolve fext program type when checking map compatibility"), freplace prog is allowed to tail call its target prog. Then, when a freplace prog attaches to its target prog's subprog and tail calls its target prog, kernel will panic. For example: tc_bpf2bpf.c: // SPDX-License-Identifier: GPL-2.0 \#include <linux/bpf.h> \#include <bpf/bpf_helpers.h> __noinline int subprog_tc(struct __sk_buff *skb) { return skb->len * 2; } SEC("tc") int entry_tc(struct __sk_buff *skb) { return subprog_tc(skb); } char __license[] SEC("license") = "GPL"; tailcall_bpf2bpf_hierarchy_freplace.c: // SPDX-License-Identifier: GPL-2.0 \#include <linux/bpf.h> \#include <bpf/bpf_helpers.h> struct { __uint(type, BPF_MAP_TYPE_PROG_ARRAY); __uint(max_entries, 1); __uint(key_size, sizeof(__u32)); __uint(value_size, sizeof(__u32)); } jmp_table SEC(".maps"); int count = 0; static __noinline int subprog_tail(struct __sk_buff *skb) { bpf_tail_call_static(skb, &jmp_table, 0); return 0; } SEC("freplace") int entry_freplace(struct __sk_buff *skb) { count++; subprog_tail(skb); subprog_tail(skb); return count; } char __license[] SEC("license") = "GPL"; The attach target of entry_freplace is subprog_tc, and the tail callee in subprog_tail is entry_tc. Then, the infinite loop will be entry_tc -> subprog_tc -> entry_freplace -> subprog_tail --tailcall-> entry_tc, because tail_call_cnt in entry_freplace will count from zero for every time of entry_freplace execution. Kernel will panic: [ 15.310490] BUG: TASK stack guard page was hit at (____ptrval____) (stack is (____ptrval____)..(____ptrval____)) [ 15.310490] Oops: stack guard page: 0000 [#1] PREEMPT SMP NOPTI [ 15.310490] CPU: 1 PID: 89 Comm: test_progs Tainted: G OE 6.10.0-rc6-g026dcdae8d3e-dirty #72 [ 15.310490] Hardware name: QEMU Ubuntu 24.04 PC (i440FX + PIIX, 1996), BIOS 1.16.3-debian-1.16.3-2 04/01/2014 [ 15.310490] RIP: 0010:bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 [ 15.310490] Code: cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc f3 0f 1e fa 0f 1f 44 00 00 0f 1f 00 55 48 89 e5 f3 0f 1e fa <50> 50 53 41 55 48 89 fb 49 bd 00 2a 46 82 98 9c ff ff 48 89 df 4c [ 15.310490] RSP: 0018:ffffb500c0aa0000 EFLAGS: 00000202 [ 15.310490] RAX: ffffb500c0aa0028 RBX: ffff9c98808b7e00 RCX: 0000000000008cb5 [ 15.310490] RDX: 0000000000000000 RSI: ffff9c9882462a00 RDI: ffff9c98808b7e00 [ 15.310490] RBP: ffffb500c0aa0000 R08: 0000000000000000 R09: 0000000000000000 [ 15.310490] R10: 0000000000000001 R11: 0000000000000000 R12: ffffb500c01af000 [ 15.310490] R13: ffffb500c01cd000 R14: 0000000000000000 R15: 0000000000000000 [ 15.310490] FS: 00007f133b665140(0000) GS:ffff9c98bbd00000(0000) knlGS:0000000000000000 [ 15.310490] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 15.310490] CR2: ffffb500c0a9fff8 CR3: 0000000102478000 CR4: 00000000000006f0 [ 15.310490] Call Trace: [ 15.310490] <#DF> [ 15.310490] ? die+0x36/0x90 [ 15.310490] ? handle_stack_overflow+0x4d/0x60 [ 15.310490] ? exc_double_fault+0x117/0x1a0 [ 15.310490] ? asm_exc_double_fault+0x23/0x30 [ 15.310490] ? bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 [ 15.310490] </#DF> [ 15.310490] <TASK> [ 15.310490] bpf_prog_85781a698094722f_entry+0x4c/0x64 [ 15.310490] bpf_prog_1c515f389a9059b4_entry2+0x19/0x1b [ 15.310490] ... [ 15.310490] bpf_prog_85781a698094722f_entry+0x4c/0x64 [ 15.310490] bpf_prog_1c515f389a9059b4_entry2+0x19/0x1b [ 15.310490] bpf_test_run+0x210/0x370 [ 15.310490] ? bpf_test_run+0x128/0x370 [ 15.310490] bpf_prog_test_run_skb+0x388/0x7a0 [ 15.310490] __sys_bpf+0xdbf/0x2c40 [ 15.310490] ? clockevents_program_event+0x52/0xf0 [ 15.310490] ? lock_release+0xbf/0x290 [ 15.310490] __x64_sys_bpf+0x1e/0x30 [ 15.310490] do_syscall_64+0x68/0x140 [ 15.310490] entry_SYSCALL_64_after_hwframe+0x76/0x7e [ 15.310490] RIP: 0033:0x7f133b52725d [ 15.310490] Code: ff c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48 89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73 01 c3 48 8b 0d 8b bb 0d 00 f7 d8 64 89 01 48 [ 15.310490] RSP: 002b:00007ffddbc10258 EFLAGS: 00000206 ORIG_RAX: 0000000000000141 [ 15.310490] RAX: ffffffffffffffda RBX: 00007ffddbc10828 RCX: 00007f133b52725d [ 15.310490] RDX: 0000000000000050 RSI: 00007ffddbc102a0 RDI: 000000000000000a [ 15.310490] RBP: 00007ffddbc10270 R08: 0000000000000000 R09: 00007ffddbc102a0 [ 15.310490] R10: 0000000000000064 R11: 0000000000000206 R12: 0000000000000004 [ 15.310490] R13: 0000000000000000 R14: 0000558ec4c24890 R15: 00007f133b6ed000 [ 15.310490] </TASK> [ 15.310490] Modules linked in: bpf_testmod(OE) [ 15.310490] ---[ end trace 0000000000000000 ]--- [ 15.310490] RIP: 0010:bpf_prog_3a140cef239a4b4f_subprog_tail+0x14/0x53 [ 15.310490] Code: cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc cc f3 0f 1e fa 0f 1f 44 00 00 0f 1f 00 55 48 89 e5 f3 0f 1e fa <50> 50 53 41 55 48 89 fb 49 bd 00 2a 46 82 98 9c ff ff 48 89 df 4c [ 15.310490] RSP: 0018:ffffb500c0aa0000 EFLAGS: 00000202 [ 15.310490] RAX: ffffb500c0aa0028 RBX: ffff9c98808b7e00 RCX: 0000000000008cb5 [ 15.310490] RDX: 0000000000000000 RSI: ffff9c9882462a00 RDI: ffff9c98808b7e00 [ 15.310490] RBP: ffffb500c0aa0000 R08: 0000000000000000 R09: 0000000000000000 [ 15.310490] R10: 0000000000000001 R11: 0000000000000000 R12: ffffb500c01af000 [ 15.310490] R13: ffffb500c01cd000 R14: 0000000000000000 R15: 0000000000000000 [ 15.310490] FS: 00007f133b665140(0000) GS:ffff9c98bbd00000(0000) knlGS:0000000000000000 [ 15.310490] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 15.310490] CR2: ffffb500c0a9fff8 CR3: 0000000102478000 CR4: 00000000000006f0 [ 15.310490] Kernel panic - not syncing: Fatal exception in interrupt [ 15.310490] Kernel Offset: 0x30000000 from 0xffffffff81000000 (relocation range: 0xffffffff80000000-0xffffffffbfffffff) This patch fixes the issue by initializing tail_call_cnt at the prologue of entry_tc. Next, when call subprog_tc, the tail_call_cnt_ptr is propagated to subprog_tc by rax. Next, when jump to entry_freplace, the tail_call_cnt_ptr will be reused to count tailcall in freplace prog. Next, when call subprog_tail, the tail_call_cnt_ptr is propagated to subprog_tail by rax. Next, while tail calling to entry_tc, the tail_call_cnt on the stack of entry_tc increments via the tail_call_cnt_ptr. The whole procedure shows as the following JITed prog dumping. bpftool p d j n entry_tc: int entry_tc(struct __sk_buff * skb): bpf_prog_1c515f389a9059b4_entry_tc: ; return subprog_tc(skb); 0: endbr64 4: xorq %rax, %rax ;; rax = 0 (tail_call_cnt) 7: nopl (%rax,%rax) c: pushq %rbp d: movq %rsp, %rbp 10: endbr64 14: cmpq $33, %rax ;; if rax > 33, rax = tcc_ptr 18: ja 0x20 ;; if rax > 33 goto 0x20 ---+ 1a: pushq %rax ;; [rbp - 8] = rax = 0 | 1b: movq %rsp, %rax ;; rax = rbp - 8 | 1e: jmp 0x21 ;; ---------+ | 20: pushq %rax ;; <--------|---------------+ 21: pushq %rax ;; <--------+ [rbp - 16] = rax 22: movq -16(%rbp), %rax ;; rax = tcc_ptr 29: callq 0x70 ;; call subprog_tc() ; return subprog_tc(skb); 2e: leave 2f: retq int subprog_tc(struct __sk_buff * skb): bpf_prog_1e8f76e2374a0607_subprog_tc: ; return skb->len * 2; 0: endbr64 4: nopl (%rax) ;; do not touch tail_call_cnt 7: jmp 0x108 ;; jump to entry_freplace() c: pushq %rbp d: movq %rsp, %rbp 10: endbr64 14: pushq %rax 15: pushq %rax 16: movl 112(%rdi), %eax ; return skb->len * 2; 19: shll %eax ; return skb->len * 2; 1b: leave 1c: retq bpftool p d j n entry_freplace: int entry_freplace(struct __sk_buff * skb): bpf_prog_85781a698094722f_entry_freplace: ; int entry_freplace(struct __sk_buff *skb) 0: endbr64 4: nopl (%rax) ;; do not touch tail_call_cnt 7: nopl (%rax,%rax) c: pushq %rbp d: movq %rsp, %rbp 10: endbr64 14: cmpq $33, %rax ;; if rax > 33, rax = tcc_ptr 18: ja 0x20 ;; if rax > 33 goto 0x20 ---+ 1a: pushq %rax ;; [rbp - 8] = rax = 0 | 1b: movq %rsp, %rax ;; rax = rbp - 8 | 1e: jmp 0x21 ;; ---------+ | 20: pushq %rax ;; <--------|---------------+ 21: pushq %rax ;; <--------+ [rbp - 16] = rax 22: pushq %rbx ;; callee saved 23: pushq %r13 ;; callee saved 25: movq %rdi, %rbx ;; rbx = skb (callee saved) ; count++; 28: movabsq $-123406219759616, %r13 32: movl (%r13), %edi 36: addl $1, %edi 39: movl %edi, (%r13) ; subprog_tail(skb); 3d: movq %rbx, %rdi ;; rdi = skb 40: movq -16(%rbp), %rax ;; rax = tcc_ptr 47: callq 0x80 ;; call subprog_tail() ; subprog_tail(skb); 4c: movq %rbx, %rdi ;; rdi = skb 4f: movq -16(%rbp), %rax ;; rax = tcc_ptr 56: callq 0x80 ;; call subprog_tail() ; return count; 5b: movl (%r13), %eax ; return count; 5f: popq %r13 61: popq %rbx 62: leave 63: retq int subprog_tail(struct __sk_buff * skb): bpf_prog_3a140cef239a4b4f_subprog_tail: ; int subprog_tail(struct __sk_buff *skb) 0: endbr64 4: nopl (%rax) ;; do not touch tail_call_cnt 7: nopl (%rax,%rax) c: pushq %rbp d: movq %rsp, %rbp 10: endbr64 14: pushq %rax ;; [rbp - 8] = rax (tcc_ptr) 15: pushq %rax ;; [rbp - 16] = rax (tcc_ptr) 16: pushq %rbx ;; callee saved 17: pushq %r13 ;; callee saved 19: movq %rdi, %rbx ;; rbx = skb ; asm volatile("r1 = %[ctx]\n\t" 1c: movabsq $-128494642337280, %r13 ;; r13 = jmp_table 26: movq %rbx, %rdi ;; 1st arg, skb 29: movq %r13, %rsi ;; 2nd arg, jmp_table 2c: xorl %edx, %edx ;; 3rd arg, index = 0 2e: movq -16(%rbp), %rax ;; rax = [rbp - 16] (tcc_ptr) 35: cmpq $33, (%rax) 39: jae 0x4e ;; if *tcc_ptr >= 33 goto 0x4e --------+ 3b: nopl (%rax,%rax) ;; jmp bypass, toggled by poking | 40: addq $1, (%rax) ;; (*tcc_ptr)++ | 44: popq %r13 ;; callee saved | 46: popq %rbx ;; callee saved | 47: popq %rax ;; undo rbp-16 push | 48: popq %rax ;; undo rbp-8 push | 49: jmp 0xfffffffffffffe18 ;; tail call target, toggled by poking | ; return 0; ;; | 4e: popq %r13 ;; restore callee saved <--------------+ 50: popq %rbx ;; restore callee saved 51: leave 52: retq As a result, the tail_call_cnt is stored on the stack of entry_tc. And the tail_call_cnt_ptr is propagated between subprog_tc, entry_freplace, subprog_tail and entry_tc. Furthermore, trampoline is required to propagate tail_call_cnt/tail_call_cnt_ptr always, no matter whether there is tailcall at run time. So, it reuses trampoline flag "BIT(7)" to tell trampoline to propagate the tail_call_cnt/tail_call_cnt_ptr, as BPF_TRAMP_F_TAIL_CALL_CTX is not used by any other arch BPF JIT. However, the bad effect is that it requires initializing tail_call_cnt at prologue always no matter whether the prog is tail_call_reachable, because it is unable to confirm itself or its subprog[s] whether to be attached by freplace prog. And, when call subprog, tail_call_cnt_ptr is required to be propagated to subprog always. Fixes: 1c123c567fb1 ("bpf: Resolve fext program type when checking map compatibility") Signed-off-by: Leon Hwang <leon.hwang@linux.dev> --- arch/x86/net/bpf_jit_comp.c | 44 +++++++++++++++++++++---------------- include/linux/bpf.h | 6 ++--- kernel/bpf/trampoline.c | 4 ++-- kernel/bpf/verifier.c | 4 ++-- 4 files changed, 32 insertions(+), 26 deletions(-)