@@ -1186,6 +1186,7 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image, u8 *rw_image
const s32 imm32 = insn->imm;
u32 dst_reg = insn->dst_reg;
u32 src_reg = insn->src_reg;
+ int adjust_off = 0;
u8 b2 = 0, b3 = 0;
u8 *start_of_ldx;
s64 jmp_offset;
@@ -1290,6 +1291,7 @@ static int do_jit(struct bpf_prog *bpf_prog, int *addrs, u8 *image, u8 *rw_image
emit_mov_imm64(&prog, dst_reg, insn[1].imm, insn[0].imm);
insn++;
i++;
+ adjust_off = 1;
break;
/* dst %= src, dst /= src, dst %= imm32, dst /= imm32 */
@@ -2073,6 +2075,17 @@ st: if (is_imm8(insn->off))
return -EFAULT;
}
memcpy(rw_image + proglen, temp, ilen);
+
+ if (bpf_prog->aux->xlated_to_jit) {
+ int off;
+
+ off = i - 1 - adjust_off;
+ if (bpf_prog->aux->func_idx)
+ off += bpf_prog->aux->func_info[bpf_prog->aux->func_idx].insn_off;
+
+ bpf_prog->aux->xlated_to_jit[off].off = proglen;
+ bpf_prog->aux->xlated_to_jit[off].len = ilen;
+ }
}
proglen += ilen;
addrs[i] = proglen;
@@ -1520,6 +1520,13 @@ struct bpf_prog_aux {
};
/* an array of original indexes for all xlated instructions */
u32 *orig_idx;
+ /* for every xlated instruction point to all generated jited
+ * instructions, if allocated
+ */
+ struct {
+ u32 off; /* local offset in the jitted code */
+ u32 len; /* the total len of generated jit code */
+ } *xlated_to_jit;
};
struct bpf_prog {
@@ -6430,6 +6430,11 @@ struct sk_reuseport_md {
#define BPF_TAG_SIZE 8
+struct xlated_to_jit {
+ __u32 off;
+ __u32 len;
+};
+
struct bpf_prog_info {
__u32 type;
__u32 id;
@@ -6472,6 +6477,8 @@ struct bpf_prog_info {
__u32 attach_btf_id;
__u32 orig_idx_len;
__aligned_u64 orig_idx;
+ __u32 xlated_to_jit_len;
+ __aligned_u64 xlated_to_jit;
} __attribute__((aligned(8)));
struct bpf_map_info {
@@ -493,6 +493,26 @@ static int bpf_prog_realloc_orig_idx(struct bpf_prog *prog, u32 off, u32 patch_l
return 0;
}
+static void adjust_func_info(struct bpf_prog *prog, u32 off, u32 insn_delta)
+{
+ int i;
+
+ if (insn_delta == 0)
+ return;
+
+ for (i = 0; i < prog->aux->func_info_cnt; i++) {
+ if (prog->aux->func_info[i].insn_off <= off)
+ continue;
+ prog->aux->func_info[i].insn_off += insn_delta;
+ }
+}
+
+static void bpf_prog_adj_orig_idx_after_remove(struct bpf_prog *prog, u32 off, u32 len)
+{
+ memmove(prog->aux->orig_idx + off, prog->aux->orig_idx + off + len,
+ sizeof(*prog->aux->orig_idx) * (prog->len - off));
+}
+
struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off,
const struct bpf_insn *patch, u32 len)
{
@@ -554,6 +574,7 @@ struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off,
BUG_ON(bpf_adj_branches(prog_adj, off, off + 1, off + len, false));
bpf_adj_linfo(prog_adj, off, insn_delta);
+ adjust_func_info(prog_adj, off, insn_delta);
return prog_adj;
}
@@ -574,6 +595,8 @@ int bpf_remove_insns(struct bpf_prog *prog, u32 off, u32 cnt)
if (err)
return err;
+ bpf_prog_adj_orig_idx_after_remove(prog, off, cnt);
+
return 0;
}
@@ -2808,6 +2831,8 @@ static void bpf_prog_free_deferred(struct work_struct *work)
}
if (aux->orig_idx)
kfree(aux->orig_idx);
+ if (aux->xlated_to_jit)
+ kfree(aux->xlated_to_jit);
}
void bpf_prog_free(struct bpf_prog *fp)
@@ -4490,6 +4490,31 @@ static int bpf_prog_get_info_by_fd(struct file *file,
return -EFAULT;
}
+ ulen = info.xlated_to_jit_len;
+ if (prog->aux->xlated_to_jit)
+ info.xlated_to_jit_len = prog->len * sizeof(struct xlated_to_jit);
+ else
+ info.xlated_to_jit_len = 0;
+ if (info.xlated_to_jit_len && ulen) {
+ struct xlated_to_jit *xlated_to_jit;
+ int i;
+
+ xlated_to_jit = kzalloc(info.xlated_to_jit_len, GFP_KERNEL);
+ if (!xlated_to_jit)
+ return -ENOMEM;
+ for (i = 0; i < prog->len; i++) {
+ xlated_to_jit[i].off = prog->aux->xlated_to_jit[i].off;
+ xlated_to_jit[i].len = prog->aux->xlated_to_jit[i].len;
+ }
+ if (copy_to_user(u64_to_user_ptr(info.xlated_to_jit),
+ xlated_to_jit,
+ min_t(u32, info.xlated_to_jit_len, ulen))) {
+ kfree(xlated_to_jit);
+ return -EFAULT;
+ }
+ kfree(xlated_to_jit);
+ }
+
if (bpf_prog_is_offloaded(prog->aux)) {
err = bpf_prog_offload_info_fill(&info, prog);
if (err)
@@ -18951,6 +18951,7 @@ static int jit_subprogs(struct bpf_verifier_env *env)
func[i]->aux->exception_cb = env->subprog_info[i].is_exception_cb;
if (!i)
func[i]->aux->exception_boundary = env->seen_exception;
+ func[i]->aux->xlated_to_jit = prog->aux->xlated_to_jit;
func[i] = bpf_int_jit_compile(func[i]);
if (!func[i]->jited) {
err = -ENOTSUPP;
@@ -20780,6 +20781,7 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr, bpfptr_t uattr, __u3
int len, ret = -EINVAL, err;
u32 log_true_size;
bool is_priv;
+ u32 size;
/* no program is valid */
if (ARRAY_SIZE(bpf_verifier_ops) == 0)
@@ -20930,6 +20932,13 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr, bpfptr_t uattr, __u3
: false;
}
+ if (ret == 0) {
+ size = array_size(sizeof(*env->prog->aux->xlated_to_jit), env->prog->len);
+ env->prog->aux->xlated_to_jit = kzalloc(size, GFP_KERNEL);
+ if (!env->prog->aux->xlated_to_jit)
+ ret = -ENOMEM;
+ }
+
if (ret == 0)
ret = fixup_call_args(env);
@@ -6430,6 +6430,11 @@ struct sk_reuseport_md {
#define BPF_TAG_SIZE 8
+struct xlated_to_jit {
+ __u32 off;
+ __u32 len;
+};
+
struct bpf_prog_info {
__u32 type;
__u32 id;
@@ -6472,6 +6477,8 @@ struct bpf_prog_info {
__u32 attach_btf_id;
__u32 orig_idx_len;
__aligned_u64 orig_idx;
+ __u32 xlated_to_jit_len;
+ __aligned_u64 xlated_to_jit;
} __attribute__((aligned(8)));
struct bpf_map_info {
Allow users to get the exact xlated -> jitted instructions mapping. This is done by using a new field xlated_to_jit in bpf_prog_info which can return up to prog->len struct xlated_to_jit { u32 off; u32 len; }; elements. The xlated_to_jit[insn_off] contains jitted offset within a function and the length of the resulting jitted instruction. Example: Original: Xlated: Jitted: 0: nopl (%rax,%rax) 5: nop 7: pushq %rbp 8: movq %rsp, %rbp 0: call 0x76 0: r0 = 0xfffffbeef b: movabsq $-1923847220, %rax 2: r0 = *(u64 *)(r0 +0) 15: movq (%rax), %rax 1: r1 = 0x9 ll 3: r1 = map[id:666][0]+9 19: movabsq $-102223334445559, %rdi 3: r2 = 0x6 5: r2 = 6 23: movl $6, %esi 4: r3 = r0 6: r3 = r0 28: movq %rax, %rdx 5: call 0x6 7: call bpf_trace_printk 2b: callq 0xffffffffcdead4dc 6: call pc+2 8: call pc+2 30: callq 0x7c 7: r0 = 0x0 9: r0 = 0 35: xorl %eax, %eax 8: exit 10: exit 37: leave 38: jmp 0xffffffffcbeeffbc --- --- --- 0: nopl (%rax,%rax) 5: nop 7: pushq %rbp 8: movq %rsp, %rbp 9: goto +0x1 11: goto pc+1 b: jmp 0xf 10: goto +0x1 12: goto pc+1 d: jmp 0x11 11: goto -0x2 13: goto pc-2 f: jmp 0xd 12: r0 = 0x0 14: r0 = 0 11: xorl %eax, %eax 13: exit 15: exit 13: leave 14: jmp 0xffffffffcbffbeef Here the xlated_to_jit array will be of length 16 (11 + 6) and equal to 0: (0xb, 10) 1: (0,0) /* undefined, as the previous instruction is 16 bytes */ 2: (0x15, 4) 3: (0x19, 10) 4: (0,0) /* undefined, as the previous instruction is 16 bytes */ 5: (0x23, 5) 6: (0x28, 3) 7: (0x2b, 5) 8: (0x30, 5) 9: (0x35, 2) 10: (0x37, 6) 11: (0xb, 2) 12: (0xd, 2) 13: (0xf, 2) 14: (0x11, 2) 15: (0x13, 6) The prologues are "unmapped": no mapping exists for xlated -> [0,b) Signed-off-by: Anton Protopopov <aspsk@isovalent.com> --- arch/x86/net/bpf_jit_comp.c | 13 +++++++++++++ include/linux/bpf.h | 7 +++++++ include/uapi/linux/bpf.h | 7 +++++++ kernel/bpf/core.c | 25 +++++++++++++++++++++++++ kernel/bpf/syscall.c | 25 +++++++++++++++++++++++++ kernel/bpf/verifier.c | 9 +++++++++ tools/include/uapi/linux/bpf.h | 7 +++++++ 7 files changed, 93 insertions(+)