@@ -1522,6 +1522,8 @@ struct bpf_prog_aux {
struct work_struct work;
struct rcu_head rcu;
};
+ /* an array of original indexes for all xlated instructions */
+ u32 *orig_idx;
};
struct bpf_prog {
@@ -553,7 +553,6 @@ struct bpf_insn_aux_data {
u8 alu_state; /* used in combination with alu_limit */
/* below fields are initialized once */
- unsigned int orig_idx; /* original instruction index */
bool jmp_point;
bool prune_point;
/* ensure we check state equivalence and save state checkpoint and
@@ -6533,6 +6533,8 @@ struct bpf_prog_info {
__u32 verified_insns;
__u32 attach_btf_obj_id;
__u32 attach_btf_id;
+ __u32 orig_idx_len;
+ __aligned_u64 orig_idx;
} __attribute__((aligned(8)));
struct bpf_map_info {
@@ -469,6 +469,30 @@ static void bpf_adj_linfo(struct bpf_prog *prog, u32 off, u32 delta)
linfo[i].insn_off += delta;
}
+static int bpf_prog_realloc_orig_idx(struct bpf_prog *prog, u32 off, u32 patch_len)
+{
+ u32 *old_idx = prog->aux->orig_idx, *new_idx;
+ u32 new_prog_len = prog->len + patch_len - 1;
+ int i;
+
+ if (patch_len <= 1)
+ return 0;
+
+ new_idx = kzalloc(array_size(new_prog_len, sizeof(u32)), GFP_KERNEL);
+ if (!new_idx)
+ return -ENOMEM;
+
+ memcpy(new_idx, old_idx, sizeof(*old_idx) * off);
+ for (i = off; i < off + patch_len; i++)
+ new_idx[i] = old_idx[off];
+ memcpy(new_idx + off + patch_len, old_idx + off + 1,
+ sizeof(*old_idx) * (prog->len - off));
+
+ prog->aux->orig_idx = new_idx;
+ kfree(old_idx);
+ return 0;
+}
+
struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off,
const struct bpf_insn *patch, u32 len)
{
@@ -494,6 +518,10 @@ struct bpf_prog *bpf_patch_insn_single(struct bpf_prog *prog, u32 off,
(err = bpf_adj_branches(prog, off, off + 1, off + len, true)))
return ERR_PTR(err);
+ err = bpf_prog_realloc_orig_idx(prog, off, len);
+ if (err)
+ return ERR_PTR(err);
+
/* Several new instructions need to be inserted. Make room
* for them. Likely, there's no need for a new allocation as
* last page could have large enough tailroom.
@@ -2778,6 +2806,7 @@ static void bpf_prog_free_deferred(struct work_struct *work)
} else {
bpf_jit_free(aux->prog);
}
+ kfree(aux->orig_idx);
}
void bpf_prog_free(struct bpf_prog *fp)
@@ -2625,6 +2625,18 @@ static bool is_perfmon_prog_type(enum bpf_prog_type prog_type)
}
}
+static void *bpf_prog_alloc_orig_idx(u32 insn_cnt)
+{
+ u32 *orig_idx;
+ int i;
+
+ orig_idx = kcalloc(insn_cnt, sizeof(*orig_idx), GFP_KERNEL);
+ if (orig_idx)
+ for (i = 0; i < insn_cnt; i++)
+ orig_idx[i] = i;
+ return orig_idx;
+}
+
/* last field in 'union bpf_attr' used by this command */
#define BPF_PROG_LOAD_LAST_FIELD prog_token_fd
@@ -2760,6 +2772,12 @@ static int bpf_prog_load(union bpf_attr *attr, bpfptr_t uattr, u32 uattr_size)
goto put_token;
}
+ prog->aux->orig_idx = bpf_prog_alloc_orig_idx(attr->insn_cnt);
+ if (!prog->aux->orig_idx) {
+ err = -ENOMEM;
+ goto free_prog;
+ }
+
prog->expected_attach_type = attr->expected_attach_type;
prog->aux->attach_btf = attach_btf;
prog->aux->attach_btf_id = attr->attach_btf_id;
@@ -4541,6 +4559,18 @@ static int bpf_prog_get_info_by_fd(struct file *file,
return -EFAULT;
}
+ ulen = info.orig_idx_len;
+ if (prog->aux->orig_idx)
+ info.orig_idx_len = prog->len * sizeof(*prog->aux->orig_idx);
+ else
+ info.orig_idx_len = 0;
+ if (info.orig_idx_len && ulen) {
+ if (copy_to_user(u64_to_user_ptr(info.orig_idx),
+ prog->aux->orig_idx,
+ min_t(u32, info.orig_idx_len, ulen)))
+ return -EFAULT;
+ }
+
if (bpf_prog_is_offloaded(prog->aux)) {
err = bpf_prog_offload_info_fill(&info, prog);
if (err)
@@ -18289,7 +18289,7 @@ static struct bpf_prog *bpf_patch_insn_data(struct bpf_verifier_env *env, u32 of
if (PTR_ERR(new_prog) == -ERANGE)
verbose(env,
"insn %d cannot be patched due to 16-bit range\n",
- env->insn_aux_data[off].orig_idx);
+ env->prog->aux->orig_idx[off]);
vfree(new_data);
return NULL;
}
@@ -20829,7 +20829,7 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr, bpfptr_t uattr, __u3
{
u64 start_time = ktime_get_ns();
struct bpf_verifier_env *env;
- int i, len, ret = -EINVAL, err;
+ int len, ret = -EINVAL, err;
u32 log_true_size;
bool is_priv;
@@ -20852,8 +20852,6 @@ int bpf_check(struct bpf_prog **prog, union bpf_attr *attr, bpfptr_t uattr, __u3
ret = -ENOMEM;
if (!env->insn_aux_data)
goto err_free_env;
- for (i = 0; i < len; i++)
- env->insn_aux_data[i].orig_idx = i;
env->prog = *prog;
env->ops = bpf_verifier_ops[env->prog->type];
env->fd_array = make_bpfptr(attr->fd_array, uattr.is_kernel);
@@ -6533,6 +6533,8 @@ struct bpf_prog_info {
__u32 verified_insns;
__u32 attach_btf_obj_id;
__u32 attach_btf_id;
+ __u32 orig_idx_len;
+ __aligned_u64 orig_idx;
} __attribute__((aligned(8)));
struct bpf_map_info {
On bpf(BPF_PROG_LOAD) syscall a user-supplied program is translated by the verifier into an "xlated" program. During this process the original instruction offsets might be adjusted and/or individual instructions might be replaced by a new set of instructions: User-supplied prog: ---> Xlated prog: -- func 0 -- -- func 0 -- insn insn' ... ... insn insn' -- func 1 -- -- func 1 -- insn insn' ... ... insn insn' -- func N -- -- func N -- insn insn' ... ... insn insn' We want to provide users (and ourselves) with the off(insn') -> off(insn) mapping so that when an xlated program is returned to the userspace by the bpf_prog_get_info_by_fd() function, users can determine the real offsets of instructions of interest. Since commit 9e4c24e7ee7d ("bpf: verifier: record original instruction index") the verifier saves the original instruction index in env->insn_aux_data. This information was, however, lost when we patched instructions. Also, the information about original index was kept in the verifier env only, so was inaccessible by later stages, like constants blinding during the jit stage. To address the above issues save the information about the original indexes in a separate array inside the prog->aux so that it doesn't depend on the verifier environment and can be adjusted, and accessed, during later stages. To let users access the information after the program was loaded, add new fields, orig_idx_len and orig_idx to struct bpf_prog_info and patch the bpf_prog_get_info_by_fd function correspondingly. Example mapping would be something like this: Original prog: Xlated prog: 0: r1 = 0x0 0: r1 = 0 1: *(u32 *)(r10 - 0x4) = r1 1: *(u32 *)(r10 -4) = r1 2: r2 = r10 2: r2 = r10 3: r2 += -0x4 3: r2 += -4 4: r1 = 0x0 ll 4: r1 = map[id:88] 6: call 0x1 6: r1 += 272 7: r0 = *(u32 *)(r2 +0) 8: if r0 >= 0x1 goto pc+3 9: r0 <<= 3 10: r0 += r1 11: goto pc+1 12: r0 = 0 7: r6 = r0 13: r6 = r0 8: if r6 == 0x0 goto +0x2 14: if r6 == 0x0 goto pc+4 9: call 0x76 15: r0 = 0xffffffff8d2079c0 17: r0 = *(u64 *)(r0 +0) 10: *(u64 *)(r6 + 0x0) = r0 18: *(u64 *)(r6 +0) = r0 11: r0 = 0x0 19: r0 = 0x0 12: exit 20: exit Here the orig_idx array has length 21 and is equal to (0, 1, 2, 3, 4, 0/*undefined*/, 6, 6, 6, 6, 6, 6, 6, 7, 8, 9, 9, 10, 11, 12) The item 6 is undefined because the r1=0ll occupies 16 bytes. Signed-off-by: Anton Protopopov <aspsk@isovalent.com> --- include/linux/bpf.h | 2 ++ include/linux/bpf_verifier.h | 1 - include/uapi/linux/bpf.h | 2 ++ kernel/bpf/core.c | 29 +++++++++++++++++++++++++++++ kernel/bpf/syscall.c | 30 ++++++++++++++++++++++++++++++ kernel/bpf/verifier.c | 6 ++---- tools/include/uapi/linux/bpf.h | 2 ++ 7 files changed, 67 insertions(+), 5 deletions(-)