Message ID | 20210625200523.726854-1-kafai@fb.com (mailing list archive) |
---|---|
State | Changes Requested |
Delegated to: | BPF |
Headers | show |
Series | bpf: Allow bpf tcp iter to do bpf_setsockopt | expand |
Context | Check | Description |
---|---|---|
netdev/cover_letter | success | Link |
netdev/fixes_present | success | Link |
netdev/patch_count | success | Link |
netdev/tree_selection | success | Clearly marked for bpf-next |
netdev/subject_prefix | success | Link |
netdev/cc_maintainers | warning | 8 maintainers not CCed: dsahern@kernel.org kpsingh@kernel.org yoshfuji@linux-ipv6.org andrii@kernel.org john.fastabend@gmail.com songliubraving@fb.com davem@davemloft.net kuba@kernel.org |
netdev/source_inline | success | Was 0 now: 0 |
netdev/verify_signedoff | success | Link |
netdev/module_param | success | Was 0 now: 0 |
netdev/build_32bit | fail | Errors and warnings before: 16 this patch: 19 |
netdev/kdoc | success | Errors and warnings before: 0 this patch: 0 |
netdev/verify_fixes | success | Link |
netdev/checkpatch | success | total: 0 errors, 0 warnings, 0 checks, 297 lines checked |
netdev/build_allmodconfig_warn | fail | Errors and warnings before: 16 this patch: 19 |
netdev/header_inline | success | Link |
On 6/25/21 1:05 PM, Martin KaFai Lau wrote: > This patch does batching and lock_sock for the bpf tcp iter. > It does not affect the proc fs iteration. > > With bpf-tcp-cc, new algo rollout happens more often. Instead of > restarting the application to pick up the new tcp-cc, the next patch > will allow bpf iter with CAP_NET_ADMIN to do setsockopt(TCP_CONGESTION). > This requires locking the sock. > > Also, unlike the proc iteration (cat /proc/net/tcp[6]), the bpf iter > can inspect all fields of a tcp_sock. It will be useful to have a > consistent view on some of the fields (e.g. the ones reported in > tcp_get_info() that also acquires the sock lock). > > Double lock: locking the bucket first and then locking the sock could > lead to deadlock. This patch takes a batching approach similar to > inet_diag. While holding the bucket lock, it batch a number of sockets > into an array first and then unlock the bucket. Before doing show(), > it then calls lock_sock_fast(). > > In a machine with ~400k connections, the maximum number of > sk in a bucket of the established hashtable is 7. 0.02% of > the established connections fall into this bucket size. > > For listen hash (port+addr lhash2), the bucket is usually very > small also except for the SO_REUSEPORT use case which the > userspace could have one SO_REUSEPORT socket per thread. > > While batching is used, it can also minimize the chance of missing > sock in the setsockopt use case if the whole bucket is batched. > This patch will start with a batch array with INIT_BATCH_SZ (16) > which will be enough for the most common cases. bpf_iter_tcp_batch() > will try to realloc to a larger array to handle exception case (e.g. > the SO_REUSEPORT case in the lhash2). > > Signed-off-by: Martin KaFai Lau <kafai@fb.com> > --- > net/ipv4/tcp_ipv4.c | 236 ++++++++++++++++++++++++++++++++++++++++++-- > 1 file changed, 230 insertions(+), 6 deletions(-) > > diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c > index 0d851289a89e..856144d33f52 100644 > --- a/net/ipv4/tcp_ipv4.c > +++ b/net/ipv4/tcp_ipv4.c > @@ -2687,6 +2687,15 @@ static int tcp4_seq_show(struct seq_file *seq, void *v) > } > > #ifdef CONFIG_BPF_SYSCALL > +struct bpf_tcp_iter_state { > + struct tcp_iter_state state; > + unsigned int cur_sk; > + unsigned int end_sk; > + unsigned int max_sk; > + struct sock **batch; > + bool st_bucket_done; > +}; > + > struct bpf_iter__tcp { > __bpf_md_ptr(struct bpf_iter_meta *, meta); > __bpf_md_ptr(struct sock_common *, sk_common); > @@ -2705,16 +2714,203 @@ static int tcp_prog_seq_show(struct bpf_prog *prog, struct bpf_iter_meta *meta, > return bpf_iter_run_prog(prog, &ctx); > } > > +static void bpf_iter_tcp_put_batch(struct bpf_tcp_iter_state *iter) > +{ > + while (iter->cur_sk < iter->end_sk) > + sock_put(iter->batch[iter->cur_sk++]); > +} > + > +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, > + unsigned int new_batch_sz) > +{ > + struct sock **new_batch; > + > + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); Since we return -ENOMEM below, should we have __GFP_NOWARN in kvmalloc flags? > + if (!new_batch) > + return -ENOMEM; > + > + bpf_iter_tcp_put_batch(iter); > + kvfree(iter->batch); > + iter->batch = new_batch; > + iter->max_sk = new_batch_sz; > + > + return 0; > +} > + [...] > + > static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) > { > struct bpf_iter_meta meta; > struct bpf_prog *prog; > struct sock *sk = v; > + bool slow; > uid_t uid; > + int ret; > > if (v == SEQ_START_TOKEN) > return 0; > > + if (sk_fullsock(sk)) > + slow = lock_sock_fast(sk); > + > + if (unlikely(sk_unhashed(sk))) { > + ret = SEQ_SKIP; > + goto unlock; > + } I am not a tcp expert. Maybe a dummy question. Is it possible to do setsockopt() for listening socket? What will happen if the listening sock is unhashed after the above check? > + > if (sk->sk_state == TCP_TIME_WAIT) { > uid = 0; > } else if (sk->sk_state == TCP_NEW_SYN_RECV) { > @@ -2728,11 +2924,18 @@ static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) > > meta.seq = seq; > prog = bpf_iter_get_info(&meta, false); > - return tcp_prog_seq_show(prog, &meta, v, uid); > + ret = tcp_prog_seq_show(prog, &meta, v, uid); > + > +unlock: > + if (sk_fullsock(sk)) > + unlock_sock_fast(sk, slow); > + return ret; > + > } > > static void bpf_iter_tcp_seq_stop(struct seq_file *seq, void *v) > { > + struct bpf_tcp_iter_state *iter = seq->private; > struct bpf_iter_meta meta; > struct bpf_prog *prog; > > @@ -2743,13 +2946,16 @@ static void bpf_iter_tcp_seq_stop(struct seq_file *seq, void *v) > (void)tcp_prog_seq_show(prog, &meta, v, 0); > } > > - tcp_seq_stop(seq, v); > + if (iter->cur_sk < iter->end_sk) { > + bpf_iter_tcp_put_batch(iter); > + iter->st_bucket_done = false; > + } > } > [...]
On Tue, Jun 29, 2021 at 10:27:17AM -0700, Yonghong Song wrote: [ ... ] > > +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, > > + unsigned int new_batch_sz) > > +{ > > + struct sock **new_batch; > > + > > + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); > > Since we return -ENOMEM below, should we have __GFP_NOWARN in kvmalloc > flags? will add in v2. > > > + if (!new_batch) > > + return -ENOMEM; > > + > > + bpf_iter_tcp_put_batch(iter); > > + kvfree(iter->batch); > > + iter->batch = new_batch; > > + iter->max_sk = new_batch_sz; > > + > > + return 0; > > +} > > + > [...] > > + > > static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) > > { > > struct bpf_iter_meta meta; > > struct bpf_prog *prog; > > struct sock *sk = v; > > + bool slow; > > uid_t uid; > > + int ret; > > if (v == SEQ_START_TOKEN) > > return 0; > > + if (sk_fullsock(sk)) > > + slow = lock_sock_fast(sk); > > + > > + if (unlikely(sk_unhashed(sk))) { > > + ret = SEQ_SKIP; > > + goto unlock; > > + } > > I am not a tcp expert. Maybe a dummy question. > Is it possible to do setsockopt() for listening socket? > What will happen if the listening sock is unhashed after the > above check? It won't happen because the sk has been locked before doing the unhashed check. Thanks for the review.
On 6/29/21 10:44 AM, Martin KaFai Lau wrote: > On Tue, Jun 29, 2021 at 10:27:17AM -0700, Yonghong Song wrote: > [ ... ] > >>> +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, >>> + unsigned int new_batch_sz) >>> +{ >>> + struct sock **new_batch; >>> + >>> + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); >> >> Since we return -ENOMEM below, should we have __GFP_NOWARN in kvmalloc >> flags? > will add in v2. > >> >>> + if (!new_batch) >>> + return -ENOMEM; >>> + >>> + bpf_iter_tcp_put_batch(iter); >>> + kvfree(iter->batch); >>> + iter->batch = new_batch; >>> + iter->max_sk = new_batch_sz; >>> + >>> + return 0; >>> +} >>> + >> [...] >>> + >>> static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) >>> { >>> struct bpf_iter_meta meta; >>> struct bpf_prog *prog; >>> struct sock *sk = v; >>> + bool slow; >>> uid_t uid; >>> + int ret; >>> if (v == SEQ_START_TOKEN) >>> return 0; >>> + if (sk_fullsock(sk)) >>> + slow = lock_sock_fast(sk); >>> + >>> + if (unlikely(sk_unhashed(sk))) { >>> + ret = SEQ_SKIP; >>> + goto unlock; >>> + } >> >> I am not a tcp expert. Maybe a dummy question. >> Is it possible to do setsockopt() for listening socket? >> What will happen if the listening sock is unhashed after the >> above check? > It won't happen because the sk has been locked before doing the > unhashed check. Ya, that is true. I guess I probably mean TCP_TIME_WAIT and TCP_NEW_SYN_RECV sockets. We cannot do setsockopt() for TCP_TIME_WAIT sockets since user space shouldn't be able to access the socket any more. But how about TCP_NEW_SYN_RECV sockets? > > Thanks for the review. >
On Tue, Jun 29, 2021 at 10:57:46AM -0700, Yonghong Song wrote: > > > On 6/29/21 10:44 AM, Martin KaFai Lau wrote: > > On Tue, Jun 29, 2021 at 10:27:17AM -0700, Yonghong Song wrote: > > [ ... ] > > > > > > +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, > > > > + unsigned int new_batch_sz) > > > > +{ > > > > + struct sock **new_batch; > > > > + > > > > + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); > > > > > > Since we return -ENOMEM below, should we have __GFP_NOWARN in kvmalloc > > > flags? > > will add in v2. > > > > > > > > > + if (!new_batch) > > > > + return -ENOMEM; > > > > + > > > > + bpf_iter_tcp_put_batch(iter); > > > > + kvfree(iter->batch); > > > > + iter->batch = new_batch; > > > > + iter->max_sk = new_batch_sz; > > > > + > > > > + return 0; > > > > +} > > > > + > > > [...] > > > > + > > > > static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) > > > > { > > > > struct bpf_iter_meta meta; > > > > struct bpf_prog *prog; > > > > struct sock *sk = v; > > > > + bool slow; > > > > uid_t uid; > > > > + int ret; > > > > if (v == SEQ_START_TOKEN) > > > > return 0; > > > > + if (sk_fullsock(sk)) > > > > + slow = lock_sock_fast(sk); > > > > + > > > > + if (unlikely(sk_unhashed(sk))) { > > > > + ret = SEQ_SKIP; > > > > + goto unlock; > > > > + } > > > > > > I am not a tcp expert. Maybe a dummy question. > > > Is it possible to do setsockopt() for listening socket? > > > What will happen if the listening sock is unhashed after the > > > above check? > > It won't happen because the sk has been locked before doing the > > unhashed check. > > Ya, that is true. I guess I probably mean TCP_TIME_WAIT and > TCP_NEW_SYN_RECV sockets. We cannot do setsockopt() for > TCP_TIME_WAIT sockets since user space shouldn't be able > to access the socket any more. > > But how about TCP_NEW_SYN_RECV sockets? _bpf_setsockopt() will return -EINVAL for non fullsock.
On 6/29/21 11:06 AM, Martin KaFai Lau wrote: > On Tue, Jun 29, 2021 at 10:57:46AM -0700, Yonghong Song wrote: >> >> >> On 6/29/21 10:44 AM, Martin KaFai Lau wrote: >>> On Tue, Jun 29, 2021 at 10:27:17AM -0700, Yonghong Song wrote: >>> [ ... ] >>> >>>>> +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, >>>>> + unsigned int new_batch_sz) >>>>> +{ >>>>> + struct sock **new_batch; >>>>> + >>>>> + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); >>>> >>>> Since we return -ENOMEM below, should we have __GFP_NOWARN in kvmalloc >>>> flags? >>> will add in v2. >>> >>>> >>>>> + if (!new_batch) >>>>> + return -ENOMEM; >>>>> + >>>>> + bpf_iter_tcp_put_batch(iter); >>>>> + kvfree(iter->batch); >>>>> + iter->batch = new_batch; >>>>> + iter->max_sk = new_batch_sz; >>>>> + >>>>> + return 0; >>>>> +} >>>>> + >>>> [...] >>>>> + >>>>> static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) >>>>> { >>>>> struct bpf_iter_meta meta; >>>>> struct bpf_prog *prog; >>>>> struct sock *sk = v; >>>>> + bool slow; >>>>> uid_t uid; >>>>> + int ret; >>>>> if (v == SEQ_START_TOKEN) >>>>> return 0; >>>>> + if (sk_fullsock(sk)) >>>>> + slow = lock_sock_fast(sk); >>>>> + >>>>> + if (unlikely(sk_unhashed(sk))) { >>>>> + ret = SEQ_SKIP; >>>>> + goto unlock; >>>>> + } >>>> >>>> I am not a tcp expert. Maybe a dummy question. >>>> Is it possible to do setsockopt() for listening socket? >>>> What will happen if the listening sock is unhashed after the >>>> above check? >>> It won't happen because the sk has been locked before doing the >>> unhashed check. >> >> Ya, that is true. I guess I probably mean TCP_TIME_WAIT and >> TCP_NEW_SYN_RECV sockets. We cannot do setsockopt() for >> TCP_TIME_WAIT sockets since user space shouldn't be able >> to access the socket any more. >> >> But how about TCP_NEW_SYN_RECV sockets? > _bpf_setsockopt() will return -EINVAL for non fullsock. That makes sense. I think whether we could block calling bpf_setsockopt() for unsupported sockets outside bpf program. But indeed letting bpf to do filtering in such cases should be simpler.
diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c index 0d851289a89e..856144d33f52 100644 --- a/net/ipv4/tcp_ipv4.c +++ b/net/ipv4/tcp_ipv4.c @@ -2687,6 +2687,15 @@ static int tcp4_seq_show(struct seq_file *seq, void *v) } #ifdef CONFIG_BPF_SYSCALL +struct bpf_tcp_iter_state { + struct tcp_iter_state state; + unsigned int cur_sk; + unsigned int end_sk; + unsigned int max_sk; + struct sock **batch; + bool st_bucket_done; +}; + struct bpf_iter__tcp { __bpf_md_ptr(struct bpf_iter_meta *, meta); __bpf_md_ptr(struct sock_common *, sk_common); @@ -2705,16 +2714,203 @@ static int tcp_prog_seq_show(struct bpf_prog *prog, struct bpf_iter_meta *meta, return bpf_iter_run_prog(prog, &ctx); } +static void bpf_iter_tcp_put_batch(struct bpf_tcp_iter_state *iter) +{ + while (iter->cur_sk < iter->end_sk) + sock_put(iter->batch[iter->cur_sk++]); +} + +static int bpf_iter_tcp_realloc_batch(struct bpf_tcp_iter_state *iter, + unsigned int new_batch_sz) +{ + struct sock **new_batch; + + new_batch = kvmalloc(sizeof(*new_batch) * new_batch_sz, GFP_USER); + if (!new_batch) + return -ENOMEM; + + bpf_iter_tcp_put_batch(iter); + kvfree(iter->batch); + iter->batch = new_batch; + iter->max_sk = new_batch_sz; + + return 0; +} + +static unsigned int bpf_iter_tcp_listening_batch(struct seq_file *seq, + struct sock *start_sk) +{ + struct bpf_tcp_iter_state *iter = seq->private; + struct tcp_iter_state *st = &iter->state; + struct inet_connection_sock *icsk; + unsigned int expected = 1; + struct sock *sk; + + sock_hold(start_sk); + iter->batch[iter->end_sk++] = start_sk; + + icsk = inet_csk(start_sk); + inet_lhash2_for_each_icsk_continue(icsk) { + sk = (struct sock *)icsk; + if (seq_sk_match(seq, sk)) { + if (iter->end_sk < iter->max_sk) { + sock_hold(sk); + iter->batch[iter->end_sk++] = sk; + } + expected++; + } + } + spin_unlock(&tcp_hashinfo.lhash2[st->bucket].lock); + + return expected; +} + +static unsigned int bpf_iter_tcp_established_batch(struct seq_file *seq, + struct sock *start_sk) +{ + struct bpf_tcp_iter_state *iter = seq->private; + struct tcp_iter_state *st = &iter->state; + struct hlist_nulls_node *node; + unsigned int expected = 1; + struct sock *sk; + + sock_hold(start_sk); + iter->batch[iter->end_sk++] = start_sk; + + sk = sk_nulls_next(start_sk); + sk_nulls_for_each_from(sk, node) { + if (seq_sk_match(seq, sk)) { + if (iter->end_sk < iter->max_sk) { + sock_hold(sk); + iter->batch[iter->end_sk++] = sk; + } + expected++; + } + } + spin_unlock_bh(inet_ehash_lockp(&tcp_hashinfo, st->bucket)); + + return expected; +} + +static struct sock *bpf_iter_tcp_batch(struct seq_file *seq) +{ + struct bpf_tcp_iter_state *iter = seq->private; + struct tcp_iter_state *st = &iter->state; + unsigned int expected; + bool resized = false; + struct sock *sk; + + /* The st->bucket is done. Directly advance to the next + * bucket instead of having the tcp_seek_last_pos() to skip + * one by one in the current bucket and eventually find out + * it has to advance to the next bucket. + */ + if (iter->st_bucket_done) { + st->offset = 0; + st->bucket++; + if (st->state == TCP_SEQ_STATE_LISTENING && + st->bucket > tcp_hashinfo.lhash2_mask) { + st->state = TCP_SEQ_STATE_ESTABLISHED; + st->bucket = 0; + } + } + +again: + /* Get a new batch */ + iter->cur_sk = 0; + iter->end_sk = 0; + iter->st_bucket_done = false; + + sk = tcp_seek_last_pos(seq); + if (!sk) + return NULL; /* Done */ + + if (st->state == TCP_SEQ_STATE_LISTENING) + expected = bpf_iter_tcp_listening_batch(seq, sk); + else + expected = bpf_iter_tcp_established_batch(seq, sk); + + if (iter->end_sk == expected) { + iter->st_bucket_done = true; + return sk; + } + + if (!resized && !bpf_iter_tcp_realloc_batch(iter, expected * 3 / 2)) { + resized = true; + goto again; + } + + return sk; +} + +static void *bpf_iter_tcp_seq_start(struct seq_file *seq, loff_t *pos) +{ + /* bpf iter does not support lseek, so it always + * continue from where it was stop()-ped. + */ + if (*pos) + return bpf_iter_tcp_batch(seq); + + return SEQ_START_TOKEN; +} + +static void *bpf_iter_tcp_seq_next(struct seq_file *seq, void *v, loff_t *pos) +{ + struct bpf_tcp_iter_state *iter = seq->private; + struct tcp_iter_state *st = &iter->state; + struct sock *sk; + + /* Whenever seq_next() is called, the iter->cur_sk is + * done with seq_show(), so advance to the next sk in + * the batch. + */ + if (iter->cur_sk < iter->end_sk) { + /* Keeping st->num consistent in tcp_iter_state. + * bpf_iter_tcp does not use st->num. + * meta.seq_num is used instead. + */ + st->num++; + /* Move st->offset to the next sk in the bucket such that + * the future start() will resume at st->offset in + * st->bucket. See tcp_seek_last_pos(). + */ + st->offset++; + sock_put(iter->batch[iter->cur_sk++]); + } + + if (iter->cur_sk < iter->end_sk) + sk = iter->batch[iter->cur_sk]; + else + sk = bpf_iter_tcp_batch(seq); + + ++*pos; + /* Keeping st->last_pos consistent in tcp_iter_state. + * bpf iter does not do lseek, so st->last_pos always equals to *pos. + */ + st->last_pos = *pos; + return sk; +} + static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) { struct bpf_iter_meta meta; struct bpf_prog *prog; struct sock *sk = v; + bool slow; uid_t uid; + int ret; if (v == SEQ_START_TOKEN) return 0; + if (sk_fullsock(sk)) + slow = lock_sock_fast(sk); + + if (unlikely(sk_unhashed(sk))) { + ret = SEQ_SKIP; + goto unlock; + } + if (sk->sk_state == TCP_TIME_WAIT) { uid = 0; } else if (sk->sk_state == TCP_NEW_SYN_RECV) { @@ -2728,11 +2924,18 @@ static int bpf_iter_tcp_seq_show(struct seq_file *seq, void *v) meta.seq = seq; prog = bpf_iter_get_info(&meta, false); - return tcp_prog_seq_show(prog, &meta, v, uid); + ret = tcp_prog_seq_show(prog, &meta, v, uid); + +unlock: + if (sk_fullsock(sk)) + unlock_sock_fast(sk, slow); + return ret; + } static void bpf_iter_tcp_seq_stop(struct seq_file *seq, void *v) { + struct bpf_tcp_iter_state *iter = seq->private; struct bpf_iter_meta meta; struct bpf_prog *prog; @@ -2743,13 +2946,16 @@ static void bpf_iter_tcp_seq_stop(struct seq_file *seq, void *v) (void)tcp_prog_seq_show(prog, &meta, v, 0); } - tcp_seq_stop(seq, v); + if (iter->cur_sk < iter->end_sk) { + bpf_iter_tcp_put_batch(iter); + iter->st_bucket_done = false; + } } static const struct seq_operations bpf_iter_tcp_seq_ops = { .show = bpf_iter_tcp_seq_show, - .start = tcp_seq_start, - .next = tcp_seq_next, + .start = bpf_iter_tcp_seq_start, + .next = bpf_iter_tcp_seq_next, .stop = bpf_iter_tcp_seq_stop, }; #endif @@ -3017,21 +3223,39 @@ static struct pernet_operations __net_initdata tcp_sk_ops = { DEFINE_BPF_ITER_FUNC(tcp, struct bpf_iter_meta *meta, struct sock_common *sk_common, uid_t uid) +#define INIT_BATCH_SZ 16 + static int bpf_iter_init_tcp(void *priv_data, struct bpf_iter_aux_info *aux) { - return bpf_iter_init_seq_net(priv_data, aux); + struct bpf_tcp_iter_state *iter = priv_data; + int err; + + err = bpf_iter_init_seq_net(priv_data, aux); + if (err) + return err; + + err = bpf_iter_tcp_realloc_batch(iter, INIT_BATCH_SZ); + if (err) { + bpf_iter_fini_seq_net(priv_data); + return err; + } + + return 0; } static void bpf_iter_fini_tcp(void *priv_data) { + struct bpf_tcp_iter_state *iter = priv_data; + bpf_iter_fini_seq_net(priv_data); + kvfree(iter->batch); } static const struct bpf_iter_seq_info tcp_seq_info = { .seq_ops = &bpf_iter_tcp_seq_ops, .init_seq_private = bpf_iter_init_tcp, .fini_seq_private = bpf_iter_fini_tcp, - .seq_priv_size = sizeof(struct tcp_iter_state), + .seq_priv_size = sizeof(struct bpf_tcp_iter_state), }; static struct bpf_iter_reg tcp_reg_info = {
This patch does batching and lock_sock for the bpf tcp iter. It does not affect the proc fs iteration. With bpf-tcp-cc, new algo rollout happens more often. Instead of restarting the application to pick up the new tcp-cc, the next patch will allow bpf iter with CAP_NET_ADMIN to do setsockopt(TCP_CONGESTION). This requires locking the sock. Also, unlike the proc iteration (cat /proc/net/tcp[6]), the bpf iter can inspect all fields of a tcp_sock. It will be useful to have a consistent view on some of the fields (e.g. the ones reported in tcp_get_info() that also acquires the sock lock). Double lock: locking the bucket first and then locking the sock could lead to deadlock. This patch takes a batching approach similar to inet_diag. While holding the bucket lock, it batch a number of sockets into an array first and then unlock the bucket. Before doing show(), it then calls lock_sock_fast(). In a machine with ~400k connections, the maximum number of sk in a bucket of the established hashtable is 7. 0.02% of the established connections fall into this bucket size. For listen hash (port+addr lhash2), the bucket is usually very small also except for the SO_REUSEPORT use case which the userspace could have one SO_REUSEPORT socket per thread. While batching is used, it can also minimize the chance of missing sock in the setsockopt use case if the whole bucket is batched. This patch will start with a batch array with INIT_BATCH_SZ (16) which will be enough for the most common cases. bpf_iter_tcp_batch() will try to realloc to a larger array to handle exception case (e.g. the SO_REUSEPORT case in the lhash2). Signed-off-by: Martin KaFai Lau <kafai@fb.com> --- net/ipv4/tcp_ipv4.c | 236 ++++++++++++++++++++++++++++++++++++++++++-- 1 file changed, 230 insertions(+), 6 deletions(-)