Message ID | 20240909225952.30324-1-yunwei356@gmail.com (mailing list archive) |
---|---|
State | Superseded |
Commit | 41d0c4677feee1ea063e0f2c2af72dc953b1f1cc |
Delegated to: | BPF |
Headers | show |
Series | [bpf-next,v2] libbpf: fix some typos in comments | expand |
Hello: This patch was applied to bpf/bpf-next.git (master) by Andrii Nakryiko <andrii@kernel.org>: On Mon, 9 Sep 2024 22:59:52 +0000 you wrote: > Fix some spelling errors in the code comments of libbpf: > > betwen -> between > paremeters -> parameters > knowning -> knowing > definiton -> definition > compatiblity -> compatibility > overriden -> overridden > occured -> occurred > proccess -> process > managment -> management > nessary -> necessary > > [...] Here is the summary with links: - [bpf-next,v2] libbpf: fix some typos in comments https://git.kernel.org/bpf/bpf-next/c/41d0c4677fee You are awesome, thank you!
Hi Alexei and all, Thanks a lot for reviewing my first patch. Sorry about the --in-reply-to mistake and name. I've fixed it and sent the v2. Best, Yusheng Zheng On Mon, Sep 9, 2024 at 11:00 PM Yusheng Zheng <yunwei356@gmail.com> wrote: > > Fix some spelling errors in the code comments of libbpf: > > betwen -> between > paremeters -> parameters > knowning -> knowing > definiton -> definition > compatiblity -> compatibility > overriden -> overridden > occured -> occurred > proccess -> process > managment -> management > nessary -> necessary > > Signed-off-by: Yusheng Zheng <yunwei356@gmail.com> > --- > tools/lib/bpf/bpf_helpers.h | 2 +- > tools/lib/bpf/bpf_tracing.h | 2 +- > tools/lib/bpf/btf.c | 2 +- > tools/lib/bpf/btf.h | 2 +- > tools/lib/bpf/btf_dump.c | 2 +- > tools/lib/bpf/libbpf.h | 10 +++++----- > tools/lib/bpf/libbpf_legacy.h | 4 ++-- > tools/lib/bpf/skel_internal.h | 2 +- > 8 files changed, 13 insertions(+), 13 deletions(-) > > diff --git a/tools/lib/bpf/bpf_helpers.h b/tools/lib/bpf/bpf_helpers.h > index 305c62817dd3..80bc0242e8dc 100644 > --- a/tools/lib/bpf/bpf_helpers.h > +++ b/tools/lib/bpf/bpf_helpers.h > @@ -341,7 +341,7 @@ extern void bpf_iter_num_destroy(struct bpf_iter_num *it) __weak __ksym; > * I.e., it looks almost like high-level for each loop in other languages, > * supports continue/break, and is verifiable by BPF verifier. > * > - * For iterating integers, the difference betwen bpf_for_each(num, i, N, M) > + * For iterating integers, the difference between bpf_for_each(num, i, N, M) > * and bpf_for(i, N, M) is in that bpf_for() provides additional proof to > * verifier that i is in [N, M) range, and in bpf_for_each() case i is `int > * *`, not just `int`. So for integers bpf_for() is more convenient. > diff --git a/tools/lib/bpf/bpf_tracing.h b/tools/lib/bpf/bpf_tracing.h > index 4eab132a963e..8ea6797a2570 100644 > --- a/tools/lib/bpf/bpf_tracing.h > +++ b/tools/lib/bpf/bpf_tracing.h > @@ -808,7 +808,7 @@ struct pt_regs; > * tp_btf/fentry/fexit BPF programs. It hides the underlying platform-specific > * low-level way of getting kprobe input arguments from struct pt_regs, and > * provides a familiar typed and named function arguments syntax and > - * semantics of accessing kprobe input paremeters. > + * semantics of accessing kprobe input parameters. > * > * Original struct pt_regs* context is preserved as 'ctx' argument. This might > * be necessary when using BPF helpers like bpf_perf_event_output(). > diff --git a/tools/lib/bpf/btf.c b/tools/lib/bpf/btf.c > index 8d51e73d55a8..3c131039c523 100644 > --- a/tools/lib/bpf/btf.c > +++ b/tools/lib/bpf/btf.c > @@ -4230,7 +4230,7 @@ static bool btf_dedup_identical_structs(struct btf_dedup *d, __u32 id1, __u32 id > * consists of portions of the graph that come from multiple compilation units. > * This is due to the fact that types within single compilation unit are always > * deduplicated and FWDs are already resolved, if referenced struct/union > - * definiton is available. So, if we had unresolved FWD and found corresponding > + * definition is available. So, if we had unresolved FWD and found corresponding > * STRUCT/UNION, they will be from different compilation units. This > * consequently means that when we "link" FWD to corresponding STRUCT/UNION, > * type graph will likely have at least two different BTF types that describe > diff --git a/tools/lib/bpf/btf.h b/tools/lib/bpf/btf.h > index b68d216837a9..4e349ad79ee6 100644 > --- a/tools/lib/bpf/btf.h > +++ b/tools/lib/bpf/btf.h > @@ -286,7 +286,7 @@ LIBBPF_API void btf_dump__free(struct btf_dump *d); > LIBBPF_API int btf_dump__dump_type(struct btf_dump *d, __u32 id); > > struct btf_dump_emit_type_decl_opts { > - /* size of this struct, for forward/backward compatiblity */ > + /* size of this struct, for forward/backward compatibility */ > size_t sz; > /* optional field name for type declaration, e.g.: > * - struct my_struct <FNAME> > diff --git a/tools/lib/bpf/btf_dump.c b/tools/lib/bpf/btf_dump.c > index 894860111ddb..0a7327541c17 100644 > --- a/tools/lib/bpf/btf_dump.c > +++ b/tools/lib/bpf/btf_dump.c > @@ -304,7 +304,7 @@ int btf_dump__dump_type(struct btf_dump *d, __u32 id) > * definition, in which case they have to be declared inline as part of field > * type declaration; or as a top-level anonymous enum, typically used for > * declaring global constants. It's impossible to distinguish between two > - * without knowning whether given enum type was referenced from other type: > + * without knowing whether given enum type was referenced from other type: > * top-level anonymous enum won't be referenced by anything, while embedded > * one will. > */ > diff --git a/tools/lib/bpf/libbpf.h b/tools/lib/bpf/libbpf.h > index 64a6a3d323e3..6917653ef9fa 100644 > --- a/tools/lib/bpf/libbpf.h > +++ b/tools/lib/bpf/libbpf.h > @@ -152,7 +152,7 @@ struct bpf_object_open_opts { > * log_buf and log_level settings. > * > * If specified, this log buffer will be passed for: > - * - each BPF progral load (BPF_PROG_LOAD) attempt, unless overriden > + * - each BPF progral load (BPF_PROG_LOAD) attempt, unless overridden > * with bpf_program__set_log() on per-program level, to get > * BPF verifier log output. > * - during BPF object's BTF load into kernel (BPF_BTF_LOAD) to get > @@ -455,7 +455,7 @@ LIBBPF_API int bpf_link__destroy(struct bpf_link *link); > /** > * @brief **bpf_program__attach()** is a generic function for attaching > * a BPF program based on auto-detection of program type, attach type, > - * and extra paremeters, where applicable. > + * and extra parameters, where applicable. > * > * @param prog BPF program to attach > * @return Reference to the newly created BPF link; or NULL is returned on error, > @@ -679,7 +679,7 @@ struct bpf_uprobe_opts { > /** > * @brief **bpf_program__attach_uprobe()** attaches a BPF program > * to the userspace function which is found by binary path and > - * offset. You can optionally specify a particular proccess to attach > + * offset. You can optionally specify a particular process to attach > * to. You can also optionally attach the program to the function > * exit instead of entry. > * > @@ -1593,11 +1593,11 @@ LIBBPF_API int perf_buffer__buffer_fd(const struct perf_buffer *pb, size_t buf_i > * memory region of the ring buffer. > * This ring buffer can be used to implement a custom events consumer. > * The ring buffer starts with the *struct perf_event_mmap_page*, which > - * holds the ring buffer managment fields, when accessing the header > + * holds the ring buffer management fields, when accessing the header > * structure it's important to be SMP aware. > * You can refer to *perf_event_read_simple* for a simple example. > * @param pb the perf buffer structure > - * @param buf_idx the buffer index to retreive > + * @param buf_idx the buffer index to retrieve > * @param buf (out) gets the base pointer of the mmap()'ed memory > * @param buf_size (out) gets the size of the mmap()'ed region > * @return 0 on success, negative error code for failure > diff --git a/tools/lib/bpf/libbpf_legacy.h b/tools/lib/bpf/libbpf_legacy.h > index 1e1be467bede..60b2600be88a 100644 > --- a/tools/lib/bpf/libbpf_legacy.h > +++ b/tools/lib/bpf/libbpf_legacy.h > @@ -76,7 +76,7 @@ enum libbpf_strict_mode { > * first BPF program or map creation operation. This is done only if > * kernel is too old to support memcg-based memory accounting for BPF > * subsystem. By default, RLIMIT_MEMLOCK limit is set to RLIM_INFINITY, > - * but it can be overriden with libbpf_set_memlock_rlim() API. > + * but it can be overridden with libbpf_set_memlock_rlim() API. > * Note that libbpf_set_memlock_rlim() needs to be called before > * the very first bpf_prog_load(), bpf_map_create() or bpf_object__load() > * operation. > @@ -97,7 +97,7 @@ LIBBPF_API int libbpf_set_strict_mode(enum libbpf_strict_mode mode); > * @brief **libbpf_get_error()** extracts the error code from the passed > * pointer > * @param ptr pointer returned from libbpf API function > - * @return error code; or 0 if no error occured > + * @return error code; or 0 if no error occurred > * > * Note, as of libbpf 1.0 this function is not necessary and not recommended > * to be used. Libbpf doesn't return error code embedded into the pointer > diff --git a/tools/lib/bpf/skel_internal.h b/tools/lib/bpf/skel_internal.h > index 1e82ab06c3eb..0875452521e9 100644 > --- a/tools/lib/bpf/skel_internal.h > +++ b/tools/lib/bpf/skel_internal.h > @@ -107,7 +107,7 @@ static inline void skel_free(const void *p) > * The loader program will perform probe_read_kernel() from maps.rodata.initial_value. > * skel_finalize_map_data() sets skel->rodata to point to actual value in a bpf map and > * does maps.rodata.initial_value = ~0ULL to signal skel_free_map_data() that kvfree > - * is not nessary. > + * is not necessary. > * > * For user space: > * skel_prep_map_data() mmaps anon memory into skel->rodata that can be accessed directly. > -- > 2.43.0 >
diff --git a/tools/lib/bpf/bpf_helpers.h b/tools/lib/bpf/bpf_helpers.h index 305c62817dd3..80bc0242e8dc 100644 --- a/tools/lib/bpf/bpf_helpers.h +++ b/tools/lib/bpf/bpf_helpers.h @@ -341,7 +341,7 @@ extern void bpf_iter_num_destroy(struct bpf_iter_num *it) __weak __ksym; * I.e., it looks almost like high-level for each loop in other languages, * supports continue/break, and is verifiable by BPF verifier. * - * For iterating integers, the difference betwen bpf_for_each(num, i, N, M) + * For iterating integers, the difference between bpf_for_each(num, i, N, M) * and bpf_for(i, N, M) is in that bpf_for() provides additional proof to * verifier that i is in [N, M) range, and in bpf_for_each() case i is `int * *`, not just `int`. So for integers bpf_for() is more convenient. diff --git a/tools/lib/bpf/bpf_tracing.h b/tools/lib/bpf/bpf_tracing.h index 4eab132a963e..8ea6797a2570 100644 --- a/tools/lib/bpf/bpf_tracing.h +++ b/tools/lib/bpf/bpf_tracing.h @@ -808,7 +808,7 @@ struct pt_regs; * tp_btf/fentry/fexit BPF programs. It hides the underlying platform-specific * low-level way of getting kprobe input arguments from struct pt_regs, and * provides a familiar typed and named function arguments syntax and - * semantics of accessing kprobe input paremeters. + * semantics of accessing kprobe input parameters. * * Original struct pt_regs* context is preserved as 'ctx' argument. This might * be necessary when using BPF helpers like bpf_perf_event_output(). diff --git a/tools/lib/bpf/btf.c b/tools/lib/bpf/btf.c index 8d51e73d55a8..3c131039c523 100644 --- a/tools/lib/bpf/btf.c +++ b/tools/lib/bpf/btf.c @@ -4230,7 +4230,7 @@ static bool btf_dedup_identical_structs(struct btf_dedup *d, __u32 id1, __u32 id * consists of portions of the graph that come from multiple compilation units. * This is due to the fact that types within single compilation unit are always * deduplicated and FWDs are already resolved, if referenced struct/union - * definiton is available. So, if we had unresolved FWD and found corresponding + * definition is available. So, if we had unresolved FWD and found corresponding * STRUCT/UNION, they will be from different compilation units. This * consequently means that when we "link" FWD to corresponding STRUCT/UNION, * type graph will likely have at least two different BTF types that describe diff --git a/tools/lib/bpf/btf.h b/tools/lib/bpf/btf.h index b68d216837a9..4e349ad79ee6 100644 --- a/tools/lib/bpf/btf.h +++ b/tools/lib/bpf/btf.h @@ -286,7 +286,7 @@ LIBBPF_API void btf_dump__free(struct btf_dump *d); LIBBPF_API int btf_dump__dump_type(struct btf_dump *d, __u32 id); struct btf_dump_emit_type_decl_opts { - /* size of this struct, for forward/backward compatiblity */ + /* size of this struct, for forward/backward compatibility */ size_t sz; /* optional field name for type declaration, e.g.: * - struct my_struct <FNAME> diff --git a/tools/lib/bpf/btf_dump.c b/tools/lib/bpf/btf_dump.c index 894860111ddb..0a7327541c17 100644 --- a/tools/lib/bpf/btf_dump.c +++ b/tools/lib/bpf/btf_dump.c @@ -304,7 +304,7 @@ int btf_dump__dump_type(struct btf_dump *d, __u32 id) * definition, in which case they have to be declared inline as part of field * type declaration; or as a top-level anonymous enum, typically used for * declaring global constants. It's impossible to distinguish between two - * without knowning whether given enum type was referenced from other type: + * without knowing whether given enum type was referenced from other type: * top-level anonymous enum won't be referenced by anything, while embedded * one will. */ diff --git a/tools/lib/bpf/libbpf.h b/tools/lib/bpf/libbpf.h index 64a6a3d323e3..6917653ef9fa 100644 --- a/tools/lib/bpf/libbpf.h +++ b/tools/lib/bpf/libbpf.h @@ -152,7 +152,7 @@ struct bpf_object_open_opts { * log_buf and log_level settings. * * If specified, this log buffer will be passed for: - * - each BPF progral load (BPF_PROG_LOAD) attempt, unless overriden + * - each BPF progral load (BPF_PROG_LOAD) attempt, unless overridden * with bpf_program__set_log() on per-program level, to get * BPF verifier log output. * - during BPF object's BTF load into kernel (BPF_BTF_LOAD) to get @@ -455,7 +455,7 @@ LIBBPF_API int bpf_link__destroy(struct bpf_link *link); /** * @brief **bpf_program__attach()** is a generic function for attaching * a BPF program based on auto-detection of program type, attach type, - * and extra paremeters, where applicable. + * and extra parameters, where applicable. * * @param prog BPF program to attach * @return Reference to the newly created BPF link; or NULL is returned on error, @@ -679,7 +679,7 @@ struct bpf_uprobe_opts { /** * @brief **bpf_program__attach_uprobe()** attaches a BPF program * to the userspace function which is found by binary path and - * offset. You can optionally specify a particular proccess to attach + * offset. You can optionally specify a particular process to attach * to. You can also optionally attach the program to the function * exit instead of entry. * @@ -1593,11 +1593,11 @@ LIBBPF_API int perf_buffer__buffer_fd(const struct perf_buffer *pb, size_t buf_i * memory region of the ring buffer. * This ring buffer can be used to implement a custom events consumer. * The ring buffer starts with the *struct perf_event_mmap_page*, which - * holds the ring buffer managment fields, when accessing the header + * holds the ring buffer management fields, when accessing the header * structure it's important to be SMP aware. * You can refer to *perf_event_read_simple* for a simple example. * @param pb the perf buffer structure - * @param buf_idx the buffer index to retreive + * @param buf_idx the buffer index to retrieve * @param buf (out) gets the base pointer of the mmap()'ed memory * @param buf_size (out) gets the size of the mmap()'ed region * @return 0 on success, negative error code for failure diff --git a/tools/lib/bpf/libbpf_legacy.h b/tools/lib/bpf/libbpf_legacy.h index 1e1be467bede..60b2600be88a 100644 --- a/tools/lib/bpf/libbpf_legacy.h +++ b/tools/lib/bpf/libbpf_legacy.h @@ -76,7 +76,7 @@ enum libbpf_strict_mode { * first BPF program or map creation operation. This is done only if * kernel is too old to support memcg-based memory accounting for BPF * subsystem. By default, RLIMIT_MEMLOCK limit is set to RLIM_INFINITY, - * but it can be overriden with libbpf_set_memlock_rlim() API. + * but it can be overridden with libbpf_set_memlock_rlim() API. * Note that libbpf_set_memlock_rlim() needs to be called before * the very first bpf_prog_load(), bpf_map_create() or bpf_object__load() * operation. @@ -97,7 +97,7 @@ LIBBPF_API int libbpf_set_strict_mode(enum libbpf_strict_mode mode); * @brief **libbpf_get_error()** extracts the error code from the passed * pointer * @param ptr pointer returned from libbpf API function - * @return error code; or 0 if no error occured + * @return error code; or 0 if no error occurred * * Note, as of libbpf 1.0 this function is not necessary and not recommended * to be used. Libbpf doesn't return error code embedded into the pointer diff --git a/tools/lib/bpf/skel_internal.h b/tools/lib/bpf/skel_internal.h index 1e82ab06c3eb..0875452521e9 100644 --- a/tools/lib/bpf/skel_internal.h +++ b/tools/lib/bpf/skel_internal.h @@ -107,7 +107,7 @@ static inline void skel_free(const void *p) * The loader program will perform probe_read_kernel() from maps.rodata.initial_value. * skel_finalize_map_data() sets skel->rodata to point to actual value in a bpf map and * does maps.rodata.initial_value = ~0ULL to signal skel_free_map_data() that kvfree - * is not nessary. + * is not necessary. * * For user space: * skel_prep_map_data() mmaps anon memory into skel->rodata that can be accessed directly.
Fix some spelling errors in the code comments of libbpf: betwen -> between paremeters -> parameters knowning -> knowing definiton -> definition compatiblity -> compatibility overriden -> overridden occured -> occurred proccess -> process managment -> management nessary -> necessary Signed-off-by: Yusheng Zheng <yunwei356@gmail.com> --- tools/lib/bpf/bpf_helpers.h | 2 +- tools/lib/bpf/bpf_tracing.h | 2 +- tools/lib/bpf/btf.c | 2 +- tools/lib/bpf/btf.h | 2 +- tools/lib/bpf/btf_dump.c | 2 +- tools/lib/bpf/libbpf.h | 10 +++++----- tools/lib/bpf/libbpf_legacy.h | 4 ++-- tools/lib/bpf/skel_internal.h | 2 +- 8 files changed, 13 insertions(+), 13 deletions(-)