@@ -558,6 +558,79 @@ static void test_uprobe_usdt(void)
uprobe_syscall_executed__destroy(skel);
}
+static volatile bool race_stop;
+
+static void *worker_trigger(void *arg)
+{
+ unsigned long rounds = 0;
+
+ while (!race_stop) {
+ uprobe_test();
+ rounds++;
+ }
+
+ printf("tid %d trigger rounds: %lu\n", gettid(), rounds);
+ return NULL;
+}
+
+static void *worker_attach(void *arg)
+{
+ struct uprobe_syscall_executed *skel;
+ unsigned long rounds = 0;
+
+ skel = uprobe_syscall_executed__open_and_load();
+ if (!ASSERT_OK_PTR(skel, "uprobe_syscall_executed__open_and_load"))
+ return NULL;
+
+ while (!race_stop) {
+ skel->links.test_uprobe_multi = bpf_program__attach_uprobe_multi(skel->progs.test_uprobe_multi,
+ -1, "/proc/self/exe", "uprobe_test", NULL);
+ if (!ASSERT_OK_PTR(skel->links.test_uprobe_multi, "bpf_program__attach_uprobe_multi"))
+ break;
+ bpf_link__destroy(skel->links.test_uprobe_multi);
+ skel->links.test_uprobe_multi = NULL;
+ rounds++;
+ }
+
+ printf("tid %d attach rounds: %lu hits: %d\n", gettid(), rounds, skel->bss->executed);
+ uprobe_syscall_executed__destroy(skel);
+ return NULL;
+}
+
+static void test_uprobe_race(void)
+{
+ int err, i, nr_cpus, nr;
+ pthread_t *threads;
+
+ nr_cpus = libbpf_num_possible_cpus();
+ if (!ASSERT_GE(nr_cpus, 0, "nr_cpus"))
+ return;
+
+ nr = nr_cpus * 2;
+ threads = malloc(sizeof(*threads) * nr);
+ if (!ASSERT_OK_PTR(threads, "malloc"))
+ return;
+
+ for (i = 0; i < nr_cpus; i++) {
+ err = pthread_create(&threads[i], NULL, worker_trigger, NULL);
+ if (!ASSERT_OK(err, "pthread_create"))
+ goto cleanup;
+ }
+
+ for (; i < nr; i++) {
+ err = pthread_create(&threads[i], NULL, worker_attach, NULL);
+ if (!ASSERT_OK(err, "pthread_create"))
+ goto cleanup;
+ }
+
+ sleep(4);
+
+cleanup:
+ race_stop = true;
+ for (nr = i, i = 0; i < nr; i++)
+ pthread_join(threads[i], NULL);
+}
+
static void __test_uprobe_syscall(void)
{
if (test__start_subtest("uretprobe_regs_equal"))
@@ -574,6 +647,8 @@ static void __test_uprobe_syscall(void)
test_uprobe_multi();
if (test__start_subtest("uprobe_usdt"))
test_uprobe_usdt();
+ if (test__start_subtest("uprobe_race"))
+ test_uprobe_race();
}
#else
static void __test_uprobe_syscall(void)
Adding test that makes sure parallel execution of the uprobe and attach/detach of optimized uprobe on it works properly. Signed-off-by: Jiri Olsa <jolsa@kernel.org> --- .../selftests/bpf/prog_tests/uprobe_syscall.c | 75 +++++++++++++++++++ 1 file changed, 75 insertions(+)