// SPDX-License-Identifier: GPL-2.0 #include #include #include #include "uprobe_multi.skel.h" #include "uprobe_multi_bench.skel.h" #include "uprobe_multi_usdt.skel.h" #include "uprobe_multi_consumers.skel.h" #include "uprobe_multi_pid_filter.skel.h" #include "uprobe_multi_session.skel.h" #include "uprobe_multi_session_single.skel.h" #include "uprobe_multi_session_cookie.skel.h" #include "uprobe_multi_session_recursive.skel.h" #include "uprobe_multi_verifier.skel.h" #include "bpf/libbpf_internal.h" #include "testing_helpers.h" #include "../sdt.h" static char test_data[] = "test_data"; noinline void uprobe_multi_func_1(void) { asm volatile (""); } noinline void uprobe_multi_func_2(void) { asm volatile (""); } noinline void uprobe_multi_func_3(void) { asm volatile (""); } noinline void usdt_trigger(void) { STAP_PROBE(test, pid_filter_usdt); } noinline void uprobe_session_recursive(int i) { if (i) uprobe_session_recursive(i - 1); } struct child { int go[2]; int c2p[2]; /* child -> parent channel */ int pid; int tid; pthread_t thread; char stack[65536]; }; static void release_child(struct child *child) { int child_status; if (!child) return; close(child->go[1]); close(child->go[0]); if (child->thread) pthread_join(child->thread, NULL); close(child->c2p[0]); close(child->c2p[1]); if (child->pid > 0) waitpid(child->pid, &child_status, 0); } static void kick_child(struct child *child) { char c = 1; if (child) { write(child->go[1], &c, 1); release_child(child); } fflush(NULL); } static int child_func(void *arg) { struct child *child = arg; int err, c; close(child->go[1]); /* wait for parent's kick */ err = read(child->go[0], &c, 1); if (err != 1) exit(err); uprobe_multi_func_1(); uprobe_multi_func_2(); uprobe_multi_func_3(); usdt_trigger(); exit(errno); } static int spawn_child_flag(struct child *child, bool clone_vm) { /* pipe to notify child to execute the trigger functions */ if (pipe(child->go)) return -1; if (clone_vm) { child->pid = child->tid = clone(child_func, child->stack + sizeof(child->stack)/2, CLONE_VM|SIGCHLD, child); } else { child->pid = child->tid = fork(); } if (child->pid < 0) { release_child(child); errno = EINVAL; return -1; } /* fork-ed child */ if (!clone_vm && child->pid == 0) child_func(child); return 0; } static int spawn_child(struct child *child) { return spawn_child_flag(child, false); } static void *child_thread(void *ctx) { struct child *child = ctx; int c = 0, err; child->tid = sys_gettid(); /* let parent know we are ready */ err = write(child->c2p[1], &c, 1); if (err != 1) pthread_exit(&err); /* wait for parent's kick */ err = read(child->go[0], &c, 1); if (err != 1) pthread_exit(&err); uprobe_multi_func_1(); uprobe_multi_func_2(); uprobe_multi_func_3(); usdt_trigger(); err = 0; pthread_exit(&err); } static int spawn_thread(struct child *child) { int c, err; /* pipe to notify child to execute the trigger functions */ if (pipe(child->go)) return -1; /* pipe to notify parent that child thread is ready */ if (pipe(child->c2p)) { close(child->go[0]); close(child->go[1]); return -1; } child->pid = getpid(); err = pthread_create(&child->thread, NULL, child_thread, child); if (err) { err = -errno; close(child->go[0]); close(child->go[1]); close(child->c2p[0]); close(child->c2p[1]); errno = -err; return -1; } err = read(child->c2p[0], &c, 1); if (!ASSERT_EQ(err, 1, "child_thread_ready")) return -1; return 0; } static void uprobe_multi_test_run(struct uprobe_multi *skel, struct child *child) { skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1; skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2; skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3; skel->bss->user_ptr = test_data; /* * Disable pid check in bpf program if we are pid filter test, * because the probe should be executed only by child->pid * passed at the probe attach. */ skel->bss->pid = child ? 0 : getpid(); skel->bss->expect_pid = child ? child->pid : 0; /* trigger all probes, if we are testing child *process*, just to make * sure that PID filtering doesn't let through activations from wrong * PIDs; when we test child *thread*, we don't want to do this to * avoid double counting number of triggering events */ if (!child || !child->thread) { uprobe_multi_func_1(); uprobe_multi_func_2(); uprobe_multi_func_3(); usdt_trigger(); } if (child) kick_child(child); /* * There are 2 entry and 2 exit probe called for each uprobe_multi_func_[123] * function and each sleepable probe (6) increments uprobe_multi_sleep_result. */ ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 2, "uprobe_multi_func_1_result"); ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 2, "uprobe_multi_func_2_result"); ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 2, "uprobe_multi_func_3_result"); ASSERT_EQ(skel->bss->uretprobe_multi_func_1_result, 2, "uretprobe_multi_func_1_result"); ASSERT_EQ(skel->bss->uretprobe_multi_func_2_result, 2, "uretprobe_multi_func_2_result"); ASSERT_EQ(skel->bss->uretprobe_multi_func_3_result, 2, "uretprobe_multi_func_3_result"); ASSERT_EQ(skel->bss->uprobe_multi_sleep_result, 6, "uprobe_multi_sleep_result"); ASSERT_FALSE(skel->bss->bad_pid_seen, "bad_pid_seen"); if (child) { ASSERT_EQ(skel->bss->child_pid, child->pid, "uprobe_multi_child_pid"); ASSERT_EQ(skel->bss->child_tid, child->tid, "uprobe_multi_child_tid"); } } static void test_skel_api(void) { struct uprobe_multi *skel = NULL; int err; skel = uprobe_multi__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load")) goto cleanup; err = uprobe_multi__attach(skel); if (!ASSERT_OK(err, "uprobe_multi__attach")) goto cleanup; uprobe_multi_test_run(skel, NULL); cleanup: uprobe_multi__destroy(skel); } static void __test_attach_api(const char *binary, const char *pattern, struct bpf_uprobe_multi_opts *opts, struct child *child) { pid_t pid = child ? child->pid : -1; struct uprobe_multi *skel = NULL; skel = uprobe_multi__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load")) goto cleanup; opts->retprobe = false; skel->links.uprobe = bpf_program__attach_uprobe_multi(skel->progs.uprobe, pid, binary, pattern, opts); if (!ASSERT_OK_PTR(skel->links.uprobe, "bpf_program__attach_uprobe_multi")) goto cleanup; opts->retprobe = true; skel->links.uretprobe = bpf_program__attach_uprobe_multi(skel->progs.uretprobe, pid, binary, pattern, opts); if (!ASSERT_OK_PTR(skel->links.uretprobe, "bpf_program__attach_uprobe_multi")) goto cleanup; opts->retprobe = false; skel->links.uprobe_sleep = bpf_program__attach_uprobe_multi(skel->progs.uprobe_sleep, pid, binary, pattern, opts); if (!ASSERT_OK_PTR(skel->links.uprobe_sleep, "bpf_program__attach_uprobe_multi")) goto cleanup; opts->retprobe = true; skel->links.uretprobe_sleep = bpf_program__attach_uprobe_multi(skel->progs.uretprobe_sleep, pid, binary, pattern, opts); if (!ASSERT_OK_PTR(skel->links.uretprobe_sleep, "bpf_program__attach_uprobe_multi")) goto cleanup; opts->retprobe = false; skel->links.uprobe_extra = bpf_program__attach_uprobe_multi(skel->progs.uprobe_extra, -1, binary, pattern, opts); if (!ASSERT_OK_PTR(skel->links.uprobe_extra, "bpf_program__attach_uprobe_multi")) goto cleanup; /* Attach (uprobe-backed) USDTs */ skel->links.usdt_pid = bpf_program__attach_usdt(skel->progs.usdt_pid, pid, binary, "test", "pid_filter_usdt", NULL); if (!ASSERT_OK_PTR(skel->links.usdt_pid, "attach_usdt_pid")) goto cleanup; skel->links.usdt_extra = bpf_program__attach_usdt(skel->progs.usdt_extra, -1, binary, "test", "pid_filter_usdt", NULL); if (!ASSERT_OK_PTR(skel->links.usdt_extra, "attach_usdt_extra")) goto cleanup; uprobe_multi_test_run(skel, child); ASSERT_FALSE(skel->bss->bad_pid_seen_usdt, "bad_pid_seen_usdt"); if (child) { ASSERT_EQ(skel->bss->child_pid_usdt, child->pid, "usdt_multi_child_pid"); ASSERT_EQ(skel->bss->child_tid_usdt, child->tid, "usdt_multi_child_tid"); } cleanup: uprobe_multi__destroy(skel); } static void test_attach_api(const char *binary, const char *pattern, struct bpf_uprobe_multi_opts *opts) { static struct child child; /* no pid filter */ __test_attach_api(binary, pattern, opts, NULL); /* pid filter */ if (!ASSERT_OK(spawn_child(&child), "spawn_child")) return; __test_attach_api(binary, pattern, opts, &child); /* pid filter (thread) */ if (!ASSERT_OK(spawn_thread(&child), "spawn_thread")) return; __test_attach_api(binary, pattern, opts, &child); } static void test_attach_api_pattern(void) { LIBBPF_OPTS(bpf_uprobe_multi_opts, opts); test_attach_api("/proc/self/exe", "uprobe_multi_func_*", &opts); test_attach_api("/proc/self/exe", "uprobe_multi_func_?", &opts); } static void test_attach_api_syms(void) { LIBBPF_OPTS(bpf_uprobe_multi_opts, opts); const char *syms[3] = { "uprobe_multi_func_1", "uprobe_multi_func_2", "uprobe_multi_func_3", }; opts.syms = syms; opts.cnt = ARRAY_SIZE(syms); test_attach_api("/proc/self/exe", NULL, &opts); } static void test_attach_api_fails(void) { LIBBPF_OPTS(bpf_link_create_opts, opts); const char *path = "/proc/self/exe"; struct uprobe_multi *skel = NULL; int prog_fd, link_fd = -1; unsigned long offset = 0; skel = uprobe_multi__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load")) goto cleanup; prog_fd = bpf_program__fd(skel->progs.uprobe_extra); /* abnormal cnt */ opts.uprobe_multi.path = path; opts.uprobe_multi.offsets = &offset; opts.uprobe_multi.cnt = INT_MAX; link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -E2BIG, "big cnt")) goto cleanup; /* cnt is 0 */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.offsets = (unsigned long *) &offset, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EINVAL, "cnt_is_zero")) goto cleanup; /* negative offset */ offset = -1; opts.uprobe_multi.path = path; opts.uprobe_multi.offsets = (unsigned long *) &offset; opts.uprobe_multi.cnt = 1; link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EINVAL, "offset_is_negative")) goto cleanup; /* offsets is NULL */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EINVAL, "offsets_is_null")) goto cleanup; /* wrong offsets pointer */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.offsets = (unsigned long *) 1, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EFAULT, "offsets_is_wrong")) goto cleanup; /* path is NULL */ offset = 1; LIBBPF_OPTS_RESET(opts, .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EINVAL, "path_is_null")) goto cleanup; /* wrong path pointer */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = (const char *) 1, .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EFAULT, "path_is_wrong")) goto cleanup; /* wrong path type */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = "/", .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EBADF, "path_is_wrong_type")) goto cleanup; /* wrong cookies pointer */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cookies = (__u64 *) 1ULL, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EFAULT, "cookies_is_wrong")) goto cleanup; /* wrong ref_ctr_offsets pointer */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cookies = (__u64 *) &offset, .uprobe_multi.ref_ctr_offsets = (unsigned long *) 1, .uprobe_multi.cnt = 1, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EFAULT, "ref_ctr_offsets_is_wrong")) goto cleanup; /* wrong flags */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.flags = 1 << 31, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; if (!ASSERT_EQ(link_fd, -EINVAL, "wrong_flags")) goto cleanup; /* wrong pid */ LIBBPF_OPTS_RESET(opts, .uprobe_multi.path = path, .uprobe_multi.offsets = (unsigned long *) &offset, .uprobe_multi.cnt = 1, .uprobe_multi.pid = -2, ); link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) goto cleanup; ASSERT_EQ(link_fd, -EINVAL, "pid_is_wrong"); cleanup: if (link_fd >= 0) close(link_fd); uprobe_multi__destroy(skel); } #ifdef __x86_64__ noinline void uprobe_multi_error_func(void) { /* * If --fcf-protection=branch is enabled the gcc generates endbr as * first instruction, so marking the exact address of int3 with the * symbol to be used in the attach_uprobe_fail_trap test below. */ asm volatile ( ".globl uprobe_multi_error_func_int3; \n" "uprobe_multi_error_func_int3: \n" "int3 \n" ); } /* * Attaching uprobe on uprobe_multi_error_func results in error * because it already starts with int3 instruction. */ static void attach_uprobe_fail_trap(struct uprobe_multi *skel) { LIBBPF_OPTS(bpf_uprobe_multi_opts, opts); const char *syms[4] = { "uprobe_multi_func_1", "uprobe_multi_func_2", "uprobe_multi_func_3", "uprobe_multi_error_func_int3", }; opts.syms = syms; opts.cnt = ARRAY_SIZE(syms); skel->links.uprobe = bpf_program__attach_uprobe_multi(skel->progs.uprobe, -1, "/proc/self/exe", NULL, &opts); if (!ASSERT_ERR_PTR(skel->links.uprobe, "bpf_program__attach_uprobe_multi")) { bpf_link__destroy(skel->links.uprobe); skel->links.uprobe = NULL; } } #else static void attach_uprobe_fail_trap(struct uprobe_multi *skel) { } #endif short sema_1 __used, sema_2 __used; static void attach_uprobe_fail_refctr(struct uprobe_multi *skel) { unsigned long *tmp_offsets = NULL, *tmp_ref_ctr_offsets = NULL; unsigned long offsets[3], ref_ctr_offsets[3]; LIBBPF_OPTS(bpf_link_create_opts, opts); const char *path = "/proc/self/exe"; const char *syms[3] = { "uprobe_multi_func_1", "uprobe_multi_func_2", }; const char *sema[3] = { "sema_1", "sema_2", }; int prog_fd, link_fd, err; prog_fd = bpf_program__fd(skel->progs.uprobe_extra); err = elf_resolve_syms_offsets("/proc/self/exe", 2, (const char **) &syms, &tmp_offsets, STT_FUNC); if (!ASSERT_OK(err, "elf_resolve_syms_offsets_func")) return; err = elf_resolve_syms_offsets("/proc/self/exe", 2, (const char **) &sema, &tmp_ref_ctr_offsets, STT_OBJECT); if (!ASSERT_OK(err, "elf_resolve_syms_offsets_sema")) goto cleanup; /* * We attach to 3 uprobes on 2 functions, so 2 uprobes share single function, * but with different ref_ctr_offset which is not allowed and results in fail. */ offsets[0] = tmp_offsets[0]; /* uprobe_multi_func_1 */ offsets[1] = tmp_offsets[1]; /* uprobe_multi_func_2 */ offsets[2] = tmp_offsets[1]; /* uprobe_multi_func_2 */ ref_ctr_offsets[0] = tmp_ref_ctr_offsets[0]; /* sema_1 */ ref_ctr_offsets[1] = tmp_ref_ctr_offsets[1]; /* sema_2 */ ref_ctr_offsets[2] = tmp_ref_ctr_offsets[0]; /* sema_1, error */ opts.uprobe_multi.path = path; opts.uprobe_multi.offsets = (const unsigned long *) &offsets; opts.uprobe_multi.ref_ctr_offsets = (const unsigned long *) &ref_ctr_offsets; opts.uprobe_multi.cnt = 3; link_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_ERR(link_fd, "link_fd")) close(link_fd); cleanup: free(tmp_ref_ctr_offsets); free(tmp_offsets); } static void test_attach_uprobe_fails(void) { struct uprobe_multi *skel = NULL; skel = uprobe_multi__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load")) return; /* attach fails due to adding uprobe on trap instruction, x86_64 only */ attach_uprobe_fail_trap(skel); /* attach fail due to wrong ref_ctr_offs on one of the uprobes */ attach_uprobe_fail_refctr(skel); uprobe_multi__destroy(skel); } static void __test_link_api(struct child *child) { int prog_fd, link1_fd = -1, link2_fd = -1, link3_fd = -1, link4_fd = -1; LIBBPF_OPTS(bpf_link_create_opts, opts); const char *path = "/proc/self/exe"; struct uprobe_multi *skel = NULL; unsigned long *offsets = NULL; const char *syms[3] = { "uprobe_multi_func_1", "uprobe_multi_func_2", "uprobe_multi_func_3", }; int link_extra_fd = -1; int err; err = elf_resolve_syms_offsets(path, 3, syms, (unsigned long **) &offsets, STT_FUNC); if (!ASSERT_OK(err, "elf_resolve_syms_offsets")) return; opts.uprobe_multi.path = path; opts.uprobe_multi.offsets = offsets; opts.uprobe_multi.cnt = ARRAY_SIZE(syms); opts.uprobe_multi.pid = child ? child->pid : 0; skel = uprobe_multi__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open_and_load")) goto cleanup; opts.kprobe_multi.flags = 0; prog_fd = bpf_program__fd(skel->progs.uprobe); link1_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_GE(link1_fd, 0, "link1_fd")) goto cleanup; opts.kprobe_multi.flags = BPF_F_UPROBE_MULTI_RETURN; prog_fd = bpf_program__fd(skel->progs.uretprobe); link2_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_GE(link2_fd, 0, "link2_fd")) goto cleanup; opts.kprobe_multi.flags = 0; prog_fd = bpf_program__fd(skel->progs.uprobe_sleep); link3_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_GE(link3_fd, 0, "link3_fd")) goto cleanup; opts.kprobe_multi.flags = BPF_F_UPROBE_MULTI_RETURN; prog_fd = bpf_program__fd(skel->progs.uretprobe_sleep); link4_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_GE(link4_fd, 0, "link4_fd")) goto cleanup; opts.kprobe_multi.flags = 0; opts.uprobe_multi.pid = 0; prog_fd = bpf_program__fd(skel->progs.uprobe_extra); link_extra_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_UPROBE_MULTI, &opts); if (!ASSERT_GE(link_extra_fd, 0, "link_extra_fd")) goto cleanup; uprobe_multi_test_run(skel, child); cleanup: if (link1_fd >= 0) close(link1_fd); if (link2_fd >= 0) close(link2_fd); if (link3_fd >= 0) close(link3_fd); if (link4_fd >= 0) close(link4_fd); if (link_extra_fd >= 0) close(link_extra_fd); uprobe_multi__destroy(skel); free(offsets); } static void test_link_api(void) { static struct child child; /* no pid filter */ __test_link_api(NULL); /* pid filter */ if (!ASSERT_OK(spawn_child(&child), "spawn_child")) return; __test_link_api(&child); /* pid filter (thread) */ if (!ASSERT_OK(spawn_thread(&child), "spawn_thread")) return; __test_link_api(&child); } static struct bpf_program * get_program(struct uprobe_multi_consumers *skel, int prog) { switch (prog) { case 0: return skel->progs.uprobe_0; case 1: return skel->progs.uprobe_1; case 2: return skel->progs.uprobe_2; case 3: return skel->progs.uprobe_3; default: ASSERT_FAIL("get_program"); return NULL; } } static struct bpf_link ** get_link(struct uprobe_multi_consumers *skel, int link) { switch (link) { case 0: return &skel->links.uprobe_0; case 1: return &skel->links.uprobe_1; case 2: return &skel->links.uprobe_2; case 3: return &skel->links.uprobe_3; default: ASSERT_FAIL("get_link"); return NULL; } } static int uprobe_attach(struct uprobe_multi_consumers *skel, int idx, unsigned long offset) { struct bpf_program *prog = get_program(skel, idx); struct bpf_link **link = get_link(skel, idx); LIBBPF_OPTS(bpf_uprobe_multi_opts, opts); if (!prog || !link) return -1; opts.offsets = &offset; opts.cnt = 1; /* * bit/prog: 0 uprobe entry * bit/prog: 1 uprobe return * bit/prog: 2 uprobe session without return * bit/prog: 3 uprobe session with return */ opts.retprobe = idx == 1; opts.session = idx == 2 || idx == 3; *link = bpf_program__attach_uprobe_multi(prog, 0, "/proc/self/exe", NULL, &opts); if (!ASSERT_OK_PTR(*link, "bpf_program__attach_uprobe_multi")) return -1; return 0; } static void uprobe_detach(struct uprobe_multi_consumers *skel, int idx) { struct bpf_link **link = get_link(skel, idx); bpf_link__destroy(*link); *link = NULL; } static bool test_bit(int bit, unsigned long val) { return val & (1 << bit); } noinline int uprobe_consumer_test(struct uprobe_multi_consumers *skel, unsigned long before, unsigned long after, unsigned long offset) { int idx; /* detach uprobe for each unset programs in 'before' state ... */ for (idx = 0; idx < 4; idx++) { if (test_bit(idx, before) && !test_bit(idx, after)) uprobe_detach(skel, idx); } /* ... and attach all new programs in 'after' state */ for (idx = 0; idx < 4; idx++) { if (!test_bit(idx, before) && test_bit(idx, after)) { if (!ASSERT_OK(uprobe_attach(skel, idx, offset), "uprobe_attach_after")) return -1; } } return 0; } /* * We generate 16 consumer_testX functions that will have uprobe installed on * and will be called in separate threads. All function pointer are stored in * "consumers" section and each thread will pick one function based on index. */ extern const void *__start_consumers; #define __CONSUMER_TEST(func) \ noinline int func(struct uprobe_multi_consumers *skel, unsigned long before, \ unsigned long after, unsigned long offset) \ { \ return uprobe_consumer_test(skel, before, after, offset); \ } \ void *__ ## func __used __attribute__((section("consumers"))) = (void *) func; #define CONSUMER_TEST(func) __CONSUMER_TEST(func) #define C1 CONSUMER_TEST(__PASTE(consumer_test, __COUNTER__)) #define C4 C1 C1 C1 C1 #define C16 C4 C4 C4 C4 C16 typedef int (*test_t)(struct uprobe_multi_consumers *, unsigned long, unsigned long, unsigned long); static int consumer_test(struct uprobe_multi_consumers *skel, unsigned long before, unsigned long after, test_t test, unsigned long offset) { int err, idx, ret = -1; printf("consumer_test before %lu after %lu\n", before, after); /* 'before' is each, we attach uprobe for every set idx */ for (idx = 0; idx < 4; idx++) { if (test_bit(idx, before)) { if (!ASSERT_OK(uprobe_attach(skel, idx, offset), "uprobe_attach_before")) goto cleanup; } } err = test(skel, before, after, offset); if (!ASSERT_EQ(err, 0, "uprobe_consumer_test")) goto cleanup; for (idx = 0; idx < 4; idx++) { bool uret_stays, uret_survives; const char *fmt = "BUG"; __u64 val = 0; switch (idx) { case 0: /* * uprobe entry * +1 if define in 'before' */ if (test_bit(idx, before)) val++; fmt = "prog 0: uprobe"; break; case 1: /* * To trigger uretprobe consumer, the uretprobe under test either stayed from * before to after (uret_stays + test_bit) or uretprobe instance survived and * we have uretprobe active in after (uret_survives + test_bit) */ uret_stays = before & after & 0b0110; uret_survives = ((before & 0b0110) && (after & 0b0110) && (before & 0b1001)); if ((uret_stays || uret_survives) && test_bit(idx, after)) val++; fmt = "prog 1: uretprobe"; break; case 2: /* * session with return * +1 if defined in 'before' * +1 if defined in 'after' */ if (test_bit(idx, before)) { val++; if (test_bit(idx, after)) val++; } fmt = "prog 2: session with return"; break; case 3: /* * session without return * +1 if defined in 'before' */ if (test_bit(idx, before)) val++; fmt = "prog 3: session with NO return"; break; } if (!ASSERT_EQ(skel->bss->uprobe_result[idx], val, fmt)) goto cleanup; skel->bss->uprobe_result[idx] = 0; } ret = 0; cleanup: for (idx = 0; idx < 4; idx++) uprobe_detach(skel, idx); return ret; } #define CONSUMER_MAX 16 /* * Each thread runs 1/16 of the load by running test for single * 'before' number (based on thread index) and full scale of * 'after' numbers. */ static void *consumer_thread(void *arg) { unsigned long idx = (unsigned long) arg; struct uprobe_multi_consumers *skel; unsigned long offset; const void *func; int after; skel = uprobe_multi_consumers__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_consumers__open_and_load")) return NULL; func = *((&__start_consumers) + idx); offset = get_uprobe_offset(func); if (!ASSERT_GE(offset, 0, "uprobe_offset")) goto out; for (after = 0; after < CONSUMER_MAX; after++) if (consumer_test(skel, idx, after, func, offset)) goto out; out: uprobe_multi_consumers__destroy(skel); return NULL; } static void test_consumers(void) { pthread_t pt[CONSUMER_MAX]; unsigned long idx; int err; /* * The idea of this test is to try all possible combinations of * uprobes consumers attached on single function. * * - 1 uprobe entry consumer * - 1 uprobe exit consumer * - 1 uprobe session with return * - 1 uprobe session without return * * The test uses 4 uprobes attached on single function, but that * translates into single uprobe with 4 consumers in kernel. * * The before/after values present the state of attached consumers * before and after the probed function: * * bit/prog 0 : uprobe entry * bit/prog 1 : uprobe return * * For example for: * * before = 0b01 * after = 0b10 * * it means that before we call 'uprobe_consumer_test' we attach * uprobes defined in 'before' value: * * - bit/prog 1: uprobe entry * * uprobe_consumer_test is called and inside it we attach and detach * uprobes based on 'after' value: * * - bit/prog 0: is detached * - bit/prog 1: is attached * * uprobe_consumer_test returns and we check counters values increased * by bpf programs on each uprobe to match the expected count based on * before/after bits. */ for (idx = 0; idx < CONSUMER_MAX; idx++) { err = pthread_create(&pt[idx], NULL, consumer_thread, (void *) idx); if (!ASSERT_OK(err, "pthread_create")) break; } while (idx) pthread_join(pt[--idx], NULL); } static struct bpf_program *uprobe_multi_program(struct uprobe_multi_pid_filter *skel, int idx) { switch (idx) { case 0: return skel->progs.uprobe_multi_0; case 1: return skel->progs.uprobe_multi_1; case 2: return skel->progs.uprobe_multi_2; } return NULL; } #define TASKS 3 static void run_pid_filter(struct uprobe_multi_pid_filter *skel, bool clone_vm, bool retprobe) { LIBBPF_OPTS(bpf_uprobe_multi_opts, opts, .retprobe = retprobe); struct bpf_link *link[TASKS] = {}; struct child child[TASKS] = {}; int i; memset(skel->bss->test, 0, sizeof(skel->bss->test)); for (i = 0; i < TASKS; i++) { if (!ASSERT_OK(spawn_child_flag(&child[i], clone_vm), "spawn_child")) goto cleanup; skel->bss->pids[i] = child[i].pid; } for (i = 0; i < TASKS; i++) { link[i] = bpf_program__attach_uprobe_multi(uprobe_multi_program(skel, i), child[i].pid, "/proc/self/exe", "uprobe_multi_func_1", &opts); if (!ASSERT_OK_PTR(link[i], "bpf_program__attach_uprobe_multi")) goto cleanup; } for (i = 0; i < TASKS; i++) kick_child(&child[i]); for (i = 0; i < TASKS; i++) { ASSERT_EQ(skel->bss->test[i][0], 1, "pid"); ASSERT_EQ(skel->bss->test[i][1], 0, "unknown"); } cleanup: for (i = 0; i < TASKS; i++) bpf_link__destroy(link[i]); for (i = 0; i < TASKS; i++) release_child(&child[i]); } static void test_pid_filter_process(bool clone_vm) { struct uprobe_multi_pid_filter *skel; skel = uprobe_multi_pid_filter__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_pid_filter__open_and_load")) return; run_pid_filter(skel, clone_vm, false); run_pid_filter(skel, clone_vm, true); uprobe_multi_pid_filter__destroy(skel); } static void test_session_skel_api(void) { struct uprobe_multi_session *skel = NULL; LIBBPF_OPTS(bpf_kprobe_multi_opts, opts); struct bpf_link *link = NULL; int err; skel = uprobe_multi_session__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_session__open_and_load")) goto cleanup; skel->bss->pid = getpid(); skel->bss->user_ptr = test_data; err = uprobe_multi_session__attach(skel); if (!ASSERT_OK(err, "uprobe_multi_session__attach")) goto cleanup; /* trigger all probes */ skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1; skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2; skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3; uprobe_multi_func_1(); uprobe_multi_func_2(); uprobe_multi_func_3(); /* * We expect 2 for uprobe_multi_func_2 because it runs both entry/return probe, * uprobe_multi_func_[13] run just the entry probe. All expected numbers are * doubled, because we run extra test for sleepable session. */ ASSERT_EQ(skel->bss->uprobe_session_result[0], 2, "uprobe_multi_func_1_result"); ASSERT_EQ(skel->bss->uprobe_session_result[1], 4, "uprobe_multi_func_2_result"); ASSERT_EQ(skel->bss->uprobe_session_result[2], 2, "uprobe_multi_func_3_result"); /* We expect increase in 3 entry and 1 return session calls -> 4 */ ASSERT_EQ(skel->bss->uprobe_multi_sleep_result, 4, "uprobe_multi_sleep_result"); cleanup: bpf_link__destroy(link); uprobe_multi_session__destroy(skel); } static void test_session_single_skel_api(void) { struct uprobe_multi_session_single *skel = NULL; LIBBPF_OPTS(bpf_kprobe_multi_opts, opts); int err; skel = uprobe_multi_session_single__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_session_single__open_and_load")) goto cleanup; skel->bss->pid = getpid(); err = uprobe_multi_session_single__attach(skel); if (!ASSERT_OK(err, "uprobe_multi_session_single__attach")) goto cleanup; uprobe_multi_func_1(); /* * We expect consumer 0 and 2 to trigger just entry handler (value 1) * and consumer 1 to hit both (value 2). */ ASSERT_EQ(skel->bss->uprobe_session_result[0], 1, "uprobe_session_result_0"); ASSERT_EQ(skel->bss->uprobe_session_result[1], 2, "uprobe_session_result_1"); ASSERT_EQ(skel->bss->uprobe_session_result[2], 1, "uprobe_session_result_2"); cleanup: uprobe_multi_session_single__destroy(skel); } static void test_session_cookie_skel_api(void) { struct uprobe_multi_session_cookie *skel = NULL; int err; skel = uprobe_multi_session_cookie__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_session_cookie__open_and_load")) goto cleanup; skel->bss->pid = getpid(); err = uprobe_multi_session_cookie__attach(skel); if (!ASSERT_OK(err, "uprobe_multi_session_cookie__attach")) goto cleanup; /* trigger all probes */ uprobe_multi_func_1(); uprobe_multi_func_2(); uprobe_multi_func_3(); ASSERT_EQ(skel->bss->test_uprobe_1_result, 1, "test_uprobe_1_result"); ASSERT_EQ(skel->bss->test_uprobe_2_result, 2, "test_uprobe_2_result"); ASSERT_EQ(skel->bss->test_uprobe_3_result, 3, "test_uprobe_3_result"); cleanup: uprobe_multi_session_cookie__destroy(skel); } static void test_session_recursive_skel_api(void) { struct uprobe_multi_session_recursive *skel = NULL; int i, err; skel = uprobe_multi_session_recursive__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_session_recursive__open_and_load")) goto cleanup; skel->bss->pid = getpid(); err = uprobe_multi_session_recursive__attach(skel); if (!ASSERT_OK(err, "uprobe_multi_session_recursive__attach")) goto cleanup; for (i = 0; i < ARRAY_SIZE(skel->bss->test_uprobe_cookie_entry); i++) skel->bss->test_uprobe_cookie_entry[i] = i + 1; uprobe_session_recursive(5); /* * entry uprobe: * uprobe_session_recursive(5) { *cookie = 1, return 0 * uprobe_session_recursive(4) { *cookie = 2, return 1 * uprobe_session_recursive(3) { *cookie = 3, return 0 * uprobe_session_recursive(2) { *cookie = 4, return 1 * uprobe_session_recursive(1) { *cookie = 5, return 0 * uprobe_session_recursive(0) { *cookie = 6, return 1 * return uprobe: * } i = 0 not executed * } i = 1 test_uprobe_cookie_return[0] = 5 * } i = 2 not executed * } i = 3 test_uprobe_cookie_return[1] = 3 * } i = 4 not executed * } i = 5 test_uprobe_cookie_return[2] = 1 */ ASSERT_EQ(skel->bss->idx_entry, 6, "idx_entry"); ASSERT_EQ(skel->bss->idx_return, 3, "idx_return"); ASSERT_EQ(skel->bss->test_uprobe_cookie_return[0], 5, "test_uprobe_cookie_return[0]"); ASSERT_EQ(skel->bss->test_uprobe_cookie_return[1], 3, "test_uprobe_cookie_return[1]"); ASSERT_EQ(skel->bss->test_uprobe_cookie_return[2], 1, "test_uprobe_cookie_return[2]"); cleanup: uprobe_multi_session_recursive__destroy(skel); } static void test_bench_attach_uprobe(void) { long attach_start_ns = 0, attach_end_ns = 0; struct uprobe_multi_bench *skel = NULL; long detach_start_ns, detach_end_ns; double attach_delta, detach_delta; int err; skel = uprobe_multi_bench__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi_bench__open_and_load")) goto cleanup; attach_start_ns = get_time_ns(); err = uprobe_multi_bench__attach(skel); if (!ASSERT_OK(err, "uprobe_multi_bench__attach")) goto cleanup; attach_end_ns = get_time_ns(); system("./uprobe_multi bench"); ASSERT_EQ(skel->bss->count, 50000, "uprobes_count"); cleanup: detach_start_ns = get_time_ns(); uprobe_multi_bench__destroy(skel); detach_end_ns = get_time_ns(); attach_delta = (attach_end_ns - attach_start_ns) / 1000000000.0; detach_delta = (detach_end_ns - detach_start_ns) / 1000000000.0; printf("%s: attached in %7.3lfs\n", __func__, attach_delta); printf("%s: detached in %7.3lfs\n", __func__, detach_delta); } static void test_bench_attach_usdt(void) { long attach_start_ns = 0, attach_end_ns = 0; struct uprobe_multi_usdt *skel = NULL; long detach_start_ns, detach_end_ns; double attach_delta, detach_delta; skel = uprobe_multi_usdt__open_and_load(); if (!ASSERT_OK_PTR(skel, "uprobe_multi__open")) goto cleanup; attach_start_ns = get_time_ns(); skel->links.usdt0 = bpf_program__attach_usdt(skel->progs.usdt0, -1, "./uprobe_multi", "test", "usdt", NULL); if (!ASSERT_OK_PTR(skel->links.usdt0, "bpf_program__attach_usdt")) goto cleanup; attach_end_ns = get_time_ns(); system("./uprobe_multi usdt"); ASSERT_EQ(skel->bss->count, 50000, "usdt_count"); cleanup: detach_start_ns = get_time_ns(); uprobe_multi_usdt__destroy(skel); detach_end_ns = get_time_ns(); attach_delta = (attach_end_ns - attach_start_ns) / 1000000000.0; detach_delta = (detach_end_ns - detach_start_ns) / 1000000000.0; printf("%s: attached in %7.3lfs\n", __func__, attach_delta); printf("%s: detached in %7.3lfs\n", __func__, detach_delta); } void test_uprobe_multi_test(void) { if (test__start_subtest("skel_api")) test_skel_api(); if (test__start_subtest("attach_api_pattern")) test_attach_api_pattern(); if (test__start_subtest("attach_api_syms")) test_attach_api_syms(); if (test__start_subtest("link_api")) test_link_api(); if (test__start_subtest("bench_uprobe")) test_bench_attach_uprobe(); if (test__start_subtest("bench_usdt")) test_bench_attach_usdt(); if (test__start_subtest("attach_api_fails")) test_attach_api_fails(); if (test__start_subtest("attach_uprobe_fails")) test_attach_uprobe_fails(); if (test__start_subtest("consumers")) test_consumers(); if (test__start_subtest("filter_fork")) test_pid_filter_process(false); if (test__start_subtest("filter_clone_vm")) test_pid_filter_process(true); if (test__start_subtest("session")) test_session_skel_api(); if (test__start_subtest("session_single")) test_session_single_skel_api(); if (test__start_subtest("session_cookie")) test_session_cookie_skel_api(); if (test__start_subtest("session_cookie_recursive")) test_session_recursive_skel_api(); RUN_TESTS(uprobe_multi_verifier); }