1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright (c) 2021 Facebook */
3 #define _GNU_SOURCE
4 #include <pthread.h>
5 #include <sched.h>
6 #include <sys/syscall.h>
7 #include <sys/mman.h>
8 #include <unistd.h>
9 #include <test_progs.h>
10 #include <network_helpers.h>
11 #include <bpf/btf.h>
12 #include "test_bpf_cookie.skel.h"
13 #include "kprobe_multi.skel.h"
14 #include "uprobe_multi.skel.h"
15
16 /* uprobe attach point */
trigger_func(void)17 static noinline void trigger_func(void)
18 {
19 asm volatile ("");
20 }
21
kprobe_subtest(struct test_bpf_cookie * skel)22 static void kprobe_subtest(struct test_bpf_cookie *skel)
23 {
24 DECLARE_LIBBPF_OPTS(bpf_kprobe_opts, opts);
25 struct bpf_link *link1 = NULL, *link2 = NULL;
26 struct bpf_link *retlink1 = NULL, *retlink2 = NULL;
27
28 /* attach two kprobes */
29 opts.bpf_cookie = 0x1;
30 opts.retprobe = false;
31 link1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe,
32 SYS_NANOSLEEP_KPROBE_NAME, &opts);
33 if (!ASSERT_OK_PTR(link1, "link1"))
34 goto cleanup;
35
36 opts.bpf_cookie = 0x2;
37 opts.retprobe = false;
38 link2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kprobe,
39 SYS_NANOSLEEP_KPROBE_NAME, &opts);
40 if (!ASSERT_OK_PTR(link2, "link2"))
41 goto cleanup;
42
43 /* attach two kretprobes */
44 opts.bpf_cookie = 0x10;
45 opts.retprobe = true;
46 retlink1 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe,
47 SYS_NANOSLEEP_KPROBE_NAME, &opts);
48 if (!ASSERT_OK_PTR(retlink1, "retlink1"))
49 goto cleanup;
50
51 opts.bpf_cookie = 0x20;
52 opts.retprobe = true;
53 retlink2 = bpf_program__attach_kprobe_opts(skel->progs.handle_kretprobe,
54 SYS_NANOSLEEP_KPROBE_NAME, &opts);
55 if (!ASSERT_OK_PTR(retlink2, "retlink2"))
56 goto cleanup;
57
58 /* trigger kprobe && kretprobe */
59 usleep(1);
60
61 ASSERT_EQ(skel->bss->kprobe_res, 0x1 | 0x2, "kprobe_res");
62 ASSERT_EQ(skel->bss->kretprobe_res, 0x10 | 0x20, "kretprobe_res");
63
64 cleanup:
65 bpf_link__destroy(link1);
66 bpf_link__destroy(link2);
67 bpf_link__destroy(retlink1);
68 bpf_link__destroy(retlink2);
69 }
70
kprobe_multi_test_run(struct kprobe_multi * skel)71 static void kprobe_multi_test_run(struct kprobe_multi *skel)
72 {
73 LIBBPF_OPTS(bpf_test_run_opts, topts);
74 int err, prog_fd;
75
76 prog_fd = bpf_program__fd(skel->progs.trigger);
77 err = bpf_prog_test_run_opts(prog_fd, &topts);
78 ASSERT_OK(err, "test_run");
79 ASSERT_EQ(topts.retval, 0, "test_run");
80
81 ASSERT_EQ(skel->bss->kprobe_test1_result, 1, "kprobe_test1_result");
82 ASSERT_EQ(skel->bss->kprobe_test2_result, 1, "kprobe_test2_result");
83 ASSERT_EQ(skel->bss->kprobe_test3_result, 1, "kprobe_test3_result");
84 ASSERT_EQ(skel->bss->kprobe_test4_result, 1, "kprobe_test4_result");
85 ASSERT_EQ(skel->bss->kprobe_test5_result, 1, "kprobe_test5_result");
86 ASSERT_EQ(skel->bss->kprobe_test6_result, 1, "kprobe_test6_result");
87 ASSERT_EQ(skel->bss->kprobe_test7_result, 1, "kprobe_test7_result");
88 ASSERT_EQ(skel->bss->kprobe_test8_result, 1, "kprobe_test8_result");
89
90 ASSERT_EQ(skel->bss->kretprobe_test1_result, 1, "kretprobe_test1_result");
91 ASSERT_EQ(skel->bss->kretprobe_test2_result, 1, "kretprobe_test2_result");
92 ASSERT_EQ(skel->bss->kretprobe_test3_result, 1, "kretprobe_test3_result");
93 ASSERT_EQ(skel->bss->kretprobe_test4_result, 1, "kretprobe_test4_result");
94 ASSERT_EQ(skel->bss->kretprobe_test5_result, 1, "kretprobe_test5_result");
95 ASSERT_EQ(skel->bss->kretprobe_test6_result, 1, "kretprobe_test6_result");
96 ASSERT_EQ(skel->bss->kretprobe_test7_result, 1, "kretprobe_test7_result");
97 ASSERT_EQ(skel->bss->kretprobe_test8_result, 1, "kretprobe_test8_result");
98 }
99
kprobe_multi_link_api_subtest(void)100 static void kprobe_multi_link_api_subtest(void)
101 {
102 int prog_fd, link1_fd = -1, link2_fd = -1;
103 struct kprobe_multi *skel = NULL;
104 LIBBPF_OPTS(bpf_link_create_opts, opts);
105 unsigned long long addrs[8];
106 __u64 cookies[8];
107
108 if (!ASSERT_OK(load_kallsyms(), "load_kallsyms"))
109 goto cleanup;
110
111 skel = kprobe_multi__open_and_load();
112 if (!ASSERT_OK_PTR(skel, "fentry_raw_skel_load"))
113 goto cleanup;
114
115 skel->bss->pid = getpid();
116 skel->bss->test_cookie = true;
117
118 #define GET_ADDR(__sym, __addr) ({ \
119 __addr = ksym_get_addr(__sym); \
120 if (!ASSERT_NEQ(__addr, 0, "ksym_get_addr " #__sym)) \
121 goto cleanup; \
122 })
123
124 GET_ADDR("bpf_fentry_test1", addrs[0]);
125 GET_ADDR("bpf_fentry_test3", addrs[1]);
126 GET_ADDR("bpf_fentry_test4", addrs[2]);
127 GET_ADDR("bpf_fentry_test5", addrs[3]);
128 GET_ADDR("bpf_fentry_test6", addrs[4]);
129 GET_ADDR("bpf_fentry_test7", addrs[5]);
130 GET_ADDR("bpf_fentry_test2", addrs[6]);
131 GET_ADDR("bpf_fentry_test8", addrs[7]);
132
133 #undef GET_ADDR
134
135 cookies[0] = 1; /* bpf_fentry_test1 */
136 cookies[1] = 2; /* bpf_fentry_test3 */
137 cookies[2] = 3; /* bpf_fentry_test4 */
138 cookies[3] = 4; /* bpf_fentry_test5 */
139 cookies[4] = 5; /* bpf_fentry_test6 */
140 cookies[5] = 6; /* bpf_fentry_test7 */
141 cookies[6] = 7; /* bpf_fentry_test2 */
142 cookies[7] = 8; /* bpf_fentry_test8 */
143
144 opts.kprobe_multi.addrs = (const unsigned long *) &addrs;
145 opts.kprobe_multi.cnt = ARRAY_SIZE(addrs);
146 opts.kprobe_multi.cookies = (const __u64 *) &cookies;
147 prog_fd = bpf_program__fd(skel->progs.test_kprobe);
148
149 link1_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, &opts);
150 if (!ASSERT_GE(link1_fd, 0, "link1_fd"))
151 goto cleanup;
152
153 cookies[0] = 8; /* bpf_fentry_test1 */
154 cookies[1] = 7; /* bpf_fentry_test3 */
155 cookies[2] = 6; /* bpf_fentry_test4 */
156 cookies[3] = 5; /* bpf_fentry_test5 */
157 cookies[4] = 4; /* bpf_fentry_test6 */
158 cookies[5] = 3; /* bpf_fentry_test7 */
159 cookies[6] = 2; /* bpf_fentry_test2 */
160 cookies[7] = 1; /* bpf_fentry_test8 */
161
162 opts.kprobe_multi.flags = BPF_F_KPROBE_MULTI_RETURN;
163 prog_fd = bpf_program__fd(skel->progs.test_kretprobe);
164
165 link2_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_KPROBE_MULTI, &opts);
166 if (!ASSERT_GE(link2_fd, 0, "link2_fd"))
167 goto cleanup;
168
169 kprobe_multi_test_run(skel);
170
171 cleanup:
172 close(link1_fd);
173 close(link2_fd);
174 kprobe_multi__destroy(skel);
175 }
176
kprobe_multi_attach_api_subtest(void)177 static void kprobe_multi_attach_api_subtest(void)
178 {
179 struct bpf_link *link1 = NULL, *link2 = NULL;
180 LIBBPF_OPTS(bpf_kprobe_multi_opts, opts);
181 LIBBPF_OPTS(bpf_test_run_opts, topts);
182 struct kprobe_multi *skel = NULL;
183 const char *syms[8] = {
184 "bpf_fentry_test1",
185 "bpf_fentry_test3",
186 "bpf_fentry_test4",
187 "bpf_fentry_test5",
188 "bpf_fentry_test6",
189 "bpf_fentry_test7",
190 "bpf_fentry_test2",
191 "bpf_fentry_test8",
192 };
193 __u64 cookies[8];
194
195 skel = kprobe_multi__open_and_load();
196 if (!ASSERT_OK_PTR(skel, "fentry_raw_skel_load"))
197 goto cleanup;
198
199 skel->bss->pid = getpid();
200 skel->bss->test_cookie = true;
201
202 cookies[0] = 1; /* bpf_fentry_test1 */
203 cookies[1] = 2; /* bpf_fentry_test3 */
204 cookies[2] = 3; /* bpf_fentry_test4 */
205 cookies[3] = 4; /* bpf_fentry_test5 */
206 cookies[4] = 5; /* bpf_fentry_test6 */
207 cookies[5] = 6; /* bpf_fentry_test7 */
208 cookies[6] = 7; /* bpf_fentry_test2 */
209 cookies[7] = 8; /* bpf_fentry_test8 */
210
211 opts.syms = syms;
212 opts.cnt = ARRAY_SIZE(syms);
213 opts.cookies = cookies;
214
215 link1 = bpf_program__attach_kprobe_multi_opts(skel->progs.test_kprobe,
216 NULL, &opts);
217 if (!ASSERT_OK_PTR(link1, "bpf_program__attach_kprobe_multi_opts"))
218 goto cleanup;
219
220 cookies[0] = 8; /* bpf_fentry_test1 */
221 cookies[1] = 7; /* bpf_fentry_test3 */
222 cookies[2] = 6; /* bpf_fentry_test4 */
223 cookies[3] = 5; /* bpf_fentry_test5 */
224 cookies[4] = 4; /* bpf_fentry_test6 */
225 cookies[5] = 3; /* bpf_fentry_test7 */
226 cookies[6] = 2; /* bpf_fentry_test2 */
227 cookies[7] = 1; /* bpf_fentry_test8 */
228
229 opts.retprobe = true;
230
231 link2 = bpf_program__attach_kprobe_multi_opts(skel->progs.test_kretprobe,
232 NULL, &opts);
233 if (!ASSERT_OK_PTR(link2, "bpf_program__attach_kprobe_multi_opts"))
234 goto cleanup;
235
236 kprobe_multi_test_run(skel);
237
238 cleanup:
239 bpf_link__destroy(link2);
240 bpf_link__destroy(link1);
241 kprobe_multi__destroy(skel);
242 }
243
244 /* defined in prog_tests/uprobe_multi_test.c */
245 void uprobe_multi_func_1(void);
246 void uprobe_multi_func_2(void);
247 void uprobe_multi_func_3(void);
248
uprobe_multi_test_run(struct uprobe_multi * skel)249 static void uprobe_multi_test_run(struct uprobe_multi *skel)
250 {
251 skel->bss->uprobe_multi_func_1_addr = (__u64) uprobe_multi_func_1;
252 skel->bss->uprobe_multi_func_2_addr = (__u64) uprobe_multi_func_2;
253 skel->bss->uprobe_multi_func_3_addr = (__u64) uprobe_multi_func_3;
254
255 skel->bss->pid = getpid();
256 skel->bss->test_cookie = true;
257
258 uprobe_multi_func_1();
259 uprobe_multi_func_2();
260 uprobe_multi_func_3();
261
262 ASSERT_EQ(skel->bss->uprobe_multi_func_1_result, 1, "uprobe_multi_func_1_result");
263 ASSERT_EQ(skel->bss->uprobe_multi_func_2_result, 1, "uprobe_multi_func_2_result");
264 ASSERT_EQ(skel->bss->uprobe_multi_func_3_result, 1, "uprobe_multi_func_3_result");
265
266 ASSERT_EQ(skel->bss->uretprobe_multi_func_1_result, 1, "uretprobe_multi_func_1_result");
267 ASSERT_EQ(skel->bss->uretprobe_multi_func_2_result, 1, "uretprobe_multi_func_2_result");
268 ASSERT_EQ(skel->bss->uretprobe_multi_func_3_result, 1, "uretprobe_multi_func_3_result");
269 }
270
uprobe_multi_attach_api_subtest(void)271 static void uprobe_multi_attach_api_subtest(void)
272 {
273 struct bpf_link *link1 = NULL, *link2 = NULL;
274 struct uprobe_multi *skel = NULL;
275 LIBBPF_OPTS(bpf_uprobe_multi_opts, opts);
276 const char *syms[3] = {
277 "uprobe_multi_func_1",
278 "uprobe_multi_func_2",
279 "uprobe_multi_func_3",
280 };
281 __u64 cookies[3];
282
283 cookies[0] = 3; /* uprobe_multi_func_1 */
284 cookies[1] = 1; /* uprobe_multi_func_2 */
285 cookies[2] = 2; /* uprobe_multi_func_3 */
286
287 opts.syms = syms;
288 opts.cnt = ARRAY_SIZE(syms);
289 opts.cookies = &cookies[0];
290
291 skel = uprobe_multi__open_and_load();
292 if (!ASSERT_OK_PTR(skel, "uprobe_multi"))
293 goto cleanup;
294
295 link1 = bpf_program__attach_uprobe_multi(skel->progs.uprobe, -1,
296 "/proc/self/exe", NULL, &opts);
297 if (!ASSERT_OK_PTR(link1, "bpf_program__attach_uprobe_multi"))
298 goto cleanup;
299
300 cookies[0] = 2; /* uprobe_multi_func_1 */
301 cookies[1] = 3; /* uprobe_multi_func_2 */
302 cookies[2] = 1; /* uprobe_multi_func_3 */
303
304 opts.retprobe = true;
305 link2 = bpf_program__attach_uprobe_multi(skel->progs.uretprobe, -1,
306 "/proc/self/exe", NULL, &opts);
307 if (!ASSERT_OK_PTR(link2, "bpf_program__attach_uprobe_multi_retprobe"))
308 goto cleanup;
309
310 uprobe_multi_test_run(skel);
311
312 cleanup:
313 bpf_link__destroy(link2);
314 bpf_link__destroy(link1);
315 uprobe_multi__destroy(skel);
316 }
317
uprobe_subtest(struct test_bpf_cookie * skel)318 static void uprobe_subtest(struct test_bpf_cookie *skel)
319 {
320 DECLARE_LIBBPF_OPTS(bpf_uprobe_opts, opts);
321 struct bpf_link *link1 = NULL, *link2 = NULL;
322 struct bpf_link *retlink1 = NULL, *retlink2 = NULL;
323 ssize_t uprobe_offset;
324
325 uprobe_offset = get_uprobe_offset(&trigger_func);
326 if (!ASSERT_GE(uprobe_offset, 0, "uprobe_offset"))
327 goto cleanup;
328
329 /* attach two uprobes */
330 opts.bpf_cookie = 0x100;
331 opts.retprobe = false;
332 link1 = bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe, 0 /* self pid */,
333 "/proc/self/exe", uprobe_offset, &opts);
334 if (!ASSERT_OK_PTR(link1, "link1"))
335 goto cleanup;
336
337 opts.bpf_cookie = 0x200;
338 opts.retprobe = false;
339 link2 = bpf_program__attach_uprobe_opts(skel->progs.handle_uprobe, -1 /* any pid */,
340 "/proc/self/exe", uprobe_offset, &opts);
341 if (!ASSERT_OK_PTR(link2, "link2"))
342 goto cleanup;
343
344 /* attach two uretprobes */
345 opts.bpf_cookie = 0x1000;
346 opts.retprobe = true;
347 retlink1 = bpf_program__attach_uprobe_opts(skel->progs.handle_uretprobe, -1 /* any pid */,
348 "/proc/self/exe", uprobe_offset, &opts);
349 if (!ASSERT_OK_PTR(retlink1, "retlink1"))
350 goto cleanup;
351
352 opts.bpf_cookie = 0x2000;
353 opts.retprobe = true;
354 retlink2 = bpf_program__attach_uprobe_opts(skel->progs.handle_uretprobe, 0 /* self pid */,
355 "/proc/self/exe", uprobe_offset, &opts);
356 if (!ASSERT_OK_PTR(retlink2, "retlink2"))
357 goto cleanup;
358
359 /* trigger uprobe && uretprobe */
360 trigger_func();
361
362 ASSERT_EQ(skel->bss->uprobe_res, 0x100 | 0x200, "uprobe_res");
363 ASSERT_EQ(skel->bss->uretprobe_res, 0x1000 | 0x2000, "uretprobe_res");
364
365 cleanup:
366 bpf_link__destroy(link1);
367 bpf_link__destroy(link2);
368 bpf_link__destroy(retlink1);
369 bpf_link__destroy(retlink2);
370 }
371
tp_subtest(struct test_bpf_cookie * skel)372 static void tp_subtest(struct test_bpf_cookie *skel)
373 {
374 DECLARE_LIBBPF_OPTS(bpf_tracepoint_opts, opts);
375 struct bpf_link *link1 = NULL, *link2 = NULL, *link3 = NULL;
376
377 /* attach first tp prog */
378 opts.bpf_cookie = 0x10000;
379 link1 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp1,
380 "syscalls", "sys_enter_nanosleep", &opts);
381 if (!ASSERT_OK_PTR(link1, "link1"))
382 goto cleanup;
383
384 /* attach second tp prog */
385 opts.bpf_cookie = 0x20000;
386 link2 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp2,
387 "syscalls", "sys_enter_nanosleep", &opts);
388 if (!ASSERT_OK_PTR(link2, "link2"))
389 goto cleanup;
390
391 /* trigger tracepoints */
392 usleep(1);
393
394 ASSERT_EQ(skel->bss->tp_res, 0x10000 | 0x20000, "tp_res1");
395
396 /* now we detach first prog and will attach third one, which causes
397 * two internal calls to bpf_prog_array_copy(), shuffling
398 * bpf_prog_array_items around. We test here that we don't lose track
399 * of associated bpf_cookies.
400 */
401 bpf_link__destroy(link1);
402 link1 = NULL;
403 kern_sync_rcu();
404 skel->bss->tp_res = 0;
405
406 /* attach third tp prog */
407 opts.bpf_cookie = 0x40000;
408 link3 = bpf_program__attach_tracepoint_opts(skel->progs.handle_tp3,
409 "syscalls", "sys_enter_nanosleep", &opts);
410 if (!ASSERT_OK_PTR(link3, "link3"))
411 goto cleanup;
412
413 /* trigger tracepoints */
414 usleep(1);
415
416 ASSERT_EQ(skel->bss->tp_res, 0x20000 | 0x40000, "tp_res2");
417
418 cleanup:
419 bpf_link__destroy(link1);
420 bpf_link__destroy(link2);
421 bpf_link__destroy(link3);
422 }
423
burn_cpu(void)424 static void burn_cpu(void)
425 {
426 volatile int j = 0;
427 cpu_set_t cpu_set;
428 int i, err;
429
430 /* generate some branches on cpu 0 */
431 CPU_ZERO(&cpu_set);
432 CPU_SET(0, &cpu_set);
433 err = pthread_setaffinity_np(pthread_self(), sizeof(cpu_set), &cpu_set);
434 ASSERT_OK(err, "set_thread_affinity");
435
436 /* spin the loop for a while (random high number) */
437 for (i = 0; i < 1000000; ++i)
438 ++j;
439 }
440
pe_subtest(struct test_bpf_cookie * skel)441 static void pe_subtest(struct test_bpf_cookie *skel)
442 {
443 DECLARE_LIBBPF_OPTS(bpf_perf_event_opts, opts);
444 struct bpf_link *link = NULL;
445 struct perf_event_attr attr;
446 int pfd = -1;
447
448 /* create perf event */
449 memset(&attr, 0, sizeof(attr));
450 attr.size = sizeof(attr);
451 attr.type = PERF_TYPE_SOFTWARE;
452 attr.config = PERF_COUNT_SW_CPU_CLOCK;
453 attr.freq = 1;
454 attr.sample_freq = 1000;
455 pfd = syscall(__NR_perf_event_open, &attr, -1, 0, -1, PERF_FLAG_FD_CLOEXEC);
456 if (!ASSERT_GE(pfd, 0, "perf_fd"))
457 goto cleanup;
458
459 opts.bpf_cookie = 0x100000;
460 link = bpf_program__attach_perf_event_opts(skel->progs.handle_pe, pfd, &opts);
461 if (!ASSERT_OK_PTR(link, "link1"))
462 goto cleanup;
463
464 burn_cpu(); /* trigger BPF prog */
465
466 ASSERT_EQ(skel->bss->pe_res, 0x100000, "pe_res1");
467
468 /* prevent bpf_link__destroy() closing pfd itself */
469 bpf_link__disconnect(link);
470 /* close BPF link's FD explicitly */
471 close(bpf_link__fd(link));
472 /* free up memory used by struct bpf_link */
473 bpf_link__destroy(link);
474 link = NULL;
475 kern_sync_rcu();
476 skel->bss->pe_res = 0;
477
478 opts.bpf_cookie = 0x200000;
479 link = bpf_program__attach_perf_event_opts(skel->progs.handle_pe, pfd, &opts);
480 if (!ASSERT_OK_PTR(link, "link2"))
481 goto cleanup;
482
483 burn_cpu(); /* trigger BPF prog */
484
485 ASSERT_EQ(skel->bss->pe_res, 0x200000, "pe_res2");
486
487 cleanup:
488 close(pfd);
489 bpf_link__destroy(link);
490 }
491
tracing_subtest(struct test_bpf_cookie * skel)492 static void tracing_subtest(struct test_bpf_cookie *skel)
493 {
494 __u64 cookie;
495 int prog_fd;
496 int fentry_fd = -1, fexit_fd = -1, fmod_ret_fd = -1;
497 LIBBPF_OPTS(bpf_test_run_opts, opts);
498 LIBBPF_OPTS(bpf_link_create_opts, link_opts);
499
500 skel->bss->fentry_res = 0;
501 skel->bss->fexit_res = 0;
502
503 cookie = 0x10000000000000L;
504 prog_fd = bpf_program__fd(skel->progs.fentry_test1);
505 link_opts.tracing.cookie = cookie;
506 fentry_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_FENTRY, &link_opts);
507 if (!ASSERT_GE(fentry_fd, 0, "fentry.link_create"))
508 goto cleanup;
509
510 cookie = 0x20000000000000L;
511 prog_fd = bpf_program__fd(skel->progs.fexit_test1);
512 link_opts.tracing.cookie = cookie;
513 fexit_fd = bpf_link_create(prog_fd, 0, BPF_TRACE_FEXIT, &link_opts);
514 if (!ASSERT_GE(fexit_fd, 0, "fexit.link_create"))
515 goto cleanup;
516
517 cookie = 0x30000000000000L;
518 prog_fd = bpf_program__fd(skel->progs.fmod_ret_test);
519 link_opts.tracing.cookie = cookie;
520 fmod_ret_fd = bpf_link_create(prog_fd, 0, BPF_MODIFY_RETURN, &link_opts);
521 if (!ASSERT_GE(fmod_ret_fd, 0, "fmod_ret.link_create"))
522 goto cleanup;
523
524 prog_fd = bpf_program__fd(skel->progs.fentry_test1);
525 bpf_prog_test_run_opts(prog_fd, &opts);
526
527 prog_fd = bpf_program__fd(skel->progs.fmod_ret_test);
528 bpf_prog_test_run_opts(prog_fd, &opts);
529
530 ASSERT_EQ(skel->bss->fentry_res, 0x10000000000000L, "fentry_res");
531 ASSERT_EQ(skel->bss->fexit_res, 0x20000000000000L, "fexit_res");
532 ASSERT_EQ(skel->bss->fmod_ret_res, 0x30000000000000L, "fmod_ret_res");
533
534 cleanup:
535 if (fentry_fd >= 0)
536 close(fentry_fd);
537 if (fexit_fd >= 0)
538 close(fexit_fd);
539 if (fmod_ret_fd >= 0)
540 close(fmod_ret_fd);
541 }
542
543 int stack_mprotect(void);
544
lsm_subtest(struct test_bpf_cookie * skel)545 static void lsm_subtest(struct test_bpf_cookie *skel)
546 {
547 __u64 cookie;
548 int prog_fd;
549 int lsm_fd = -1;
550 LIBBPF_OPTS(bpf_link_create_opts, link_opts);
551 int err;
552
553 skel->bss->lsm_res = 0;
554
555 cookie = 0x90000000000090L;
556 prog_fd = bpf_program__fd(skel->progs.test_int_hook);
557 link_opts.tracing.cookie = cookie;
558 lsm_fd = bpf_link_create(prog_fd, 0, BPF_LSM_MAC, &link_opts);
559 if (!ASSERT_GE(lsm_fd, 0, "lsm.link_create"))
560 goto cleanup;
561
562 err = stack_mprotect();
563 if (!ASSERT_EQ(err, -1, "stack_mprotect") ||
564 !ASSERT_EQ(errno, EPERM, "stack_mprotect"))
565 goto cleanup;
566
567 usleep(1);
568
569 ASSERT_EQ(skel->bss->lsm_res, 0x90000000000090L, "fentry_res");
570
571 cleanup:
572 if (lsm_fd >= 0)
573 close(lsm_fd);
574 }
575
test_bpf_cookie(void)576 void test_bpf_cookie(void)
577 {
578 struct test_bpf_cookie *skel;
579
580 skel = test_bpf_cookie__open_and_load();
581 if (!ASSERT_OK_PTR(skel, "skel_open"))
582 return;
583
584 skel->bss->my_tid = syscall(SYS_gettid);
585
586 if (test__start_subtest("kprobe"))
587 kprobe_subtest(skel);
588 if (test__start_subtest("multi_kprobe_link_api"))
589 kprobe_multi_link_api_subtest();
590 if (test__start_subtest("multi_kprobe_attach_api"))
591 kprobe_multi_attach_api_subtest();
592 if (test__start_subtest("uprobe"))
593 uprobe_subtest(skel);
594 if (test__start_subtest("multi_uprobe_attach_api"))
595 uprobe_multi_attach_api_subtest();
596 if (test__start_subtest("tracepoint"))
597 tp_subtest(skel);
598 if (test__start_subtest("perf_event"))
599 pe_subtest(skel);
600 if (test__start_subtest("trampoline"))
601 tracing_subtest(skel);
602 if (test__start_subtest("lsm"))
603 lsm_subtest(skel);
604
605 test_bpf_cookie__destroy(skel);
606 }
607