Skip to content
Navigation Menu
Toggle navigation
Sign in
In this repository
All GitHub Enterprise
↵
Jump to
↵
No suggested jump to results
In this repository
All GitHub Enterprise
↵
Jump to
↵
In this organization
All GitHub Enterprise
↵
Jump to
↵
In this repository
All GitHub Enterprise
↵
Jump to
↵
Sign in
Reseting focus
You signed in with another tab or window.
Reload
to refresh your session.
You signed out in another tab or window.
Reload
to refresh your session.
You switched accounts on another tab or window.
Reload
to refresh your session.
Dismiss alert
{{ message }}
mariux64
/
linux
Public
Notifications
You must be signed in to change notification settings
Fork
0
Star
0
Code
Issues
2
Pull requests
0
Actions
Projects
0
Wiki
Security
Insights
Additional navigation options
Code
Issues
Pull requests
Actions
Projects
Wiki
Security
Insights
Files
e67b2c7
Breadcrumbs
linux
/
tools
/
testing
/
selftests
/
bpf
/
prog_tests
/
get_stack_raw_tp.c
Blame
Blame
Latest commit
History
History
139 lines (120 loc) · 3.7 KB
Breadcrumbs
linux
/
tools
/
testing
/
selftests
/
bpf
/
prog_tests
/
get_stack_raw_tp.c
Top
File metadata and controls
Code
Blame
139 lines (120 loc) · 3.7 KB
Raw
// SPDX-License-Identifier: GPL-2.0 #include <test_progs.h> #define MAX_CNT_RAWTP 10ull #define MAX_STACK_RAWTP 100 struct get_stack_trace_t { int pid; int kern_stack_size; int user_stack_size; int user_stack_buildid_size; __u64 kern_stack[MAX_STACK_RAWTP]; __u64 user_stack[MAX_STACK_RAWTP]; struct bpf_stack_build_id user_stack_buildid[MAX_STACK_RAWTP]; }; static int get_stack_print_output(void *data, int size) { bool good_kern_stack = false, good_user_stack = false; const char *nonjit_func = "___bpf_prog_run"; struct get_stack_trace_t *e = data; int i, num_stack; static __u64 cnt; struct ksym *ks; cnt++; if (size < sizeof(struct get_stack_trace_t)) { __u64 *raw_data = data; bool found = false; num_stack = size / sizeof(__u64); /* If jit is enabled, we do not have a good way to * verify the sanity of the kernel stack. So we * just assume it is good if the stack is not empty. * This could be improved in the future. */ if (jit_enabled) { found = num_stack > 0; } else { for (i = 0; i < num_stack; i++) { ks = ksym_search(raw_data[i]); if (ks && (strcmp(ks->name, nonjit_func) == 0)) { found = true; break; } } } if (found) { good_kern_stack = true; good_user_stack = true; } } else { num_stack = e->kern_stack_size / sizeof(__u64); if (jit_enabled) { good_kern_stack = num_stack > 0; } else { for (i = 0; i < num_stack; i++) { ks = ksym_search(e->kern_stack[i]); if (ks && (strcmp(ks->name, nonjit_func) == 0)) { good_kern_stack = true; break; } } } if (e->user_stack_size > 0 && e->user_stack_buildid_size > 0) good_user_stack = true; } if (!good_kern_stack || !good_user_stack) return LIBBPF_PERF_EVENT_ERROR; if (cnt == MAX_CNT_RAWTP) return LIBBPF_PERF_EVENT_DONE; return LIBBPF_PERF_EVENT_CONT; } void test_get_stack_raw_tp(void) { const char *file = "./test_get_stack_rawtp.o"; int i, efd, err, prog_fd, pmu_fd, perfmap_fd; struct perf_event_attr attr = {}; struct timespec tv = {0, 10}; __u32 key = 0, duration = 0; struct bpf_object *obj; err = bpf_prog_load(file, BPF_PROG_TYPE_RAW_TRACEPOINT, &obj, &prog_fd); if (CHECK(err, "prog_load raw tp", "err %d errno %d\n", err, errno)) return; efd = bpf_raw_tracepoint_open("sys_enter", prog_fd); if (CHECK(efd < 0, "raw_tp_open", "err %d errno %d\n", efd, errno)) goto close_prog; perfmap_fd = bpf_find_map(__func__, obj, "perfmap"); if (CHECK(perfmap_fd < 0, "bpf_find_map", "err %d errno %d\n", perfmap_fd, errno)) goto close_prog; err = load_kallsyms(); if (CHECK(err < 0, "load_kallsyms", "err %d errno %d\n", err, errno)) goto close_prog; attr.sample_type = PERF_SAMPLE_RAW; attr.type = PERF_TYPE_SOFTWARE; attr.config = PERF_COUNT_SW_BPF_OUTPUT; pmu_fd = syscall(__NR_perf_event_open, &attr, getpid()/*pid*/, -1/*cpu*/, -1/*group_fd*/, 0); if (CHECK(pmu_fd < 0, "perf_event_open", "err %d errno %d\n", pmu_fd, errno)) goto close_prog; err = bpf_map_update_elem(perfmap_fd, &key, &pmu_fd, BPF_ANY); if (CHECK(err < 0, "bpf_map_update_elem", "err %d errno %d\n", err, errno)) goto close_prog; err = ioctl(pmu_fd, PERF_EVENT_IOC_ENABLE, 0); if (CHECK(err < 0, "ioctl PERF_EVENT_IOC_ENABLE", "err %d errno %d\n", err, errno)) goto close_prog; err = perf_event_mmap(pmu_fd); if (CHECK(err < 0, "perf_event_mmap", "err %d errno %d\n", err, errno)) goto close_prog; /* trigger some syscall action */ for (i = 0; i < MAX_CNT_RAWTP; i++) nanosleep(&tv, NULL); err = perf_event_poller(pmu_fd, get_stack_print_output); if (CHECK(err < 0, "perf_event_poller", "err %d errno %d\n", err, errno)) goto close_prog; goto close_prog_noerr; close_prog: error_cnt++; close_prog_noerr: bpf_object__close(obj); }
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
You can’t perform that action at this time.