1 /* SPDX-License-Identifier: GPL-2.0 */
2 #include <stdlib.h>
3
4 #include <bpf/bpf.h>
5 #include <linux/err.h>
6 #include <internal/xyarray.h>
7
8 #include "util/debug.h"
9 #include "util/evsel.h"
10
11 #include "util/bpf-filter.h"
12 #include <util/bpf-filter-flex.h>
13 #include <util/bpf-filter-bison.h>
14
15 #include "bpf_skel/sample-filter.h"
16 #include "bpf_skel/sample_filter.skel.h"
17
18 #define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y))
19
20 #define __PERF_SAMPLE_TYPE(st, opt) { st, #st, opt }
21 #define PERF_SAMPLE_TYPE(_st, opt) __PERF_SAMPLE_TYPE(PERF_SAMPLE_##_st, opt)
22
23 static const struct perf_sample_info {
24 u64 type;
25 const char *name;
26 const char *option;
27 } sample_table[] = {
28 /* default sample flags */
29 PERF_SAMPLE_TYPE(IP, NULL),
30 PERF_SAMPLE_TYPE(TID, NULL),
31 PERF_SAMPLE_TYPE(PERIOD, NULL),
32 /* flags mostly set by default, but still have options */
33 PERF_SAMPLE_TYPE(ID, "--sample-identifier"),
34 PERF_SAMPLE_TYPE(CPU, "--sample-cpu"),
35 PERF_SAMPLE_TYPE(TIME, "-T"),
36 /* optional sample flags */
37 PERF_SAMPLE_TYPE(ADDR, "-d"),
38 PERF_SAMPLE_TYPE(DATA_SRC, "-d"),
39 PERF_SAMPLE_TYPE(PHYS_ADDR, "--phys-data"),
40 PERF_SAMPLE_TYPE(WEIGHT, "-W"),
41 PERF_SAMPLE_TYPE(WEIGHT_STRUCT, "-W"),
42 PERF_SAMPLE_TYPE(TRANSACTION, "--transaction"),
43 PERF_SAMPLE_TYPE(CODE_PAGE_SIZE, "--code-page-size"),
44 PERF_SAMPLE_TYPE(DATA_PAGE_SIZE, "--data-page-size"),
45 };
46
get_sample_info(u64 flags)47 static const struct perf_sample_info *get_sample_info(u64 flags)
48 {
49 size_t i;
50
51 for (i = 0; i < ARRAY_SIZE(sample_table); i++) {
52 if (sample_table[i].type == flags)
53 return &sample_table[i];
54 }
55 return NULL;
56 }
57
check_sample_flags(struct evsel * evsel,struct perf_bpf_filter_expr * expr)58 static int check_sample_flags(struct evsel *evsel, struct perf_bpf_filter_expr *expr)
59 {
60 const struct perf_sample_info *info;
61
62 if (evsel->core.attr.sample_type & expr->sample_flags)
63 return 0;
64
65 if (expr->op == PBF_OP_GROUP_BEGIN) {
66 struct perf_bpf_filter_expr *group;
67
68 list_for_each_entry(group, &expr->groups, list) {
69 if (check_sample_flags(evsel, group) < 0)
70 return -1;
71 }
72 return 0;
73 }
74
75 info = get_sample_info(expr->sample_flags);
76 if (info == NULL) {
77 pr_err("Error: %s event does not have sample flags %lx\n",
78 evsel__name(evsel), expr->sample_flags);
79 return -1;
80 }
81
82 pr_err("Error: %s event does not have %s\n", evsel__name(evsel), info->name);
83 if (info->option)
84 pr_err(" Hint: please add %s option to perf record\n", info->option);
85 return -1;
86 }
87
perf_bpf_filter__prepare(struct evsel * evsel)88 int perf_bpf_filter__prepare(struct evsel *evsel)
89 {
90 int i, x, y, fd;
91 struct sample_filter_bpf *skel;
92 struct bpf_program *prog;
93 struct bpf_link *link;
94 struct perf_bpf_filter_expr *expr;
95
96 skel = sample_filter_bpf__open_and_load();
97 if (!skel) {
98 pr_err("Failed to load perf sample-filter BPF skeleton\n");
99 return -1;
100 }
101
102 i = 0;
103 fd = bpf_map__fd(skel->maps.filters);
104 list_for_each_entry(expr, &evsel->bpf_filters, list) {
105 struct perf_bpf_filter_entry entry = {
106 .op = expr->op,
107 .part = expr->part,
108 .flags = expr->sample_flags,
109 .value = expr->val,
110 };
111
112 if (check_sample_flags(evsel, expr) < 0)
113 return -1;
114
115 bpf_map_update_elem(fd, &i, &entry, BPF_ANY);
116 i++;
117
118 if (expr->op == PBF_OP_GROUP_BEGIN) {
119 struct perf_bpf_filter_expr *group;
120
121 list_for_each_entry(group, &expr->groups, list) {
122 struct perf_bpf_filter_entry group_entry = {
123 .op = group->op,
124 .part = group->part,
125 .flags = group->sample_flags,
126 .value = group->val,
127 };
128 bpf_map_update_elem(fd, &i, &group_entry, BPF_ANY);
129 i++;
130 }
131
132 memset(&entry, 0, sizeof(entry));
133 entry.op = PBF_OP_GROUP_END;
134 bpf_map_update_elem(fd, &i, &entry, BPF_ANY);
135 i++;
136 }
137 }
138
139 if (i > MAX_FILTERS) {
140 pr_err("Too many filters: %d (max = %d)\n", i, MAX_FILTERS);
141 return -1;
142 }
143 prog = skel->progs.perf_sample_filter;
144 for (x = 0; x < xyarray__max_x(evsel->core.fd); x++) {
145 for (y = 0; y < xyarray__max_y(evsel->core.fd); y++) {
146 link = bpf_program__attach_perf_event(prog, FD(evsel, x, y));
147 if (IS_ERR(link)) {
148 pr_err("Failed to attach perf sample-filter program\n");
149 return PTR_ERR(link);
150 }
151 }
152 }
153 evsel->bpf_skel = skel;
154 return 0;
155 }
156
perf_bpf_filter__destroy(struct evsel * evsel)157 int perf_bpf_filter__destroy(struct evsel *evsel)
158 {
159 struct perf_bpf_filter_expr *expr, *tmp;
160
161 list_for_each_entry_safe(expr, tmp, &evsel->bpf_filters, list) {
162 list_del(&expr->list);
163 free(expr);
164 }
165 sample_filter_bpf__destroy(evsel->bpf_skel);
166 return 0;
167 }
168
perf_bpf_filter__lost_count(struct evsel * evsel)169 u64 perf_bpf_filter__lost_count(struct evsel *evsel)
170 {
171 struct sample_filter_bpf *skel = evsel->bpf_skel;
172
173 return skel ? skel->bss->dropped : 0;
174 }
175
perf_bpf_filter_expr__new(unsigned long sample_flags,int part,enum perf_bpf_filter_op op,unsigned long val)176 struct perf_bpf_filter_expr *perf_bpf_filter_expr__new(unsigned long sample_flags, int part,
177 enum perf_bpf_filter_op op,
178 unsigned long val)
179 {
180 struct perf_bpf_filter_expr *expr;
181
182 expr = malloc(sizeof(*expr));
183 if (expr != NULL) {
184 expr->sample_flags = sample_flags;
185 expr->part = part;
186 expr->op = op;
187 expr->val = val;
188 INIT_LIST_HEAD(&expr->groups);
189 }
190 return expr;
191 }
192
perf_bpf_filter__parse(struct list_head * expr_head,const char * str)193 int perf_bpf_filter__parse(struct list_head *expr_head, const char *str)
194 {
195 YY_BUFFER_STATE buffer;
196 int ret;
197
198 buffer = perf_bpf_filter__scan_string(str);
199
200 ret = perf_bpf_filter_parse(expr_head);
201
202 perf_bpf_filter__flush_buffer(buffer);
203 perf_bpf_filter__delete_buffer(buffer);
204 perf_bpf_filter_lex_destroy();
205
206 return ret;
207 }
208