xref: /openbmc/linux/kernel/trace/trace_eprobe.c (revision bfad37c5)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * event probes
4  *
5  * Part of this code was copied from kernel/trace/trace_kprobe.c written by
6  * Masami Hiramatsu <mhiramat@kernel.org>
7  *
8  * Copyright (C) 2021, VMware Inc, Steven Rostedt <rostedt@goodmis.org>
9  * Copyright (C) 2021, VMware Inc, Tzvetomir Stoyanov tz.stoyanov@gmail.com>
10  *
11  */
12 #include <linux/module.h>
13 #include <linux/mutex.h>
14 #include <linux/ftrace.h>
15 
16 #include "trace_dynevent.h"
17 #include "trace_probe.h"
18 #include "trace_probe_tmpl.h"
19 
20 #define EPROBE_EVENT_SYSTEM "eprobes"
21 
22 struct trace_eprobe {
23 	/* tracepoint system */
24 	const char *event_system;
25 
26 	/* tracepoint event */
27 	const char *event_name;
28 
29 	struct trace_event_call *event;
30 
31 	struct dyn_event	devent;
32 	struct trace_probe	tp;
33 };
34 
35 struct eprobe_data {
36 	struct trace_event_file	*file;
37 	struct trace_eprobe	*ep;
38 };
39 
40 static int __trace_eprobe_create(int argc, const char *argv[]);
41 
42 static void trace_event_probe_cleanup(struct trace_eprobe *ep)
43 {
44 	if (!ep)
45 		return;
46 	trace_probe_cleanup(&ep->tp);
47 	kfree(ep->event_name);
48 	kfree(ep->event_system);
49 	if (ep->event)
50 		trace_event_put_ref(ep->event);
51 	kfree(ep);
52 }
53 
54 static struct trace_eprobe *to_trace_eprobe(struct dyn_event *ev)
55 {
56 	return container_of(ev, struct trace_eprobe, devent);
57 }
58 
59 static int eprobe_dyn_event_create(const char *raw_command)
60 {
61 	return trace_probe_create(raw_command, __trace_eprobe_create);
62 }
63 
64 static int eprobe_dyn_event_show(struct seq_file *m, struct dyn_event *ev)
65 {
66 	struct trace_eprobe *ep = to_trace_eprobe(ev);
67 	int i;
68 
69 	seq_printf(m, "e:%s/%s", trace_probe_group_name(&ep->tp),
70 				trace_probe_name(&ep->tp));
71 	seq_printf(m, " %s.%s", ep->event_system, ep->event_name);
72 
73 	for (i = 0; i < ep->tp.nr_args; i++)
74 		seq_printf(m, " %s=%s", ep->tp.args[i].name, ep->tp.args[i].comm);
75 	seq_putc(m, '\n');
76 
77 	return 0;
78 }
79 
80 static int unregister_trace_eprobe(struct trace_eprobe *ep)
81 {
82 	/* If other probes are on the event, just unregister eprobe */
83 	if (trace_probe_has_sibling(&ep->tp))
84 		goto unreg;
85 
86 	/* Enabled event can not be unregistered */
87 	if (trace_probe_is_enabled(&ep->tp))
88 		return -EBUSY;
89 
90 	/* Will fail if probe is being used by ftrace or perf */
91 	if (trace_probe_unregister_event_call(&ep->tp))
92 		return -EBUSY;
93 
94 unreg:
95 	dyn_event_remove(&ep->devent);
96 	trace_probe_unlink(&ep->tp);
97 
98 	return 0;
99 }
100 
101 static int eprobe_dyn_event_release(struct dyn_event *ev)
102 {
103 	struct trace_eprobe *ep = to_trace_eprobe(ev);
104 	int ret = unregister_trace_eprobe(ep);
105 
106 	if (!ret)
107 		trace_event_probe_cleanup(ep);
108 	return ret;
109 }
110 
111 static bool eprobe_dyn_event_is_busy(struct dyn_event *ev)
112 {
113 	struct trace_eprobe *ep = to_trace_eprobe(ev);
114 
115 	return trace_probe_is_enabled(&ep->tp);
116 }
117 
118 static bool eprobe_dyn_event_match(const char *system, const char *event,
119 			int argc, const char **argv, struct dyn_event *ev)
120 {
121 	struct trace_eprobe *ep = to_trace_eprobe(ev);
122 
123 	return strcmp(trace_probe_name(&ep->tp), event) == 0 &&
124 	    (!system || strcmp(trace_probe_group_name(&ep->tp), system) == 0) &&
125 	    trace_probe_match_command_args(&ep->tp, argc, argv);
126 }
127 
128 static struct dyn_event_operations eprobe_dyn_event_ops = {
129 	.create = eprobe_dyn_event_create,
130 	.show = eprobe_dyn_event_show,
131 	.is_busy = eprobe_dyn_event_is_busy,
132 	.free = eprobe_dyn_event_release,
133 	.match = eprobe_dyn_event_match,
134 };
135 
136 static struct trace_eprobe *alloc_event_probe(const char *group,
137 					      const char *this_event,
138 					      struct trace_event_call *event,
139 					      int nargs)
140 {
141 	struct trace_eprobe *ep;
142 	const char *event_name;
143 	const char *sys_name;
144 	int ret = -ENOMEM;
145 
146 	if (!event)
147 		return ERR_PTR(-ENODEV);
148 
149 	sys_name = event->class->system;
150 	event_name = trace_event_name(event);
151 
152 	ep = kzalloc(struct_size(ep, tp.args, nargs), GFP_KERNEL);
153 	if (!ep) {
154 		trace_event_put_ref(event);
155 		goto error;
156 	}
157 	ep->event = event;
158 	ep->event_name = kstrdup(event_name, GFP_KERNEL);
159 	if (!ep->event_name)
160 		goto error;
161 	ep->event_system = kstrdup(sys_name, GFP_KERNEL);
162 	if (!ep->event_system)
163 		goto error;
164 
165 	ret = trace_probe_init(&ep->tp, this_event, group, false);
166 	if (ret < 0)
167 		goto error;
168 
169 	dyn_event_init(&ep->devent, &eprobe_dyn_event_ops);
170 	return ep;
171 error:
172 	trace_event_probe_cleanup(ep);
173 	return ERR_PTR(ret);
174 }
175 
176 static int trace_eprobe_tp_arg_update(struct trace_eprobe *ep, int i)
177 {
178 	struct probe_arg *parg = &ep->tp.args[i];
179 	struct ftrace_event_field *field;
180 	struct list_head *head;
181 
182 	head = trace_get_fields(ep->event);
183 	list_for_each_entry(field, head, link) {
184 		if (!strcmp(parg->code->data, field->name)) {
185 			kfree(parg->code->data);
186 			parg->code->data = field;
187 			return 0;
188 		}
189 	}
190 	kfree(parg->code->data);
191 	parg->code->data = NULL;
192 	return -ENOENT;
193 }
194 
195 static int eprobe_event_define_fields(struct trace_event_call *event_call)
196 {
197 	int ret;
198 	struct eprobe_trace_entry_head field;
199 	struct trace_probe *tp;
200 
201 	tp = trace_probe_primary_from_call(event_call);
202 	if (WARN_ON_ONCE(!tp))
203 		return -ENOENT;
204 
205 	DEFINE_FIELD(unsigned int, type, FIELD_STRING_TYPE, 0);
206 
207 	return traceprobe_define_arg_fields(event_call, sizeof(field), tp);
208 }
209 
210 static struct trace_event_fields eprobe_fields_array[] = {
211 	{ .type = TRACE_FUNCTION_TYPE,
212 	  .define_fields = eprobe_event_define_fields },
213 	{}
214 };
215 
216 /* Event entry printers */
217 static enum print_line_t
218 print_eprobe_event(struct trace_iterator *iter, int flags,
219 		   struct trace_event *event)
220 {
221 	struct eprobe_trace_entry_head *field;
222 	struct trace_event_call *pevent;
223 	struct trace_event *probed_event;
224 	struct trace_seq *s = &iter->seq;
225 	struct trace_probe *tp;
226 
227 	field = (struct eprobe_trace_entry_head *)iter->ent;
228 	tp = trace_probe_primary_from_call(
229 		container_of(event, struct trace_event_call, event));
230 	if (WARN_ON_ONCE(!tp))
231 		goto out;
232 
233 	trace_seq_printf(s, "%s: (", trace_probe_name(tp));
234 
235 	probed_event = ftrace_find_event(field->type);
236 	if (probed_event) {
237 		pevent = container_of(probed_event, struct trace_event_call, event);
238 		trace_seq_printf(s, "%s.%s", pevent->class->system,
239 				 trace_event_name(pevent));
240 	} else {
241 		trace_seq_printf(s, "%u", field->type);
242 	}
243 
244 	trace_seq_putc(s, ')');
245 
246 	if (print_probe_args(s, tp->args, tp->nr_args,
247 			     (u8 *)&field[1], field) < 0)
248 		goto out;
249 
250 	trace_seq_putc(s, '\n');
251  out:
252 	return trace_handle_return(s);
253 }
254 
255 static unsigned long get_event_field(struct fetch_insn *code, void *rec)
256 {
257 	struct ftrace_event_field *field = code->data;
258 	unsigned long val;
259 	void *addr;
260 
261 	addr = rec + field->offset;
262 
263 	switch (field->size) {
264 	case 1:
265 		if (field->is_signed)
266 			val = *(char *)addr;
267 		else
268 			val = *(unsigned char *)addr;
269 		break;
270 	case 2:
271 		if (field->is_signed)
272 			val = *(short *)addr;
273 		else
274 			val = *(unsigned short *)addr;
275 		break;
276 	case 4:
277 		if (field->is_signed)
278 			val = *(int *)addr;
279 		else
280 			val = *(unsigned int *)addr;
281 		break;
282 	default:
283 		if (field->is_signed)
284 			val = *(long *)addr;
285 		else
286 			val = *(unsigned long *)addr;
287 		break;
288 	}
289 	return val;
290 }
291 
292 static int get_eprobe_size(struct trace_probe *tp, void *rec)
293 {
294 	struct probe_arg *arg;
295 	int i, len, ret = 0;
296 
297 	for (i = 0; i < tp->nr_args; i++) {
298 		arg = tp->args + i;
299 		if (unlikely(arg->dynamic)) {
300 			unsigned long val;
301 
302 			val = get_event_field(arg->code, rec);
303 			len = process_fetch_insn_bottom(arg->code + 1, val, NULL, NULL);
304 			if (len > 0)
305 				ret += len;
306 		}
307 	}
308 
309 	return ret;
310 }
311 
312 /* Kprobe specific fetch functions */
313 
314 /* Note that we don't verify it, since the code does not come from user space */
315 static int
316 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest,
317 		   void *base)
318 {
319 	unsigned long val;
320 
321 	val = get_event_field(code, rec);
322 	return process_fetch_insn_bottom(code + 1, val, dest, base);
323 }
324 NOKPROBE_SYMBOL(process_fetch_insn)
325 
326 /* Return the length of string -- including null terminal byte */
327 static nokprobe_inline int
328 fetch_store_strlen_user(unsigned long addr)
329 {
330 	const void __user *uaddr =  (__force const void __user *)addr;
331 
332 	return strnlen_user_nofault(uaddr, MAX_STRING_SIZE);
333 }
334 
335 /* Return the length of string -- including null terminal byte */
336 static nokprobe_inline int
337 fetch_store_strlen(unsigned long addr)
338 {
339 	int ret, len = 0;
340 	u8 c;
341 
342 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
343 	if (addr < TASK_SIZE)
344 		return fetch_store_strlen_user(addr);
345 #endif
346 
347 	do {
348 		ret = copy_from_kernel_nofault(&c, (u8 *)addr + len, 1);
349 		len++;
350 	} while (c && ret == 0 && len < MAX_STRING_SIZE);
351 
352 	return (ret < 0) ? ret : len;
353 }
354 
355 /*
356  * Fetch a null-terminated string from user. Caller MUST set *(u32 *)buf
357  * with max length and relative data location.
358  */
359 static nokprobe_inline int
360 fetch_store_string_user(unsigned long addr, void *dest, void *base)
361 {
362 	const void __user *uaddr =  (__force const void __user *)addr;
363 	int maxlen = get_loc_len(*(u32 *)dest);
364 	void *__dest;
365 	long ret;
366 
367 	if (unlikely(!maxlen))
368 		return -ENOMEM;
369 
370 	__dest = get_loc_data(dest, base);
371 
372 	ret = strncpy_from_user_nofault(__dest, uaddr, maxlen);
373 	if (ret >= 0)
374 		*(u32 *)dest = make_data_loc(ret, __dest - base);
375 
376 	return ret;
377 }
378 
379 /*
380  * Fetch a null-terminated string. Caller MUST set *(u32 *)buf with max
381  * length and relative data location.
382  */
383 static nokprobe_inline int
384 fetch_store_string(unsigned long addr, void *dest, void *base)
385 {
386 	int maxlen = get_loc_len(*(u32 *)dest);
387 	void *__dest;
388 	long ret;
389 
390 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
391 	if ((unsigned long)addr < TASK_SIZE)
392 		return fetch_store_string_user(addr, dest, base);
393 #endif
394 
395 	if (unlikely(!maxlen))
396 		return -ENOMEM;
397 
398 	__dest = get_loc_data(dest, base);
399 
400 	/*
401 	 * Try to get string again, since the string can be changed while
402 	 * probing.
403 	 */
404 	ret = strncpy_from_kernel_nofault(__dest, (void *)addr, maxlen);
405 	if (ret >= 0)
406 		*(u32 *)dest = make_data_loc(ret, __dest - base);
407 
408 	return ret;
409 }
410 
411 static nokprobe_inline int
412 probe_mem_read_user(void *dest, void *src, size_t size)
413 {
414 	const void __user *uaddr =  (__force const void __user *)src;
415 
416 	return copy_from_user_nofault(dest, uaddr, size);
417 }
418 
419 static nokprobe_inline int
420 probe_mem_read(void *dest, void *src, size_t size)
421 {
422 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
423 	if ((unsigned long)src < TASK_SIZE)
424 		return probe_mem_read_user(dest, src, size);
425 #endif
426 	return copy_from_kernel_nofault(dest, src, size);
427 }
428 
429 /* eprobe handler */
430 static inline void
431 __eprobe_trace_func(struct eprobe_data *edata, void *rec)
432 {
433 	struct eprobe_trace_entry_head *entry;
434 	struct trace_event_call *call = trace_probe_event_call(&edata->ep->tp);
435 	struct trace_event_buffer fbuffer;
436 	int dsize;
437 
438 	if (WARN_ON_ONCE(call != edata->file->event_call))
439 		return;
440 
441 	if (trace_trigger_soft_disabled(edata->file))
442 		return;
443 
444 	fbuffer.trace_ctx = tracing_gen_ctx();
445 	fbuffer.trace_file = edata->file;
446 
447 	dsize = get_eprobe_size(&edata->ep->tp, rec);
448 	fbuffer.regs = NULL;
449 
450 	fbuffer.event =
451 		trace_event_buffer_lock_reserve(&fbuffer.buffer, edata->file,
452 					call->event.type,
453 					sizeof(*entry) + edata->ep->tp.size + dsize,
454 					fbuffer.trace_ctx);
455 	if (!fbuffer.event)
456 		return;
457 
458 	entry = fbuffer.entry = ring_buffer_event_data(fbuffer.event);
459 	if (edata->ep->event)
460 		entry->type = edata->ep->event->event.type;
461 	else
462 		entry->type = 0;
463 	store_trace_args(&entry[1], &edata->ep->tp, rec, sizeof(*entry), dsize);
464 
465 	trace_event_buffer_commit(&fbuffer);
466 }
467 
468 /*
469  * The event probe implementation uses event triggers to get access to
470  * the event it is attached to, but is not an actual trigger. The below
471  * functions are just stubs to fulfill what is needed to use the trigger
472  * infrastructure.
473  */
474 static int eprobe_trigger_init(struct event_trigger_ops *ops,
475 			       struct event_trigger_data *data)
476 {
477 	return 0;
478 }
479 
480 static void eprobe_trigger_free(struct event_trigger_ops *ops,
481 				struct event_trigger_data *data)
482 {
483 
484 }
485 
486 static int eprobe_trigger_print(struct seq_file *m,
487 				struct event_trigger_ops *ops,
488 				struct event_trigger_data *data)
489 {
490 	/* Do not print eprobe event triggers */
491 	return 0;
492 }
493 
494 static void eprobe_trigger_func(struct event_trigger_data *data,
495 				struct trace_buffer *buffer, void *rec,
496 				struct ring_buffer_event *rbe)
497 {
498 	struct eprobe_data *edata = data->private_data;
499 
500 	__eprobe_trace_func(edata, rec);
501 }
502 
503 static struct event_trigger_ops eprobe_trigger_ops = {
504 	.func			= eprobe_trigger_func,
505 	.print			= eprobe_trigger_print,
506 	.init			= eprobe_trigger_init,
507 	.free			= eprobe_trigger_free,
508 };
509 
510 static int eprobe_trigger_cmd_func(struct event_command *cmd_ops,
511 				   struct trace_event_file *file,
512 				   char *glob, char *cmd, char *param)
513 {
514 	return -1;
515 }
516 
517 static int eprobe_trigger_reg_func(char *glob, struct event_trigger_ops *ops,
518 				 struct event_trigger_data *data,
519 				 struct trace_event_file *file)
520 {
521 	return -1;
522 }
523 
524 static void eprobe_trigger_unreg_func(char *glob, struct event_trigger_ops *ops,
525 				    struct event_trigger_data *data,
526 				    struct trace_event_file *file)
527 {
528 
529 }
530 
531 static struct event_trigger_ops *eprobe_trigger_get_ops(char *cmd,
532 							char *param)
533 {
534 	return &eprobe_trigger_ops;
535 }
536 
537 static struct event_command event_trigger_cmd = {
538 	.name			= "eprobe",
539 	.trigger_type		= ETT_EVENT_EPROBE,
540 	.flags			= EVENT_CMD_FL_NEEDS_REC,
541 	.func			= eprobe_trigger_cmd_func,
542 	.reg			= eprobe_trigger_reg_func,
543 	.unreg			= eprobe_trigger_unreg_func,
544 	.unreg_all		= NULL,
545 	.get_trigger_ops	= eprobe_trigger_get_ops,
546 	.set_filter		= NULL,
547 };
548 
549 static struct event_trigger_data *
550 new_eprobe_trigger(struct trace_eprobe *ep, struct trace_event_file *file)
551 {
552 	struct event_trigger_data *trigger;
553 	struct eprobe_data *edata;
554 
555 	edata = kzalloc(sizeof(*edata), GFP_KERNEL);
556 	trigger = kzalloc(sizeof(*trigger), GFP_KERNEL);
557 	if (!trigger || !edata) {
558 		kfree(edata);
559 		kfree(trigger);
560 		return ERR_PTR(-ENOMEM);
561 	}
562 
563 	trigger->flags = EVENT_TRIGGER_FL_PROBE;
564 	trigger->count = -1;
565 	trigger->ops = &eprobe_trigger_ops;
566 
567 	/*
568 	 * EVENT PROBE triggers are not registered as commands with
569 	 * register_event_command(), as they are not controlled by the user
570 	 * from the trigger file
571 	 */
572 	trigger->cmd_ops = &event_trigger_cmd;
573 
574 	INIT_LIST_HEAD(&trigger->list);
575 	RCU_INIT_POINTER(trigger->filter, NULL);
576 
577 	edata->file = file;
578 	edata->ep = ep;
579 	trigger->private_data = edata;
580 
581 	return trigger;
582 }
583 
584 static int enable_eprobe(struct trace_eprobe *ep,
585 			 struct trace_event_file *eprobe_file)
586 {
587 	struct event_trigger_data *trigger;
588 	struct trace_event_file *file;
589 	struct trace_array *tr = eprobe_file->tr;
590 
591 	file = find_event_file(tr, ep->event_system, ep->event_name);
592 	if (!file)
593 		return -ENOENT;
594 	trigger = new_eprobe_trigger(ep, eprobe_file);
595 	if (IS_ERR(trigger))
596 		return PTR_ERR(trigger);
597 
598 	list_add_tail_rcu(&trigger->list, &file->triggers);
599 
600 	trace_event_trigger_enable_disable(file, 1);
601 	update_cond_flag(file);
602 
603 	return 0;
604 }
605 
606 static struct trace_event_functions eprobe_funcs = {
607 	.trace		= print_eprobe_event
608 };
609 
610 static int disable_eprobe(struct trace_eprobe *ep,
611 			  struct trace_array *tr)
612 {
613 	struct event_trigger_data *trigger;
614 	struct trace_event_file *file;
615 	struct eprobe_data *edata;
616 
617 	file = find_event_file(tr, ep->event_system, ep->event_name);
618 	if (!file)
619 		return -ENOENT;
620 
621 	list_for_each_entry(trigger, &file->triggers, list) {
622 		if (!(trigger->flags & EVENT_TRIGGER_FL_PROBE))
623 			continue;
624 		edata = trigger->private_data;
625 		if (edata->ep == ep)
626 			break;
627 	}
628 	if (list_entry_is_head(trigger, &file->triggers, list))
629 		return -ENODEV;
630 
631 	list_del_rcu(&trigger->list);
632 
633 	trace_event_trigger_enable_disable(file, 0);
634 	update_cond_flag(file);
635 	return 0;
636 }
637 
638 static int enable_trace_eprobe(struct trace_event_call *call,
639 			       struct trace_event_file *file)
640 {
641 	struct trace_probe *pos, *tp;
642 	struct trace_eprobe *ep;
643 	bool enabled;
644 	int ret = 0;
645 
646 	tp = trace_probe_primary_from_call(call);
647 	if (WARN_ON_ONCE(!tp))
648 		return -ENODEV;
649 	enabled = trace_probe_is_enabled(tp);
650 
651 	/* This also changes "enabled" state */
652 	if (file) {
653 		ret = trace_probe_add_file(tp, file);
654 		if (ret)
655 			return ret;
656 	} else
657 		trace_probe_set_flag(tp, TP_FLAG_PROFILE);
658 
659 	if (enabled)
660 		return 0;
661 
662 	list_for_each_entry(pos, trace_probe_probe_list(tp), list) {
663 		ep = container_of(pos, struct trace_eprobe, tp);
664 		ret = enable_eprobe(ep, file);
665 		if (ret)
666 			break;
667 		enabled = true;
668 	}
669 
670 	if (ret) {
671 		/* Failed to enable one of them. Roll back all */
672 		if (enabled)
673 			disable_eprobe(ep, file->tr);
674 		if (file)
675 			trace_probe_remove_file(tp, file);
676 		else
677 			trace_probe_clear_flag(tp, TP_FLAG_PROFILE);
678 	}
679 
680 	return ret;
681 }
682 
683 static int disable_trace_eprobe(struct trace_event_call *call,
684 				struct trace_event_file *file)
685 {
686 	struct trace_probe *pos, *tp;
687 	struct trace_eprobe *ep;
688 
689 	tp = trace_probe_primary_from_call(call);
690 	if (WARN_ON_ONCE(!tp))
691 		return -ENODEV;
692 
693 	if (file) {
694 		if (!trace_probe_get_file_link(tp, file))
695 			return -ENOENT;
696 		if (!trace_probe_has_single_file(tp))
697 			goto out;
698 		trace_probe_clear_flag(tp, TP_FLAG_TRACE);
699 	} else
700 		trace_probe_clear_flag(tp, TP_FLAG_PROFILE);
701 
702 	if (!trace_probe_is_enabled(tp)) {
703 		list_for_each_entry(pos, trace_probe_probe_list(tp), list) {
704 			ep = container_of(pos, struct trace_eprobe, tp);
705 			disable_eprobe(ep, file->tr);
706 		}
707 	}
708 
709  out:
710 	if (file)
711 		/*
712 		 * Synchronization is done in below function. For perf event,
713 		 * file == NULL and perf_trace_event_unreg() calls
714 		 * tracepoint_synchronize_unregister() to ensure synchronize
715 		 * event. We don't need to care about it.
716 		 */
717 		trace_probe_remove_file(tp, file);
718 
719 	return 0;
720 }
721 
722 static int eprobe_register(struct trace_event_call *event,
723 			   enum trace_reg type, void *data)
724 {
725 	struct trace_event_file *file = data;
726 
727 	switch (type) {
728 	case TRACE_REG_REGISTER:
729 		return enable_trace_eprobe(event, file);
730 	case TRACE_REG_UNREGISTER:
731 		return disable_trace_eprobe(event, file);
732 #ifdef CONFIG_PERF_EVENTS
733 	case TRACE_REG_PERF_REGISTER:
734 	case TRACE_REG_PERF_UNREGISTER:
735 	case TRACE_REG_PERF_OPEN:
736 	case TRACE_REG_PERF_CLOSE:
737 	case TRACE_REG_PERF_ADD:
738 	case TRACE_REG_PERF_DEL:
739 		return 0;
740 #endif
741 	}
742 	return 0;
743 }
744 
745 static inline void init_trace_eprobe_call(struct trace_eprobe *ep)
746 {
747 	struct trace_event_call *call = trace_probe_event_call(&ep->tp);
748 
749 	call->flags = TRACE_EVENT_FL_EPROBE;
750 	call->event.funcs = &eprobe_funcs;
751 	call->class->fields_array = eprobe_fields_array;
752 	call->class->reg = eprobe_register;
753 }
754 
755 static struct trace_event_call *
756 find_and_get_event(const char *system, const char *event_name)
757 {
758 	struct trace_event_call *tp_event;
759 	const char *name;
760 
761 	list_for_each_entry(tp_event, &ftrace_events, list) {
762 		/* Skip other probes and ftrace events */
763 		if (tp_event->flags &
764 		    (TRACE_EVENT_FL_IGNORE_ENABLE |
765 		     TRACE_EVENT_FL_KPROBE |
766 		     TRACE_EVENT_FL_UPROBE |
767 		     TRACE_EVENT_FL_EPROBE))
768 			continue;
769 		if (!tp_event->class->system ||
770 		    strcmp(system, tp_event->class->system))
771 			continue;
772 		name = trace_event_name(tp_event);
773 		if (!name || strcmp(event_name, name))
774 			continue;
775 		if (!trace_event_try_get_ref(tp_event)) {
776 			return NULL;
777 			break;
778 		}
779 		return tp_event;
780 		break;
781 	}
782 	return NULL;
783 }
784 
785 static int trace_eprobe_tp_update_arg(struct trace_eprobe *ep, const char *argv[], int i)
786 {
787 	unsigned int flags = TPARG_FL_KERNEL | TPARG_FL_TPOINT;
788 	int ret;
789 
790 	ret = traceprobe_parse_probe_arg(&ep->tp, i, argv[i], flags);
791 	if (ret)
792 		return ret;
793 
794 	if (ep->tp.args[i].code->op == FETCH_OP_TP_ARG)
795 		ret = trace_eprobe_tp_arg_update(ep, i);
796 
797 	return ret;
798 }
799 
800 static int __trace_eprobe_create(int argc, const char *argv[])
801 {
802 	/*
803 	 * Argument syntax:
804 	 *      e[:[GRP/]ENAME] SYSTEM.EVENT [FETCHARGS]
805 	 * Fetch args:
806 	 *  <name>=$<field>[:TYPE]
807 	 */
808 	const char *event = NULL, *group = EPROBE_EVENT_SYSTEM;
809 	const char *sys_event = NULL, *sys_name = NULL;
810 	struct trace_event_call *event_call;
811 	struct trace_eprobe *ep = NULL;
812 	char buf1[MAX_EVENT_NAME_LEN];
813 	char buf2[MAX_EVENT_NAME_LEN];
814 	int ret = 0;
815 	int i;
816 
817 	if (argc < 2 || argv[0][0] != 'e')
818 		return -ECANCELED;
819 
820 	trace_probe_log_init("event_probe", argc, argv);
821 
822 	event = strchr(&argv[0][1], ':');
823 	if (event) {
824 		event++;
825 		ret = traceprobe_parse_event_name(&event, &group, buf1,
826 						  event - argv[0]);
827 		if (ret)
828 			goto parse_error;
829 	} else {
830 		strscpy(buf1, argv[1], MAX_EVENT_NAME_LEN);
831 		sanitize_event_name(buf1);
832 		event = buf1;
833 	}
834 	if (!is_good_name(event) || !is_good_name(group))
835 		goto parse_error;
836 
837 	sys_event = argv[1];
838 	ret = traceprobe_parse_event_name(&sys_event, &sys_name, buf2,
839 					  sys_event - argv[1]);
840 	if (ret || !sys_name)
841 		goto parse_error;
842 	if (!is_good_name(sys_event) || !is_good_name(sys_name))
843 		goto parse_error;
844 
845 	mutex_lock(&event_mutex);
846 	event_call = find_and_get_event(sys_name, sys_event);
847 	ep = alloc_event_probe(group, event, event_call, argc - 2);
848 	mutex_unlock(&event_mutex);
849 
850 	if (IS_ERR(ep)) {
851 		ret = PTR_ERR(ep);
852 		/* This must return -ENOMEM, else there is a bug */
853 		WARN_ON_ONCE(ret != -ENOMEM);
854 		ep = NULL;
855 		goto error;
856 	}
857 
858 	argc -= 2; argv += 2;
859 	/* parse arguments */
860 	for (i = 0; i < argc && i < MAX_TRACE_ARGS; i++) {
861 		trace_probe_log_set_index(i + 2);
862 		ret = trace_eprobe_tp_update_arg(ep, argv, i);
863 		if (ret)
864 			goto error;
865 	}
866 	ret = traceprobe_set_print_fmt(&ep->tp, PROBE_PRINT_EVENT);
867 	if (ret < 0)
868 		goto error;
869 	init_trace_eprobe_call(ep);
870 	mutex_lock(&event_mutex);
871 	ret = trace_probe_register_event_call(&ep->tp);
872 	if (ret) {
873 		if (ret == -EEXIST) {
874 			trace_probe_log_set_index(0);
875 			trace_probe_log_err(0, EVENT_EXIST);
876 		}
877 		mutex_unlock(&event_mutex);
878 		goto error;
879 	}
880 	ret = dyn_event_add(&ep->devent, &ep->tp.event->call);
881 	mutex_unlock(&event_mutex);
882 	return ret;
883 parse_error:
884 	ret = -EINVAL;
885 error:
886 	trace_event_probe_cleanup(ep);
887 	return ret;
888 }
889 
890 /*
891  * Register dynevent at core_initcall. This allows kernel to setup eprobe
892  * events in postcore_initcall without tracefs.
893  */
894 static __init int trace_events_eprobe_init_early(void)
895 {
896 	int err = 0;
897 
898 	err = dyn_event_register(&eprobe_dyn_event_ops);
899 	if (err)
900 		pr_warn("Could not register eprobe_dyn_event_ops\n");
901 
902 	return err;
903 }
904 core_initcall(trace_events_eprobe_init_early);
905