1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * event probes 4 * 5 * Part of this code was copied from kernel/trace/trace_kprobe.c written by 6 * Masami Hiramatsu <mhiramat@kernel.org> 7 * 8 * Copyright (C) 2021, VMware Inc, Steven Rostedt <rostedt@goodmis.org> 9 * Copyright (C) 2021, VMware Inc, Tzvetomir Stoyanov tz.stoyanov@gmail.com> 10 * 11 */ 12 #include <linux/module.h> 13 #include <linux/mutex.h> 14 #include <linux/ftrace.h> 15 16 #include "trace_dynevent.h" 17 #include "trace_probe.h" 18 #include "trace_probe_tmpl.h" 19 20 #define EPROBE_EVENT_SYSTEM "eprobes" 21 22 struct trace_eprobe { 23 /* tracepoint system */ 24 const char *event_system; 25 26 /* tracepoint event */ 27 const char *event_name; 28 29 struct trace_event_call *event; 30 31 struct dyn_event devent; 32 struct trace_probe tp; 33 }; 34 35 struct eprobe_data { 36 struct trace_event_file *file; 37 struct trace_eprobe *ep; 38 }; 39 40 static int __trace_eprobe_create(int argc, const char *argv[]); 41 42 static void trace_event_probe_cleanup(struct trace_eprobe *ep) 43 { 44 if (!ep) 45 return; 46 trace_probe_cleanup(&ep->tp); 47 kfree(ep->event_name); 48 kfree(ep->event_system); 49 if (ep->event) 50 trace_event_put_ref(ep->event); 51 kfree(ep); 52 } 53 54 static struct trace_eprobe *to_trace_eprobe(struct dyn_event *ev) 55 { 56 return container_of(ev, struct trace_eprobe, devent); 57 } 58 59 static int eprobe_dyn_event_create(const char *raw_command) 60 { 61 return trace_probe_create(raw_command, __trace_eprobe_create); 62 } 63 64 static int eprobe_dyn_event_show(struct seq_file *m, struct dyn_event *ev) 65 { 66 struct trace_eprobe *ep = to_trace_eprobe(ev); 67 int i; 68 69 seq_printf(m, "e:%s/%s", trace_probe_group_name(&ep->tp), 70 trace_probe_name(&ep->tp)); 71 seq_printf(m, " %s.%s", ep->event_system, ep->event_name); 72 73 for (i = 0; i < ep->tp.nr_args; i++) 74 seq_printf(m, " %s=%s", ep->tp.args[i].name, ep->tp.args[i].comm); 75 seq_putc(m, '\n'); 76 77 return 0; 78 } 79 80 static int unregister_trace_eprobe(struct trace_eprobe *ep) 81 { 82 /* If other probes are on the event, just unregister eprobe */ 83 if (trace_probe_has_sibling(&ep->tp)) 84 goto unreg; 85 86 /* Enabled event can not be unregistered */ 87 if (trace_probe_is_enabled(&ep->tp)) 88 return -EBUSY; 89 90 /* Will fail if probe is being used by ftrace or perf */ 91 if (trace_probe_unregister_event_call(&ep->tp)) 92 return -EBUSY; 93 94 unreg: 95 dyn_event_remove(&ep->devent); 96 trace_probe_unlink(&ep->tp); 97 98 return 0; 99 } 100 101 static int eprobe_dyn_event_release(struct dyn_event *ev) 102 { 103 struct trace_eprobe *ep = to_trace_eprobe(ev); 104 int ret = unregister_trace_eprobe(ep); 105 106 if (!ret) 107 trace_event_probe_cleanup(ep); 108 return ret; 109 } 110 111 static bool eprobe_dyn_event_is_busy(struct dyn_event *ev) 112 { 113 struct trace_eprobe *ep = to_trace_eprobe(ev); 114 115 return trace_probe_is_enabled(&ep->tp); 116 } 117 118 static bool eprobe_dyn_event_match(const char *system, const char *event, 119 int argc, const char **argv, struct dyn_event *ev) 120 { 121 struct trace_eprobe *ep = to_trace_eprobe(ev); 122 const char *slash; 123 124 /* 125 * We match the following: 126 * event only - match all eprobes with event name 127 * system and event only - match all system/event probes 128 * 129 * The below has the above satisfied with more arguments: 130 * 131 * attached system/event - If the arg has the system and event 132 * the probe is attached to, match 133 * probes with the attachment. 134 * 135 * If any more args are given, then it requires a full match. 136 */ 137 138 /* 139 * If system exists, but this probe is not part of that system 140 * do not match. 141 */ 142 if (system && strcmp(trace_probe_group_name(&ep->tp), system) != 0) 143 return false; 144 145 /* Must match the event name */ 146 if (strcmp(trace_probe_name(&ep->tp), event) != 0) 147 return false; 148 149 /* No arguments match all */ 150 if (argc < 1) 151 return true; 152 153 /* First argument is the system/event the probe is attached to */ 154 155 slash = strchr(argv[0], '/'); 156 if (!slash) 157 slash = strchr(argv[0], '.'); 158 if (!slash) 159 return false; 160 161 if (strncmp(ep->event_system, argv[0], slash - argv[0])) 162 return false; 163 if (strcmp(ep->event_name, slash + 1)) 164 return false; 165 166 argc--; 167 argv++; 168 169 /* If there are no other args, then match */ 170 if (argc < 1) 171 return true; 172 173 return trace_probe_match_command_args(&ep->tp, argc, argv); 174 } 175 176 static struct dyn_event_operations eprobe_dyn_event_ops = { 177 .create = eprobe_dyn_event_create, 178 .show = eprobe_dyn_event_show, 179 .is_busy = eprobe_dyn_event_is_busy, 180 .free = eprobe_dyn_event_release, 181 .match = eprobe_dyn_event_match, 182 }; 183 184 static struct trace_eprobe *alloc_event_probe(const char *group, 185 const char *this_event, 186 struct trace_event_call *event, 187 int nargs) 188 { 189 struct trace_eprobe *ep; 190 const char *event_name; 191 const char *sys_name; 192 int ret = -ENOMEM; 193 194 if (!event) 195 return ERR_PTR(-ENODEV); 196 197 sys_name = event->class->system; 198 event_name = trace_event_name(event); 199 200 ep = kzalloc(struct_size(ep, tp.args, nargs), GFP_KERNEL); 201 if (!ep) { 202 trace_event_put_ref(event); 203 goto error; 204 } 205 ep->event = event; 206 ep->event_name = kstrdup(event_name, GFP_KERNEL); 207 if (!ep->event_name) 208 goto error; 209 ep->event_system = kstrdup(sys_name, GFP_KERNEL); 210 if (!ep->event_system) 211 goto error; 212 213 ret = trace_probe_init(&ep->tp, this_event, group, false); 214 if (ret < 0) 215 goto error; 216 217 dyn_event_init(&ep->devent, &eprobe_dyn_event_ops); 218 return ep; 219 error: 220 trace_event_probe_cleanup(ep); 221 return ERR_PTR(ret); 222 } 223 224 static int trace_eprobe_tp_arg_update(struct trace_eprobe *ep, int i) 225 { 226 struct probe_arg *parg = &ep->tp.args[i]; 227 struct ftrace_event_field *field; 228 struct list_head *head; 229 230 head = trace_get_fields(ep->event); 231 list_for_each_entry(field, head, link) { 232 if (!strcmp(parg->code->data, field->name)) { 233 kfree(parg->code->data); 234 parg->code->data = field; 235 return 0; 236 } 237 } 238 kfree(parg->code->data); 239 parg->code->data = NULL; 240 return -ENOENT; 241 } 242 243 static int eprobe_event_define_fields(struct trace_event_call *event_call) 244 { 245 int ret; 246 struct eprobe_trace_entry_head field; 247 struct trace_probe *tp; 248 249 tp = trace_probe_primary_from_call(event_call); 250 if (WARN_ON_ONCE(!tp)) 251 return -ENOENT; 252 253 DEFINE_FIELD(unsigned int, type, FIELD_STRING_TYPE, 0); 254 255 return traceprobe_define_arg_fields(event_call, sizeof(field), tp); 256 } 257 258 static struct trace_event_fields eprobe_fields_array[] = { 259 { .type = TRACE_FUNCTION_TYPE, 260 .define_fields = eprobe_event_define_fields }, 261 {} 262 }; 263 264 /* Event entry printers */ 265 static enum print_line_t 266 print_eprobe_event(struct trace_iterator *iter, int flags, 267 struct trace_event *event) 268 { 269 struct eprobe_trace_entry_head *field; 270 struct trace_event_call *pevent; 271 struct trace_event *probed_event; 272 struct trace_seq *s = &iter->seq; 273 struct trace_probe *tp; 274 275 field = (struct eprobe_trace_entry_head *)iter->ent; 276 tp = trace_probe_primary_from_call( 277 container_of(event, struct trace_event_call, event)); 278 if (WARN_ON_ONCE(!tp)) 279 goto out; 280 281 trace_seq_printf(s, "%s: (", trace_probe_name(tp)); 282 283 probed_event = ftrace_find_event(field->type); 284 if (probed_event) { 285 pevent = container_of(probed_event, struct trace_event_call, event); 286 trace_seq_printf(s, "%s.%s", pevent->class->system, 287 trace_event_name(pevent)); 288 } else { 289 trace_seq_printf(s, "%u", field->type); 290 } 291 292 trace_seq_putc(s, ')'); 293 294 if (print_probe_args(s, tp->args, tp->nr_args, 295 (u8 *)&field[1], field) < 0) 296 goto out; 297 298 trace_seq_putc(s, '\n'); 299 out: 300 return trace_handle_return(s); 301 } 302 303 static unsigned long get_event_field(struct fetch_insn *code, void *rec) 304 { 305 struct ftrace_event_field *field = code->data; 306 unsigned long val; 307 void *addr; 308 309 addr = rec + field->offset; 310 311 switch (field->size) { 312 case 1: 313 if (field->is_signed) 314 val = *(char *)addr; 315 else 316 val = *(unsigned char *)addr; 317 break; 318 case 2: 319 if (field->is_signed) 320 val = *(short *)addr; 321 else 322 val = *(unsigned short *)addr; 323 break; 324 case 4: 325 if (field->is_signed) 326 val = *(int *)addr; 327 else 328 val = *(unsigned int *)addr; 329 break; 330 default: 331 if (field->is_signed) 332 val = *(long *)addr; 333 else 334 val = *(unsigned long *)addr; 335 break; 336 } 337 return val; 338 } 339 340 static int get_eprobe_size(struct trace_probe *tp, void *rec) 341 { 342 struct probe_arg *arg; 343 int i, len, ret = 0; 344 345 for (i = 0; i < tp->nr_args; i++) { 346 arg = tp->args + i; 347 if (unlikely(arg->dynamic)) { 348 unsigned long val; 349 350 val = get_event_field(arg->code, rec); 351 len = process_fetch_insn_bottom(arg->code + 1, val, NULL, NULL); 352 if (len > 0) 353 ret += len; 354 } 355 } 356 357 return ret; 358 } 359 360 /* Kprobe specific fetch functions */ 361 362 /* Note that we don't verify it, since the code does not come from user space */ 363 static int 364 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, 365 void *base) 366 { 367 unsigned long val; 368 369 val = get_event_field(code, rec); 370 return process_fetch_insn_bottom(code + 1, val, dest, base); 371 } 372 NOKPROBE_SYMBOL(process_fetch_insn) 373 374 /* Return the length of string -- including null terminal byte */ 375 static nokprobe_inline int 376 fetch_store_strlen_user(unsigned long addr) 377 { 378 const void __user *uaddr = (__force const void __user *)addr; 379 380 return strnlen_user_nofault(uaddr, MAX_STRING_SIZE); 381 } 382 383 /* Return the length of string -- including null terminal byte */ 384 static nokprobe_inline int 385 fetch_store_strlen(unsigned long addr) 386 { 387 int ret, len = 0; 388 u8 c; 389 390 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE 391 if (addr < TASK_SIZE) 392 return fetch_store_strlen_user(addr); 393 #endif 394 395 do { 396 ret = copy_from_kernel_nofault(&c, (u8 *)addr + len, 1); 397 len++; 398 } while (c && ret == 0 && len < MAX_STRING_SIZE); 399 400 return (ret < 0) ? ret : len; 401 } 402 403 /* 404 * Fetch a null-terminated string from user. Caller MUST set *(u32 *)buf 405 * with max length and relative data location. 406 */ 407 static nokprobe_inline int 408 fetch_store_string_user(unsigned long addr, void *dest, void *base) 409 { 410 const void __user *uaddr = (__force const void __user *)addr; 411 int maxlen = get_loc_len(*(u32 *)dest); 412 void *__dest; 413 long ret; 414 415 if (unlikely(!maxlen)) 416 return -ENOMEM; 417 418 __dest = get_loc_data(dest, base); 419 420 ret = strncpy_from_user_nofault(__dest, uaddr, maxlen); 421 if (ret >= 0) 422 *(u32 *)dest = make_data_loc(ret, __dest - base); 423 424 return ret; 425 } 426 427 /* 428 * Fetch a null-terminated string. Caller MUST set *(u32 *)buf with max 429 * length and relative data location. 430 */ 431 static nokprobe_inline int 432 fetch_store_string(unsigned long addr, void *dest, void *base) 433 { 434 int maxlen = get_loc_len(*(u32 *)dest); 435 void *__dest; 436 long ret; 437 438 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE 439 if ((unsigned long)addr < TASK_SIZE) 440 return fetch_store_string_user(addr, dest, base); 441 #endif 442 443 if (unlikely(!maxlen)) 444 return -ENOMEM; 445 446 __dest = get_loc_data(dest, base); 447 448 /* 449 * Try to get string again, since the string can be changed while 450 * probing. 451 */ 452 ret = strncpy_from_kernel_nofault(__dest, (void *)addr, maxlen); 453 if (ret >= 0) 454 *(u32 *)dest = make_data_loc(ret, __dest - base); 455 456 return ret; 457 } 458 459 static nokprobe_inline int 460 probe_mem_read_user(void *dest, void *src, size_t size) 461 { 462 const void __user *uaddr = (__force const void __user *)src; 463 464 return copy_from_user_nofault(dest, uaddr, size); 465 } 466 467 static nokprobe_inline int 468 probe_mem_read(void *dest, void *src, size_t size) 469 { 470 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE 471 if ((unsigned long)src < TASK_SIZE) 472 return probe_mem_read_user(dest, src, size); 473 #endif 474 return copy_from_kernel_nofault(dest, src, size); 475 } 476 477 /* eprobe handler */ 478 static inline void 479 __eprobe_trace_func(struct eprobe_data *edata, void *rec) 480 { 481 struct eprobe_trace_entry_head *entry; 482 struct trace_event_call *call = trace_probe_event_call(&edata->ep->tp); 483 struct trace_event_buffer fbuffer; 484 int dsize; 485 486 if (WARN_ON_ONCE(call != edata->file->event_call)) 487 return; 488 489 if (trace_trigger_soft_disabled(edata->file)) 490 return; 491 492 dsize = get_eprobe_size(&edata->ep->tp, rec); 493 494 entry = trace_event_buffer_reserve(&fbuffer, edata->file, 495 sizeof(*entry) + edata->ep->tp.size + dsize); 496 497 if (!entry) 498 return; 499 500 entry = fbuffer.entry = ring_buffer_event_data(fbuffer.event); 501 if (edata->ep->event) 502 entry->type = edata->ep->event->event.type; 503 else 504 entry->type = 0; 505 store_trace_args(&entry[1], &edata->ep->tp, rec, sizeof(*entry), dsize); 506 507 trace_event_buffer_commit(&fbuffer); 508 } 509 510 /* 511 * The event probe implementation uses event triggers to get access to 512 * the event it is attached to, but is not an actual trigger. The below 513 * functions are just stubs to fulfill what is needed to use the trigger 514 * infrastructure. 515 */ 516 static int eprobe_trigger_init(struct event_trigger_ops *ops, 517 struct event_trigger_data *data) 518 { 519 return 0; 520 } 521 522 static void eprobe_trigger_free(struct event_trigger_ops *ops, 523 struct event_trigger_data *data) 524 { 525 526 } 527 528 static int eprobe_trigger_print(struct seq_file *m, 529 struct event_trigger_ops *ops, 530 struct event_trigger_data *data) 531 { 532 /* Do not print eprobe event triggers */ 533 return 0; 534 } 535 536 static void eprobe_trigger_func(struct event_trigger_data *data, 537 struct trace_buffer *buffer, void *rec, 538 struct ring_buffer_event *rbe) 539 { 540 struct eprobe_data *edata = data->private_data; 541 542 __eprobe_trace_func(edata, rec); 543 } 544 545 static struct event_trigger_ops eprobe_trigger_ops = { 546 .trigger = eprobe_trigger_func, 547 .print = eprobe_trigger_print, 548 .init = eprobe_trigger_init, 549 .free = eprobe_trigger_free, 550 }; 551 552 static int eprobe_trigger_cmd_parse(struct event_command *cmd_ops, 553 struct trace_event_file *file, 554 char *glob, char *cmd, char *param) 555 { 556 return -1; 557 } 558 559 static int eprobe_trigger_reg_func(char *glob, 560 struct event_trigger_data *data, 561 struct trace_event_file *file) 562 { 563 return -1; 564 } 565 566 static void eprobe_trigger_unreg_func(char *glob, 567 struct event_trigger_data *data, 568 struct trace_event_file *file) 569 { 570 571 } 572 573 static struct event_trigger_ops *eprobe_trigger_get_ops(char *cmd, 574 char *param) 575 { 576 return &eprobe_trigger_ops; 577 } 578 579 static struct event_command event_trigger_cmd = { 580 .name = "eprobe", 581 .trigger_type = ETT_EVENT_EPROBE, 582 .flags = EVENT_CMD_FL_NEEDS_REC, 583 .parse = eprobe_trigger_cmd_parse, 584 .reg = eprobe_trigger_reg_func, 585 .unreg = eprobe_trigger_unreg_func, 586 .unreg_all = NULL, 587 .get_trigger_ops = eprobe_trigger_get_ops, 588 .set_filter = NULL, 589 }; 590 591 static struct event_trigger_data * 592 new_eprobe_trigger(struct trace_eprobe *ep, struct trace_event_file *file) 593 { 594 struct event_trigger_data *trigger; 595 struct eprobe_data *edata; 596 597 edata = kzalloc(sizeof(*edata), GFP_KERNEL); 598 trigger = kzalloc(sizeof(*trigger), GFP_KERNEL); 599 if (!trigger || !edata) { 600 kfree(edata); 601 kfree(trigger); 602 return ERR_PTR(-ENOMEM); 603 } 604 605 trigger->flags = EVENT_TRIGGER_FL_PROBE; 606 trigger->count = -1; 607 trigger->ops = &eprobe_trigger_ops; 608 609 /* 610 * EVENT PROBE triggers are not registered as commands with 611 * register_event_command(), as they are not controlled by the user 612 * from the trigger file 613 */ 614 trigger->cmd_ops = &event_trigger_cmd; 615 616 INIT_LIST_HEAD(&trigger->list); 617 RCU_INIT_POINTER(trigger->filter, NULL); 618 619 edata->file = file; 620 edata->ep = ep; 621 trigger->private_data = edata; 622 623 return trigger; 624 } 625 626 static int enable_eprobe(struct trace_eprobe *ep, 627 struct trace_event_file *eprobe_file) 628 { 629 struct event_trigger_data *trigger; 630 struct trace_event_file *file; 631 struct trace_array *tr = eprobe_file->tr; 632 633 file = find_event_file(tr, ep->event_system, ep->event_name); 634 if (!file) 635 return -ENOENT; 636 trigger = new_eprobe_trigger(ep, eprobe_file); 637 if (IS_ERR(trigger)) 638 return PTR_ERR(trigger); 639 640 list_add_tail_rcu(&trigger->list, &file->triggers); 641 642 trace_event_trigger_enable_disable(file, 1); 643 update_cond_flag(file); 644 645 return 0; 646 } 647 648 static struct trace_event_functions eprobe_funcs = { 649 .trace = print_eprobe_event 650 }; 651 652 static int disable_eprobe(struct trace_eprobe *ep, 653 struct trace_array *tr) 654 { 655 struct event_trigger_data *trigger; 656 struct trace_event_file *file; 657 struct eprobe_data *edata; 658 659 file = find_event_file(tr, ep->event_system, ep->event_name); 660 if (!file) 661 return -ENOENT; 662 663 list_for_each_entry(trigger, &file->triggers, list) { 664 if (!(trigger->flags & EVENT_TRIGGER_FL_PROBE)) 665 continue; 666 edata = trigger->private_data; 667 if (edata->ep == ep) 668 break; 669 } 670 if (list_entry_is_head(trigger, &file->triggers, list)) 671 return -ENODEV; 672 673 list_del_rcu(&trigger->list); 674 675 trace_event_trigger_enable_disable(file, 0); 676 update_cond_flag(file); 677 678 /* Make sure nothing is using the edata or trigger */ 679 tracepoint_synchronize_unregister(); 680 681 kfree(edata); 682 kfree(trigger); 683 684 return 0; 685 } 686 687 static int enable_trace_eprobe(struct trace_event_call *call, 688 struct trace_event_file *file) 689 { 690 struct trace_probe *pos, *tp; 691 struct trace_eprobe *ep; 692 bool enabled; 693 int ret = 0; 694 695 tp = trace_probe_primary_from_call(call); 696 if (WARN_ON_ONCE(!tp)) 697 return -ENODEV; 698 enabled = trace_probe_is_enabled(tp); 699 700 /* This also changes "enabled" state */ 701 if (file) { 702 ret = trace_probe_add_file(tp, file); 703 if (ret) 704 return ret; 705 } else 706 trace_probe_set_flag(tp, TP_FLAG_PROFILE); 707 708 if (enabled) 709 return 0; 710 711 list_for_each_entry(pos, trace_probe_probe_list(tp), list) { 712 ep = container_of(pos, struct trace_eprobe, tp); 713 ret = enable_eprobe(ep, file); 714 if (ret) 715 break; 716 enabled = true; 717 } 718 719 if (ret) { 720 /* Failed to enable one of them. Roll back all */ 721 if (enabled) 722 disable_eprobe(ep, file->tr); 723 if (file) 724 trace_probe_remove_file(tp, file); 725 else 726 trace_probe_clear_flag(tp, TP_FLAG_PROFILE); 727 } 728 729 return ret; 730 } 731 732 static int disable_trace_eprobe(struct trace_event_call *call, 733 struct trace_event_file *file) 734 { 735 struct trace_probe *pos, *tp; 736 struct trace_eprobe *ep; 737 738 tp = trace_probe_primary_from_call(call); 739 if (WARN_ON_ONCE(!tp)) 740 return -ENODEV; 741 742 if (file) { 743 if (!trace_probe_get_file_link(tp, file)) 744 return -ENOENT; 745 if (!trace_probe_has_single_file(tp)) 746 goto out; 747 trace_probe_clear_flag(tp, TP_FLAG_TRACE); 748 } else 749 trace_probe_clear_flag(tp, TP_FLAG_PROFILE); 750 751 if (!trace_probe_is_enabled(tp)) { 752 list_for_each_entry(pos, trace_probe_probe_list(tp), list) { 753 ep = container_of(pos, struct trace_eprobe, tp); 754 disable_eprobe(ep, file->tr); 755 } 756 } 757 758 out: 759 if (file) 760 /* 761 * Synchronization is done in below function. For perf event, 762 * file == NULL and perf_trace_event_unreg() calls 763 * tracepoint_synchronize_unregister() to ensure synchronize 764 * event. We don't need to care about it. 765 */ 766 trace_probe_remove_file(tp, file); 767 768 return 0; 769 } 770 771 static int eprobe_register(struct trace_event_call *event, 772 enum trace_reg type, void *data) 773 { 774 struct trace_event_file *file = data; 775 776 switch (type) { 777 case TRACE_REG_REGISTER: 778 return enable_trace_eprobe(event, file); 779 case TRACE_REG_UNREGISTER: 780 return disable_trace_eprobe(event, file); 781 #ifdef CONFIG_PERF_EVENTS 782 case TRACE_REG_PERF_REGISTER: 783 case TRACE_REG_PERF_UNREGISTER: 784 case TRACE_REG_PERF_OPEN: 785 case TRACE_REG_PERF_CLOSE: 786 case TRACE_REG_PERF_ADD: 787 case TRACE_REG_PERF_DEL: 788 return 0; 789 #endif 790 } 791 return 0; 792 } 793 794 static inline void init_trace_eprobe_call(struct trace_eprobe *ep) 795 { 796 struct trace_event_call *call = trace_probe_event_call(&ep->tp); 797 798 call->flags = TRACE_EVENT_FL_EPROBE; 799 call->event.funcs = &eprobe_funcs; 800 call->class->fields_array = eprobe_fields_array; 801 call->class->reg = eprobe_register; 802 } 803 804 static struct trace_event_call * 805 find_and_get_event(const char *system, const char *event_name) 806 { 807 struct trace_event_call *tp_event; 808 const char *name; 809 810 list_for_each_entry(tp_event, &ftrace_events, list) { 811 /* Skip other probes and ftrace events */ 812 if (tp_event->flags & 813 (TRACE_EVENT_FL_IGNORE_ENABLE | 814 TRACE_EVENT_FL_KPROBE | 815 TRACE_EVENT_FL_UPROBE | 816 TRACE_EVENT_FL_EPROBE)) 817 continue; 818 if (!tp_event->class->system || 819 strcmp(system, tp_event->class->system)) 820 continue; 821 name = trace_event_name(tp_event); 822 if (!name || strcmp(event_name, name)) 823 continue; 824 if (!trace_event_try_get_ref(tp_event)) { 825 return NULL; 826 break; 827 } 828 return tp_event; 829 break; 830 } 831 return NULL; 832 } 833 834 static int trace_eprobe_tp_update_arg(struct trace_eprobe *ep, const char *argv[], int i) 835 { 836 unsigned int flags = TPARG_FL_KERNEL | TPARG_FL_TPOINT; 837 int ret; 838 839 ret = traceprobe_parse_probe_arg(&ep->tp, i, argv[i], flags); 840 if (ret) 841 return ret; 842 843 if (ep->tp.args[i].code->op == FETCH_OP_TP_ARG) 844 ret = trace_eprobe_tp_arg_update(ep, i); 845 846 return ret; 847 } 848 849 static int __trace_eprobe_create(int argc, const char *argv[]) 850 { 851 /* 852 * Argument syntax: 853 * e[:[GRP/]ENAME] SYSTEM.EVENT [FETCHARGS] 854 * Fetch args: 855 * <name>=$<field>[:TYPE] 856 */ 857 const char *event = NULL, *group = EPROBE_EVENT_SYSTEM; 858 const char *sys_event = NULL, *sys_name = NULL; 859 struct trace_event_call *event_call; 860 struct trace_eprobe *ep = NULL; 861 char buf1[MAX_EVENT_NAME_LEN]; 862 char buf2[MAX_EVENT_NAME_LEN]; 863 int ret = 0; 864 int i; 865 866 if (argc < 2 || argv[0][0] != 'e') 867 return -ECANCELED; 868 869 trace_probe_log_init("event_probe", argc, argv); 870 871 event = strchr(&argv[0][1], ':'); 872 if (event) { 873 event++; 874 ret = traceprobe_parse_event_name(&event, &group, buf1, 875 event - argv[0]); 876 if (ret) 877 goto parse_error; 878 } else { 879 strscpy(buf1, argv[1], MAX_EVENT_NAME_LEN); 880 sanitize_event_name(buf1); 881 event = buf1; 882 } 883 if (!is_good_name(event) || !is_good_name(group)) 884 goto parse_error; 885 886 sys_event = argv[1]; 887 ret = traceprobe_parse_event_name(&sys_event, &sys_name, buf2, 888 sys_event - argv[1]); 889 if (ret || !sys_name) 890 goto parse_error; 891 if (!is_good_name(sys_event) || !is_good_name(sys_name)) 892 goto parse_error; 893 894 mutex_lock(&event_mutex); 895 event_call = find_and_get_event(sys_name, sys_event); 896 ep = alloc_event_probe(group, event, event_call, argc - 2); 897 mutex_unlock(&event_mutex); 898 899 if (IS_ERR(ep)) { 900 ret = PTR_ERR(ep); 901 /* This must return -ENOMEM or missing event, else there is a bug */ 902 WARN_ON_ONCE(ret != -ENOMEM && ret != -ENODEV); 903 ep = NULL; 904 goto error; 905 } 906 907 argc -= 2; argv += 2; 908 /* parse arguments */ 909 for (i = 0; i < argc && i < MAX_TRACE_ARGS; i++) { 910 trace_probe_log_set_index(i + 2); 911 ret = trace_eprobe_tp_update_arg(ep, argv, i); 912 if (ret) 913 goto error; 914 } 915 ret = traceprobe_set_print_fmt(&ep->tp, PROBE_PRINT_EVENT); 916 if (ret < 0) 917 goto error; 918 init_trace_eprobe_call(ep); 919 mutex_lock(&event_mutex); 920 ret = trace_probe_register_event_call(&ep->tp); 921 if (ret) { 922 if (ret == -EEXIST) { 923 trace_probe_log_set_index(0); 924 trace_probe_log_err(0, EVENT_EXIST); 925 } 926 mutex_unlock(&event_mutex); 927 goto error; 928 } 929 ret = dyn_event_add(&ep->devent, &ep->tp.event->call); 930 mutex_unlock(&event_mutex); 931 return ret; 932 parse_error: 933 ret = -EINVAL; 934 error: 935 trace_event_probe_cleanup(ep); 936 return ret; 937 } 938 939 /* 940 * Register dynevent at core_initcall. This allows kernel to setup eprobe 941 * events in postcore_initcall without tracefs. 942 */ 943 static __init int trace_events_eprobe_init_early(void) 944 { 945 int err = 0; 946 947 err = dyn_event_register(&eprobe_dyn_event_ops); 948 if (err) 949 pr_warn("Could not register eprobe_dyn_event_ops\n"); 950 951 return err; 952 } 953 core_initcall(trace_events_eprobe_init_early); 954