1 #include <traceevent/event-parse.h> 2 #include "builtin.h" 3 #include "util/color.h" 4 #include "util/debug.h" 5 #include "util/evlist.h" 6 #include "util/machine.h" 7 #include "util/session.h" 8 #include "util/thread.h" 9 #include "util/parse-options.h" 10 #include "util/strlist.h" 11 #include "util/intlist.h" 12 #include "util/thread_map.h" 13 14 #include <libaudit.h> 15 #include <stdlib.h> 16 #include <sys/mman.h> 17 #include <linux/futex.h> 18 19 static size_t syscall_arg__scnprintf_hex(char *bf, size_t size, 20 unsigned long arg, 21 u8 arg_idx __maybe_unused, 22 u8 *arg_mask __maybe_unused) 23 { 24 return scnprintf(bf, size, "%#lx", arg); 25 } 26 27 #define SCA_HEX syscall_arg__scnprintf_hex 28 29 static size_t syscall_arg__scnprintf_whence(char *bf, size_t size, 30 unsigned long arg, 31 u8 arg_idx __maybe_unused, 32 u8 *arg_mask __maybe_unused) 33 { 34 int whence = arg; 35 36 switch (whence) { 37 #define P_WHENCE(n) case SEEK_##n: return scnprintf(bf, size, #n) 38 P_WHENCE(SET); 39 P_WHENCE(CUR); 40 P_WHENCE(END); 41 #ifdef SEEK_DATA 42 P_WHENCE(DATA); 43 #endif 44 #ifdef SEEK_HOLE 45 P_WHENCE(HOLE); 46 #endif 47 #undef P_WHENCE 48 default: break; 49 } 50 51 return scnprintf(bf, size, "%#x", whence); 52 } 53 54 #define SCA_WHENCE syscall_arg__scnprintf_whence 55 56 static size_t syscall_arg__scnprintf_mmap_prot(char *bf, size_t size, 57 unsigned long arg, 58 u8 arg_idx __maybe_unused, 59 u8 *arg_mask __maybe_unused) 60 { 61 int printed = 0, prot = arg; 62 63 if (prot == PROT_NONE) 64 return scnprintf(bf, size, "NONE"); 65 #define P_MMAP_PROT(n) \ 66 if (prot & PROT_##n) { \ 67 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", #n); \ 68 prot &= ~PROT_##n; \ 69 } 70 71 P_MMAP_PROT(EXEC); 72 P_MMAP_PROT(READ); 73 P_MMAP_PROT(WRITE); 74 #ifdef PROT_SEM 75 P_MMAP_PROT(SEM); 76 #endif 77 P_MMAP_PROT(GROWSDOWN); 78 P_MMAP_PROT(GROWSUP); 79 #undef P_MMAP_PROT 80 81 if (prot) 82 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", prot); 83 84 return printed; 85 } 86 87 #define SCA_MMAP_PROT syscall_arg__scnprintf_mmap_prot 88 89 static size_t syscall_arg__scnprintf_mmap_flags(char *bf, size_t size, 90 unsigned long arg, u8 arg_idx __maybe_unused, 91 u8 *arg_mask __maybe_unused) 92 { 93 int printed = 0, flags = arg; 94 95 #define P_MMAP_FLAG(n) \ 96 if (flags & MAP_##n) { \ 97 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", #n); \ 98 flags &= ~MAP_##n; \ 99 } 100 101 P_MMAP_FLAG(SHARED); 102 P_MMAP_FLAG(PRIVATE); 103 P_MMAP_FLAG(32BIT); 104 P_MMAP_FLAG(ANONYMOUS); 105 P_MMAP_FLAG(DENYWRITE); 106 P_MMAP_FLAG(EXECUTABLE); 107 P_MMAP_FLAG(FILE); 108 P_MMAP_FLAG(FIXED); 109 P_MMAP_FLAG(GROWSDOWN); 110 #ifdef MAP_HUGETLB 111 P_MMAP_FLAG(HUGETLB); 112 #endif 113 P_MMAP_FLAG(LOCKED); 114 P_MMAP_FLAG(NONBLOCK); 115 P_MMAP_FLAG(NORESERVE); 116 P_MMAP_FLAG(POPULATE); 117 P_MMAP_FLAG(STACK); 118 #ifdef MAP_UNINITIALIZED 119 P_MMAP_FLAG(UNINITIALIZED); 120 #endif 121 #undef P_MMAP_FLAG 122 123 if (flags) 124 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags); 125 126 return printed; 127 } 128 129 #define SCA_MMAP_FLAGS syscall_arg__scnprintf_mmap_flags 130 131 static size_t syscall_arg__scnprintf_madvise_behavior(char *bf, size_t size, 132 unsigned long arg, u8 arg_idx __maybe_unused, 133 u8 *arg_mask __maybe_unused) 134 { 135 int behavior = arg; 136 137 switch (behavior) { 138 #define P_MADV_BHV(n) case MADV_##n: return scnprintf(bf, size, #n) 139 P_MADV_BHV(NORMAL); 140 P_MADV_BHV(RANDOM); 141 P_MADV_BHV(SEQUENTIAL); 142 P_MADV_BHV(WILLNEED); 143 P_MADV_BHV(DONTNEED); 144 P_MADV_BHV(REMOVE); 145 P_MADV_BHV(DONTFORK); 146 P_MADV_BHV(DOFORK); 147 P_MADV_BHV(HWPOISON); 148 #ifdef MADV_SOFT_OFFLINE 149 P_MADV_BHV(SOFT_OFFLINE); 150 #endif 151 P_MADV_BHV(MERGEABLE); 152 P_MADV_BHV(UNMERGEABLE); 153 #ifdef MADV_HUGEPAGE 154 P_MADV_BHV(HUGEPAGE); 155 #endif 156 #ifdef MADV_NOHUGEPAGE 157 P_MADV_BHV(NOHUGEPAGE); 158 #endif 159 #ifdef MADV_DONTDUMP 160 P_MADV_BHV(DONTDUMP); 161 #endif 162 #ifdef MADV_DODUMP 163 P_MADV_BHV(DODUMP); 164 #endif 165 #undef P_MADV_PHV 166 default: break; 167 } 168 169 return scnprintf(bf, size, "%#x", behavior); 170 } 171 172 #define SCA_MADV_BHV syscall_arg__scnprintf_madvise_behavior 173 174 static size_t syscall_arg__scnprintf_futex_op(char *bf, size_t size, unsigned long arg, 175 u8 arg_idx __maybe_unused, u8 *arg_mask) 176 { 177 enum syscall_futex_args { 178 SCF_UADDR = (1 << 0), 179 SCF_OP = (1 << 1), 180 SCF_VAL = (1 << 2), 181 SCF_TIMEOUT = (1 << 3), 182 SCF_UADDR2 = (1 << 4), 183 SCF_VAL3 = (1 << 5), 184 }; 185 int op = arg; 186 int cmd = op & FUTEX_CMD_MASK; 187 size_t printed = 0; 188 189 switch (cmd) { 190 #define P_FUTEX_OP(n) case FUTEX_##n: printed = scnprintf(bf, size, #n); 191 P_FUTEX_OP(WAIT); *arg_mask |= SCF_VAL3|SCF_UADDR2; break; 192 P_FUTEX_OP(WAKE); *arg_mask |= SCF_VAL3|SCF_UADDR2|SCF_TIMEOUT; break; 193 P_FUTEX_OP(FD); *arg_mask |= SCF_VAL3|SCF_UADDR2|SCF_TIMEOUT; break; 194 P_FUTEX_OP(REQUEUE); *arg_mask |= SCF_VAL3|SCF_TIMEOUT; break; 195 P_FUTEX_OP(CMP_REQUEUE); *arg_mask |= SCF_TIMEOUT; break; 196 P_FUTEX_OP(CMP_REQUEUE_PI); *arg_mask |= SCF_TIMEOUT; break; 197 P_FUTEX_OP(WAKE_OP); break; 198 P_FUTEX_OP(LOCK_PI); *arg_mask |= SCF_VAL3|SCF_UADDR2|SCF_TIMEOUT; break; 199 P_FUTEX_OP(UNLOCK_PI); *arg_mask |= SCF_VAL3|SCF_UADDR2|SCF_TIMEOUT; break; 200 P_FUTEX_OP(TRYLOCK_PI); *arg_mask |= SCF_VAL3|SCF_UADDR2; break; 201 P_FUTEX_OP(WAIT_BITSET); *arg_mask |= SCF_UADDR2; break; 202 P_FUTEX_OP(WAKE_BITSET); *arg_mask |= SCF_UADDR2; break; 203 P_FUTEX_OP(WAIT_REQUEUE_PI); break; 204 default: printed = scnprintf(bf, size, "%#x", cmd); break; 205 } 206 207 if (op & FUTEX_PRIVATE_FLAG) 208 printed += scnprintf(bf + printed, size - printed, "|PRIV"); 209 210 if (op & FUTEX_CLOCK_REALTIME) 211 printed += scnprintf(bf + printed, size - printed, "|CLKRT"); 212 213 return printed; 214 } 215 216 #define SCA_FUTEX_OP syscall_arg__scnprintf_futex_op 217 218 static size_t syscall_arg__scnprintf_open_flags(char *bf, size_t size, 219 unsigned long arg, 220 u8 arg_idx, u8 *arg_mask) 221 { 222 int printed = 0, flags = arg; 223 224 if (!(flags & O_CREAT)) 225 *arg_mask |= 1 << (arg_idx + 1); /* Mask the mode parm */ 226 227 if (flags == 0) 228 return scnprintf(bf, size, "RDONLY"); 229 #define P_FLAG(n) \ 230 if (flags & O_##n) { \ 231 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", #n); \ 232 flags &= ~O_##n; \ 233 } 234 235 P_FLAG(APPEND); 236 P_FLAG(ASYNC); 237 P_FLAG(CLOEXEC); 238 P_FLAG(CREAT); 239 P_FLAG(DIRECT); 240 P_FLAG(DIRECTORY); 241 P_FLAG(EXCL); 242 P_FLAG(LARGEFILE); 243 P_FLAG(NOATIME); 244 P_FLAG(NOCTTY); 245 #ifdef O_NONBLOCK 246 P_FLAG(NONBLOCK); 247 #elif O_NDELAY 248 P_FLAG(NDELAY); 249 #endif 250 #ifdef O_PATH 251 P_FLAG(PATH); 252 #endif 253 P_FLAG(RDWR); 254 #ifdef O_DSYNC 255 if ((flags & O_SYNC) == O_SYNC) 256 printed += scnprintf(bf + printed, size - printed, "%s%s", printed ? "|" : "", "SYNC"); 257 else { 258 P_FLAG(DSYNC); 259 } 260 #else 261 P_FLAG(SYNC); 262 #endif 263 P_FLAG(TRUNC); 264 P_FLAG(WRONLY); 265 #undef P_FLAG 266 267 if (flags) 268 printed += scnprintf(bf + printed, size - printed, "%s%#x", printed ? "|" : "", flags); 269 270 return printed; 271 } 272 273 #define SCA_OPEN_FLAGS syscall_arg__scnprintf_open_flags 274 275 static struct syscall_fmt { 276 const char *name; 277 const char *alias; 278 size_t (*arg_scnprintf[6])(char *bf, size_t size, unsigned long arg, u8 arg_idx, u8 *arg_mask); 279 bool errmsg; 280 bool timeout; 281 bool hexret; 282 } syscall_fmts[] = { 283 { .name = "access", .errmsg = true, }, 284 { .name = "arch_prctl", .errmsg = true, .alias = "prctl", }, 285 { .name = "brk", .hexret = true, 286 .arg_scnprintf = { [0] = SCA_HEX, /* brk */ }, }, 287 { .name = "mmap", .hexret = true, }, 288 { .name = "connect", .errmsg = true, }, 289 { .name = "fstat", .errmsg = true, .alias = "newfstat", }, 290 { .name = "fstatat", .errmsg = true, .alias = "newfstatat", }, 291 { .name = "futex", .errmsg = true, 292 .arg_scnprintf = { [1] = SCA_FUTEX_OP, /* op */ }, }, 293 { .name = "ioctl", .errmsg = true, 294 .arg_scnprintf = { [2] = SCA_HEX, /* arg */ }, }, 295 { .name = "lseek", .errmsg = true, 296 .arg_scnprintf = { [2] = SCA_WHENCE, /* whence */ }, }, 297 { .name = "lstat", .errmsg = true, .alias = "newlstat", }, 298 { .name = "madvise", .errmsg = true, 299 .arg_scnprintf = { [0] = SCA_HEX, /* start */ 300 [2] = SCA_MADV_BHV, /* behavior */ }, }, 301 { .name = "mmap", .hexret = true, 302 .arg_scnprintf = { [0] = SCA_HEX, /* addr */ 303 [2] = SCA_MMAP_PROT, /* prot */ 304 [3] = SCA_MMAP_FLAGS, /* flags */ }, }, 305 { .name = "mprotect", .errmsg = true, 306 .arg_scnprintf = { [0] = SCA_HEX, /* start */ 307 [2] = SCA_MMAP_PROT, /* prot */ }, }, 308 { .name = "mremap", .hexret = true, 309 .arg_scnprintf = { [0] = SCA_HEX, /* addr */ 310 [4] = SCA_HEX, /* new_addr */ }, }, 311 { .name = "munmap", .errmsg = true, 312 .arg_scnprintf = { [0] = SCA_HEX, /* addr */ }, }, 313 { .name = "open", .errmsg = true, 314 .arg_scnprintf = { [1] = SCA_OPEN_FLAGS, /* flags */ }, }, 315 { .name = "open_by_handle_at", .errmsg = true, 316 .arg_scnprintf = { [2] = SCA_OPEN_FLAGS, /* flags */ }, }, 317 { .name = "openat", .errmsg = true, 318 .arg_scnprintf = { [2] = SCA_OPEN_FLAGS, /* flags */ }, }, 319 { .name = "poll", .errmsg = true, .timeout = true, }, 320 { .name = "ppoll", .errmsg = true, .timeout = true, }, 321 { .name = "pread", .errmsg = true, .alias = "pread64", }, 322 { .name = "pwrite", .errmsg = true, .alias = "pwrite64", }, 323 { .name = "read", .errmsg = true, }, 324 { .name = "recvfrom", .errmsg = true, }, 325 { .name = "select", .errmsg = true, .timeout = true, }, 326 { .name = "socket", .errmsg = true, }, 327 { .name = "stat", .errmsg = true, .alias = "newstat", }, 328 { .name = "uname", .errmsg = true, .alias = "newuname", }, 329 }; 330 331 static int syscall_fmt__cmp(const void *name, const void *fmtp) 332 { 333 const struct syscall_fmt *fmt = fmtp; 334 return strcmp(name, fmt->name); 335 } 336 337 static struct syscall_fmt *syscall_fmt__find(const char *name) 338 { 339 const int nmemb = ARRAY_SIZE(syscall_fmts); 340 return bsearch(name, syscall_fmts, nmemb, sizeof(struct syscall_fmt), syscall_fmt__cmp); 341 } 342 343 struct syscall { 344 struct event_format *tp_format; 345 const char *name; 346 bool filtered; 347 struct syscall_fmt *fmt; 348 size_t (**arg_scnprintf)(char *bf, size_t size, 349 unsigned long arg, u8 arg_idx, u8 *args_mask); 350 }; 351 352 static size_t fprintf_duration(unsigned long t, FILE *fp) 353 { 354 double duration = (double)t / NSEC_PER_MSEC; 355 size_t printed = fprintf(fp, "("); 356 357 if (duration >= 1.0) 358 printed += color_fprintf(fp, PERF_COLOR_RED, "%6.3f ms", duration); 359 else if (duration >= 0.01) 360 printed += color_fprintf(fp, PERF_COLOR_YELLOW, "%6.3f ms", duration); 361 else 362 printed += color_fprintf(fp, PERF_COLOR_NORMAL, "%6.3f ms", duration); 363 return printed + fprintf(fp, "): "); 364 } 365 366 struct thread_trace { 367 u64 entry_time; 368 u64 exit_time; 369 bool entry_pending; 370 unsigned long nr_events; 371 char *entry_str; 372 double runtime_ms; 373 }; 374 375 static struct thread_trace *thread_trace__new(void) 376 { 377 return zalloc(sizeof(struct thread_trace)); 378 } 379 380 static struct thread_trace *thread__trace(struct thread *thread, FILE *fp) 381 { 382 struct thread_trace *ttrace; 383 384 if (thread == NULL) 385 goto fail; 386 387 if (thread->priv == NULL) 388 thread->priv = thread_trace__new(); 389 390 if (thread->priv == NULL) 391 goto fail; 392 393 ttrace = thread->priv; 394 ++ttrace->nr_events; 395 396 return ttrace; 397 fail: 398 color_fprintf(fp, PERF_COLOR_RED, 399 "WARNING: not enough memory, dropping samples!\n"); 400 return NULL; 401 } 402 403 struct trace { 404 struct perf_tool tool; 405 int audit_machine; 406 struct { 407 int max; 408 struct syscall *table; 409 } syscalls; 410 struct perf_record_opts opts; 411 struct machine host; 412 u64 base_time; 413 FILE *output; 414 unsigned long nr_events; 415 struct strlist *ev_qualifier; 416 bool not_ev_qualifier; 417 struct intlist *tid_list; 418 struct intlist *pid_list; 419 bool sched; 420 bool multiple_threads; 421 double duration_filter; 422 double runtime_ms; 423 }; 424 425 static bool trace__filter_duration(struct trace *trace, double t) 426 { 427 return t < (trace->duration_filter * NSEC_PER_MSEC); 428 } 429 430 static size_t trace__fprintf_tstamp(struct trace *trace, u64 tstamp, FILE *fp) 431 { 432 double ts = (double)(tstamp - trace->base_time) / NSEC_PER_MSEC; 433 434 return fprintf(fp, "%10.3f ", ts); 435 } 436 437 static bool done = false; 438 439 static void sig_handler(int sig __maybe_unused) 440 { 441 done = true; 442 } 443 444 static size_t trace__fprintf_entry_head(struct trace *trace, struct thread *thread, 445 u64 duration, u64 tstamp, FILE *fp) 446 { 447 size_t printed = trace__fprintf_tstamp(trace, tstamp, fp); 448 printed += fprintf_duration(duration, fp); 449 450 if (trace->multiple_threads) 451 printed += fprintf(fp, "%d ", thread->tid); 452 453 return printed; 454 } 455 456 static int trace__process_event(struct trace *trace, struct machine *machine, 457 union perf_event *event) 458 { 459 int ret = 0; 460 461 switch (event->header.type) { 462 case PERF_RECORD_LOST: 463 color_fprintf(trace->output, PERF_COLOR_RED, 464 "LOST %" PRIu64 " events!\n", event->lost.lost); 465 ret = machine__process_lost_event(machine, event); 466 default: 467 ret = machine__process_event(machine, event); 468 break; 469 } 470 471 return ret; 472 } 473 474 static int trace__tool_process(struct perf_tool *tool, 475 union perf_event *event, 476 struct perf_sample *sample __maybe_unused, 477 struct machine *machine) 478 { 479 struct trace *trace = container_of(tool, struct trace, tool); 480 return trace__process_event(trace, machine, event); 481 } 482 483 static int trace__symbols_init(struct trace *trace, struct perf_evlist *evlist) 484 { 485 int err = symbol__init(); 486 487 if (err) 488 return err; 489 490 machine__init(&trace->host, "", HOST_KERNEL_ID); 491 machine__create_kernel_maps(&trace->host); 492 493 if (perf_target__has_task(&trace->opts.target)) { 494 err = perf_event__synthesize_thread_map(&trace->tool, evlist->threads, 495 trace__tool_process, 496 &trace->host); 497 } else { 498 err = perf_event__synthesize_threads(&trace->tool, trace__tool_process, 499 &trace->host); 500 } 501 502 if (err) 503 symbol__exit(); 504 505 return err; 506 } 507 508 static int syscall__set_arg_fmts(struct syscall *sc) 509 { 510 struct format_field *field; 511 int idx = 0; 512 513 sc->arg_scnprintf = calloc(sc->tp_format->format.nr_fields - 1, sizeof(void *)); 514 if (sc->arg_scnprintf == NULL) 515 return -1; 516 517 for (field = sc->tp_format->format.fields->next; field; field = field->next) { 518 if (sc->fmt && sc->fmt->arg_scnprintf[idx]) 519 sc->arg_scnprintf[idx] = sc->fmt->arg_scnprintf[idx]; 520 else if (field->flags & FIELD_IS_POINTER) 521 sc->arg_scnprintf[idx] = syscall_arg__scnprintf_hex; 522 ++idx; 523 } 524 525 return 0; 526 } 527 528 static int trace__read_syscall_info(struct trace *trace, int id) 529 { 530 char tp_name[128]; 531 struct syscall *sc; 532 const char *name = audit_syscall_to_name(id, trace->audit_machine); 533 534 if (name == NULL) 535 return -1; 536 537 if (id > trace->syscalls.max) { 538 struct syscall *nsyscalls = realloc(trace->syscalls.table, (id + 1) * sizeof(*sc)); 539 540 if (nsyscalls == NULL) 541 return -1; 542 543 if (trace->syscalls.max != -1) { 544 memset(nsyscalls + trace->syscalls.max + 1, 0, 545 (id - trace->syscalls.max) * sizeof(*sc)); 546 } else { 547 memset(nsyscalls, 0, (id + 1) * sizeof(*sc)); 548 } 549 550 trace->syscalls.table = nsyscalls; 551 trace->syscalls.max = id; 552 } 553 554 sc = trace->syscalls.table + id; 555 sc->name = name; 556 557 if (trace->ev_qualifier) { 558 bool in = strlist__find(trace->ev_qualifier, name) != NULL; 559 560 if (!(in ^ trace->not_ev_qualifier)) { 561 sc->filtered = true; 562 /* 563 * No need to do read tracepoint information since this will be 564 * filtered out. 565 */ 566 return 0; 567 } 568 } 569 570 sc->fmt = syscall_fmt__find(sc->name); 571 572 snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->name); 573 sc->tp_format = event_format__new("syscalls", tp_name); 574 575 if (sc->tp_format == NULL && sc->fmt && sc->fmt->alias) { 576 snprintf(tp_name, sizeof(tp_name), "sys_enter_%s", sc->fmt->alias); 577 sc->tp_format = event_format__new("syscalls", tp_name); 578 } 579 580 if (sc->tp_format == NULL) 581 return -1; 582 583 return syscall__set_arg_fmts(sc); 584 } 585 586 static size_t syscall__scnprintf_args(struct syscall *sc, char *bf, size_t size, 587 unsigned long *args) 588 { 589 int i = 0; 590 size_t printed = 0; 591 592 if (sc->tp_format != NULL) { 593 struct format_field *field; 594 u8 mask = 0, bit = 1; 595 596 for (field = sc->tp_format->format.fields->next; field; 597 field = field->next, ++i, bit <<= 1) { 598 if (mask & bit) 599 continue; 600 601 printed += scnprintf(bf + printed, size - printed, 602 "%s%s: ", printed ? ", " : "", field->name); 603 604 if (sc->arg_scnprintf && sc->arg_scnprintf[i]) { 605 printed += sc->arg_scnprintf[i](bf + printed, size - printed, 606 args[i], i, &mask); 607 } else { 608 printed += scnprintf(bf + printed, size - printed, 609 "%ld", args[i]); 610 } 611 } 612 } else { 613 while (i < 6) { 614 printed += scnprintf(bf + printed, size - printed, 615 "%sarg%d: %ld", 616 printed ? ", " : "", i, args[i]); 617 ++i; 618 } 619 } 620 621 return printed; 622 } 623 624 typedef int (*tracepoint_handler)(struct trace *trace, struct perf_evsel *evsel, 625 struct perf_sample *sample); 626 627 static struct syscall *trace__syscall_info(struct trace *trace, 628 struct perf_evsel *evsel, 629 struct perf_sample *sample) 630 { 631 int id = perf_evsel__intval(evsel, sample, "id"); 632 633 if (id < 0) { 634 635 /* 636 * XXX: Noticed on x86_64, reproduced as far back as 3.0.36, haven't tried 637 * before that, leaving at a higher verbosity level till that is 638 * explained. Reproduced with plain ftrace with: 639 * 640 * echo 1 > /t/events/raw_syscalls/sys_exit/enable 641 * grep "NR -1 " /t/trace_pipe 642 * 643 * After generating some load on the machine. 644 */ 645 if (verbose > 1) { 646 static u64 n; 647 fprintf(trace->output, "Invalid syscall %d id, skipping (%s, %" PRIu64 ") ...\n", 648 id, perf_evsel__name(evsel), ++n); 649 } 650 return NULL; 651 } 652 653 if ((id > trace->syscalls.max || trace->syscalls.table[id].name == NULL) && 654 trace__read_syscall_info(trace, id)) 655 goto out_cant_read; 656 657 if ((id > trace->syscalls.max || trace->syscalls.table[id].name == NULL)) 658 goto out_cant_read; 659 660 return &trace->syscalls.table[id]; 661 662 out_cant_read: 663 if (verbose) { 664 fprintf(trace->output, "Problems reading syscall %d", id); 665 if (id <= trace->syscalls.max && trace->syscalls.table[id].name != NULL) 666 fprintf(trace->output, "(%s)", trace->syscalls.table[id].name); 667 fputs(" information\n", trace->output); 668 } 669 return NULL; 670 } 671 672 static int trace__sys_enter(struct trace *trace, struct perf_evsel *evsel, 673 struct perf_sample *sample) 674 { 675 char *msg; 676 void *args; 677 size_t printed = 0; 678 struct thread *thread; 679 struct syscall *sc = trace__syscall_info(trace, evsel, sample); 680 struct thread_trace *ttrace; 681 682 if (sc == NULL) 683 return -1; 684 685 if (sc->filtered) 686 return 0; 687 688 thread = machine__findnew_thread(&trace->host, sample->pid, 689 sample->tid); 690 ttrace = thread__trace(thread, trace->output); 691 if (ttrace == NULL) 692 return -1; 693 694 args = perf_evsel__rawptr(evsel, sample, "args"); 695 if (args == NULL) { 696 fprintf(trace->output, "Problems reading syscall arguments\n"); 697 return -1; 698 } 699 700 ttrace = thread->priv; 701 702 if (ttrace->entry_str == NULL) { 703 ttrace->entry_str = malloc(1024); 704 if (!ttrace->entry_str) 705 return -1; 706 } 707 708 ttrace->entry_time = sample->time; 709 msg = ttrace->entry_str; 710 printed += scnprintf(msg + printed, 1024 - printed, "%s(", sc->name); 711 712 printed += syscall__scnprintf_args(sc, msg + printed, 1024 - printed, args); 713 714 if (!strcmp(sc->name, "exit_group") || !strcmp(sc->name, "exit")) { 715 if (!trace->duration_filter) { 716 trace__fprintf_entry_head(trace, thread, 1, sample->time, trace->output); 717 fprintf(trace->output, "%-70s\n", ttrace->entry_str); 718 } 719 } else 720 ttrace->entry_pending = true; 721 722 return 0; 723 } 724 725 static int trace__sys_exit(struct trace *trace, struct perf_evsel *evsel, 726 struct perf_sample *sample) 727 { 728 int ret; 729 u64 duration = 0; 730 struct thread *thread; 731 struct syscall *sc = trace__syscall_info(trace, evsel, sample); 732 struct thread_trace *ttrace; 733 734 if (sc == NULL) 735 return -1; 736 737 if (sc->filtered) 738 return 0; 739 740 thread = machine__findnew_thread(&trace->host, sample->pid, 741 sample->tid); 742 ttrace = thread__trace(thread, trace->output); 743 if (ttrace == NULL) 744 return -1; 745 746 ret = perf_evsel__intval(evsel, sample, "ret"); 747 748 ttrace = thread->priv; 749 750 ttrace->exit_time = sample->time; 751 752 if (ttrace->entry_time) { 753 duration = sample->time - ttrace->entry_time; 754 if (trace__filter_duration(trace, duration)) 755 goto out; 756 } else if (trace->duration_filter) 757 goto out; 758 759 trace__fprintf_entry_head(trace, thread, duration, sample->time, trace->output); 760 761 if (ttrace->entry_pending) { 762 fprintf(trace->output, "%-70s", ttrace->entry_str); 763 } else { 764 fprintf(trace->output, " ... ["); 765 color_fprintf(trace->output, PERF_COLOR_YELLOW, "continued"); 766 fprintf(trace->output, "]: %s()", sc->name); 767 } 768 769 if (sc->fmt == NULL) { 770 signed_print: 771 fprintf(trace->output, ") = %d", ret); 772 } else if (ret < 0 && sc->fmt->errmsg) { 773 char bf[256]; 774 const char *emsg = strerror_r(-ret, bf, sizeof(bf)), 775 *e = audit_errno_to_name(-ret); 776 777 fprintf(trace->output, ") = -1 %s %s", e, emsg); 778 } else if (ret == 0 && sc->fmt->timeout) 779 fprintf(trace->output, ") = 0 Timeout"); 780 else if (sc->fmt->hexret) 781 fprintf(trace->output, ") = %#x", ret); 782 else 783 goto signed_print; 784 785 fputc('\n', trace->output); 786 out: 787 ttrace->entry_pending = false; 788 789 return 0; 790 } 791 792 static int trace__sched_stat_runtime(struct trace *trace, struct perf_evsel *evsel, 793 struct perf_sample *sample) 794 { 795 u64 runtime = perf_evsel__intval(evsel, sample, "runtime"); 796 double runtime_ms = (double)runtime / NSEC_PER_MSEC; 797 struct thread *thread = machine__findnew_thread(&trace->host, 798 sample->pid, 799 sample->tid); 800 struct thread_trace *ttrace = thread__trace(thread, trace->output); 801 802 if (ttrace == NULL) 803 goto out_dump; 804 805 ttrace->runtime_ms += runtime_ms; 806 trace->runtime_ms += runtime_ms; 807 return 0; 808 809 out_dump: 810 fprintf(trace->output, "%s: comm=%s,pid=%u,runtime=%" PRIu64 ",vruntime=%" PRIu64 ")\n", 811 evsel->name, 812 perf_evsel__strval(evsel, sample, "comm"), 813 (pid_t)perf_evsel__intval(evsel, sample, "pid"), 814 runtime, 815 perf_evsel__intval(evsel, sample, "vruntime")); 816 return 0; 817 } 818 819 static bool skip_sample(struct trace *trace, struct perf_sample *sample) 820 { 821 if ((trace->pid_list && intlist__find(trace->pid_list, sample->pid)) || 822 (trace->tid_list && intlist__find(trace->tid_list, sample->tid))) 823 return false; 824 825 if (trace->pid_list || trace->tid_list) 826 return true; 827 828 return false; 829 } 830 831 static int trace__process_sample(struct perf_tool *tool, 832 union perf_event *event __maybe_unused, 833 struct perf_sample *sample, 834 struct perf_evsel *evsel, 835 struct machine *machine __maybe_unused) 836 { 837 struct trace *trace = container_of(tool, struct trace, tool); 838 int err = 0; 839 840 tracepoint_handler handler = evsel->handler.func; 841 842 if (skip_sample(trace, sample)) 843 return 0; 844 845 if (trace->base_time == 0) 846 trace->base_time = sample->time; 847 848 if (handler) 849 handler(trace, evsel, sample); 850 851 return err; 852 } 853 854 static bool 855 perf_session__has_tp(struct perf_session *session, const char *name) 856 { 857 struct perf_evsel *evsel; 858 859 evsel = perf_evlist__find_tracepoint_by_name(session->evlist, name); 860 861 return evsel != NULL; 862 } 863 864 static int parse_target_str(struct trace *trace) 865 { 866 if (trace->opts.target.pid) { 867 trace->pid_list = intlist__new(trace->opts.target.pid); 868 if (trace->pid_list == NULL) { 869 pr_err("Error parsing process id string\n"); 870 return -EINVAL; 871 } 872 } 873 874 if (trace->opts.target.tid) { 875 trace->tid_list = intlist__new(trace->opts.target.tid); 876 if (trace->tid_list == NULL) { 877 pr_err("Error parsing thread id string\n"); 878 return -EINVAL; 879 } 880 } 881 882 return 0; 883 } 884 885 static int trace__run(struct trace *trace, int argc, const char **argv) 886 { 887 struct perf_evlist *evlist = perf_evlist__new(); 888 struct perf_evsel *evsel; 889 int err = -1, i; 890 unsigned long before; 891 const bool forks = argc > 0; 892 893 if (evlist == NULL) { 894 fprintf(trace->output, "Not enough memory to run!\n"); 895 goto out; 896 } 897 898 if (perf_evlist__add_newtp(evlist, "raw_syscalls", "sys_enter", trace__sys_enter) || 899 perf_evlist__add_newtp(evlist, "raw_syscalls", "sys_exit", trace__sys_exit)) { 900 fprintf(trace->output, "Couldn't read the raw_syscalls tracepoints information!\n"); 901 goto out_delete_evlist; 902 } 903 904 if (trace->sched && 905 perf_evlist__add_newtp(evlist, "sched", "sched_stat_runtime", 906 trace__sched_stat_runtime)) { 907 fprintf(trace->output, "Couldn't read the sched_stat_runtime tracepoint information!\n"); 908 goto out_delete_evlist; 909 } 910 911 err = perf_evlist__create_maps(evlist, &trace->opts.target); 912 if (err < 0) { 913 fprintf(trace->output, "Problems parsing the target to trace, check your options!\n"); 914 goto out_delete_evlist; 915 } 916 917 err = trace__symbols_init(trace, evlist); 918 if (err < 0) { 919 fprintf(trace->output, "Problems initializing symbol libraries!\n"); 920 goto out_delete_maps; 921 } 922 923 perf_evlist__config(evlist, &trace->opts); 924 925 signal(SIGCHLD, sig_handler); 926 signal(SIGINT, sig_handler); 927 928 if (forks) { 929 err = perf_evlist__prepare_workload(evlist, &trace->opts.target, 930 argv, false, false); 931 if (err < 0) { 932 fprintf(trace->output, "Couldn't run the workload!\n"); 933 goto out_delete_maps; 934 } 935 } 936 937 err = perf_evlist__open(evlist); 938 if (err < 0) { 939 fprintf(trace->output, "Couldn't create the events: %s\n", strerror(errno)); 940 goto out_delete_maps; 941 } 942 943 err = perf_evlist__mmap(evlist, UINT_MAX, false); 944 if (err < 0) { 945 fprintf(trace->output, "Couldn't mmap the events: %s\n", strerror(errno)); 946 goto out_close_evlist; 947 } 948 949 perf_evlist__enable(evlist); 950 951 if (forks) 952 perf_evlist__start_workload(evlist); 953 954 trace->multiple_threads = evlist->threads->map[0] == -1 || evlist->threads->nr > 1; 955 again: 956 before = trace->nr_events; 957 958 for (i = 0; i < evlist->nr_mmaps; i++) { 959 union perf_event *event; 960 961 while ((event = perf_evlist__mmap_read(evlist, i)) != NULL) { 962 const u32 type = event->header.type; 963 tracepoint_handler handler; 964 struct perf_sample sample; 965 966 ++trace->nr_events; 967 968 err = perf_evlist__parse_sample(evlist, event, &sample); 969 if (err) { 970 fprintf(trace->output, "Can't parse sample, err = %d, skipping...\n", err); 971 continue; 972 } 973 974 if (trace->base_time == 0) 975 trace->base_time = sample.time; 976 977 if (type != PERF_RECORD_SAMPLE) { 978 trace__process_event(trace, &trace->host, event); 979 continue; 980 } 981 982 evsel = perf_evlist__id2evsel(evlist, sample.id); 983 if (evsel == NULL) { 984 fprintf(trace->output, "Unknown tp ID %" PRIu64 ", skipping...\n", sample.id); 985 continue; 986 } 987 988 if (sample.raw_data == NULL) { 989 fprintf(trace->output, "%s sample with no payload for tid: %d, cpu %d, raw_size=%d, skipping...\n", 990 perf_evsel__name(evsel), sample.tid, 991 sample.cpu, sample.raw_size); 992 continue; 993 } 994 995 handler = evsel->handler.func; 996 handler(trace, evsel, &sample); 997 } 998 } 999 1000 if (trace->nr_events == before) { 1001 if (done) 1002 goto out_unmap_evlist; 1003 1004 poll(evlist->pollfd, evlist->nr_fds, -1); 1005 } 1006 1007 if (done) 1008 perf_evlist__disable(evlist); 1009 1010 goto again; 1011 1012 out_unmap_evlist: 1013 perf_evlist__munmap(evlist); 1014 out_close_evlist: 1015 perf_evlist__close(evlist); 1016 out_delete_maps: 1017 perf_evlist__delete_maps(evlist); 1018 out_delete_evlist: 1019 perf_evlist__delete(evlist); 1020 out: 1021 return err; 1022 } 1023 1024 static int trace__replay(struct trace *trace) 1025 { 1026 const struct perf_evsel_str_handler handlers[] = { 1027 { "raw_syscalls:sys_enter", trace__sys_enter, }, 1028 { "raw_syscalls:sys_exit", trace__sys_exit, }, 1029 }; 1030 1031 struct perf_session *session; 1032 int err = -1; 1033 1034 trace->tool.sample = trace__process_sample; 1035 trace->tool.mmap = perf_event__process_mmap; 1036 trace->tool.comm = perf_event__process_comm; 1037 trace->tool.exit = perf_event__process_exit; 1038 trace->tool.fork = perf_event__process_fork; 1039 trace->tool.attr = perf_event__process_attr; 1040 trace->tool.tracing_data = perf_event__process_tracing_data; 1041 trace->tool.build_id = perf_event__process_build_id; 1042 1043 trace->tool.ordered_samples = true; 1044 trace->tool.ordering_requires_timestamps = true; 1045 1046 /* add tid to output */ 1047 trace->multiple_threads = true; 1048 1049 if (symbol__init() < 0) 1050 return -1; 1051 1052 session = perf_session__new(input_name, O_RDONLY, 0, false, 1053 &trace->tool); 1054 if (session == NULL) 1055 return -ENOMEM; 1056 1057 err = perf_session__set_tracepoints_handlers(session, handlers); 1058 if (err) 1059 goto out; 1060 1061 if (!perf_session__has_tp(session, "raw_syscalls:sys_enter")) { 1062 pr_err("Data file does not have raw_syscalls:sys_enter events\n"); 1063 goto out; 1064 } 1065 1066 if (!perf_session__has_tp(session, "raw_syscalls:sys_exit")) { 1067 pr_err("Data file does not have raw_syscalls:sys_exit events\n"); 1068 goto out; 1069 } 1070 1071 err = parse_target_str(trace); 1072 if (err != 0) 1073 goto out; 1074 1075 setup_pager(); 1076 1077 err = perf_session__process_events(session, &trace->tool); 1078 if (err) 1079 pr_err("Failed to process events, error %d", err); 1080 1081 out: 1082 perf_session__delete(session); 1083 1084 return err; 1085 } 1086 1087 static size_t trace__fprintf_threads_header(FILE *fp) 1088 { 1089 size_t printed; 1090 1091 printed = fprintf(fp, "\n _____________________________________________________________________\n"); 1092 printed += fprintf(fp," __) Summary of events (__\n\n"); 1093 printed += fprintf(fp," [ task - pid ] [ events ] [ ratio ] [ runtime ]\n"); 1094 printed += fprintf(fp," _____________________________________________________________________\n\n"); 1095 1096 return printed; 1097 } 1098 1099 static size_t trace__fprintf_thread_summary(struct trace *trace, FILE *fp) 1100 { 1101 size_t printed = trace__fprintf_threads_header(fp); 1102 struct rb_node *nd; 1103 1104 for (nd = rb_first(&trace->host.threads); nd; nd = rb_next(nd)) { 1105 struct thread *thread = rb_entry(nd, struct thread, rb_node); 1106 struct thread_trace *ttrace = thread->priv; 1107 const char *color; 1108 double ratio; 1109 1110 if (ttrace == NULL) 1111 continue; 1112 1113 ratio = (double)ttrace->nr_events / trace->nr_events * 100.0; 1114 1115 color = PERF_COLOR_NORMAL; 1116 if (ratio > 50.0) 1117 color = PERF_COLOR_RED; 1118 else if (ratio > 25.0) 1119 color = PERF_COLOR_GREEN; 1120 else if (ratio > 5.0) 1121 color = PERF_COLOR_YELLOW; 1122 1123 printed += color_fprintf(fp, color, "%20s", thread->comm); 1124 printed += fprintf(fp, " - %-5d :%11lu [", thread->tid, ttrace->nr_events); 1125 printed += color_fprintf(fp, color, "%5.1f%%", ratio); 1126 printed += fprintf(fp, " ] %10.3f ms\n", ttrace->runtime_ms); 1127 } 1128 1129 return printed; 1130 } 1131 1132 static int trace__set_duration(const struct option *opt, const char *str, 1133 int unset __maybe_unused) 1134 { 1135 struct trace *trace = opt->value; 1136 1137 trace->duration_filter = atof(str); 1138 return 0; 1139 } 1140 1141 static int trace__open_output(struct trace *trace, const char *filename) 1142 { 1143 struct stat st; 1144 1145 if (!stat(filename, &st) && st.st_size) { 1146 char oldname[PATH_MAX]; 1147 1148 scnprintf(oldname, sizeof(oldname), "%s.old", filename); 1149 unlink(oldname); 1150 rename(filename, oldname); 1151 } 1152 1153 trace->output = fopen(filename, "w"); 1154 1155 return trace->output == NULL ? -errno : 0; 1156 } 1157 1158 int cmd_trace(int argc, const char **argv, const char *prefix __maybe_unused) 1159 { 1160 const char * const trace_usage[] = { 1161 "perf trace [<options>] [<command>]", 1162 "perf trace [<options>] -- <command> [<options>]", 1163 NULL 1164 }; 1165 struct trace trace = { 1166 .audit_machine = audit_detect_machine(), 1167 .syscalls = { 1168 . max = -1, 1169 }, 1170 .opts = { 1171 .target = { 1172 .uid = UINT_MAX, 1173 .uses_mmap = true, 1174 }, 1175 .user_freq = UINT_MAX, 1176 .user_interval = ULLONG_MAX, 1177 .no_delay = true, 1178 .mmap_pages = 1024, 1179 }, 1180 .output = stdout, 1181 }; 1182 const char *output_name = NULL; 1183 const char *ev_qualifier_str = NULL; 1184 const struct option trace_options[] = { 1185 OPT_STRING('e', "expr", &ev_qualifier_str, "expr", 1186 "list of events to trace"), 1187 OPT_STRING('o', "output", &output_name, "file", "output file name"), 1188 OPT_STRING('i', "input", &input_name, "file", "Analyze events in file"), 1189 OPT_STRING('p', "pid", &trace.opts.target.pid, "pid", 1190 "trace events on existing process id"), 1191 OPT_STRING('t', "tid", &trace.opts.target.tid, "tid", 1192 "trace events on existing thread id"), 1193 OPT_BOOLEAN('a', "all-cpus", &trace.opts.target.system_wide, 1194 "system-wide collection from all CPUs"), 1195 OPT_STRING('C', "cpu", &trace.opts.target.cpu_list, "cpu", 1196 "list of cpus to monitor"), 1197 OPT_BOOLEAN(0, "no-inherit", &trace.opts.no_inherit, 1198 "child tasks do not inherit counters"), 1199 OPT_UINTEGER('m', "mmap-pages", &trace.opts.mmap_pages, 1200 "number of mmap data pages"), 1201 OPT_STRING('u', "uid", &trace.opts.target.uid_str, "user", 1202 "user to profile"), 1203 OPT_CALLBACK(0, "duration", &trace, "float", 1204 "show only events with duration > N.M ms", 1205 trace__set_duration), 1206 OPT_BOOLEAN(0, "sched", &trace.sched, "show blocking scheduler events"), 1207 OPT_INCR('v', "verbose", &verbose, "be more verbose"), 1208 OPT_END() 1209 }; 1210 int err; 1211 char bf[BUFSIZ]; 1212 1213 argc = parse_options(argc, argv, trace_options, trace_usage, 0); 1214 1215 if (output_name != NULL) { 1216 err = trace__open_output(&trace, output_name); 1217 if (err < 0) { 1218 perror("failed to create output file"); 1219 goto out; 1220 } 1221 } 1222 1223 if (ev_qualifier_str != NULL) { 1224 const char *s = ev_qualifier_str; 1225 1226 trace.not_ev_qualifier = *s == '!'; 1227 if (trace.not_ev_qualifier) 1228 ++s; 1229 trace.ev_qualifier = strlist__new(true, s); 1230 if (trace.ev_qualifier == NULL) { 1231 fputs("Not enough memory to parse event qualifier", 1232 trace.output); 1233 err = -ENOMEM; 1234 goto out_close; 1235 } 1236 } 1237 1238 err = perf_target__validate(&trace.opts.target); 1239 if (err) { 1240 perf_target__strerror(&trace.opts.target, err, bf, sizeof(bf)); 1241 fprintf(trace.output, "%s", bf); 1242 goto out_close; 1243 } 1244 1245 err = perf_target__parse_uid(&trace.opts.target); 1246 if (err) { 1247 perf_target__strerror(&trace.opts.target, err, bf, sizeof(bf)); 1248 fprintf(trace.output, "%s", bf); 1249 goto out_close; 1250 } 1251 1252 if (!argc && perf_target__none(&trace.opts.target)) 1253 trace.opts.target.system_wide = true; 1254 1255 if (input_name) 1256 err = trace__replay(&trace); 1257 else 1258 err = trace__run(&trace, argc, argv); 1259 1260 if (trace.sched && !err) 1261 trace__fprintf_thread_summary(&trace, trace.output); 1262 1263 out_close: 1264 if (output_name != NULL) 1265 fclose(trace.output); 1266 out: 1267 return err; 1268 } 1269