1 // SPDX-License-Identifier: GPL-2.0 2 #include <errno.h> 3 #include <inttypes.h> 4 #include <regex.h> 5 #include <stdlib.h> 6 #include <linux/mman.h> 7 #include <linux/time64.h> 8 #include "debug.h" 9 #include "dso.h" 10 #include "sort.h" 11 #include "hist.h" 12 #include "cacheline.h" 13 #include "comm.h" 14 #include "map.h" 15 #include "maps.h" 16 #include "symbol.h" 17 #include "map_symbol.h" 18 #include "branch.h" 19 #include "thread.h" 20 #include "evsel.h" 21 #include "evlist.h" 22 #include "srcline.h" 23 #include "strlist.h" 24 #include "strbuf.h" 25 #include <traceevent/event-parse.h> 26 #include "mem-events.h" 27 #include "annotate.h" 28 #include "event.h" 29 #include "time-utils.h" 30 #include "cgroup.h" 31 #include "machine.h" 32 #include <linux/kernel.h> 33 #include <linux/string.h> 34 35 regex_t parent_regex; 36 const char default_parent_pattern[] = "^sys_|^do_page_fault"; 37 const char *parent_pattern = default_parent_pattern; 38 const char *default_sort_order = "comm,dso,symbol"; 39 const char default_branch_sort_order[] = "comm,dso_from,symbol_from,symbol_to,cycles"; 40 const char default_mem_sort_order[] = "local_weight,mem,sym,dso,symbol_daddr,dso_daddr,snoop,tlb,locked,blocked,local_ins_lat,local_p_stage_cyc"; 41 const char default_top_sort_order[] = "dso,symbol"; 42 const char default_diff_sort_order[] = "dso,symbol"; 43 const char default_tracepoint_sort_order[] = "trace"; 44 const char *sort_order; 45 const char *field_order; 46 regex_t ignore_callees_regex; 47 int have_ignore_callees = 0; 48 enum sort_mode sort__mode = SORT_MODE__NORMAL; 49 static const char *const dynamic_headers[] = {"local_ins_lat", "ins_lat", "local_p_stage_cyc", "p_stage_cyc"}; 50 static const char *const arch_specific_sort_keys[] = {"local_p_stage_cyc", "p_stage_cyc"}; 51 52 /* 53 * Replaces all occurrences of a char used with the: 54 * 55 * -t, --field-separator 56 * 57 * option, that uses a special separator character and don't pad with spaces, 58 * replacing all occurrences of this separator in symbol names (and other 59 * output) with a '.' character, that thus it's the only non valid separator. 60 */ 61 static int repsep_snprintf(char *bf, size_t size, const char *fmt, ...) 62 { 63 int n; 64 va_list ap; 65 66 va_start(ap, fmt); 67 n = vsnprintf(bf, size, fmt, ap); 68 if (symbol_conf.field_sep && n > 0) { 69 char *sep = bf; 70 71 while (1) { 72 sep = strchr(sep, *symbol_conf.field_sep); 73 if (sep == NULL) 74 break; 75 *sep = '.'; 76 } 77 } 78 va_end(ap); 79 80 if (n >= (int)size) 81 return size - 1; 82 return n; 83 } 84 85 static int64_t cmp_null(const void *l, const void *r) 86 { 87 if (!l && !r) 88 return 0; 89 else if (!l) 90 return -1; 91 else 92 return 1; 93 } 94 95 /* --sort pid */ 96 97 static int64_t 98 sort__thread_cmp(struct hist_entry *left, struct hist_entry *right) 99 { 100 return right->thread->tid - left->thread->tid; 101 } 102 103 static int hist_entry__thread_snprintf(struct hist_entry *he, char *bf, 104 size_t size, unsigned int width) 105 { 106 const char *comm = thread__comm_str(he->thread); 107 108 width = max(7U, width) - 8; 109 return repsep_snprintf(bf, size, "%7d:%-*.*s", he->thread->tid, 110 width, width, comm ?: ""); 111 } 112 113 static int hist_entry__thread_filter(struct hist_entry *he, int type, const void *arg) 114 { 115 const struct thread *th = arg; 116 117 if (type != HIST_FILTER__THREAD) 118 return -1; 119 120 return th && he->thread != th; 121 } 122 123 struct sort_entry sort_thread = { 124 .se_header = " Pid:Command", 125 .se_cmp = sort__thread_cmp, 126 .se_snprintf = hist_entry__thread_snprintf, 127 .se_filter = hist_entry__thread_filter, 128 .se_width_idx = HISTC_THREAD, 129 }; 130 131 /* --sort comm */ 132 133 /* 134 * We can't use pointer comparison in functions below, 135 * because it gives different results based on pointer 136 * values, which could break some sorting assumptions. 137 */ 138 static int64_t 139 sort__comm_cmp(struct hist_entry *left, struct hist_entry *right) 140 { 141 return strcmp(comm__str(right->comm), comm__str(left->comm)); 142 } 143 144 static int64_t 145 sort__comm_collapse(struct hist_entry *left, struct hist_entry *right) 146 { 147 return strcmp(comm__str(right->comm), comm__str(left->comm)); 148 } 149 150 static int64_t 151 sort__comm_sort(struct hist_entry *left, struct hist_entry *right) 152 { 153 return strcmp(comm__str(right->comm), comm__str(left->comm)); 154 } 155 156 static int hist_entry__comm_snprintf(struct hist_entry *he, char *bf, 157 size_t size, unsigned int width) 158 { 159 return repsep_snprintf(bf, size, "%-*.*s", width, width, comm__str(he->comm)); 160 } 161 162 struct sort_entry sort_comm = { 163 .se_header = "Command", 164 .se_cmp = sort__comm_cmp, 165 .se_collapse = sort__comm_collapse, 166 .se_sort = sort__comm_sort, 167 .se_snprintf = hist_entry__comm_snprintf, 168 .se_filter = hist_entry__thread_filter, 169 .se_width_idx = HISTC_COMM, 170 }; 171 172 /* --sort dso */ 173 174 static int64_t _sort__dso_cmp(struct map *map_l, struct map *map_r) 175 { 176 struct dso *dso_l = map_l ? map_l->dso : NULL; 177 struct dso *dso_r = map_r ? map_r->dso : NULL; 178 const char *dso_name_l, *dso_name_r; 179 180 if (!dso_l || !dso_r) 181 return cmp_null(dso_r, dso_l); 182 183 if (verbose > 0) { 184 dso_name_l = dso_l->long_name; 185 dso_name_r = dso_r->long_name; 186 } else { 187 dso_name_l = dso_l->short_name; 188 dso_name_r = dso_r->short_name; 189 } 190 191 return strcmp(dso_name_l, dso_name_r); 192 } 193 194 static int64_t 195 sort__dso_cmp(struct hist_entry *left, struct hist_entry *right) 196 { 197 return _sort__dso_cmp(right->ms.map, left->ms.map); 198 } 199 200 static int _hist_entry__dso_snprintf(struct map *map, char *bf, 201 size_t size, unsigned int width) 202 { 203 if (map && map->dso) { 204 const char *dso_name = verbose > 0 ? map->dso->long_name : 205 map->dso->short_name; 206 return repsep_snprintf(bf, size, "%-*.*s", width, width, dso_name); 207 } 208 209 return repsep_snprintf(bf, size, "%-*.*s", width, width, "[unknown]"); 210 } 211 212 static int hist_entry__dso_snprintf(struct hist_entry *he, char *bf, 213 size_t size, unsigned int width) 214 { 215 return _hist_entry__dso_snprintf(he->ms.map, bf, size, width); 216 } 217 218 static int hist_entry__dso_filter(struct hist_entry *he, int type, const void *arg) 219 { 220 const struct dso *dso = arg; 221 222 if (type != HIST_FILTER__DSO) 223 return -1; 224 225 return dso && (!he->ms.map || he->ms.map->dso != dso); 226 } 227 228 struct sort_entry sort_dso = { 229 .se_header = "Shared Object", 230 .se_cmp = sort__dso_cmp, 231 .se_snprintf = hist_entry__dso_snprintf, 232 .se_filter = hist_entry__dso_filter, 233 .se_width_idx = HISTC_DSO, 234 }; 235 236 /* --sort symbol */ 237 238 static int64_t _sort__addr_cmp(u64 left_ip, u64 right_ip) 239 { 240 return (int64_t)(right_ip - left_ip); 241 } 242 243 int64_t _sort__sym_cmp(struct symbol *sym_l, struct symbol *sym_r) 244 { 245 if (!sym_l || !sym_r) 246 return cmp_null(sym_l, sym_r); 247 248 if (sym_l == sym_r) 249 return 0; 250 251 if (sym_l->inlined || sym_r->inlined) { 252 int ret = strcmp(sym_l->name, sym_r->name); 253 254 if (ret) 255 return ret; 256 if ((sym_l->start <= sym_r->end) && (sym_l->end >= sym_r->start)) 257 return 0; 258 } 259 260 if (sym_l->start != sym_r->start) 261 return (int64_t)(sym_r->start - sym_l->start); 262 263 return (int64_t)(sym_r->end - sym_l->end); 264 } 265 266 static int64_t 267 sort__sym_cmp(struct hist_entry *left, struct hist_entry *right) 268 { 269 int64_t ret; 270 271 if (!left->ms.sym && !right->ms.sym) 272 return _sort__addr_cmp(left->ip, right->ip); 273 274 /* 275 * comparing symbol address alone is not enough since it's a 276 * relative address within a dso. 277 */ 278 if (!hists__has(left->hists, dso) || hists__has(right->hists, dso)) { 279 ret = sort__dso_cmp(left, right); 280 if (ret != 0) 281 return ret; 282 } 283 284 return _sort__sym_cmp(left->ms.sym, right->ms.sym); 285 } 286 287 static int64_t 288 sort__sym_sort(struct hist_entry *left, struct hist_entry *right) 289 { 290 if (!left->ms.sym || !right->ms.sym) 291 return cmp_null(left->ms.sym, right->ms.sym); 292 293 return strcmp(right->ms.sym->name, left->ms.sym->name); 294 } 295 296 static int _hist_entry__sym_snprintf(struct map_symbol *ms, 297 u64 ip, char level, char *bf, size_t size, 298 unsigned int width) 299 { 300 struct symbol *sym = ms->sym; 301 struct map *map = ms->map; 302 size_t ret = 0; 303 304 if (verbose > 0) { 305 char o = map ? dso__symtab_origin(map->dso) : '!'; 306 u64 rip = ip; 307 308 if (map && map->dso && map->dso->kernel 309 && map->dso->adjust_symbols) 310 rip = map->unmap_ip(map, ip); 311 312 ret += repsep_snprintf(bf, size, "%-#*llx %c ", 313 BITS_PER_LONG / 4 + 2, rip, o); 314 } 315 316 ret += repsep_snprintf(bf + ret, size - ret, "[%c] ", level); 317 if (sym && map) { 318 if (sym->type == STT_OBJECT) { 319 ret += repsep_snprintf(bf + ret, size - ret, "%s", sym->name); 320 ret += repsep_snprintf(bf + ret, size - ret, "+0x%llx", 321 ip - map->unmap_ip(map, sym->start)); 322 } else { 323 ret += repsep_snprintf(bf + ret, size - ret, "%.*s", 324 width - ret, 325 sym->name); 326 if (sym->inlined) 327 ret += repsep_snprintf(bf + ret, size - ret, 328 " (inlined)"); 329 } 330 } else { 331 size_t len = BITS_PER_LONG / 4; 332 ret += repsep_snprintf(bf + ret, size - ret, "%-#.*llx", 333 len, ip); 334 } 335 336 return ret; 337 } 338 339 int hist_entry__sym_snprintf(struct hist_entry *he, char *bf, size_t size, unsigned int width) 340 { 341 return _hist_entry__sym_snprintf(&he->ms, he->ip, 342 he->level, bf, size, width); 343 } 344 345 static int hist_entry__sym_filter(struct hist_entry *he, int type, const void *arg) 346 { 347 const char *sym = arg; 348 349 if (type != HIST_FILTER__SYMBOL) 350 return -1; 351 352 return sym && (!he->ms.sym || !strstr(he->ms.sym->name, sym)); 353 } 354 355 struct sort_entry sort_sym = { 356 .se_header = "Symbol", 357 .se_cmp = sort__sym_cmp, 358 .se_sort = sort__sym_sort, 359 .se_snprintf = hist_entry__sym_snprintf, 360 .se_filter = hist_entry__sym_filter, 361 .se_width_idx = HISTC_SYMBOL, 362 }; 363 364 /* --sort srcline */ 365 366 char *hist_entry__srcline(struct hist_entry *he) 367 { 368 return map__srcline(he->ms.map, he->ip, he->ms.sym); 369 } 370 371 static int64_t 372 sort__srcline_cmp(struct hist_entry *left, struct hist_entry *right) 373 { 374 if (!left->srcline) 375 left->srcline = hist_entry__srcline(left); 376 if (!right->srcline) 377 right->srcline = hist_entry__srcline(right); 378 379 return strcmp(right->srcline, left->srcline); 380 } 381 382 static int hist_entry__srcline_snprintf(struct hist_entry *he, char *bf, 383 size_t size, unsigned int width) 384 { 385 if (!he->srcline) 386 he->srcline = hist_entry__srcline(he); 387 388 return repsep_snprintf(bf, size, "%-.*s", width, he->srcline); 389 } 390 391 struct sort_entry sort_srcline = { 392 .se_header = "Source:Line", 393 .se_cmp = sort__srcline_cmp, 394 .se_snprintf = hist_entry__srcline_snprintf, 395 .se_width_idx = HISTC_SRCLINE, 396 }; 397 398 /* --sort srcline_from */ 399 400 static char *addr_map_symbol__srcline(struct addr_map_symbol *ams) 401 { 402 return map__srcline(ams->ms.map, ams->al_addr, ams->ms.sym); 403 } 404 405 static int64_t 406 sort__srcline_from_cmp(struct hist_entry *left, struct hist_entry *right) 407 { 408 if (!left->branch_info->srcline_from) 409 left->branch_info->srcline_from = addr_map_symbol__srcline(&left->branch_info->from); 410 411 if (!right->branch_info->srcline_from) 412 right->branch_info->srcline_from = addr_map_symbol__srcline(&right->branch_info->from); 413 414 return strcmp(right->branch_info->srcline_from, left->branch_info->srcline_from); 415 } 416 417 static int hist_entry__srcline_from_snprintf(struct hist_entry *he, char *bf, 418 size_t size, unsigned int width) 419 { 420 return repsep_snprintf(bf, size, "%-*.*s", width, width, he->branch_info->srcline_from); 421 } 422 423 struct sort_entry sort_srcline_from = { 424 .se_header = "From Source:Line", 425 .se_cmp = sort__srcline_from_cmp, 426 .se_snprintf = hist_entry__srcline_from_snprintf, 427 .se_width_idx = HISTC_SRCLINE_FROM, 428 }; 429 430 /* --sort srcline_to */ 431 432 static int64_t 433 sort__srcline_to_cmp(struct hist_entry *left, struct hist_entry *right) 434 { 435 if (!left->branch_info->srcline_to) 436 left->branch_info->srcline_to = addr_map_symbol__srcline(&left->branch_info->to); 437 438 if (!right->branch_info->srcline_to) 439 right->branch_info->srcline_to = addr_map_symbol__srcline(&right->branch_info->to); 440 441 return strcmp(right->branch_info->srcline_to, left->branch_info->srcline_to); 442 } 443 444 static int hist_entry__srcline_to_snprintf(struct hist_entry *he, char *bf, 445 size_t size, unsigned int width) 446 { 447 return repsep_snprintf(bf, size, "%-*.*s", width, width, he->branch_info->srcline_to); 448 } 449 450 struct sort_entry sort_srcline_to = { 451 .se_header = "To Source:Line", 452 .se_cmp = sort__srcline_to_cmp, 453 .se_snprintf = hist_entry__srcline_to_snprintf, 454 .se_width_idx = HISTC_SRCLINE_TO, 455 }; 456 457 static int hist_entry__sym_ipc_snprintf(struct hist_entry *he, char *bf, 458 size_t size, unsigned int width) 459 { 460 461 struct symbol *sym = he->ms.sym; 462 struct annotation *notes; 463 double ipc = 0.0, coverage = 0.0; 464 char tmp[64]; 465 466 if (!sym) 467 return repsep_snprintf(bf, size, "%-*s", width, "-"); 468 469 notes = symbol__annotation(sym); 470 471 if (notes->hit_cycles) 472 ipc = notes->hit_insn / ((double)notes->hit_cycles); 473 474 if (notes->total_insn) { 475 coverage = notes->cover_insn * 100.0 / 476 ((double)notes->total_insn); 477 } 478 479 snprintf(tmp, sizeof(tmp), "%-5.2f [%5.1f%%]", ipc, coverage); 480 return repsep_snprintf(bf, size, "%-*s", width, tmp); 481 } 482 483 struct sort_entry sort_sym_ipc = { 484 .se_header = "IPC [IPC Coverage]", 485 .se_cmp = sort__sym_cmp, 486 .se_snprintf = hist_entry__sym_ipc_snprintf, 487 .se_width_idx = HISTC_SYMBOL_IPC, 488 }; 489 490 static int hist_entry__sym_ipc_null_snprintf(struct hist_entry *he 491 __maybe_unused, 492 char *bf, size_t size, 493 unsigned int width) 494 { 495 char tmp[64]; 496 497 snprintf(tmp, sizeof(tmp), "%-5s %2s", "-", "-"); 498 return repsep_snprintf(bf, size, "%-*s", width, tmp); 499 } 500 501 struct sort_entry sort_sym_ipc_null = { 502 .se_header = "IPC [IPC Coverage]", 503 .se_cmp = sort__sym_cmp, 504 .se_snprintf = hist_entry__sym_ipc_null_snprintf, 505 .se_width_idx = HISTC_SYMBOL_IPC, 506 }; 507 508 /* --sort srcfile */ 509 510 static char no_srcfile[1]; 511 512 static char *hist_entry__get_srcfile(struct hist_entry *e) 513 { 514 char *sf, *p; 515 struct map *map = e->ms.map; 516 517 if (!map) 518 return no_srcfile; 519 520 sf = __get_srcline(map->dso, map__rip_2objdump(map, e->ip), 521 e->ms.sym, false, true, true, e->ip); 522 if (!strcmp(sf, SRCLINE_UNKNOWN)) 523 return no_srcfile; 524 p = strchr(sf, ':'); 525 if (p && *sf) { 526 *p = 0; 527 return sf; 528 } 529 free(sf); 530 return no_srcfile; 531 } 532 533 static int64_t 534 sort__srcfile_cmp(struct hist_entry *left, struct hist_entry *right) 535 { 536 if (!left->srcfile) 537 left->srcfile = hist_entry__get_srcfile(left); 538 if (!right->srcfile) 539 right->srcfile = hist_entry__get_srcfile(right); 540 541 return strcmp(right->srcfile, left->srcfile); 542 } 543 544 static int hist_entry__srcfile_snprintf(struct hist_entry *he, char *bf, 545 size_t size, unsigned int width) 546 { 547 if (!he->srcfile) 548 he->srcfile = hist_entry__get_srcfile(he); 549 550 return repsep_snprintf(bf, size, "%-.*s", width, he->srcfile); 551 } 552 553 struct sort_entry sort_srcfile = { 554 .se_header = "Source File", 555 .se_cmp = sort__srcfile_cmp, 556 .se_snprintf = hist_entry__srcfile_snprintf, 557 .se_width_idx = HISTC_SRCFILE, 558 }; 559 560 /* --sort parent */ 561 562 static int64_t 563 sort__parent_cmp(struct hist_entry *left, struct hist_entry *right) 564 { 565 struct symbol *sym_l = left->parent; 566 struct symbol *sym_r = right->parent; 567 568 if (!sym_l || !sym_r) 569 return cmp_null(sym_l, sym_r); 570 571 return strcmp(sym_r->name, sym_l->name); 572 } 573 574 static int hist_entry__parent_snprintf(struct hist_entry *he, char *bf, 575 size_t size, unsigned int width) 576 { 577 return repsep_snprintf(bf, size, "%-*.*s", width, width, 578 he->parent ? he->parent->name : "[other]"); 579 } 580 581 struct sort_entry sort_parent = { 582 .se_header = "Parent symbol", 583 .se_cmp = sort__parent_cmp, 584 .se_snprintf = hist_entry__parent_snprintf, 585 .se_width_idx = HISTC_PARENT, 586 }; 587 588 /* --sort cpu */ 589 590 static int64_t 591 sort__cpu_cmp(struct hist_entry *left, struct hist_entry *right) 592 { 593 return right->cpu - left->cpu; 594 } 595 596 static int hist_entry__cpu_snprintf(struct hist_entry *he, char *bf, 597 size_t size, unsigned int width) 598 { 599 return repsep_snprintf(bf, size, "%*.*d", width, width, he->cpu); 600 } 601 602 struct sort_entry sort_cpu = { 603 .se_header = "CPU", 604 .se_cmp = sort__cpu_cmp, 605 .se_snprintf = hist_entry__cpu_snprintf, 606 .se_width_idx = HISTC_CPU, 607 }; 608 609 /* --sort cgroup_id */ 610 611 static int64_t _sort__cgroup_dev_cmp(u64 left_dev, u64 right_dev) 612 { 613 return (int64_t)(right_dev - left_dev); 614 } 615 616 static int64_t _sort__cgroup_inode_cmp(u64 left_ino, u64 right_ino) 617 { 618 return (int64_t)(right_ino - left_ino); 619 } 620 621 static int64_t 622 sort__cgroup_id_cmp(struct hist_entry *left, struct hist_entry *right) 623 { 624 int64_t ret; 625 626 ret = _sort__cgroup_dev_cmp(right->cgroup_id.dev, left->cgroup_id.dev); 627 if (ret != 0) 628 return ret; 629 630 return _sort__cgroup_inode_cmp(right->cgroup_id.ino, 631 left->cgroup_id.ino); 632 } 633 634 static int hist_entry__cgroup_id_snprintf(struct hist_entry *he, 635 char *bf, size_t size, 636 unsigned int width __maybe_unused) 637 { 638 return repsep_snprintf(bf, size, "%lu/0x%lx", he->cgroup_id.dev, 639 he->cgroup_id.ino); 640 } 641 642 struct sort_entry sort_cgroup_id = { 643 .se_header = "cgroup id (dev/inode)", 644 .se_cmp = sort__cgroup_id_cmp, 645 .se_snprintf = hist_entry__cgroup_id_snprintf, 646 .se_width_idx = HISTC_CGROUP_ID, 647 }; 648 649 /* --sort cgroup */ 650 651 static int64_t 652 sort__cgroup_cmp(struct hist_entry *left, struct hist_entry *right) 653 { 654 return right->cgroup - left->cgroup; 655 } 656 657 static int hist_entry__cgroup_snprintf(struct hist_entry *he, 658 char *bf, size_t size, 659 unsigned int width __maybe_unused) 660 { 661 const char *cgrp_name = "N/A"; 662 663 if (he->cgroup) { 664 struct cgroup *cgrp = cgroup__find(he->ms.maps->machine->env, 665 he->cgroup); 666 if (cgrp != NULL) 667 cgrp_name = cgrp->name; 668 else 669 cgrp_name = "unknown"; 670 } 671 672 return repsep_snprintf(bf, size, "%s", cgrp_name); 673 } 674 675 struct sort_entry sort_cgroup = { 676 .se_header = "Cgroup", 677 .se_cmp = sort__cgroup_cmp, 678 .se_snprintf = hist_entry__cgroup_snprintf, 679 .se_width_idx = HISTC_CGROUP, 680 }; 681 682 /* --sort socket */ 683 684 static int64_t 685 sort__socket_cmp(struct hist_entry *left, struct hist_entry *right) 686 { 687 return right->socket - left->socket; 688 } 689 690 static int hist_entry__socket_snprintf(struct hist_entry *he, char *bf, 691 size_t size, unsigned int width) 692 { 693 return repsep_snprintf(bf, size, "%*.*d", width, width-3, he->socket); 694 } 695 696 static int hist_entry__socket_filter(struct hist_entry *he, int type, const void *arg) 697 { 698 int sk = *(const int *)arg; 699 700 if (type != HIST_FILTER__SOCKET) 701 return -1; 702 703 return sk >= 0 && he->socket != sk; 704 } 705 706 struct sort_entry sort_socket = { 707 .se_header = "Socket", 708 .se_cmp = sort__socket_cmp, 709 .se_snprintf = hist_entry__socket_snprintf, 710 .se_filter = hist_entry__socket_filter, 711 .se_width_idx = HISTC_SOCKET, 712 }; 713 714 /* --sort time */ 715 716 static int64_t 717 sort__time_cmp(struct hist_entry *left, struct hist_entry *right) 718 { 719 return right->time - left->time; 720 } 721 722 static int hist_entry__time_snprintf(struct hist_entry *he, char *bf, 723 size_t size, unsigned int width) 724 { 725 char he_time[32]; 726 727 if (symbol_conf.nanosecs) 728 timestamp__scnprintf_nsec(he->time, he_time, 729 sizeof(he_time)); 730 else 731 timestamp__scnprintf_usec(he->time, he_time, 732 sizeof(he_time)); 733 734 return repsep_snprintf(bf, size, "%-.*s", width, he_time); 735 } 736 737 struct sort_entry sort_time = { 738 .se_header = "Time", 739 .se_cmp = sort__time_cmp, 740 .se_snprintf = hist_entry__time_snprintf, 741 .se_width_idx = HISTC_TIME, 742 }; 743 744 /* --sort trace */ 745 746 static char *get_trace_output(struct hist_entry *he) 747 { 748 struct trace_seq seq; 749 struct evsel *evsel; 750 struct tep_record rec = { 751 .data = he->raw_data, 752 .size = he->raw_size, 753 }; 754 755 evsel = hists_to_evsel(he->hists); 756 757 trace_seq_init(&seq); 758 if (symbol_conf.raw_trace) { 759 tep_print_fields(&seq, he->raw_data, he->raw_size, 760 evsel->tp_format); 761 } else { 762 tep_print_event(evsel->tp_format->tep, 763 &seq, &rec, "%s", TEP_PRINT_INFO); 764 } 765 /* 766 * Trim the buffer, it starts at 4KB and we're not going to 767 * add anything more to this buffer. 768 */ 769 return realloc(seq.buffer, seq.len + 1); 770 } 771 772 static int64_t 773 sort__trace_cmp(struct hist_entry *left, struct hist_entry *right) 774 { 775 struct evsel *evsel; 776 777 evsel = hists_to_evsel(left->hists); 778 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) 779 return 0; 780 781 if (left->trace_output == NULL) 782 left->trace_output = get_trace_output(left); 783 if (right->trace_output == NULL) 784 right->trace_output = get_trace_output(right); 785 786 return strcmp(right->trace_output, left->trace_output); 787 } 788 789 static int hist_entry__trace_snprintf(struct hist_entry *he, char *bf, 790 size_t size, unsigned int width) 791 { 792 struct evsel *evsel; 793 794 evsel = hists_to_evsel(he->hists); 795 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) 796 return scnprintf(bf, size, "%-.*s", width, "N/A"); 797 798 if (he->trace_output == NULL) 799 he->trace_output = get_trace_output(he); 800 return repsep_snprintf(bf, size, "%-.*s", width, he->trace_output); 801 } 802 803 struct sort_entry sort_trace = { 804 .se_header = "Trace output", 805 .se_cmp = sort__trace_cmp, 806 .se_snprintf = hist_entry__trace_snprintf, 807 .se_width_idx = HISTC_TRACE, 808 }; 809 810 /* sort keys for branch stacks */ 811 812 static int64_t 813 sort__dso_from_cmp(struct hist_entry *left, struct hist_entry *right) 814 { 815 if (!left->branch_info || !right->branch_info) 816 return cmp_null(left->branch_info, right->branch_info); 817 818 return _sort__dso_cmp(left->branch_info->from.ms.map, 819 right->branch_info->from.ms.map); 820 } 821 822 static int hist_entry__dso_from_snprintf(struct hist_entry *he, char *bf, 823 size_t size, unsigned int width) 824 { 825 if (he->branch_info) 826 return _hist_entry__dso_snprintf(he->branch_info->from.ms.map, 827 bf, size, width); 828 else 829 return repsep_snprintf(bf, size, "%-*.*s", width, width, "N/A"); 830 } 831 832 static int hist_entry__dso_from_filter(struct hist_entry *he, int type, 833 const void *arg) 834 { 835 const struct dso *dso = arg; 836 837 if (type != HIST_FILTER__DSO) 838 return -1; 839 840 return dso && (!he->branch_info || !he->branch_info->from.ms.map || 841 he->branch_info->from.ms.map->dso != dso); 842 } 843 844 static int64_t 845 sort__dso_to_cmp(struct hist_entry *left, struct hist_entry *right) 846 { 847 if (!left->branch_info || !right->branch_info) 848 return cmp_null(left->branch_info, right->branch_info); 849 850 return _sort__dso_cmp(left->branch_info->to.ms.map, 851 right->branch_info->to.ms.map); 852 } 853 854 static int hist_entry__dso_to_snprintf(struct hist_entry *he, char *bf, 855 size_t size, unsigned int width) 856 { 857 if (he->branch_info) 858 return _hist_entry__dso_snprintf(he->branch_info->to.ms.map, 859 bf, size, width); 860 else 861 return repsep_snprintf(bf, size, "%-*.*s", width, width, "N/A"); 862 } 863 864 static int hist_entry__dso_to_filter(struct hist_entry *he, int type, 865 const void *arg) 866 { 867 const struct dso *dso = arg; 868 869 if (type != HIST_FILTER__DSO) 870 return -1; 871 872 return dso && (!he->branch_info || !he->branch_info->to.ms.map || 873 he->branch_info->to.ms.map->dso != dso); 874 } 875 876 static int64_t 877 sort__sym_from_cmp(struct hist_entry *left, struct hist_entry *right) 878 { 879 struct addr_map_symbol *from_l = &left->branch_info->from; 880 struct addr_map_symbol *from_r = &right->branch_info->from; 881 882 if (!left->branch_info || !right->branch_info) 883 return cmp_null(left->branch_info, right->branch_info); 884 885 from_l = &left->branch_info->from; 886 from_r = &right->branch_info->from; 887 888 if (!from_l->ms.sym && !from_r->ms.sym) 889 return _sort__addr_cmp(from_l->addr, from_r->addr); 890 891 return _sort__sym_cmp(from_l->ms.sym, from_r->ms.sym); 892 } 893 894 static int64_t 895 sort__sym_to_cmp(struct hist_entry *left, struct hist_entry *right) 896 { 897 struct addr_map_symbol *to_l, *to_r; 898 899 if (!left->branch_info || !right->branch_info) 900 return cmp_null(left->branch_info, right->branch_info); 901 902 to_l = &left->branch_info->to; 903 to_r = &right->branch_info->to; 904 905 if (!to_l->ms.sym && !to_r->ms.sym) 906 return _sort__addr_cmp(to_l->addr, to_r->addr); 907 908 return _sort__sym_cmp(to_l->ms.sym, to_r->ms.sym); 909 } 910 911 static int hist_entry__sym_from_snprintf(struct hist_entry *he, char *bf, 912 size_t size, unsigned int width) 913 { 914 if (he->branch_info) { 915 struct addr_map_symbol *from = &he->branch_info->from; 916 917 return _hist_entry__sym_snprintf(&from->ms, from->al_addr, 918 he->level, bf, size, width); 919 } 920 921 return repsep_snprintf(bf, size, "%-*.*s", width, width, "N/A"); 922 } 923 924 static int hist_entry__sym_to_snprintf(struct hist_entry *he, char *bf, 925 size_t size, unsigned int width) 926 { 927 if (he->branch_info) { 928 struct addr_map_symbol *to = &he->branch_info->to; 929 930 return _hist_entry__sym_snprintf(&to->ms, to->al_addr, 931 he->level, bf, size, width); 932 } 933 934 return repsep_snprintf(bf, size, "%-*.*s", width, width, "N/A"); 935 } 936 937 static int hist_entry__sym_from_filter(struct hist_entry *he, int type, 938 const void *arg) 939 { 940 const char *sym = arg; 941 942 if (type != HIST_FILTER__SYMBOL) 943 return -1; 944 945 return sym && !(he->branch_info && he->branch_info->from.ms.sym && 946 strstr(he->branch_info->from.ms.sym->name, sym)); 947 } 948 949 static int hist_entry__sym_to_filter(struct hist_entry *he, int type, 950 const void *arg) 951 { 952 const char *sym = arg; 953 954 if (type != HIST_FILTER__SYMBOL) 955 return -1; 956 957 return sym && !(he->branch_info && he->branch_info->to.ms.sym && 958 strstr(he->branch_info->to.ms.sym->name, sym)); 959 } 960 961 struct sort_entry sort_dso_from = { 962 .se_header = "Source Shared Object", 963 .se_cmp = sort__dso_from_cmp, 964 .se_snprintf = hist_entry__dso_from_snprintf, 965 .se_filter = hist_entry__dso_from_filter, 966 .se_width_idx = HISTC_DSO_FROM, 967 }; 968 969 struct sort_entry sort_dso_to = { 970 .se_header = "Target Shared Object", 971 .se_cmp = sort__dso_to_cmp, 972 .se_snprintf = hist_entry__dso_to_snprintf, 973 .se_filter = hist_entry__dso_to_filter, 974 .se_width_idx = HISTC_DSO_TO, 975 }; 976 977 struct sort_entry sort_sym_from = { 978 .se_header = "Source Symbol", 979 .se_cmp = sort__sym_from_cmp, 980 .se_snprintf = hist_entry__sym_from_snprintf, 981 .se_filter = hist_entry__sym_from_filter, 982 .se_width_idx = HISTC_SYMBOL_FROM, 983 }; 984 985 struct sort_entry sort_sym_to = { 986 .se_header = "Target Symbol", 987 .se_cmp = sort__sym_to_cmp, 988 .se_snprintf = hist_entry__sym_to_snprintf, 989 .se_filter = hist_entry__sym_to_filter, 990 .se_width_idx = HISTC_SYMBOL_TO, 991 }; 992 993 static int64_t 994 sort__mispredict_cmp(struct hist_entry *left, struct hist_entry *right) 995 { 996 unsigned char mp, p; 997 998 if (!left->branch_info || !right->branch_info) 999 return cmp_null(left->branch_info, right->branch_info); 1000 1001 mp = left->branch_info->flags.mispred != right->branch_info->flags.mispred; 1002 p = left->branch_info->flags.predicted != right->branch_info->flags.predicted; 1003 return mp || p; 1004 } 1005 1006 static int hist_entry__mispredict_snprintf(struct hist_entry *he, char *bf, 1007 size_t size, unsigned int width){ 1008 static const char *out = "N/A"; 1009 1010 if (he->branch_info) { 1011 if (he->branch_info->flags.predicted) 1012 out = "N"; 1013 else if (he->branch_info->flags.mispred) 1014 out = "Y"; 1015 } 1016 1017 return repsep_snprintf(bf, size, "%-*.*s", width, width, out); 1018 } 1019 1020 static int64_t 1021 sort__cycles_cmp(struct hist_entry *left, struct hist_entry *right) 1022 { 1023 if (!left->branch_info || !right->branch_info) 1024 return cmp_null(left->branch_info, right->branch_info); 1025 1026 return left->branch_info->flags.cycles - 1027 right->branch_info->flags.cycles; 1028 } 1029 1030 static int hist_entry__cycles_snprintf(struct hist_entry *he, char *bf, 1031 size_t size, unsigned int width) 1032 { 1033 if (!he->branch_info) 1034 return scnprintf(bf, size, "%-.*s", width, "N/A"); 1035 if (he->branch_info->flags.cycles == 0) 1036 return repsep_snprintf(bf, size, "%-*s", width, "-"); 1037 return repsep_snprintf(bf, size, "%-*hd", width, 1038 he->branch_info->flags.cycles); 1039 } 1040 1041 struct sort_entry sort_cycles = { 1042 .se_header = "Basic Block Cycles", 1043 .se_cmp = sort__cycles_cmp, 1044 .se_snprintf = hist_entry__cycles_snprintf, 1045 .se_width_idx = HISTC_CYCLES, 1046 }; 1047 1048 /* --sort daddr_sym */ 1049 int64_t 1050 sort__daddr_cmp(struct hist_entry *left, struct hist_entry *right) 1051 { 1052 uint64_t l = 0, r = 0; 1053 1054 if (left->mem_info) 1055 l = left->mem_info->daddr.addr; 1056 if (right->mem_info) 1057 r = right->mem_info->daddr.addr; 1058 1059 return (int64_t)(r - l); 1060 } 1061 1062 static int hist_entry__daddr_snprintf(struct hist_entry *he, char *bf, 1063 size_t size, unsigned int width) 1064 { 1065 uint64_t addr = 0; 1066 struct map_symbol *ms = NULL; 1067 1068 if (he->mem_info) { 1069 addr = he->mem_info->daddr.addr; 1070 ms = &he->mem_info->daddr.ms; 1071 } 1072 return _hist_entry__sym_snprintf(ms, addr, he->level, bf, size, width); 1073 } 1074 1075 int64_t 1076 sort__iaddr_cmp(struct hist_entry *left, struct hist_entry *right) 1077 { 1078 uint64_t l = 0, r = 0; 1079 1080 if (left->mem_info) 1081 l = left->mem_info->iaddr.addr; 1082 if (right->mem_info) 1083 r = right->mem_info->iaddr.addr; 1084 1085 return (int64_t)(r - l); 1086 } 1087 1088 static int hist_entry__iaddr_snprintf(struct hist_entry *he, char *bf, 1089 size_t size, unsigned int width) 1090 { 1091 uint64_t addr = 0; 1092 struct map_symbol *ms = NULL; 1093 1094 if (he->mem_info) { 1095 addr = he->mem_info->iaddr.addr; 1096 ms = &he->mem_info->iaddr.ms; 1097 } 1098 return _hist_entry__sym_snprintf(ms, addr, he->level, bf, size, width); 1099 } 1100 1101 static int64_t 1102 sort__dso_daddr_cmp(struct hist_entry *left, struct hist_entry *right) 1103 { 1104 struct map *map_l = NULL; 1105 struct map *map_r = NULL; 1106 1107 if (left->mem_info) 1108 map_l = left->mem_info->daddr.ms.map; 1109 if (right->mem_info) 1110 map_r = right->mem_info->daddr.ms.map; 1111 1112 return _sort__dso_cmp(map_l, map_r); 1113 } 1114 1115 static int hist_entry__dso_daddr_snprintf(struct hist_entry *he, char *bf, 1116 size_t size, unsigned int width) 1117 { 1118 struct map *map = NULL; 1119 1120 if (he->mem_info) 1121 map = he->mem_info->daddr.ms.map; 1122 1123 return _hist_entry__dso_snprintf(map, bf, size, width); 1124 } 1125 1126 static int64_t 1127 sort__locked_cmp(struct hist_entry *left, struct hist_entry *right) 1128 { 1129 union perf_mem_data_src data_src_l; 1130 union perf_mem_data_src data_src_r; 1131 1132 if (left->mem_info) 1133 data_src_l = left->mem_info->data_src; 1134 else 1135 data_src_l.mem_lock = PERF_MEM_LOCK_NA; 1136 1137 if (right->mem_info) 1138 data_src_r = right->mem_info->data_src; 1139 else 1140 data_src_r.mem_lock = PERF_MEM_LOCK_NA; 1141 1142 return (int64_t)(data_src_r.mem_lock - data_src_l.mem_lock); 1143 } 1144 1145 static int hist_entry__locked_snprintf(struct hist_entry *he, char *bf, 1146 size_t size, unsigned int width) 1147 { 1148 char out[10]; 1149 1150 perf_mem__lck_scnprintf(out, sizeof(out), he->mem_info); 1151 return repsep_snprintf(bf, size, "%.*s", width, out); 1152 } 1153 1154 static int64_t 1155 sort__tlb_cmp(struct hist_entry *left, struct hist_entry *right) 1156 { 1157 union perf_mem_data_src data_src_l; 1158 union perf_mem_data_src data_src_r; 1159 1160 if (left->mem_info) 1161 data_src_l = left->mem_info->data_src; 1162 else 1163 data_src_l.mem_dtlb = PERF_MEM_TLB_NA; 1164 1165 if (right->mem_info) 1166 data_src_r = right->mem_info->data_src; 1167 else 1168 data_src_r.mem_dtlb = PERF_MEM_TLB_NA; 1169 1170 return (int64_t)(data_src_r.mem_dtlb - data_src_l.mem_dtlb); 1171 } 1172 1173 static int hist_entry__tlb_snprintf(struct hist_entry *he, char *bf, 1174 size_t size, unsigned int width) 1175 { 1176 char out[64]; 1177 1178 perf_mem__tlb_scnprintf(out, sizeof(out), he->mem_info); 1179 return repsep_snprintf(bf, size, "%-*s", width, out); 1180 } 1181 1182 static int64_t 1183 sort__lvl_cmp(struct hist_entry *left, struct hist_entry *right) 1184 { 1185 union perf_mem_data_src data_src_l; 1186 union perf_mem_data_src data_src_r; 1187 1188 if (left->mem_info) 1189 data_src_l = left->mem_info->data_src; 1190 else 1191 data_src_l.mem_lvl = PERF_MEM_LVL_NA; 1192 1193 if (right->mem_info) 1194 data_src_r = right->mem_info->data_src; 1195 else 1196 data_src_r.mem_lvl = PERF_MEM_LVL_NA; 1197 1198 return (int64_t)(data_src_r.mem_lvl - data_src_l.mem_lvl); 1199 } 1200 1201 static int hist_entry__lvl_snprintf(struct hist_entry *he, char *bf, 1202 size_t size, unsigned int width) 1203 { 1204 char out[64]; 1205 1206 perf_mem__lvl_scnprintf(out, sizeof(out), he->mem_info); 1207 return repsep_snprintf(bf, size, "%-*s", width, out); 1208 } 1209 1210 static int64_t 1211 sort__snoop_cmp(struct hist_entry *left, struct hist_entry *right) 1212 { 1213 union perf_mem_data_src data_src_l; 1214 union perf_mem_data_src data_src_r; 1215 1216 if (left->mem_info) 1217 data_src_l = left->mem_info->data_src; 1218 else 1219 data_src_l.mem_snoop = PERF_MEM_SNOOP_NA; 1220 1221 if (right->mem_info) 1222 data_src_r = right->mem_info->data_src; 1223 else 1224 data_src_r.mem_snoop = PERF_MEM_SNOOP_NA; 1225 1226 return (int64_t)(data_src_r.mem_snoop - data_src_l.mem_snoop); 1227 } 1228 1229 static int hist_entry__snoop_snprintf(struct hist_entry *he, char *bf, 1230 size_t size, unsigned int width) 1231 { 1232 char out[64]; 1233 1234 perf_mem__snp_scnprintf(out, sizeof(out), he->mem_info); 1235 return repsep_snprintf(bf, size, "%-*s", width, out); 1236 } 1237 1238 int64_t 1239 sort__dcacheline_cmp(struct hist_entry *left, struct hist_entry *right) 1240 { 1241 u64 l, r; 1242 struct map *l_map, *r_map; 1243 int rc; 1244 1245 if (!left->mem_info) return -1; 1246 if (!right->mem_info) return 1; 1247 1248 /* group event types together */ 1249 if (left->cpumode > right->cpumode) return -1; 1250 if (left->cpumode < right->cpumode) return 1; 1251 1252 l_map = left->mem_info->daddr.ms.map; 1253 r_map = right->mem_info->daddr.ms.map; 1254 1255 /* if both are NULL, jump to sort on al_addr instead */ 1256 if (!l_map && !r_map) 1257 goto addr; 1258 1259 if (!l_map) return -1; 1260 if (!r_map) return 1; 1261 1262 rc = dso__cmp_id(l_map->dso, r_map->dso); 1263 if (rc) 1264 return rc; 1265 /* 1266 * Addresses with no major/minor numbers are assumed to be 1267 * anonymous in userspace. Sort those on pid then address. 1268 * 1269 * The kernel and non-zero major/minor mapped areas are 1270 * assumed to be unity mapped. Sort those on address. 1271 */ 1272 1273 if ((left->cpumode != PERF_RECORD_MISC_KERNEL) && 1274 (!(l_map->flags & MAP_SHARED)) && 1275 !l_map->dso->id.maj && !l_map->dso->id.min && 1276 !l_map->dso->id.ino && !l_map->dso->id.ino_generation) { 1277 /* userspace anonymous */ 1278 1279 if (left->thread->pid_ > right->thread->pid_) return -1; 1280 if (left->thread->pid_ < right->thread->pid_) return 1; 1281 } 1282 1283 addr: 1284 /* al_addr does all the right addr - start + offset calculations */ 1285 l = cl_address(left->mem_info->daddr.al_addr); 1286 r = cl_address(right->mem_info->daddr.al_addr); 1287 1288 if (l > r) return -1; 1289 if (l < r) return 1; 1290 1291 return 0; 1292 } 1293 1294 static int hist_entry__dcacheline_snprintf(struct hist_entry *he, char *bf, 1295 size_t size, unsigned int width) 1296 { 1297 1298 uint64_t addr = 0; 1299 struct map_symbol *ms = NULL; 1300 char level = he->level; 1301 1302 if (he->mem_info) { 1303 struct map *map = he->mem_info->daddr.ms.map; 1304 1305 addr = cl_address(he->mem_info->daddr.al_addr); 1306 ms = &he->mem_info->daddr.ms; 1307 1308 /* print [s] for shared data mmaps */ 1309 if ((he->cpumode != PERF_RECORD_MISC_KERNEL) && 1310 map && !(map->prot & PROT_EXEC) && 1311 (map->flags & MAP_SHARED) && 1312 (map->dso->id.maj || map->dso->id.min || 1313 map->dso->id.ino || map->dso->id.ino_generation)) 1314 level = 's'; 1315 else if (!map) 1316 level = 'X'; 1317 } 1318 return _hist_entry__sym_snprintf(ms, addr, level, bf, size, width); 1319 } 1320 1321 struct sort_entry sort_mispredict = { 1322 .se_header = "Branch Mispredicted", 1323 .se_cmp = sort__mispredict_cmp, 1324 .se_snprintf = hist_entry__mispredict_snprintf, 1325 .se_width_idx = HISTC_MISPREDICT, 1326 }; 1327 1328 static int64_t 1329 sort__weight_cmp(struct hist_entry *left, struct hist_entry *right) 1330 { 1331 return left->weight - right->weight; 1332 } 1333 1334 static int hist_entry__local_weight_snprintf(struct hist_entry *he, char *bf, 1335 size_t size, unsigned int width) 1336 { 1337 return repsep_snprintf(bf, size, "%-*llu", width, he->weight); 1338 } 1339 1340 struct sort_entry sort_local_weight = { 1341 .se_header = "Local Weight", 1342 .se_cmp = sort__weight_cmp, 1343 .se_snprintf = hist_entry__local_weight_snprintf, 1344 .se_width_idx = HISTC_LOCAL_WEIGHT, 1345 }; 1346 1347 static int hist_entry__global_weight_snprintf(struct hist_entry *he, char *bf, 1348 size_t size, unsigned int width) 1349 { 1350 return repsep_snprintf(bf, size, "%-*llu", width, 1351 he->weight * he->stat.nr_events); 1352 } 1353 1354 struct sort_entry sort_global_weight = { 1355 .se_header = "Weight", 1356 .se_cmp = sort__weight_cmp, 1357 .se_snprintf = hist_entry__global_weight_snprintf, 1358 .se_width_idx = HISTC_GLOBAL_WEIGHT, 1359 }; 1360 1361 static int64_t 1362 sort__ins_lat_cmp(struct hist_entry *left, struct hist_entry *right) 1363 { 1364 return left->ins_lat - right->ins_lat; 1365 } 1366 1367 static int hist_entry__local_ins_lat_snprintf(struct hist_entry *he, char *bf, 1368 size_t size, unsigned int width) 1369 { 1370 return repsep_snprintf(bf, size, "%-*u", width, he->ins_lat); 1371 } 1372 1373 struct sort_entry sort_local_ins_lat = { 1374 .se_header = "Local INSTR Latency", 1375 .se_cmp = sort__ins_lat_cmp, 1376 .se_snprintf = hist_entry__local_ins_lat_snprintf, 1377 .se_width_idx = HISTC_LOCAL_INS_LAT, 1378 }; 1379 1380 static int hist_entry__global_ins_lat_snprintf(struct hist_entry *he, char *bf, 1381 size_t size, unsigned int width) 1382 { 1383 return repsep_snprintf(bf, size, "%-*u", width, 1384 he->ins_lat * he->stat.nr_events); 1385 } 1386 1387 struct sort_entry sort_global_ins_lat = { 1388 .se_header = "INSTR Latency", 1389 .se_cmp = sort__ins_lat_cmp, 1390 .se_snprintf = hist_entry__global_ins_lat_snprintf, 1391 .se_width_idx = HISTC_GLOBAL_INS_LAT, 1392 }; 1393 1394 static int64_t 1395 sort__p_stage_cyc_cmp(struct hist_entry *left, struct hist_entry *right) 1396 { 1397 return left->p_stage_cyc - right->p_stage_cyc; 1398 } 1399 1400 static int hist_entry__global_p_stage_cyc_snprintf(struct hist_entry *he, char *bf, 1401 size_t size, unsigned int width) 1402 { 1403 return repsep_snprintf(bf, size, "%-*u", width, 1404 he->p_stage_cyc * he->stat.nr_events); 1405 } 1406 1407 1408 static int hist_entry__p_stage_cyc_snprintf(struct hist_entry *he, char *bf, 1409 size_t size, unsigned int width) 1410 { 1411 return repsep_snprintf(bf, size, "%-*u", width, he->p_stage_cyc); 1412 } 1413 1414 struct sort_entry sort_local_p_stage_cyc = { 1415 .se_header = "Local Pipeline Stage Cycle", 1416 .se_cmp = sort__p_stage_cyc_cmp, 1417 .se_snprintf = hist_entry__p_stage_cyc_snprintf, 1418 .se_width_idx = HISTC_LOCAL_P_STAGE_CYC, 1419 }; 1420 1421 struct sort_entry sort_global_p_stage_cyc = { 1422 .se_header = "Pipeline Stage Cycle", 1423 .se_cmp = sort__p_stage_cyc_cmp, 1424 .se_snprintf = hist_entry__global_p_stage_cyc_snprintf, 1425 .se_width_idx = HISTC_GLOBAL_P_STAGE_CYC, 1426 }; 1427 1428 struct sort_entry sort_mem_daddr_sym = { 1429 .se_header = "Data Symbol", 1430 .se_cmp = sort__daddr_cmp, 1431 .se_snprintf = hist_entry__daddr_snprintf, 1432 .se_width_idx = HISTC_MEM_DADDR_SYMBOL, 1433 }; 1434 1435 struct sort_entry sort_mem_iaddr_sym = { 1436 .se_header = "Code Symbol", 1437 .se_cmp = sort__iaddr_cmp, 1438 .se_snprintf = hist_entry__iaddr_snprintf, 1439 .se_width_idx = HISTC_MEM_IADDR_SYMBOL, 1440 }; 1441 1442 struct sort_entry sort_mem_daddr_dso = { 1443 .se_header = "Data Object", 1444 .se_cmp = sort__dso_daddr_cmp, 1445 .se_snprintf = hist_entry__dso_daddr_snprintf, 1446 .se_width_idx = HISTC_MEM_DADDR_DSO, 1447 }; 1448 1449 struct sort_entry sort_mem_locked = { 1450 .se_header = "Locked", 1451 .se_cmp = sort__locked_cmp, 1452 .se_snprintf = hist_entry__locked_snprintf, 1453 .se_width_idx = HISTC_MEM_LOCKED, 1454 }; 1455 1456 struct sort_entry sort_mem_tlb = { 1457 .se_header = "TLB access", 1458 .se_cmp = sort__tlb_cmp, 1459 .se_snprintf = hist_entry__tlb_snprintf, 1460 .se_width_idx = HISTC_MEM_TLB, 1461 }; 1462 1463 struct sort_entry sort_mem_lvl = { 1464 .se_header = "Memory access", 1465 .se_cmp = sort__lvl_cmp, 1466 .se_snprintf = hist_entry__lvl_snprintf, 1467 .se_width_idx = HISTC_MEM_LVL, 1468 }; 1469 1470 struct sort_entry sort_mem_snoop = { 1471 .se_header = "Snoop", 1472 .se_cmp = sort__snoop_cmp, 1473 .se_snprintf = hist_entry__snoop_snprintf, 1474 .se_width_idx = HISTC_MEM_SNOOP, 1475 }; 1476 1477 struct sort_entry sort_mem_dcacheline = { 1478 .se_header = "Data Cacheline", 1479 .se_cmp = sort__dcacheline_cmp, 1480 .se_snprintf = hist_entry__dcacheline_snprintf, 1481 .se_width_idx = HISTC_MEM_DCACHELINE, 1482 }; 1483 1484 static int64_t 1485 sort__blocked_cmp(struct hist_entry *left, struct hist_entry *right) 1486 { 1487 union perf_mem_data_src data_src_l; 1488 union perf_mem_data_src data_src_r; 1489 1490 if (left->mem_info) 1491 data_src_l = left->mem_info->data_src; 1492 else 1493 data_src_l.mem_blk = PERF_MEM_BLK_NA; 1494 1495 if (right->mem_info) 1496 data_src_r = right->mem_info->data_src; 1497 else 1498 data_src_r.mem_blk = PERF_MEM_BLK_NA; 1499 1500 return (int64_t)(data_src_r.mem_blk - data_src_l.mem_blk); 1501 } 1502 1503 static int hist_entry__blocked_snprintf(struct hist_entry *he, char *bf, 1504 size_t size, unsigned int width) 1505 { 1506 char out[16]; 1507 1508 perf_mem__blk_scnprintf(out, sizeof(out), he->mem_info); 1509 return repsep_snprintf(bf, size, "%.*s", width, out); 1510 } 1511 1512 struct sort_entry sort_mem_blocked = { 1513 .se_header = "Blocked", 1514 .se_cmp = sort__blocked_cmp, 1515 .se_snprintf = hist_entry__blocked_snprintf, 1516 .se_width_idx = HISTC_MEM_BLOCKED, 1517 }; 1518 1519 static int64_t 1520 sort__phys_daddr_cmp(struct hist_entry *left, struct hist_entry *right) 1521 { 1522 uint64_t l = 0, r = 0; 1523 1524 if (left->mem_info) 1525 l = left->mem_info->daddr.phys_addr; 1526 if (right->mem_info) 1527 r = right->mem_info->daddr.phys_addr; 1528 1529 return (int64_t)(r - l); 1530 } 1531 1532 static int hist_entry__phys_daddr_snprintf(struct hist_entry *he, char *bf, 1533 size_t size, unsigned int width) 1534 { 1535 uint64_t addr = 0; 1536 size_t ret = 0; 1537 size_t len = BITS_PER_LONG / 4; 1538 1539 addr = he->mem_info->daddr.phys_addr; 1540 1541 ret += repsep_snprintf(bf + ret, size - ret, "[%c] ", he->level); 1542 1543 ret += repsep_snprintf(bf + ret, size - ret, "%-#.*llx", len, addr); 1544 1545 ret += repsep_snprintf(bf + ret, size - ret, "%-*s", width - ret, ""); 1546 1547 if (ret > width) 1548 bf[width] = '\0'; 1549 1550 return width; 1551 } 1552 1553 struct sort_entry sort_mem_phys_daddr = { 1554 .se_header = "Data Physical Address", 1555 .se_cmp = sort__phys_daddr_cmp, 1556 .se_snprintf = hist_entry__phys_daddr_snprintf, 1557 .se_width_idx = HISTC_MEM_PHYS_DADDR, 1558 }; 1559 1560 static int64_t 1561 sort__data_page_size_cmp(struct hist_entry *left, struct hist_entry *right) 1562 { 1563 uint64_t l = 0, r = 0; 1564 1565 if (left->mem_info) 1566 l = left->mem_info->daddr.data_page_size; 1567 if (right->mem_info) 1568 r = right->mem_info->daddr.data_page_size; 1569 1570 return (int64_t)(r - l); 1571 } 1572 1573 static int hist_entry__data_page_size_snprintf(struct hist_entry *he, char *bf, 1574 size_t size, unsigned int width) 1575 { 1576 char str[PAGE_SIZE_NAME_LEN]; 1577 1578 return repsep_snprintf(bf, size, "%-*s", width, 1579 get_page_size_name(he->mem_info->daddr.data_page_size, str)); 1580 } 1581 1582 struct sort_entry sort_mem_data_page_size = { 1583 .se_header = "Data Page Size", 1584 .se_cmp = sort__data_page_size_cmp, 1585 .se_snprintf = hist_entry__data_page_size_snprintf, 1586 .se_width_idx = HISTC_MEM_DATA_PAGE_SIZE, 1587 }; 1588 1589 static int64_t 1590 sort__code_page_size_cmp(struct hist_entry *left, struct hist_entry *right) 1591 { 1592 uint64_t l = left->code_page_size; 1593 uint64_t r = right->code_page_size; 1594 1595 return (int64_t)(r - l); 1596 } 1597 1598 static int hist_entry__code_page_size_snprintf(struct hist_entry *he, char *bf, 1599 size_t size, unsigned int width) 1600 { 1601 char str[PAGE_SIZE_NAME_LEN]; 1602 1603 return repsep_snprintf(bf, size, "%-*s", width, 1604 get_page_size_name(he->code_page_size, str)); 1605 } 1606 1607 struct sort_entry sort_code_page_size = { 1608 .se_header = "Code Page Size", 1609 .se_cmp = sort__code_page_size_cmp, 1610 .se_snprintf = hist_entry__code_page_size_snprintf, 1611 .se_width_idx = HISTC_CODE_PAGE_SIZE, 1612 }; 1613 1614 static int64_t 1615 sort__abort_cmp(struct hist_entry *left, struct hist_entry *right) 1616 { 1617 if (!left->branch_info || !right->branch_info) 1618 return cmp_null(left->branch_info, right->branch_info); 1619 1620 return left->branch_info->flags.abort != 1621 right->branch_info->flags.abort; 1622 } 1623 1624 static int hist_entry__abort_snprintf(struct hist_entry *he, char *bf, 1625 size_t size, unsigned int width) 1626 { 1627 static const char *out = "N/A"; 1628 1629 if (he->branch_info) { 1630 if (he->branch_info->flags.abort) 1631 out = "A"; 1632 else 1633 out = "."; 1634 } 1635 1636 return repsep_snprintf(bf, size, "%-*s", width, out); 1637 } 1638 1639 struct sort_entry sort_abort = { 1640 .se_header = "Transaction abort", 1641 .se_cmp = sort__abort_cmp, 1642 .se_snprintf = hist_entry__abort_snprintf, 1643 .se_width_idx = HISTC_ABORT, 1644 }; 1645 1646 static int64_t 1647 sort__in_tx_cmp(struct hist_entry *left, struct hist_entry *right) 1648 { 1649 if (!left->branch_info || !right->branch_info) 1650 return cmp_null(left->branch_info, right->branch_info); 1651 1652 return left->branch_info->flags.in_tx != 1653 right->branch_info->flags.in_tx; 1654 } 1655 1656 static int hist_entry__in_tx_snprintf(struct hist_entry *he, char *bf, 1657 size_t size, unsigned int width) 1658 { 1659 static const char *out = "N/A"; 1660 1661 if (he->branch_info) { 1662 if (he->branch_info->flags.in_tx) 1663 out = "T"; 1664 else 1665 out = "."; 1666 } 1667 1668 return repsep_snprintf(bf, size, "%-*s", width, out); 1669 } 1670 1671 struct sort_entry sort_in_tx = { 1672 .se_header = "Branch in transaction", 1673 .se_cmp = sort__in_tx_cmp, 1674 .se_snprintf = hist_entry__in_tx_snprintf, 1675 .se_width_idx = HISTC_IN_TX, 1676 }; 1677 1678 static int64_t 1679 sort__transaction_cmp(struct hist_entry *left, struct hist_entry *right) 1680 { 1681 return left->transaction - right->transaction; 1682 } 1683 1684 static inline char *add_str(char *p, const char *str) 1685 { 1686 strcpy(p, str); 1687 return p + strlen(str); 1688 } 1689 1690 static struct txbit { 1691 unsigned flag; 1692 const char *name; 1693 int skip_for_len; 1694 } txbits[] = { 1695 { PERF_TXN_ELISION, "EL ", 0 }, 1696 { PERF_TXN_TRANSACTION, "TX ", 1 }, 1697 { PERF_TXN_SYNC, "SYNC ", 1 }, 1698 { PERF_TXN_ASYNC, "ASYNC ", 0 }, 1699 { PERF_TXN_RETRY, "RETRY ", 0 }, 1700 { PERF_TXN_CONFLICT, "CON ", 0 }, 1701 { PERF_TXN_CAPACITY_WRITE, "CAP-WRITE ", 1 }, 1702 { PERF_TXN_CAPACITY_READ, "CAP-READ ", 0 }, 1703 { 0, NULL, 0 } 1704 }; 1705 1706 int hist_entry__transaction_len(void) 1707 { 1708 int i; 1709 int len = 0; 1710 1711 for (i = 0; txbits[i].name; i++) { 1712 if (!txbits[i].skip_for_len) 1713 len += strlen(txbits[i].name); 1714 } 1715 len += 4; /* :XX<space> */ 1716 return len; 1717 } 1718 1719 static int hist_entry__transaction_snprintf(struct hist_entry *he, char *bf, 1720 size_t size, unsigned int width) 1721 { 1722 u64 t = he->transaction; 1723 char buf[128]; 1724 char *p = buf; 1725 int i; 1726 1727 buf[0] = 0; 1728 for (i = 0; txbits[i].name; i++) 1729 if (txbits[i].flag & t) 1730 p = add_str(p, txbits[i].name); 1731 if (t && !(t & (PERF_TXN_SYNC|PERF_TXN_ASYNC))) 1732 p = add_str(p, "NEITHER "); 1733 if (t & PERF_TXN_ABORT_MASK) { 1734 sprintf(p, ":%" PRIx64, 1735 (t & PERF_TXN_ABORT_MASK) >> 1736 PERF_TXN_ABORT_SHIFT); 1737 p += strlen(p); 1738 } 1739 1740 return repsep_snprintf(bf, size, "%-*s", width, buf); 1741 } 1742 1743 struct sort_entry sort_transaction = { 1744 .se_header = "Transaction ", 1745 .se_cmp = sort__transaction_cmp, 1746 .se_snprintf = hist_entry__transaction_snprintf, 1747 .se_width_idx = HISTC_TRANSACTION, 1748 }; 1749 1750 /* --sort symbol_size */ 1751 1752 static int64_t _sort__sym_size_cmp(struct symbol *sym_l, struct symbol *sym_r) 1753 { 1754 int64_t size_l = sym_l != NULL ? symbol__size(sym_l) : 0; 1755 int64_t size_r = sym_r != NULL ? symbol__size(sym_r) : 0; 1756 1757 return size_l < size_r ? -1 : 1758 size_l == size_r ? 0 : 1; 1759 } 1760 1761 static int64_t 1762 sort__sym_size_cmp(struct hist_entry *left, struct hist_entry *right) 1763 { 1764 return _sort__sym_size_cmp(right->ms.sym, left->ms.sym); 1765 } 1766 1767 static int _hist_entry__sym_size_snprintf(struct symbol *sym, char *bf, 1768 size_t bf_size, unsigned int width) 1769 { 1770 if (sym) 1771 return repsep_snprintf(bf, bf_size, "%*d", width, symbol__size(sym)); 1772 1773 return repsep_snprintf(bf, bf_size, "%*s", width, "unknown"); 1774 } 1775 1776 static int hist_entry__sym_size_snprintf(struct hist_entry *he, char *bf, 1777 size_t size, unsigned int width) 1778 { 1779 return _hist_entry__sym_size_snprintf(he->ms.sym, bf, size, width); 1780 } 1781 1782 struct sort_entry sort_sym_size = { 1783 .se_header = "Symbol size", 1784 .se_cmp = sort__sym_size_cmp, 1785 .se_snprintf = hist_entry__sym_size_snprintf, 1786 .se_width_idx = HISTC_SYM_SIZE, 1787 }; 1788 1789 /* --sort dso_size */ 1790 1791 static int64_t _sort__dso_size_cmp(struct map *map_l, struct map *map_r) 1792 { 1793 int64_t size_l = map_l != NULL ? map__size(map_l) : 0; 1794 int64_t size_r = map_r != NULL ? map__size(map_r) : 0; 1795 1796 return size_l < size_r ? -1 : 1797 size_l == size_r ? 0 : 1; 1798 } 1799 1800 static int64_t 1801 sort__dso_size_cmp(struct hist_entry *left, struct hist_entry *right) 1802 { 1803 return _sort__dso_size_cmp(right->ms.map, left->ms.map); 1804 } 1805 1806 static int _hist_entry__dso_size_snprintf(struct map *map, char *bf, 1807 size_t bf_size, unsigned int width) 1808 { 1809 if (map && map->dso) 1810 return repsep_snprintf(bf, bf_size, "%*d", width, 1811 map__size(map)); 1812 1813 return repsep_snprintf(bf, bf_size, "%*s", width, "unknown"); 1814 } 1815 1816 static int hist_entry__dso_size_snprintf(struct hist_entry *he, char *bf, 1817 size_t size, unsigned int width) 1818 { 1819 return _hist_entry__dso_size_snprintf(he->ms.map, bf, size, width); 1820 } 1821 1822 struct sort_entry sort_dso_size = { 1823 .se_header = "DSO size", 1824 .se_cmp = sort__dso_size_cmp, 1825 .se_snprintf = hist_entry__dso_size_snprintf, 1826 .se_width_idx = HISTC_DSO_SIZE, 1827 }; 1828 1829 1830 struct sort_dimension { 1831 const char *name; 1832 struct sort_entry *entry; 1833 int taken; 1834 }; 1835 1836 int __weak arch_support_sort_key(const char *sort_key __maybe_unused) 1837 { 1838 return 0; 1839 } 1840 1841 const char * __weak arch_perf_header_entry(const char *se_header) 1842 { 1843 return se_header; 1844 } 1845 1846 static void sort_dimension_add_dynamic_header(struct sort_dimension *sd) 1847 { 1848 sd->entry->se_header = arch_perf_header_entry(sd->entry->se_header); 1849 } 1850 1851 #define DIM(d, n, func) [d] = { .name = n, .entry = &(func) } 1852 1853 static struct sort_dimension common_sort_dimensions[] = { 1854 DIM(SORT_PID, "pid", sort_thread), 1855 DIM(SORT_COMM, "comm", sort_comm), 1856 DIM(SORT_DSO, "dso", sort_dso), 1857 DIM(SORT_SYM, "symbol", sort_sym), 1858 DIM(SORT_PARENT, "parent", sort_parent), 1859 DIM(SORT_CPU, "cpu", sort_cpu), 1860 DIM(SORT_SOCKET, "socket", sort_socket), 1861 DIM(SORT_SRCLINE, "srcline", sort_srcline), 1862 DIM(SORT_SRCFILE, "srcfile", sort_srcfile), 1863 DIM(SORT_LOCAL_WEIGHT, "local_weight", sort_local_weight), 1864 DIM(SORT_GLOBAL_WEIGHT, "weight", sort_global_weight), 1865 DIM(SORT_TRANSACTION, "transaction", sort_transaction), 1866 DIM(SORT_TRACE, "trace", sort_trace), 1867 DIM(SORT_SYM_SIZE, "symbol_size", sort_sym_size), 1868 DIM(SORT_DSO_SIZE, "dso_size", sort_dso_size), 1869 DIM(SORT_CGROUP, "cgroup", sort_cgroup), 1870 DIM(SORT_CGROUP_ID, "cgroup_id", sort_cgroup_id), 1871 DIM(SORT_SYM_IPC_NULL, "ipc_null", sort_sym_ipc_null), 1872 DIM(SORT_TIME, "time", sort_time), 1873 DIM(SORT_CODE_PAGE_SIZE, "code_page_size", sort_code_page_size), 1874 DIM(SORT_LOCAL_INS_LAT, "local_ins_lat", sort_local_ins_lat), 1875 DIM(SORT_GLOBAL_INS_LAT, "ins_lat", sort_global_ins_lat), 1876 DIM(SORT_LOCAL_PIPELINE_STAGE_CYC, "local_p_stage_cyc", sort_local_p_stage_cyc), 1877 DIM(SORT_GLOBAL_PIPELINE_STAGE_CYC, "p_stage_cyc", sort_global_p_stage_cyc), 1878 }; 1879 1880 #undef DIM 1881 1882 #define DIM(d, n, func) [d - __SORT_BRANCH_STACK] = { .name = n, .entry = &(func) } 1883 1884 static struct sort_dimension bstack_sort_dimensions[] = { 1885 DIM(SORT_DSO_FROM, "dso_from", sort_dso_from), 1886 DIM(SORT_DSO_TO, "dso_to", sort_dso_to), 1887 DIM(SORT_SYM_FROM, "symbol_from", sort_sym_from), 1888 DIM(SORT_SYM_TO, "symbol_to", sort_sym_to), 1889 DIM(SORT_MISPREDICT, "mispredict", sort_mispredict), 1890 DIM(SORT_IN_TX, "in_tx", sort_in_tx), 1891 DIM(SORT_ABORT, "abort", sort_abort), 1892 DIM(SORT_CYCLES, "cycles", sort_cycles), 1893 DIM(SORT_SRCLINE_FROM, "srcline_from", sort_srcline_from), 1894 DIM(SORT_SRCLINE_TO, "srcline_to", sort_srcline_to), 1895 DIM(SORT_SYM_IPC, "ipc_lbr", sort_sym_ipc), 1896 }; 1897 1898 #undef DIM 1899 1900 #define DIM(d, n, func) [d - __SORT_MEMORY_MODE] = { .name = n, .entry = &(func) } 1901 1902 static struct sort_dimension memory_sort_dimensions[] = { 1903 DIM(SORT_MEM_DADDR_SYMBOL, "symbol_daddr", sort_mem_daddr_sym), 1904 DIM(SORT_MEM_IADDR_SYMBOL, "symbol_iaddr", sort_mem_iaddr_sym), 1905 DIM(SORT_MEM_DADDR_DSO, "dso_daddr", sort_mem_daddr_dso), 1906 DIM(SORT_MEM_LOCKED, "locked", sort_mem_locked), 1907 DIM(SORT_MEM_TLB, "tlb", sort_mem_tlb), 1908 DIM(SORT_MEM_LVL, "mem", sort_mem_lvl), 1909 DIM(SORT_MEM_SNOOP, "snoop", sort_mem_snoop), 1910 DIM(SORT_MEM_DCACHELINE, "dcacheline", sort_mem_dcacheline), 1911 DIM(SORT_MEM_PHYS_DADDR, "phys_daddr", sort_mem_phys_daddr), 1912 DIM(SORT_MEM_DATA_PAGE_SIZE, "data_page_size", sort_mem_data_page_size), 1913 DIM(SORT_MEM_BLOCKED, "blocked", sort_mem_blocked), 1914 }; 1915 1916 #undef DIM 1917 1918 struct hpp_dimension { 1919 const char *name; 1920 struct perf_hpp_fmt *fmt; 1921 int taken; 1922 }; 1923 1924 #define DIM(d, n) { .name = n, .fmt = &perf_hpp__format[d], } 1925 1926 static struct hpp_dimension hpp_sort_dimensions[] = { 1927 DIM(PERF_HPP__OVERHEAD, "overhead"), 1928 DIM(PERF_HPP__OVERHEAD_SYS, "overhead_sys"), 1929 DIM(PERF_HPP__OVERHEAD_US, "overhead_us"), 1930 DIM(PERF_HPP__OVERHEAD_GUEST_SYS, "overhead_guest_sys"), 1931 DIM(PERF_HPP__OVERHEAD_GUEST_US, "overhead_guest_us"), 1932 DIM(PERF_HPP__OVERHEAD_ACC, "overhead_children"), 1933 DIM(PERF_HPP__SAMPLES, "sample"), 1934 DIM(PERF_HPP__PERIOD, "period"), 1935 }; 1936 1937 #undef DIM 1938 1939 struct hpp_sort_entry { 1940 struct perf_hpp_fmt hpp; 1941 struct sort_entry *se; 1942 }; 1943 1944 void perf_hpp__reset_sort_width(struct perf_hpp_fmt *fmt, struct hists *hists) 1945 { 1946 struct hpp_sort_entry *hse; 1947 1948 if (!perf_hpp__is_sort_entry(fmt)) 1949 return; 1950 1951 hse = container_of(fmt, struct hpp_sort_entry, hpp); 1952 hists__new_col_len(hists, hse->se->se_width_idx, strlen(fmt->name)); 1953 } 1954 1955 static int __sort__hpp_header(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, 1956 struct hists *hists, int line __maybe_unused, 1957 int *span __maybe_unused) 1958 { 1959 struct hpp_sort_entry *hse; 1960 size_t len = fmt->user_len; 1961 1962 hse = container_of(fmt, struct hpp_sort_entry, hpp); 1963 1964 if (!len) 1965 len = hists__col_len(hists, hse->se->se_width_idx); 1966 1967 return scnprintf(hpp->buf, hpp->size, "%-*.*s", len, len, fmt->name); 1968 } 1969 1970 static int __sort__hpp_width(struct perf_hpp_fmt *fmt, 1971 struct perf_hpp *hpp __maybe_unused, 1972 struct hists *hists) 1973 { 1974 struct hpp_sort_entry *hse; 1975 size_t len = fmt->user_len; 1976 1977 hse = container_of(fmt, struct hpp_sort_entry, hpp); 1978 1979 if (!len) 1980 len = hists__col_len(hists, hse->se->se_width_idx); 1981 1982 return len; 1983 } 1984 1985 static int __sort__hpp_entry(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, 1986 struct hist_entry *he) 1987 { 1988 struct hpp_sort_entry *hse; 1989 size_t len = fmt->user_len; 1990 1991 hse = container_of(fmt, struct hpp_sort_entry, hpp); 1992 1993 if (!len) 1994 len = hists__col_len(he->hists, hse->se->se_width_idx); 1995 1996 return hse->se->se_snprintf(he, hpp->buf, hpp->size, len); 1997 } 1998 1999 static int64_t __sort__hpp_cmp(struct perf_hpp_fmt *fmt, 2000 struct hist_entry *a, struct hist_entry *b) 2001 { 2002 struct hpp_sort_entry *hse; 2003 2004 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2005 return hse->se->se_cmp(a, b); 2006 } 2007 2008 static int64_t __sort__hpp_collapse(struct perf_hpp_fmt *fmt, 2009 struct hist_entry *a, struct hist_entry *b) 2010 { 2011 struct hpp_sort_entry *hse; 2012 int64_t (*collapse_fn)(struct hist_entry *, struct hist_entry *); 2013 2014 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2015 collapse_fn = hse->se->se_collapse ?: hse->se->se_cmp; 2016 return collapse_fn(a, b); 2017 } 2018 2019 static int64_t __sort__hpp_sort(struct perf_hpp_fmt *fmt, 2020 struct hist_entry *a, struct hist_entry *b) 2021 { 2022 struct hpp_sort_entry *hse; 2023 int64_t (*sort_fn)(struct hist_entry *, struct hist_entry *); 2024 2025 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2026 sort_fn = hse->se->se_sort ?: hse->se->se_cmp; 2027 return sort_fn(a, b); 2028 } 2029 2030 bool perf_hpp__is_sort_entry(struct perf_hpp_fmt *format) 2031 { 2032 return format->header == __sort__hpp_header; 2033 } 2034 2035 #define MK_SORT_ENTRY_CHK(key) \ 2036 bool perf_hpp__is_ ## key ## _entry(struct perf_hpp_fmt *fmt) \ 2037 { \ 2038 struct hpp_sort_entry *hse; \ 2039 \ 2040 if (!perf_hpp__is_sort_entry(fmt)) \ 2041 return false; \ 2042 \ 2043 hse = container_of(fmt, struct hpp_sort_entry, hpp); \ 2044 return hse->se == &sort_ ## key ; \ 2045 } 2046 2047 MK_SORT_ENTRY_CHK(trace) 2048 MK_SORT_ENTRY_CHK(srcline) 2049 MK_SORT_ENTRY_CHK(srcfile) 2050 MK_SORT_ENTRY_CHK(thread) 2051 MK_SORT_ENTRY_CHK(comm) 2052 MK_SORT_ENTRY_CHK(dso) 2053 MK_SORT_ENTRY_CHK(sym) 2054 2055 2056 static bool __sort__hpp_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) 2057 { 2058 struct hpp_sort_entry *hse_a; 2059 struct hpp_sort_entry *hse_b; 2060 2061 if (!perf_hpp__is_sort_entry(a) || !perf_hpp__is_sort_entry(b)) 2062 return false; 2063 2064 hse_a = container_of(a, struct hpp_sort_entry, hpp); 2065 hse_b = container_of(b, struct hpp_sort_entry, hpp); 2066 2067 return hse_a->se == hse_b->se; 2068 } 2069 2070 static void hse_free(struct perf_hpp_fmt *fmt) 2071 { 2072 struct hpp_sort_entry *hse; 2073 2074 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2075 free(hse); 2076 } 2077 2078 static struct hpp_sort_entry * 2079 __sort_dimension__alloc_hpp(struct sort_dimension *sd, int level) 2080 { 2081 struct hpp_sort_entry *hse; 2082 2083 hse = malloc(sizeof(*hse)); 2084 if (hse == NULL) { 2085 pr_err("Memory allocation failed\n"); 2086 return NULL; 2087 } 2088 2089 hse->se = sd->entry; 2090 hse->hpp.name = sd->entry->se_header; 2091 hse->hpp.header = __sort__hpp_header; 2092 hse->hpp.width = __sort__hpp_width; 2093 hse->hpp.entry = __sort__hpp_entry; 2094 hse->hpp.color = NULL; 2095 2096 hse->hpp.cmp = __sort__hpp_cmp; 2097 hse->hpp.collapse = __sort__hpp_collapse; 2098 hse->hpp.sort = __sort__hpp_sort; 2099 hse->hpp.equal = __sort__hpp_equal; 2100 hse->hpp.free = hse_free; 2101 2102 INIT_LIST_HEAD(&hse->hpp.list); 2103 INIT_LIST_HEAD(&hse->hpp.sort_list); 2104 hse->hpp.elide = false; 2105 hse->hpp.len = 0; 2106 hse->hpp.user_len = 0; 2107 hse->hpp.level = level; 2108 2109 return hse; 2110 } 2111 2112 static void hpp_free(struct perf_hpp_fmt *fmt) 2113 { 2114 free(fmt); 2115 } 2116 2117 static struct perf_hpp_fmt *__hpp_dimension__alloc_hpp(struct hpp_dimension *hd, 2118 int level) 2119 { 2120 struct perf_hpp_fmt *fmt; 2121 2122 fmt = memdup(hd->fmt, sizeof(*fmt)); 2123 if (fmt) { 2124 INIT_LIST_HEAD(&fmt->list); 2125 INIT_LIST_HEAD(&fmt->sort_list); 2126 fmt->free = hpp_free; 2127 fmt->level = level; 2128 } 2129 2130 return fmt; 2131 } 2132 2133 int hist_entry__filter(struct hist_entry *he, int type, const void *arg) 2134 { 2135 struct perf_hpp_fmt *fmt; 2136 struct hpp_sort_entry *hse; 2137 int ret = -1; 2138 int r; 2139 2140 perf_hpp_list__for_each_format(he->hpp_list, fmt) { 2141 if (!perf_hpp__is_sort_entry(fmt)) 2142 continue; 2143 2144 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2145 if (hse->se->se_filter == NULL) 2146 continue; 2147 2148 /* 2149 * hist entry is filtered if any of sort key in the hpp list 2150 * is applied. But it should skip non-matched filter types. 2151 */ 2152 r = hse->se->se_filter(he, type, arg); 2153 if (r >= 0) { 2154 if (ret < 0) 2155 ret = 0; 2156 ret |= r; 2157 } 2158 } 2159 2160 return ret; 2161 } 2162 2163 static int __sort_dimension__add_hpp_sort(struct sort_dimension *sd, 2164 struct perf_hpp_list *list, 2165 int level) 2166 { 2167 struct hpp_sort_entry *hse = __sort_dimension__alloc_hpp(sd, level); 2168 2169 if (hse == NULL) 2170 return -1; 2171 2172 perf_hpp_list__register_sort_field(list, &hse->hpp); 2173 return 0; 2174 } 2175 2176 static int __sort_dimension__add_hpp_output(struct sort_dimension *sd, 2177 struct perf_hpp_list *list) 2178 { 2179 struct hpp_sort_entry *hse = __sort_dimension__alloc_hpp(sd, 0); 2180 2181 if (hse == NULL) 2182 return -1; 2183 2184 perf_hpp_list__column_register(list, &hse->hpp); 2185 return 0; 2186 } 2187 2188 struct hpp_dynamic_entry { 2189 struct perf_hpp_fmt hpp; 2190 struct evsel *evsel; 2191 struct tep_format_field *field; 2192 unsigned dynamic_len; 2193 bool raw_trace; 2194 }; 2195 2196 static int hde_width(struct hpp_dynamic_entry *hde) 2197 { 2198 if (!hde->hpp.len) { 2199 int len = hde->dynamic_len; 2200 int namelen = strlen(hde->field->name); 2201 int fieldlen = hde->field->size; 2202 2203 if (namelen > len) 2204 len = namelen; 2205 2206 if (!(hde->field->flags & TEP_FIELD_IS_STRING)) { 2207 /* length for print hex numbers */ 2208 fieldlen = hde->field->size * 2 + 2; 2209 } 2210 if (fieldlen > len) 2211 len = fieldlen; 2212 2213 hde->hpp.len = len; 2214 } 2215 return hde->hpp.len; 2216 } 2217 2218 static void update_dynamic_len(struct hpp_dynamic_entry *hde, 2219 struct hist_entry *he) 2220 { 2221 char *str, *pos; 2222 struct tep_format_field *field = hde->field; 2223 size_t namelen; 2224 bool last = false; 2225 2226 if (hde->raw_trace) 2227 return; 2228 2229 /* parse pretty print result and update max length */ 2230 if (!he->trace_output) 2231 he->trace_output = get_trace_output(he); 2232 2233 namelen = strlen(field->name); 2234 str = he->trace_output; 2235 2236 while (str) { 2237 pos = strchr(str, ' '); 2238 if (pos == NULL) { 2239 last = true; 2240 pos = str + strlen(str); 2241 } 2242 2243 if (!strncmp(str, field->name, namelen)) { 2244 size_t len; 2245 2246 str += namelen + 1; 2247 len = pos - str; 2248 2249 if (len > hde->dynamic_len) 2250 hde->dynamic_len = len; 2251 break; 2252 } 2253 2254 if (last) 2255 str = NULL; 2256 else 2257 str = pos + 1; 2258 } 2259 } 2260 2261 static int __sort__hde_header(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, 2262 struct hists *hists __maybe_unused, 2263 int line __maybe_unused, 2264 int *span __maybe_unused) 2265 { 2266 struct hpp_dynamic_entry *hde; 2267 size_t len = fmt->user_len; 2268 2269 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2270 2271 if (!len) 2272 len = hde_width(hde); 2273 2274 return scnprintf(hpp->buf, hpp->size, "%*.*s", len, len, hde->field->name); 2275 } 2276 2277 static int __sort__hde_width(struct perf_hpp_fmt *fmt, 2278 struct perf_hpp *hpp __maybe_unused, 2279 struct hists *hists __maybe_unused) 2280 { 2281 struct hpp_dynamic_entry *hde; 2282 size_t len = fmt->user_len; 2283 2284 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2285 2286 if (!len) 2287 len = hde_width(hde); 2288 2289 return len; 2290 } 2291 2292 bool perf_hpp__defined_dynamic_entry(struct perf_hpp_fmt *fmt, struct hists *hists) 2293 { 2294 struct hpp_dynamic_entry *hde; 2295 2296 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2297 2298 return hists_to_evsel(hists) == hde->evsel; 2299 } 2300 2301 static int __sort__hde_entry(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp, 2302 struct hist_entry *he) 2303 { 2304 struct hpp_dynamic_entry *hde; 2305 size_t len = fmt->user_len; 2306 char *str, *pos; 2307 struct tep_format_field *field; 2308 size_t namelen; 2309 bool last = false; 2310 int ret; 2311 2312 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2313 2314 if (!len) 2315 len = hde_width(hde); 2316 2317 if (hde->raw_trace) 2318 goto raw_field; 2319 2320 if (!he->trace_output) 2321 he->trace_output = get_trace_output(he); 2322 2323 field = hde->field; 2324 namelen = strlen(field->name); 2325 str = he->trace_output; 2326 2327 while (str) { 2328 pos = strchr(str, ' '); 2329 if (pos == NULL) { 2330 last = true; 2331 pos = str + strlen(str); 2332 } 2333 2334 if (!strncmp(str, field->name, namelen)) { 2335 str += namelen + 1; 2336 str = strndup(str, pos - str); 2337 2338 if (str == NULL) 2339 return scnprintf(hpp->buf, hpp->size, 2340 "%*.*s", len, len, "ERROR"); 2341 break; 2342 } 2343 2344 if (last) 2345 str = NULL; 2346 else 2347 str = pos + 1; 2348 } 2349 2350 if (str == NULL) { 2351 struct trace_seq seq; 2352 raw_field: 2353 trace_seq_init(&seq); 2354 tep_print_field(&seq, he->raw_data, hde->field); 2355 str = seq.buffer; 2356 } 2357 2358 ret = scnprintf(hpp->buf, hpp->size, "%*.*s", len, len, str); 2359 free(str); 2360 return ret; 2361 } 2362 2363 static int64_t __sort__hde_cmp(struct perf_hpp_fmt *fmt, 2364 struct hist_entry *a, struct hist_entry *b) 2365 { 2366 struct hpp_dynamic_entry *hde; 2367 struct tep_format_field *field; 2368 unsigned offset, size; 2369 2370 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2371 2372 if (b == NULL) { 2373 update_dynamic_len(hde, a); 2374 return 0; 2375 } 2376 2377 field = hde->field; 2378 if (field->flags & TEP_FIELD_IS_DYNAMIC) { 2379 unsigned long long dyn; 2380 2381 tep_read_number_field(field, a->raw_data, &dyn); 2382 offset = dyn & 0xffff; 2383 size = (dyn >> 16) & 0xffff; 2384 if (field->flags & TEP_FIELD_IS_RELATIVE) 2385 offset += field->offset + field->size; 2386 2387 /* record max width for output */ 2388 if (size > hde->dynamic_len) 2389 hde->dynamic_len = size; 2390 } else { 2391 offset = field->offset; 2392 size = field->size; 2393 } 2394 2395 return memcmp(a->raw_data + offset, b->raw_data + offset, size); 2396 } 2397 2398 bool perf_hpp__is_dynamic_entry(struct perf_hpp_fmt *fmt) 2399 { 2400 return fmt->cmp == __sort__hde_cmp; 2401 } 2402 2403 static bool __sort__hde_equal(struct perf_hpp_fmt *a, struct perf_hpp_fmt *b) 2404 { 2405 struct hpp_dynamic_entry *hde_a; 2406 struct hpp_dynamic_entry *hde_b; 2407 2408 if (!perf_hpp__is_dynamic_entry(a) || !perf_hpp__is_dynamic_entry(b)) 2409 return false; 2410 2411 hde_a = container_of(a, struct hpp_dynamic_entry, hpp); 2412 hde_b = container_of(b, struct hpp_dynamic_entry, hpp); 2413 2414 return hde_a->field == hde_b->field; 2415 } 2416 2417 static void hde_free(struct perf_hpp_fmt *fmt) 2418 { 2419 struct hpp_dynamic_entry *hde; 2420 2421 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2422 free(hde); 2423 } 2424 2425 static struct hpp_dynamic_entry * 2426 __alloc_dynamic_entry(struct evsel *evsel, struct tep_format_field *field, 2427 int level) 2428 { 2429 struct hpp_dynamic_entry *hde; 2430 2431 hde = malloc(sizeof(*hde)); 2432 if (hde == NULL) { 2433 pr_debug("Memory allocation failed\n"); 2434 return NULL; 2435 } 2436 2437 hde->evsel = evsel; 2438 hde->field = field; 2439 hde->dynamic_len = 0; 2440 2441 hde->hpp.name = field->name; 2442 hde->hpp.header = __sort__hde_header; 2443 hde->hpp.width = __sort__hde_width; 2444 hde->hpp.entry = __sort__hde_entry; 2445 hde->hpp.color = NULL; 2446 2447 hde->hpp.cmp = __sort__hde_cmp; 2448 hde->hpp.collapse = __sort__hde_cmp; 2449 hde->hpp.sort = __sort__hde_cmp; 2450 hde->hpp.equal = __sort__hde_equal; 2451 hde->hpp.free = hde_free; 2452 2453 INIT_LIST_HEAD(&hde->hpp.list); 2454 INIT_LIST_HEAD(&hde->hpp.sort_list); 2455 hde->hpp.elide = false; 2456 hde->hpp.len = 0; 2457 hde->hpp.user_len = 0; 2458 hde->hpp.level = level; 2459 2460 return hde; 2461 } 2462 2463 struct perf_hpp_fmt *perf_hpp_fmt__dup(struct perf_hpp_fmt *fmt) 2464 { 2465 struct perf_hpp_fmt *new_fmt = NULL; 2466 2467 if (perf_hpp__is_sort_entry(fmt)) { 2468 struct hpp_sort_entry *hse, *new_hse; 2469 2470 hse = container_of(fmt, struct hpp_sort_entry, hpp); 2471 new_hse = memdup(hse, sizeof(*hse)); 2472 if (new_hse) 2473 new_fmt = &new_hse->hpp; 2474 } else if (perf_hpp__is_dynamic_entry(fmt)) { 2475 struct hpp_dynamic_entry *hde, *new_hde; 2476 2477 hde = container_of(fmt, struct hpp_dynamic_entry, hpp); 2478 new_hde = memdup(hde, sizeof(*hde)); 2479 if (new_hde) 2480 new_fmt = &new_hde->hpp; 2481 } else { 2482 new_fmt = memdup(fmt, sizeof(*fmt)); 2483 } 2484 2485 INIT_LIST_HEAD(&new_fmt->list); 2486 INIT_LIST_HEAD(&new_fmt->sort_list); 2487 2488 return new_fmt; 2489 } 2490 2491 static int parse_field_name(char *str, char **event, char **field, char **opt) 2492 { 2493 char *event_name, *field_name, *opt_name; 2494 2495 event_name = str; 2496 field_name = strchr(str, '.'); 2497 2498 if (field_name) { 2499 *field_name++ = '\0'; 2500 } else { 2501 event_name = NULL; 2502 field_name = str; 2503 } 2504 2505 opt_name = strchr(field_name, '/'); 2506 if (opt_name) 2507 *opt_name++ = '\0'; 2508 2509 *event = event_name; 2510 *field = field_name; 2511 *opt = opt_name; 2512 2513 return 0; 2514 } 2515 2516 /* find match evsel using a given event name. The event name can be: 2517 * 1. '%' + event index (e.g. '%1' for first event) 2518 * 2. full event name (e.g. sched:sched_switch) 2519 * 3. partial event name (should not contain ':') 2520 */ 2521 static struct evsel *find_evsel(struct evlist *evlist, char *event_name) 2522 { 2523 struct evsel *evsel = NULL; 2524 struct evsel *pos; 2525 bool full_name; 2526 2527 /* case 1 */ 2528 if (event_name[0] == '%') { 2529 int nr = strtol(event_name+1, NULL, 0); 2530 2531 if (nr > evlist->core.nr_entries) 2532 return NULL; 2533 2534 evsel = evlist__first(evlist); 2535 while (--nr > 0) 2536 evsel = evsel__next(evsel); 2537 2538 return evsel; 2539 } 2540 2541 full_name = !!strchr(event_name, ':'); 2542 evlist__for_each_entry(evlist, pos) { 2543 /* case 2 */ 2544 if (full_name && !strcmp(pos->name, event_name)) 2545 return pos; 2546 /* case 3 */ 2547 if (!full_name && strstr(pos->name, event_name)) { 2548 if (evsel) { 2549 pr_debug("'%s' event is ambiguous: it can be %s or %s\n", 2550 event_name, evsel->name, pos->name); 2551 return NULL; 2552 } 2553 evsel = pos; 2554 } 2555 } 2556 2557 return evsel; 2558 } 2559 2560 static int __dynamic_dimension__add(struct evsel *evsel, 2561 struct tep_format_field *field, 2562 bool raw_trace, int level) 2563 { 2564 struct hpp_dynamic_entry *hde; 2565 2566 hde = __alloc_dynamic_entry(evsel, field, level); 2567 if (hde == NULL) 2568 return -ENOMEM; 2569 2570 hde->raw_trace = raw_trace; 2571 2572 perf_hpp__register_sort_field(&hde->hpp); 2573 return 0; 2574 } 2575 2576 static int add_evsel_fields(struct evsel *evsel, bool raw_trace, int level) 2577 { 2578 int ret; 2579 struct tep_format_field *field; 2580 2581 field = evsel->tp_format->format.fields; 2582 while (field) { 2583 ret = __dynamic_dimension__add(evsel, field, raw_trace, level); 2584 if (ret < 0) 2585 return ret; 2586 2587 field = field->next; 2588 } 2589 return 0; 2590 } 2591 2592 static int add_all_dynamic_fields(struct evlist *evlist, bool raw_trace, 2593 int level) 2594 { 2595 int ret; 2596 struct evsel *evsel; 2597 2598 evlist__for_each_entry(evlist, evsel) { 2599 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) 2600 continue; 2601 2602 ret = add_evsel_fields(evsel, raw_trace, level); 2603 if (ret < 0) 2604 return ret; 2605 } 2606 return 0; 2607 } 2608 2609 static int add_all_matching_fields(struct evlist *evlist, 2610 char *field_name, bool raw_trace, int level) 2611 { 2612 int ret = -ESRCH; 2613 struct evsel *evsel; 2614 struct tep_format_field *field; 2615 2616 evlist__for_each_entry(evlist, evsel) { 2617 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) 2618 continue; 2619 2620 field = tep_find_any_field(evsel->tp_format, field_name); 2621 if (field == NULL) 2622 continue; 2623 2624 ret = __dynamic_dimension__add(evsel, field, raw_trace, level); 2625 if (ret < 0) 2626 break; 2627 } 2628 return ret; 2629 } 2630 2631 static int add_dynamic_entry(struct evlist *evlist, const char *tok, 2632 int level) 2633 { 2634 char *str, *event_name, *field_name, *opt_name; 2635 struct evsel *evsel; 2636 struct tep_format_field *field; 2637 bool raw_trace = symbol_conf.raw_trace; 2638 int ret = 0; 2639 2640 if (evlist == NULL) 2641 return -ENOENT; 2642 2643 str = strdup(tok); 2644 if (str == NULL) 2645 return -ENOMEM; 2646 2647 if (parse_field_name(str, &event_name, &field_name, &opt_name) < 0) { 2648 ret = -EINVAL; 2649 goto out; 2650 } 2651 2652 if (opt_name) { 2653 if (strcmp(opt_name, "raw")) { 2654 pr_debug("unsupported field option %s\n", opt_name); 2655 ret = -EINVAL; 2656 goto out; 2657 } 2658 raw_trace = true; 2659 } 2660 2661 if (!strcmp(field_name, "trace_fields")) { 2662 ret = add_all_dynamic_fields(evlist, raw_trace, level); 2663 goto out; 2664 } 2665 2666 if (event_name == NULL) { 2667 ret = add_all_matching_fields(evlist, field_name, raw_trace, level); 2668 goto out; 2669 } 2670 2671 evsel = find_evsel(evlist, event_name); 2672 if (evsel == NULL) { 2673 pr_debug("Cannot find event: %s\n", event_name); 2674 ret = -ENOENT; 2675 goto out; 2676 } 2677 2678 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) { 2679 pr_debug("%s is not a tracepoint event\n", event_name); 2680 ret = -EINVAL; 2681 goto out; 2682 } 2683 2684 if (!strcmp(field_name, "*")) { 2685 ret = add_evsel_fields(evsel, raw_trace, level); 2686 } else { 2687 field = tep_find_any_field(evsel->tp_format, field_name); 2688 if (field == NULL) { 2689 pr_debug("Cannot find event field for %s.%s\n", 2690 event_name, field_name); 2691 return -ENOENT; 2692 } 2693 2694 ret = __dynamic_dimension__add(evsel, field, raw_trace, level); 2695 } 2696 2697 out: 2698 free(str); 2699 return ret; 2700 } 2701 2702 static int __sort_dimension__add(struct sort_dimension *sd, 2703 struct perf_hpp_list *list, 2704 int level) 2705 { 2706 if (sd->taken) 2707 return 0; 2708 2709 if (__sort_dimension__add_hpp_sort(sd, list, level) < 0) 2710 return -1; 2711 2712 if (sd->entry->se_collapse) 2713 list->need_collapse = 1; 2714 2715 sd->taken = 1; 2716 2717 return 0; 2718 } 2719 2720 static int __hpp_dimension__add(struct hpp_dimension *hd, 2721 struct perf_hpp_list *list, 2722 int level) 2723 { 2724 struct perf_hpp_fmt *fmt; 2725 2726 if (hd->taken) 2727 return 0; 2728 2729 fmt = __hpp_dimension__alloc_hpp(hd, level); 2730 if (!fmt) 2731 return -1; 2732 2733 hd->taken = 1; 2734 perf_hpp_list__register_sort_field(list, fmt); 2735 return 0; 2736 } 2737 2738 static int __sort_dimension__add_output(struct perf_hpp_list *list, 2739 struct sort_dimension *sd) 2740 { 2741 if (sd->taken) 2742 return 0; 2743 2744 if (__sort_dimension__add_hpp_output(sd, list) < 0) 2745 return -1; 2746 2747 sd->taken = 1; 2748 return 0; 2749 } 2750 2751 static int __hpp_dimension__add_output(struct perf_hpp_list *list, 2752 struct hpp_dimension *hd) 2753 { 2754 struct perf_hpp_fmt *fmt; 2755 2756 if (hd->taken) 2757 return 0; 2758 2759 fmt = __hpp_dimension__alloc_hpp(hd, 0); 2760 if (!fmt) 2761 return -1; 2762 2763 hd->taken = 1; 2764 perf_hpp_list__column_register(list, fmt); 2765 return 0; 2766 } 2767 2768 int hpp_dimension__add_output(unsigned col) 2769 { 2770 BUG_ON(col >= PERF_HPP__MAX_INDEX); 2771 return __hpp_dimension__add_output(&perf_hpp_list, &hpp_sort_dimensions[col]); 2772 } 2773 2774 int sort_dimension__add(struct perf_hpp_list *list, const char *tok, 2775 struct evlist *evlist, 2776 int level) 2777 { 2778 unsigned int i, j; 2779 2780 /* 2781 * Check to see if there are any arch specific 2782 * sort dimensions not applicable for the current 2783 * architecture. If so, Skip that sort key since 2784 * we don't want to display it in the output fields. 2785 */ 2786 for (j = 0; j < ARRAY_SIZE(arch_specific_sort_keys); j++) { 2787 if (!strcmp(arch_specific_sort_keys[j], tok) && 2788 !arch_support_sort_key(tok)) { 2789 return 0; 2790 } 2791 } 2792 2793 for (i = 0; i < ARRAY_SIZE(common_sort_dimensions); i++) { 2794 struct sort_dimension *sd = &common_sort_dimensions[i]; 2795 2796 if (strncasecmp(tok, sd->name, strlen(tok))) 2797 continue; 2798 2799 for (j = 0; j < ARRAY_SIZE(dynamic_headers); j++) { 2800 if (!strcmp(dynamic_headers[j], sd->name)) 2801 sort_dimension_add_dynamic_header(sd); 2802 } 2803 2804 if (sd->entry == &sort_parent) { 2805 int ret = regcomp(&parent_regex, parent_pattern, REG_EXTENDED); 2806 if (ret) { 2807 char err[BUFSIZ]; 2808 2809 regerror(ret, &parent_regex, err, sizeof(err)); 2810 pr_err("Invalid regex: %s\n%s", parent_pattern, err); 2811 return -EINVAL; 2812 } 2813 list->parent = 1; 2814 } else if (sd->entry == &sort_sym) { 2815 list->sym = 1; 2816 /* 2817 * perf diff displays the performance difference amongst 2818 * two or more perf.data files. Those files could come 2819 * from different binaries. So we should not compare 2820 * their ips, but the name of symbol. 2821 */ 2822 if (sort__mode == SORT_MODE__DIFF) 2823 sd->entry->se_collapse = sort__sym_sort; 2824 2825 } else if (sd->entry == &sort_dso) { 2826 list->dso = 1; 2827 } else if (sd->entry == &sort_socket) { 2828 list->socket = 1; 2829 } else if (sd->entry == &sort_thread) { 2830 list->thread = 1; 2831 } else if (sd->entry == &sort_comm) { 2832 list->comm = 1; 2833 } 2834 2835 return __sort_dimension__add(sd, list, level); 2836 } 2837 2838 for (i = 0; i < ARRAY_SIZE(hpp_sort_dimensions); i++) { 2839 struct hpp_dimension *hd = &hpp_sort_dimensions[i]; 2840 2841 if (strncasecmp(tok, hd->name, strlen(tok))) 2842 continue; 2843 2844 return __hpp_dimension__add(hd, list, level); 2845 } 2846 2847 for (i = 0; i < ARRAY_SIZE(bstack_sort_dimensions); i++) { 2848 struct sort_dimension *sd = &bstack_sort_dimensions[i]; 2849 2850 if (strncasecmp(tok, sd->name, strlen(tok))) 2851 continue; 2852 2853 if (sort__mode != SORT_MODE__BRANCH) 2854 return -EINVAL; 2855 2856 if (sd->entry == &sort_sym_from || sd->entry == &sort_sym_to) 2857 list->sym = 1; 2858 2859 __sort_dimension__add(sd, list, level); 2860 return 0; 2861 } 2862 2863 for (i = 0; i < ARRAY_SIZE(memory_sort_dimensions); i++) { 2864 struct sort_dimension *sd = &memory_sort_dimensions[i]; 2865 2866 if (strncasecmp(tok, sd->name, strlen(tok))) 2867 continue; 2868 2869 if (sort__mode != SORT_MODE__MEMORY) 2870 return -EINVAL; 2871 2872 if (sd->entry == &sort_mem_dcacheline && cacheline_size() == 0) 2873 return -EINVAL; 2874 2875 if (sd->entry == &sort_mem_daddr_sym) 2876 list->sym = 1; 2877 2878 __sort_dimension__add(sd, list, level); 2879 return 0; 2880 } 2881 2882 if (!add_dynamic_entry(evlist, tok, level)) 2883 return 0; 2884 2885 return -ESRCH; 2886 } 2887 2888 static int setup_sort_list(struct perf_hpp_list *list, char *str, 2889 struct evlist *evlist) 2890 { 2891 char *tmp, *tok; 2892 int ret = 0; 2893 int level = 0; 2894 int next_level = 1; 2895 bool in_group = false; 2896 2897 do { 2898 tok = str; 2899 tmp = strpbrk(str, "{}, "); 2900 if (tmp) { 2901 if (in_group) 2902 next_level = level; 2903 else 2904 next_level = level + 1; 2905 2906 if (*tmp == '{') 2907 in_group = true; 2908 else if (*tmp == '}') 2909 in_group = false; 2910 2911 *tmp = '\0'; 2912 str = tmp + 1; 2913 } 2914 2915 if (*tok) { 2916 ret = sort_dimension__add(list, tok, evlist, level); 2917 if (ret == -EINVAL) { 2918 if (!cacheline_size() && !strncasecmp(tok, "dcacheline", strlen(tok))) 2919 ui__error("The \"dcacheline\" --sort key needs to know the cacheline size and it couldn't be determined on this system"); 2920 else 2921 ui__error("Invalid --sort key: `%s'", tok); 2922 break; 2923 } else if (ret == -ESRCH) { 2924 ui__error("Unknown --sort key: `%s'", tok); 2925 break; 2926 } 2927 } 2928 2929 level = next_level; 2930 } while (tmp); 2931 2932 return ret; 2933 } 2934 2935 static const char *get_default_sort_order(struct evlist *evlist) 2936 { 2937 const char *default_sort_orders[] = { 2938 default_sort_order, 2939 default_branch_sort_order, 2940 default_mem_sort_order, 2941 default_top_sort_order, 2942 default_diff_sort_order, 2943 default_tracepoint_sort_order, 2944 }; 2945 bool use_trace = true; 2946 struct evsel *evsel; 2947 2948 BUG_ON(sort__mode >= ARRAY_SIZE(default_sort_orders)); 2949 2950 if (evlist == NULL || evlist__empty(evlist)) 2951 goto out_no_evlist; 2952 2953 evlist__for_each_entry(evlist, evsel) { 2954 if (evsel->core.attr.type != PERF_TYPE_TRACEPOINT) { 2955 use_trace = false; 2956 break; 2957 } 2958 } 2959 2960 if (use_trace) { 2961 sort__mode = SORT_MODE__TRACEPOINT; 2962 if (symbol_conf.raw_trace) 2963 return "trace_fields"; 2964 } 2965 out_no_evlist: 2966 return default_sort_orders[sort__mode]; 2967 } 2968 2969 static int setup_sort_order(struct evlist *evlist) 2970 { 2971 char *new_sort_order; 2972 2973 /* 2974 * Append '+'-prefixed sort order to the default sort 2975 * order string. 2976 */ 2977 if (!sort_order || is_strict_order(sort_order)) 2978 return 0; 2979 2980 if (sort_order[1] == '\0') { 2981 ui__error("Invalid --sort key: `+'"); 2982 return -EINVAL; 2983 } 2984 2985 /* 2986 * We allocate new sort_order string, but we never free it, 2987 * because it's checked over the rest of the code. 2988 */ 2989 if (asprintf(&new_sort_order, "%s,%s", 2990 get_default_sort_order(evlist), sort_order + 1) < 0) { 2991 pr_err("Not enough memory to set up --sort"); 2992 return -ENOMEM; 2993 } 2994 2995 sort_order = new_sort_order; 2996 return 0; 2997 } 2998 2999 /* 3000 * Adds 'pre,' prefix into 'str' is 'pre' is 3001 * not already part of 'str'. 3002 */ 3003 static char *prefix_if_not_in(const char *pre, char *str) 3004 { 3005 char *n; 3006 3007 if (!str || strstr(str, pre)) 3008 return str; 3009 3010 if (asprintf(&n, "%s,%s", pre, str) < 0) 3011 n = NULL; 3012 3013 free(str); 3014 return n; 3015 } 3016 3017 static char *setup_overhead(char *keys) 3018 { 3019 if (sort__mode == SORT_MODE__DIFF) 3020 return keys; 3021 3022 keys = prefix_if_not_in("overhead", keys); 3023 3024 if (symbol_conf.cumulate_callchain) 3025 keys = prefix_if_not_in("overhead_children", keys); 3026 3027 return keys; 3028 } 3029 3030 static int __setup_sorting(struct evlist *evlist) 3031 { 3032 char *str; 3033 const char *sort_keys; 3034 int ret = 0; 3035 3036 ret = setup_sort_order(evlist); 3037 if (ret) 3038 return ret; 3039 3040 sort_keys = sort_order; 3041 if (sort_keys == NULL) { 3042 if (is_strict_order(field_order)) { 3043 /* 3044 * If user specified field order but no sort order, 3045 * we'll honor it and not add default sort orders. 3046 */ 3047 return 0; 3048 } 3049 3050 sort_keys = get_default_sort_order(evlist); 3051 } 3052 3053 str = strdup(sort_keys); 3054 if (str == NULL) { 3055 pr_err("Not enough memory to setup sort keys"); 3056 return -ENOMEM; 3057 } 3058 3059 /* 3060 * Prepend overhead fields for backward compatibility. 3061 */ 3062 if (!is_strict_order(field_order)) { 3063 str = setup_overhead(str); 3064 if (str == NULL) { 3065 pr_err("Not enough memory to setup overhead keys"); 3066 return -ENOMEM; 3067 } 3068 } 3069 3070 ret = setup_sort_list(&perf_hpp_list, str, evlist); 3071 3072 free(str); 3073 return ret; 3074 } 3075 3076 void perf_hpp__set_elide(int idx, bool elide) 3077 { 3078 struct perf_hpp_fmt *fmt; 3079 struct hpp_sort_entry *hse; 3080 3081 perf_hpp_list__for_each_format(&perf_hpp_list, fmt) { 3082 if (!perf_hpp__is_sort_entry(fmt)) 3083 continue; 3084 3085 hse = container_of(fmt, struct hpp_sort_entry, hpp); 3086 if (hse->se->se_width_idx == idx) { 3087 fmt->elide = elide; 3088 break; 3089 } 3090 } 3091 } 3092 3093 static bool __get_elide(struct strlist *list, const char *list_name, FILE *fp) 3094 { 3095 if (list && strlist__nr_entries(list) == 1) { 3096 if (fp != NULL) 3097 fprintf(fp, "# %s: %s\n", list_name, 3098 strlist__entry(list, 0)->s); 3099 return true; 3100 } 3101 return false; 3102 } 3103 3104 static bool get_elide(int idx, FILE *output) 3105 { 3106 switch (idx) { 3107 case HISTC_SYMBOL: 3108 return __get_elide(symbol_conf.sym_list, "symbol", output); 3109 case HISTC_DSO: 3110 return __get_elide(symbol_conf.dso_list, "dso", output); 3111 case HISTC_COMM: 3112 return __get_elide(symbol_conf.comm_list, "comm", output); 3113 default: 3114 break; 3115 } 3116 3117 if (sort__mode != SORT_MODE__BRANCH) 3118 return false; 3119 3120 switch (idx) { 3121 case HISTC_SYMBOL_FROM: 3122 return __get_elide(symbol_conf.sym_from_list, "sym_from", output); 3123 case HISTC_SYMBOL_TO: 3124 return __get_elide(symbol_conf.sym_to_list, "sym_to", output); 3125 case HISTC_DSO_FROM: 3126 return __get_elide(symbol_conf.dso_from_list, "dso_from", output); 3127 case HISTC_DSO_TO: 3128 return __get_elide(symbol_conf.dso_to_list, "dso_to", output); 3129 default: 3130 break; 3131 } 3132 3133 return false; 3134 } 3135 3136 void sort__setup_elide(FILE *output) 3137 { 3138 struct perf_hpp_fmt *fmt; 3139 struct hpp_sort_entry *hse; 3140 3141 perf_hpp_list__for_each_format(&perf_hpp_list, fmt) { 3142 if (!perf_hpp__is_sort_entry(fmt)) 3143 continue; 3144 3145 hse = container_of(fmt, struct hpp_sort_entry, hpp); 3146 fmt->elide = get_elide(hse->se->se_width_idx, output); 3147 } 3148 3149 /* 3150 * It makes no sense to elide all of sort entries. 3151 * Just revert them to show up again. 3152 */ 3153 perf_hpp_list__for_each_format(&perf_hpp_list, fmt) { 3154 if (!perf_hpp__is_sort_entry(fmt)) 3155 continue; 3156 3157 if (!fmt->elide) 3158 return; 3159 } 3160 3161 perf_hpp_list__for_each_format(&perf_hpp_list, fmt) { 3162 if (!perf_hpp__is_sort_entry(fmt)) 3163 continue; 3164 3165 fmt->elide = false; 3166 } 3167 } 3168 3169 int output_field_add(struct perf_hpp_list *list, char *tok) 3170 { 3171 unsigned int i; 3172 3173 for (i = 0; i < ARRAY_SIZE(common_sort_dimensions); i++) { 3174 struct sort_dimension *sd = &common_sort_dimensions[i]; 3175 3176 if (strncasecmp(tok, sd->name, strlen(tok))) 3177 continue; 3178 3179 return __sort_dimension__add_output(list, sd); 3180 } 3181 3182 for (i = 0; i < ARRAY_SIZE(hpp_sort_dimensions); i++) { 3183 struct hpp_dimension *hd = &hpp_sort_dimensions[i]; 3184 3185 if (strncasecmp(tok, hd->name, strlen(tok))) 3186 continue; 3187 3188 return __hpp_dimension__add_output(list, hd); 3189 } 3190 3191 for (i = 0; i < ARRAY_SIZE(bstack_sort_dimensions); i++) { 3192 struct sort_dimension *sd = &bstack_sort_dimensions[i]; 3193 3194 if (strncasecmp(tok, sd->name, strlen(tok))) 3195 continue; 3196 3197 if (sort__mode != SORT_MODE__BRANCH) 3198 return -EINVAL; 3199 3200 return __sort_dimension__add_output(list, sd); 3201 } 3202 3203 for (i = 0; i < ARRAY_SIZE(memory_sort_dimensions); i++) { 3204 struct sort_dimension *sd = &memory_sort_dimensions[i]; 3205 3206 if (strncasecmp(tok, sd->name, strlen(tok))) 3207 continue; 3208 3209 if (sort__mode != SORT_MODE__MEMORY) 3210 return -EINVAL; 3211 3212 return __sort_dimension__add_output(list, sd); 3213 } 3214 3215 return -ESRCH; 3216 } 3217 3218 static int setup_output_list(struct perf_hpp_list *list, char *str) 3219 { 3220 char *tmp, *tok; 3221 int ret = 0; 3222 3223 for (tok = strtok_r(str, ", ", &tmp); 3224 tok; tok = strtok_r(NULL, ", ", &tmp)) { 3225 ret = output_field_add(list, tok); 3226 if (ret == -EINVAL) { 3227 ui__error("Invalid --fields key: `%s'", tok); 3228 break; 3229 } else if (ret == -ESRCH) { 3230 ui__error("Unknown --fields key: `%s'", tok); 3231 break; 3232 } 3233 } 3234 3235 return ret; 3236 } 3237 3238 void reset_dimensions(void) 3239 { 3240 unsigned int i; 3241 3242 for (i = 0; i < ARRAY_SIZE(common_sort_dimensions); i++) 3243 common_sort_dimensions[i].taken = 0; 3244 3245 for (i = 0; i < ARRAY_SIZE(hpp_sort_dimensions); i++) 3246 hpp_sort_dimensions[i].taken = 0; 3247 3248 for (i = 0; i < ARRAY_SIZE(bstack_sort_dimensions); i++) 3249 bstack_sort_dimensions[i].taken = 0; 3250 3251 for (i = 0; i < ARRAY_SIZE(memory_sort_dimensions); i++) 3252 memory_sort_dimensions[i].taken = 0; 3253 } 3254 3255 bool is_strict_order(const char *order) 3256 { 3257 return order && (*order != '+'); 3258 } 3259 3260 static int __setup_output_field(void) 3261 { 3262 char *str, *strp; 3263 int ret = -EINVAL; 3264 3265 if (field_order == NULL) 3266 return 0; 3267 3268 strp = str = strdup(field_order); 3269 if (str == NULL) { 3270 pr_err("Not enough memory to setup output fields"); 3271 return -ENOMEM; 3272 } 3273 3274 if (!is_strict_order(field_order)) 3275 strp++; 3276 3277 if (!strlen(strp)) { 3278 ui__error("Invalid --fields key: `+'"); 3279 goto out; 3280 } 3281 3282 ret = setup_output_list(&perf_hpp_list, strp); 3283 3284 out: 3285 free(str); 3286 return ret; 3287 } 3288 3289 int setup_sorting(struct evlist *evlist) 3290 { 3291 int err; 3292 3293 err = __setup_sorting(evlist); 3294 if (err < 0) 3295 return err; 3296 3297 if (parent_pattern != default_parent_pattern) { 3298 err = sort_dimension__add(&perf_hpp_list, "parent", evlist, -1); 3299 if (err < 0) 3300 return err; 3301 } 3302 3303 reset_dimensions(); 3304 3305 /* 3306 * perf diff doesn't use default hpp output fields. 3307 */ 3308 if (sort__mode != SORT_MODE__DIFF) 3309 perf_hpp__init(); 3310 3311 err = __setup_output_field(); 3312 if (err < 0) 3313 return err; 3314 3315 /* copy sort keys to output fields */ 3316 perf_hpp__setup_output_field(&perf_hpp_list); 3317 /* and then copy output fields to sort keys */ 3318 perf_hpp__append_sort_keys(&perf_hpp_list); 3319 3320 /* setup hists-specific output fields */ 3321 if (perf_hpp__setup_hists_formats(&perf_hpp_list, evlist) < 0) 3322 return -1; 3323 3324 return 0; 3325 } 3326 3327 void reset_output_field(void) 3328 { 3329 perf_hpp_list.need_collapse = 0; 3330 perf_hpp_list.parent = 0; 3331 perf_hpp_list.sym = 0; 3332 perf_hpp_list.dso = 0; 3333 3334 field_order = NULL; 3335 sort_order = NULL; 3336 3337 reset_dimensions(); 3338 perf_hpp__reset_output_field(&perf_hpp_list); 3339 } 3340 3341 #define INDENT (3*8 + 1) 3342 3343 static void add_key(struct strbuf *sb, const char *str, int *llen) 3344 { 3345 if (*llen >= 75) { 3346 strbuf_addstr(sb, "\n\t\t\t "); 3347 *llen = INDENT; 3348 } 3349 strbuf_addf(sb, " %s", str); 3350 *llen += strlen(str) + 1; 3351 } 3352 3353 static void add_sort_string(struct strbuf *sb, struct sort_dimension *s, int n, 3354 int *llen) 3355 { 3356 int i; 3357 3358 for (i = 0; i < n; i++) 3359 add_key(sb, s[i].name, llen); 3360 } 3361 3362 static void add_hpp_sort_string(struct strbuf *sb, struct hpp_dimension *s, int n, 3363 int *llen) 3364 { 3365 int i; 3366 3367 for (i = 0; i < n; i++) 3368 add_key(sb, s[i].name, llen); 3369 } 3370 3371 char *sort_help(const char *prefix) 3372 { 3373 struct strbuf sb; 3374 char *s; 3375 int len = strlen(prefix) + INDENT; 3376 3377 strbuf_init(&sb, 300); 3378 strbuf_addstr(&sb, prefix); 3379 add_hpp_sort_string(&sb, hpp_sort_dimensions, 3380 ARRAY_SIZE(hpp_sort_dimensions), &len); 3381 add_sort_string(&sb, common_sort_dimensions, 3382 ARRAY_SIZE(common_sort_dimensions), &len); 3383 add_sort_string(&sb, bstack_sort_dimensions, 3384 ARRAY_SIZE(bstack_sort_dimensions), &len); 3385 add_sort_string(&sb, memory_sort_dimensions, 3386 ARRAY_SIZE(memory_sort_dimensions), &len); 3387 s = strbuf_detach(&sb, NULL); 3388 strbuf_release(&sb); 3389 return s; 3390 } 3391