1 /* 2 * auxtrace.h: AUX area trace support 3 * Copyright (c) 2013-2015, Intel Corporation. 4 * 5 * This program is free software; you can redistribute it and/or modify it 6 * under the terms and conditions of the GNU General Public License, 7 * version 2, as published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 12 * more details. 13 * 14 */ 15 16 #ifndef __PERF_AUXTRACE_H 17 #define __PERF_AUXTRACE_H 18 19 #include <sys/types.h> 20 #include <stdbool.h> 21 #include <stddef.h> 22 #include <linux/list.h> 23 #include <linux/perf_event.h> 24 #include <linux/types.h> 25 26 #include "../perf.h" 27 #include "event.h" 28 #include "session.h" 29 #include "debug.h" 30 31 union perf_event; 32 struct perf_session; 33 struct perf_evlist; 34 struct perf_tool; 35 struct option; 36 struct record_opts; 37 struct auxtrace_info_event; 38 struct events_stats; 39 40 enum auxtrace_type { 41 PERF_AUXTRACE_UNKNOWN, 42 PERF_AUXTRACE_INTEL_PT, 43 PERF_AUXTRACE_INTEL_BTS, 44 }; 45 46 enum itrace_period_type { 47 PERF_ITRACE_PERIOD_INSTRUCTIONS, 48 PERF_ITRACE_PERIOD_TICKS, 49 PERF_ITRACE_PERIOD_NANOSECS, 50 }; 51 52 /** 53 * struct itrace_synth_opts - AUX area tracing synthesis options. 54 * @set: indicates whether or not options have been set 55 * @inject: indicates the event (not just the sample) must be fully synthesized 56 * because 'perf inject' will write it out 57 * @instructions: whether to synthesize 'instructions' events 58 * @branches: whether to synthesize 'branches' events 59 * @transactions: whether to synthesize events for transactions 60 * @errors: whether to synthesize decoder error events 61 * @dont_decode: whether to skip decoding entirely 62 * @log: write a decoding log 63 * @calls: limit branch samples to calls (can be combined with @returns) 64 * @returns: limit branch samples to returns (can be combined with @calls) 65 * @callchain: add callchain to 'instructions' events 66 * @thread_stack: feed branches to the thread_stack 67 * @last_branch: add branch context to 'instruction' events 68 * @callchain_sz: maximum callchain size 69 * @last_branch_sz: branch context size 70 * @period: 'instructions' events period 71 * @period_type: 'instructions' events period type 72 * @initial_skip: skip N events at the beginning. 73 */ 74 struct itrace_synth_opts { 75 bool set; 76 bool inject; 77 bool instructions; 78 bool branches; 79 bool transactions; 80 bool errors; 81 bool dont_decode; 82 bool log; 83 bool calls; 84 bool returns; 85 bool callchain; 86 bool thread_stack; 87 bool last_branch; 88 unsigned int callchain_sz; 89 unsigned int last_branch_sz; 90 unsigned long long period; 91 enum itrace_period_type period_type; 92 unsigned long initial_skip; 93 }; 94 95 /** 96 * struct auxtrace_index_entry - indexes a AUX area tracing event within a 97 * perf.data file. 98 * @file_offset: offset within the perf.data file 99 * @sz: size of the event 100 */ 101 struct auxtrace_index_entry { 102 u64 file_offset; 103 u64 sz; 104 }; 105 106 #define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256 107 108 /** 109 * struct auxtrace_index - index of AUX area tracing events within a perf.data 110 * file. 111 * @list: linking a number of arrays of entries 112 * @nr: number of entries 113 * @entries: array of entries 114 */ 115 struct auxtrace_index { 116 struct list_head list; 117 size_t nr; 118 struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT]; 119 }; 120 121 /** 122 * struct auxtrace - session callbacks to allow AUX area data decoding. 123 * @process_event: lets the decoder see all session events 124 * @flush_events: process any remaining data 125 * @free_events: free resources associated with event processing 126 * @free: free resources associated with the session 127 */ 128 struct auxtrace { 129 int (*process_event)(struct perf_session *session, 130 union perf_event *event, 131 struct perf_sample *sample, 132 struct perf_tool *tool); 133 int (*process_auxtrace_event)(struct perf_session *session, 134 union perf_event *event, 135 struct perf_tool *tool); 136 int (*flush_events)(struct perf_session *session, 137 struct perf_tool *tool); 138 void (*free_events)(struct perf_session *session); 139 void (*free)(struct perf_session *session); 140 }; 141 142 /** 143 * struct auxtrace_buffer - a buffer containing AUX area tracing data. 144 * @list: buffers are queued in a list held by struct auxtrace_queue 145 * @size: size of the buffer in bytes 146 * @pid: in per-thread mode, the pid this buffer is associated with 147 * @tid: in per-thread mode, the tid this buffer is associated with 148 * @cpu: in per-cpu mode, the cpu this buffer is associated with 149 * @data: actual buffer data (can be null if the data has not been loaded) 150 * @data_offset: file offset at which the buffer can be read 151 * @mmap_addr: mmap address at which the buffer can be read 152 * @mmap_size: size of the mmap at @mmap_addr 153 * @data_needs_freeing: @data was malloc'd so free it when it is no longer 154 * needed 155 * @consecutive: the original data was split up and this buffer is consecutive 156 * to the previous buffer 157 * @offset: offset as determined by aux_head / aux_tail members of struct 158 * perf_event_mmap_page 159 * @reference: an implementation-specific reference determined when the data is 160 * recorded 161 * @buffer_nr: used to number each buffer 162 * @use_size: implementation actually only uses this number of bytes 163 * @use_data: implementation actually only uses data starting at this address 164 */ 165 struct auxtrace_buffer { 166 struct list_head list; 167 size_t size; 168 pid_t pid; 169 pid_t tid; 170 int cpu; 171 void *data; 172 off_t data_offset; 173 void *mmap_addr; 174 size_t mmap_size; 175 bool data_needs_freeing; 176 bool consecutive; 177 u64 offset; 178 u64 reference; 179 u64 buffer_nr; 180 size_t use_size; 181 void *use_data; 182 }; 183 184 /** 185 * struct auxtrace_queue - a queue of AUX area tracing data buffers. 186 * @head: head of buffer list 187 * @tid: in per-thread mode, the tid this queue is associated with 188 * @cpu: in per-cpu mode, the cpu this queue is associated with 189 * @set: %true once this queue has been dedicated to a specific thread or cpu 190 * @priv: implementation-specific data 191 */ 192 struct auxtrace_queue { 193 struct list_head head; 194 pid_t tid; 195 int cpu; 196 bool set; 197 void *priv; 198 }; 199 200 /** 201 * struct auxtrace_queues - an array of AUX area tracing queues. 202 * @queue_array: array of queues 203 * @nr_queues: number of queues 204 * @new_data: set whenever new data is queued 205 * @populated: queues have been fully populated using the auxtrace_index 206 * @next_buffer_nr: used to number each buffer 207 */ 208 struct auxtrace_queues { 209 struct auxtrace_queue *queue_array; 210 unsigned int nr_queues; 211 bool new_data; 212 bool populated; 213 u64 next_buffer_nr; 214 }; 215 216 /** 217 * struct auxtrace_heap_item - element of struct auxtrace_heap. 218 * @queue_nr: queue number 219 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected 220 * to be a timestamp 221 */ 222 struct auxtrace_heap_item { 223 unsigned int queue_nr; 224 u64 ordinal; 225 }; 226 227 /** 228 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues. 229 * @heap_array: the heap 230 * @heap_cnt: the number of elements in the heap 231 * @heap_sz: maximum number of elements (grows as needed) 232 */ 233 struct auxtrace_heap { 234 struct auxtrace_heap_item *heap_array; 235 unsigned int heap_cnt; 236 unsigned int heap_sz; 237 }; 238 239 /** 240 * struct auxtrace_mmap - records an mmap of the auxtrace buffer. 241 * @base: address of mapped area 242 * @userpg: pointer to buffer's perf_event_mmap_page 243 * @mask: %0 if @len is not a power of two, otherwise (@len - %1) 244 * @len: size of mapped area 245 * @prev: previous aux_head 246 * @idx: index of this mmap 247 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu 248 * mmap) otherwise %0 249 * @cpu: cpu number for a per-cpu mmap otherwise %-1 250 */ 251 struct auxtrace_mmap { 252 void *base; 253 void *userpg; 254 size_t mask; 255 size_t len; 256 u64 prev; 257 int idx; 258 pid_t tid; 259 int cpu; 260 }; 261 262 /** 263 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap. 264 * @mask: %0 if @len is not a power of two, otherwise (@len - %1) 265 * @offset: file offset of mapped area 266 * @len: size of mapped area 267 * @prot: mmap memory protection 268 * @idx: index of this mmap 269 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu 270 * mmap) otherwise %0 271 * @cpu: cpu number for a per-cpu mmap otherwise %-1 272 */ 273 struct auxtrace_mmap_params { 274 size_t mask; 275 off_t offset; 276 size_t len; 277 int prot; 278 int idx; 279 pid_t tid; 280 int cpu; 281 }; 282 283 /** 284 * struct auxtrace_record - callbacks for recording AUX area data. 285 * @recording_options: validate and process recording options 286 * @info_priv_size: return the size of the private data in auxtrace_info_event 287 * @info_fill: fill-in the private data in auxtrace_info_event 288 * @free: free this auxtrace record structure 289 * @snapshot_start: starting a snapshot 290 * @snapshot_finish: finishing a snapshot 291 * @find_snapshot: find data to snapshot within auxtrace mmap 292 * @parse_snapshot_options: parse snapshot options 293 * @reference: provide a 64-bit reference number for auxtrace_event 294 * @read_finish: called after reading from an auxtrace mmap 295 */ 296 struct auxtrace_record { 297 int (*recording_options)(struct auxtrace_record *itr, 298 struct perf_evlist *evlist, 299 struct record_opts *opts); 300 size_t (*info_priv_size)(struct auxtrace_record *itr, 301 struct perf_evlist *evlist); 302 int (*info_fill)(struct auxtrace_record *itr, 303 struct perf_session *session, 304 struct auxtrace_info_event *auxtrace_info, 305 size_t priv_size); 306 void (*free)(struct auxtrace_record *itr); 307 int (*snapshot_start)(struct auxtrace_record *itr); 308 int (*snapshot_finish)(struct auxtrace_record *itr); 309 int (*find_snapshot)(struct auxtrace_record *itr, int idx, 310 struct auxtrace_mmap *mm, unsigned char *data, 311 u64 *head, u64 *old); 312 int (*parse_snapshot_options)(struct auxtrace_record *itr, 313 struct record_opts *opts, 314 const char *str); 315 u64 (*reference)(struct auxtrace_record *itr); 316 int (*read_finish)(struct auxtrace_record *itr, int idx); 317 unsigned int alignment; 318 }; 319 320 #ifdef HAVE_AUXTRACE_SUPPORT 321 322 /* 323 * In snapshot mode the mmapped page is read-only which makes using 324 * __sync_val_compare_and_swap() problematic. However, snapshot mode expects 325 * the buffer is not updated while the snapshot is made (e.g. Intel PT disables 326 * the event) so there is not a race anyway. 327 */ 328 static inline u64 auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap *mm) 329 { 330 struct perf_event_mmap_page *pc = mm->userpg; 331 u64 head = ACCESS_ONCE(pc->aux_head); 332 333 /* Ensure all reads are done after we read the head */ 334 rmb(); 335 return head; 336 } 337 338 static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm) 339 { 340 struct perf_event_mmap_page *pc = mm->userpg; 341 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT) 342 u64 head = ACCESS_ONCE(pc->aux_head); 343 #else 344 u64 head = __sync_val_compare_and_swap(&pc->aux_head, 0, 0); 345 #endif 346 347 /* Ensure all reads are done after we read the head */ 348 rmb(); 349 return head; 350 } 351 352 static inline void auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail) 353 { 354 struct perf_event_mmap_page *pc = mm->userpg; 355 #if BITS_PER_LONG != 64 && defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT) 356 u64 old_tail; 357 #endif 358 359 /* Ensure all reads are done before we write the tail out */ 360 mb(); 361 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT) 362 pc->aux_tail = tail; 363 #else 364 do { 365 old_tail = __sync_val_compare_and_swap(&pc->aux_tail, 0, 0); 366 } while (!__sync_bool_compare_and_swap(&pc->aux_tail, old_tail, tail)); 367 #endif 368 } 369 370 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm, 371 struct auxtrace_mmap_params *mp, 372 void *userpg, int fd); 373 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm); 374 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp, 375 off_t auxtrace_offset, 376 unsigned int auxtrace_pages, 377 bool auxtrace_overwrite); 378 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, 379 struct perf_evlist *evlist, int idx, 380 bool per_cpu); 381 382 typedef int (*process_auxtrace_t)(struct perf_tool *tool, 383 union perf_event *event, void *data1, 384 size_t len1, void *data2, size_t len2); 385 386 int auxtrace_mmap__read(struct auxtrace_mmap *mm, struct auxtrace_record *itr, 387 struct perf_tool *tool, process_auxtrace_t fn); 388 389 int auxtrace_mmap__read_snapshot(struct auxtrace_mmap *mm, 390 struct auxtrace_record *itr, 391 struct perf_tool *tool, process_auxtrace_t fn, 392 size_t snapshot_size); 393 394 int auxtrace_queues__init(struct auxtrace_queues *queues); 395 int auxtrace_queues__add_event(struct auxtrace_queues *queues, 396 struct perf_session *session, 397 union perf_event *event, off_t data_offset, 398 struct auxtrace_buffer **buffer_ptr); 399 void auxtrace_queues__free(struct auxtrace_queues *queues); 400 int auxtrace_queues__process_index(struct auxtrace_queues *queues, 401 struct perf_session *session); 402 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue, 403 struct auxtrace_buffer *buffer); 404 void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd); 405 void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer); 406 void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer); 407 void auxtrace_buffer__free(struct auxtrace_buffer *buffer); 408 409 int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr, 410 u64 ordinal); 411 void auxtrace_heap__pop(struct auxtrace_heap *heap); 412 void auxtrace_heap__free(struct auxtrace_heap *heap); 413 414 struct auxtrace_cache_entry { 415 struct hlist_node hash; 416 u32 key; 417 }; 418 419 struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size, 420 unsigned int limit_percent); 421 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache); 422 void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c); 423 void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry); 424 int auxtrace_cache__add(struct auxtrace_cache *c, u32 key, 425 struct auxtrace_cache_entry *entry); 426 void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key); 427 428 struct auxtrace_record *auxtrace_record__init(struct perf_evlist *evlist, 429 int *err); 430 431 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, 432 struct record_opts *opts, 433 const char *str); 434 int auxtrace_record__options(struct auxtrace_record *itr, 435 struct perf_evlist *evlist, 436 struct record_opts *opts); 437 size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr, 438 struct perf_evlist *evlist); 439 int auxtrace_record__info_fill(struct auxtrace_record *itr, 440 struct perf_session *session, 441 struct auxtrace_info_event *auxtrace_info, 442 size_t priv_size); 443 void auxtrace_record__free(struct auxtrace_record *itr); 444 int auxtrace_record__snapshot_start(struct auxtrace_record *itr); 445 int auxtrace_record__snapshot_finish(struct auxtrace_record *itr); 446 int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, 447 struct auxtrace_mmap *mm, 448 unsigned char *data, u64 *head, u64 *old); 449 u64 auxtrace_record__reference(struct auxtrace_record *itr); 450 451 int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event, 452 off_t file_offset); 453 int auxtrace_index__write(int fd, struct list_head *head); 454 int auxtrace_index__process(int fd, u64 size, struct perf_session *session, 455 bool needs_swap); 456 void auxtrace_index__free(struct list_head *head); 457 458 void auxtrace_synth_error(struct auxtrace_error_event *auxtrace_error, int type, 459 int code, int cpu, pid_t pid, pid_t tid, u64 ip, 460 const char *msg); 461 462 int perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr, 463 struct perf_tool *tool, 464 struct perf_session *session, 465 perf_event__handler_t process); 466 int perf_event__process_auxtrace_info(struct perf_tool *tool, 467 union perf_event *event, 468 struct perf_session *session); 469 s64 perf_event__process_auxtrace(struct perf_tool *tool, 470 union perf_event *event, 471 struct perf_session *session); 472 int perf_event__process_auxtrace_error(struct perf_tool *tool, 473 union perf_event *event, 474 struct perf_session *session); 475 int itrace_parse_synth_opts(const struct option *opt, const char *str, 476 int unset); 477 void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts); 478 479 size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp); 480 void perf_session__auxtrace_error_inc(struct perf_session *session, 481 union perf_event *event); 482 void events_stats__auxtrace_error_warn(const struct events_stats *stats); 483 484 static inline int auxtrace__process_event(struct perf_session *session, 485 union perf_event *event, 486 struct perf_sample *sample, 487 struct perf_tool *tool) 488 { 489 if (!session->auxtrace) 490 return 0; 491 492 return session->auxtrace->process_event(session, event, sample, tool); 493 } 494 495 static inline int auxtrace__flush_events(struct perf_session *session, 496 struct perf_tool *tool) 497 { 498 if (!session->auxtrace) 499 return 0; 500 501 return session->auxtrace->flush_events(session, tool); 502 } 503 504 static inline void auxtrace__free_events(struct perf_session *session) 505 { 506 if (!session->auxtrace) 507 return; 508 509 return session->auxtrace->free_events(session); 510 } 511 512 static inline void auxtrace__free(struct perf_session *session) 513 { 514 if (!session->auxtrace) 515 return; 516 517 return session->auxtrace->free(session); 518 } 519 520 #else 521 522 static inline struct auxtrace_record * 523 auxtrace_record__init(struct perf_evlist *evlist __maybe_unused, 524 int *err) 525 { 526 *err = 0; 527 return NULL; 528 } 529 530 static inline 531 void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused) 532 { 533 } 534 535 static inline int 536 perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr __maybe_unused, 537 struct perf_tool *tool __maybe_unused, 538 struct perf_session *session __maybe_unused, 539 perf_event__handler_t process __maybe_unused) 540 { 541 return -EINVAL; 542 } 543 544 static inline 545 int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused, 546 struct perf_evlist *evlist __maybe_unused, 547 struct record_opts *opts __maybe_unused) 548 { 549 return 0; 550 } 551 552 #define perf_event__process_auxtrace_info 0 553 #define perf_event__process_auxtrace 0 554 #define perf_event__process_auxtrace_error 0 555 556 static inline 557 void perf_session__auxtrace_error_inc(struct perf_session *session 558 __maybe_unused, 559 union perf_event *event 560 __maybe_unused) 561 { 562 } 563 564 static inline 565 void events_stats__auxtrace_error_warn(const struct events_stats *stats 566 __maybe_unused) 567 { 568 } 569 570 static inline 571 int itrace_parse_synth_opts(const struct option *opt __maybe_unused, 572 const char *str __maybe_unused, 573 int unset __maybe_unused) 574 { 575 pr_err("AUX area tracing not supported\n"); 576 return -EINVAL; 577 } 578 579 static inline 580 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused, 581 struct record_opts *opts __maybe_unused, 582 const char *str) 583 { 584 if (!str) 585 return 0; 586 pr_err("AUX area tracing not supported\n"); 587 return -EINVAL; 588 } 589 590 static inline 591 int auxtrace__process_event(struct perf_session *session __maybe_unused, 592 union perf_event *event __maybe_unused, 593 struct perf_sample *sample __maybe_unused, 594 struct perf_tool *tool __maybe_unused) 595 { 596 return 0; 597 } 598 599 static inline 600 int auxtrace__flush_events(struct perf_session *session __maybe_unused, 601 struct perf_tool *tool __maybe_unused) 602 { 603 return 0; 604 } 605 606 static inline 607 void auxtrace__free_events(struct perf_session *session __maybe_unused) 608 { 609 } 610 611 static inline 612 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused) 613 { 614 } 615 616 static inline 617 void auxtrace__free(struct perf_session *session __maybe_unused) 618 { 619 } 620 621 static inline 622 int auxtrace_index__write(int fd __maybe_unused, 623 struct list_head *head __maybe_unused) 624 { 625 return -EINVAL; 626 } 627 628 static inline 629 int auxtrace_index__process(int fd __maybe_unused, 630 u64 size __maybe_unused, 631 struct perf_session *session __maybe_unused, 632 bool needs_swap __maybe_unused) 633 { 634 return -EINVAL; 635 } 636 637 static inline 638 void auxtrace_index__free(struct list_head *head __maybe_unused) 639 { 640 } 641 642 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm, 643 struct auxtrace_mmap_params *mp, 644 void *userpg, int fd); 645 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm); 646 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp, 647 off_t auxtrace_offset, 648 unsigned int auxtrace_pages, 649 bool auxtrace_overwrite); 650 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, 651 struct perf_evlist *evlist, int idx, 652 bool per_cpu); 653 654 #endif 655 656 #endif 657