xref: /openbmc/linux/tools/perf/util/auxtrace.h (revision 31e67366)
1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3  * auxtrace.h: AUX area trace support
4  * Copyright (c) 2013-2015, Intel Corporation.
5  */
6 
7 #ifndef __PERF_AUXTRACE_H
8 #define __PERF_AUXTRACE_H
9 
10 #include <sys/types.h>
11 #include <errno.h>
12 #include <stdbool.h>
13 #include <stddef.h>
14 #include <stdio.h> // FILE
15 #include <linux/list.h>
16 #include <linux/perf_event.h>
17 #include <linux/types.h>
18 #include <asm/bitsperlong.h>
19 #include <asm/barrier.h>
20 
21 union perf_event;
22 struct perf_session;
23 struct evlist;
24 struct evsel;
25 struct perf_tool;
26 struct mmap;
27 struct perf_sample;
28 struct option;
29 struct record_opts;
30 struct perf_record_auxtrace_error;
31 struct perf_record_auxtrace_info;
32 struct events_stats;
33 struct perf_pmu;
34 
35 enum auxtrace_error_type {
36        PERF_AUXTRACE_ERROR_ITRACE  = 1,
37        PERF_AUXTRACE_ERROR_MAX
38 };
39 
40 /* Auxtrace records must have the same alignment as perf event records */
41 #define PERF_AUXTRACE_RECORD_ALIGNMENT 8
42 
43 enum auxtrace_type {
44 	PERF_AUXTRACE_UNKNOWN,
45 	PERF_AUXTRACE_INTEL_PT,
46 	PERF_AUXTRACE_INTEL_BTS,
47 	PERF_AUXTRACE_CS_ETM,
48 	PERF_AUXTRACE_ARM_SPE,
49 	PERF_AUXTRACE_S390_CPUMSF,
50 };
51 
52 enum itrace_period_type {
53 	PERF_ITRACE_PERIOD_INSTRUCTIONS,
54 	PERF_ITRACE_PERIOD_TICKS,
55 	PERF_ITRACE_PERIOD_NANOSECS,
56 };
57 
58 #define AUXTRACE_ERR_FLG_OVERFLOW	(1 << ('o' - 'a'))
59 #define AUXTRACE_ERR_FLG_DATA_LOST	(1 << ('l' - 'a'))
60 
61 #define AUXTRACE_LOG_FLG_ALL_PERF_EVTS	(1 << ('a' - 'a'))
62 
63 /**
64  * struct itrace_synth_opts - AUX area tracing synthesis options.
65  * @set: indicates whether or not options have been set
66  * @default_no_sample: Default to no sampling.
67  * @inject: indicates the event (not just the sample) must be fully synthesized
68  *          because 'perf inject' will write it out
69  * @instructions: whether to synthesize 'instructions' events
70  * @branches: whether to synthesize 'branches' events
71  *            (branch misses only for Arm SPE)
72  * @transactions: whether to synthesize events for transactions
73  * @ptwrites: whether to synthesize events for ptwrites
74  * @pwr_events: whether to synthesize power events
75  * @other_events: whether to synthesize other events recorded due to the use of
76  *                aux_output
77  * @errors: whether to synthesize decoder error events
78  * @dont_decode: whether to skip decoding entirely
79  * @log: write a decoding log
80  * @calls: limit branch samples to calls (can be combined with @returns)
81  * @returns: limit branch samples to returns (can be combined with @calls)
82  * @callchain: add callchain to 'instructions' events
83  * @add_callchain: add callchain to existing event records
84  * @thread_stack: feed branches to the thread_stack
85  * @last_branch: add branch context to 'instruction' events
86  * @add_last_branch: add branch context to existing event records
87  * @flc: whether to synthesize first level cache events
88  * @llc: whether to synthesize last level cache events
89  * @tlb: whether to synthesize TLB events
90  * @remote_access: whether to synthesize remote access events
91  * @mem: whether to synthesize memory events
92  * @callchain_sz: maximum callchain size
93  * @last_branch_sz: branch context size
94  * @period: 'instructions' events period
95  * @period_type: 'instructions' events period type
96  * @initial_skip: skip N events at the beginning.
97  * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all
98  * @ptime_range: time intervals to trace or NULL
99  * @range_num: number of time intervals to trace
100  * @error_plus_flags: flags to affect what errors are reported
101  * @error_minus_flags: flags to affect what errors are reported
102  * @log_plus_flags: flags to affect what is logged
103  * @log_minus_flags: flags to affect what is logged
104  * @quick: quicker (less detailed) decoding
105  */
106 struct itrace_synth_opts {
107 	bool			set;
108 	bool			default_no_sample;
109 	bool			inject;
110 	bool			instructions;
111 	bool			branches;
112 	bool			transactions;
113 	bool			ptwrites;
114 	bool			pwr_events;
115 	bool			other_events;
116 	bool			errors;
117 	bool			dont_decode;
118 	bool			log;
119 	bool			calls;
120 	bool			returns;
121 	bool			callchain;
122 	bool			add_callchain;
123 	bool			thread_stack;
124 	bool			last_branch;
125 	bool			add_last_branch;
126 	bool			flc;
127 	bool			llc;
128 	bool			tlb;
129 	bool			remote_access;
130 	bool			mem;
131 	unsigned int		callchain_sz;
132 	unsigned int		last_branch_sz;
133 	unsigned long long	period;
134 	enum itrace_period_type	period_type;
135 	unsigned long		initial_skip;
136 	unsigned long		*cpu_bitmap;
137 	struct perf_time_interval *ptime_range;
138 	int			range_num;
139 	unsigned int		error_plus_flags;
140 	unsigned int		error_minus_flags;
141 	unsigned int		log_plus_flags;
142 	unsigned int		log_minus_flags;
143 	unsigned int		quick;
144 };
145 
146 /**
147  * struct auxtrace_index_entry - indexes a AUX area tracing event within a
148  *                               perf.data file.
149  * @file_offset: offset within the perf.data file
150  * @sz: size of the event
151  */
152 struct auxtrace_index_entry {
153 	u64			file_offset;
154 	u64			sz;
155 };
156 
157 #define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256
158 
159 /**
160  * struct auxtrace_index - index of AUX area tracing events within a perf.data
161  *                         file.
162  * @list: linking a number of arrays of entries
163  * @nr: number of entries
164  * @entries: array of entries
165  */
166 struct auxtrace_index {
167 	struct list_head	list;
168 	size_t			nr;
169 	struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT];
170 };
171 
172 /**
173  * struct auxtrace - session callbacks to allow AUX area data decoding.
174  * @process_event: lets the decoder see all session events
175  * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event
176  * @queue_data: queue an AUX sample or PERF_RECORD_AUXTRACE event for later
177  *              processing
178  * @dump_auxtrace_sample: dump AUX area sample data
179  * @flush_events: process any remaining data
180  * @free_events: free resources associated with event processing
181  * @free: free resources associated with the session
182  */
183 struct auxtrace {
184 	int (*process_event)(struct perf_session *session,
185 			     union perf_event *event,
186 			     struct perf_sample *sample,
187 			     struct perf_tool *tool);
188 	int (*process_auxtrace_event)(struct perf_session *session,
189 				      union perf_event *event,
190 				      struct perf_tool *tool);
191 	int (*queue_data)(struct perf_session *session,
192 			  struct perf_sample *sample, union perf_event *event,
193 			  u64 data_offset);
194 	void (*dump_auxtrace_sample)(struct perf_session *session,
195 				     struct perf_sample *sample);
196 	int (*flush_events)(struct perf_session *session,
197 			    struct perf_tool *tool);
198 	void (*free_events)(struct perf_session *session);
199 	void (*free)(struct perf_session *session);
200 	bool (*evsel_is_auxtrace)(struct perf_session *session,
201 				  struct evsel *evsel);
202 };
203 
204 /**
205  * struct auxtrace_buffer - a buffer containing AUX area tracing data.
206  * @list: buffers are queued in a list held by struct auxtrace_queue
207  * @size: size of the buffer in bytes
208  * @pid: in per-thread mode, the pid this buffer is associated with
209  * @tid: in per-thread mode, the tid this buffer is associated with
210  * @cpu: in per-cpu mode, the cpu this buffer is associated with
211  * @data: actual buffer data (can be null if the data has not been loaded)
212  * @data_offset: file offset at which the buffer can be read
213  * @mmap_addr: mmap address at which the buffer can be read
214  * @mmap_size: size of the mmap at @mmap_addr
215  * @data_needs_freeing: @data was malloc'd so free it when it is no longer
216  *                      needed
217  * @consecutive: the original data was split up and this buffer is consecutive
218  *               to the previous buffer
219  * @offset: offset as determined by aux_head / aux_tail members of struct
220  *          perf_event_mmap_page
221  * @reference: an implementation-specific reference determined when the data is
222  *             recorded
223  * @buffer_nr: used to number each buffer
224  * @use_size: implementation actually only uses this number of bytes
225  * @use_data: implementation actually only uses data starting at this address
226  */
227 struct auxtrace_buffer {
228 	struct list_head	list;
229 	size_t			size;
230 	pid_t			pid;
231 	pid_t			tid;
232 	int			cpu;
233 	void			*data;
234 	off_t			data_offset;
235 	void			*mmap_addr;
236 	size_t			mmap_size;
237 	bool			data_needs_freeing;
238 	bool			consecutive;
239 	u64			offset;
240 	u64			reference;
241 	u64			buffer_nr;
242 	size_t			use_size;
243 	void			*use_data;
244 };
245 
246 /**
247  * struct auxtrace_queue - a queue of AUX area tracing data buffers.
248  * @head: head of buffer list
249  * @tid: in per-thread mode, the tid this queue is associated with
250  * @cpu: in per-cpu mode, the cpu this queue is associated with
251  * @set: %true once this queue has been dedicated to a specific thread or cpu
252  * @priv: implementation-specific data
253  */
254 struct auxtrace_queue {
255 	struct list_head	head;
256 	pid_t			tid;
257 	int			cpu;
258 	bool			set;
259 	void			*priv;
260 };
261 
262 /**
263  * struct auxtrace_queues - an array of AUX area tracing queues.
264  * @queue_array: array of queues
265  * @nr_queues: number of queues
266  * @new_data: set whenever new data is queued
267  * @populated: queues have been fully populated using the auxtrace_index
268  * @next_buffer_nr: used to number each buffer
269  */
270 struct auxtrace_queues {
271 	struct auxtrace_queue	*queue_array;
272 	unsigned int		nr_queues;
273 	bool			new_data;
274 	bool			populated;
275 	u64			next_buffer_nr;
276 };
277 
278 /**
279  * struct auxtrace_heap_item - element of struct auxtrace_heap.
280  * @queue_nr: queue number
281  * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected
282  *           to be a timestamp
283  */
284 struct auxtrace_heap_item {
285 	unsigned int		queue_nr;
286 	u64			ordinal;
287 };
288 
289 /**
290  * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues.
291  * @heap_array: the heap
292  * @heap_cnt: the number of elements in the heap
293  * @heap_sz: maximum number of elements (grows as needed)
294  */
295 struct auxtrace_heap {
296 	struct auxtrace_heap_item	*heap_array;
297 	unsigned int		heap_cnt;
298 	unsigned int		heap_sz;
299 };
300 
301 /**
302  * struct auxtrace_mmap - records an mmap of the auxtrace buffer.
303  * @base: address of mapped area
304  * @userpg: pointer to buffer's perf_event_mmap_page
305  * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
306  * @len: size of mapped area
307  * @prev: previous aux_head
308  * @idx: index of this mmap
309  * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
310  *       mmap) otherwise %0
311  * @cpu: cpu number for a per-cpu mmap otherwise %-1
312  */
313 struct auxtrace_mmap {
314 	void		*base;
315 	void		*userpg;
316 	size_t		mask;
317 	size_t		len;
318 	u64		prev;
319 	int		idx;
320 	pid_t		tid;
321 	int		cpu;
322 };
323 
324 /**
325  * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap.
326  * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
327  * @offset: file offset of mapped area
328  * @len: size of mapped area
329  * @prot: mmap memory protection
330  * @idx: index of this mmap
331  * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
332  *       mmap) otherwise %0
333  * @cpu: cpu number for a per-cpu mmap otherwise %-1
334  */
335 struct auxtrace_mmap_params {
336 	size_t		mask;
337 	off_t		offset;
338 	size_t		len;
339 	int		prot;
340 	int		idx;
341 	pid_t		tid;
342 	int		cpu;
343 };
344 
345 /**
346  * struct auxtrace_record - callbacks for recording AUX area data.
347  * @recording_options: validate and process recording options
348  * @info_priv_size: return the size of the private data in auxtrace_info_event
349  * @info_fill: fill-in the private data in auxtrace_info_event
350  * @free: free this auxtrace record structure
351  * @snapshot_start: starting a snapshot
352  * @snapshot_finish: finishing a snapshot
353  * @find_snapshot: find data to snapshot within auxtrace mmap
354  * @parse_snapshot_options: parse snapshot options
355  * @reference: provide a 64-bit reference number for auxtrace_event
356  * @read_finish: called after reading from an auxtrace mmap
357  * @alignment: alignment (if any) for AUX area data
358  * @default_aux_sample_size: default sample size for --aux sample option
359  * @pmu: associated pmu
360  * @evlist: selected events list
361  */
362 struct auxtrace_record {
363 	int (*recording_options)(struct auxtrace_record *itr,
364 				 struct evlist *evlist,
365 				 struct record_opts *opts);
366 	size_t (*info_priv_size)(struct auxtrace_record *itr,
367 				 struct evlist *evlist);
368 	int (*info_fill)(struct auxtrace_record *itr,
369 			 struct perf_session *session,
370 			 struct perf_record_auxtrace_info *auxtrace_info,
371 			 size_t priv_size);
372 	void (*free)(struct auxtrace_record *itr);
373 	int (*snapshot_start)(struct auxtrace_record *itr);
374 	int (*snapshot_finish)(struct auxtrace_record *itr);
375 	int (*find_snapshot)(struct auxtrace_record *itr, int idx,
376 			     struct auxtrace_mmap *mm, unsigned char *data,
377 			     u64 *head, u64 *old);
378 	int (*parse_snapshot_options)(struct auxtrace_record *itr,
379 				      struct record_opts *opts,
380 				      const char *str);
381 	u64 (*reference)(struct auxtrace_record *itr);
382 	int (*read_finish)(struct auxtrace_record *itr, int idx);
383 	unsigned int alignment;
384 	unsigned int default_aux_sample_size;
385 	struct perf_pmu *pmu;
386 	struct evlist *evlist;
387 };
388 
389 /**
390  * struct addr_filter - address filter.
391  * @list: list node
392  * @range: true if it is a range filter
393  * @start: true if action is 'filter' or 'start'
394  * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted
395  *          to 'stop')
396  * @sym_from: symbol name for the filter address
397  * @sym_to: symbol name that determines the filter size
398  * @sym_from_idx: selects n'th from symbols with the same name (0 means global
399  *                and less than 0 means symbol must be unique)
400  * @sym_to_idx: same as @sym_from_idx but for @sym_to
401  * @addr: filter address
402  * @size: filter region size (for range filters)
403  * @filename: DSO file name or NULL for the kernel
404  * @str: allocated string that contains the other string members
405  */
406 struct addr_filter {
407 	struct list_head	list;
408 	bool			range;
409 	bool			start;
410 	const char		*action;
411 	const char		*sym_from;
412 	const char		*sym_to;
413 	int			sym_from_idx;
414 	int			sym_to_idx;
415 	u64			addr;
416 	u64			size;
417 	const char		*filename;
418 	char			*str;
419 };
420 
421 /**
422  * struct addr_filters - list of address filters.
423  * @head: list of address filters
424  * @cnt: number of address filters
425  */
426 struct addr_filters {
427 	struct list_head	head;
428 	int			cnt;
429 };
430 
431 struct auxtrace_cache;
432 
433 #ifdef HAVE_AUXTRACE_SUPPORT
434 
435 /*
436  * In snapshot mode the mmapped page is read-only which makes using
437  * __sync_val_compare_and_swap() problematic.  However, snapshot mode expects
438  * the buffer is not updated while the snapshot is made (e.g. Intel PT disables
439  * the event) so there is not a race anyway.
440  */
441 static inline u64 auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap *mm)
442 {
443 	struct perf_event_mmap_page *pc = mm->userpg;
444 	u64 head = READ_ONCE(pc->aux_head);
445 
446 	/* Ensure all reads are done after we read the head */
447 	rmb();
448 	return head;
449 }
450 
451 static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm)
452 {
453 	struct perf_event_mmap_page *pc = mm->userpg;
454 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
455 	u64 head = READ_ONCE(pc->aux_head);
456 #else
457 	u64 head = __sync_val_compare_and_swap(&pc->aux_head, 0, 0);
458 #endif
459 
460 	/* Ensure all reads are done after we read the head */
461 	rmb();
462 	return head;
463 }
464 
465 static inline void auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail)
466 {
467 	struct perf_event_mmap_page *pc = mm->userpg;
468 #if BITS_PER_LONG != 64 && defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
469 	u64 old_tail;
470 #endif
471 
472 	/* Ensure all reads are done before we write the tail out */
473 	mb();
474 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
475 	pc->aux_tail = tail;
476 #else
477 	do {
478 		old_tail = __sync_val_compare_and_swap(&pc->aux_tail, 0, 0);
479 	} while (!__sync_bool_compare_and_swap(&pc->aux_tail, old_tail, tail));
480 #endif
481 }
482 
483 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
484 			struct auxtrace_mmap_params *mp,
485 			void *userpg, int fd);
486 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
487 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
488 				off_t auxtrace_offset,
489 				unsigned int auxtrace_pages,
490 				bool auxtrace_overwrite);
491 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
492 				   struct evlist *evlist, int idx,
493 				   bool per_cpu);
494 
495 typedef int (*process_auxtrace_t)(struct perf_tool *tool,
496 				  struct mmap *map,
497 				  union perf_event *event, void *data1,
498 				  size_t len1, void *data2, size_t len2);
499 
500 int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr,
501 			struct perf_tool *tool, process_auxtrace_t fn);
502 
503 int auxtrace_mmap__read_snapshot(struct mmap *map,
504 				 struct auxtrace_record *itr,
505 				 struct perf_tool *tool, process_auxtrace_t fn,
506 				 size_t snapshot_size);
507 
508 int auxtrace_queues__init(struct auxtrace_queues *queues);
509 int auxtrace_queues__add_event(struct auxtrace_queues *queues,
510 			       struct perf_session *session,
511 			       union perf_event *event, off_t data_offset,
512 			       struct auxtrace_buffer **buffer_ptr);
513 struct auxtrace_queue *
514 auxtrace_queues__sample_queue(struct auxtrace_queues *queues,
515 			      struct perf_sample *sample,
516 			      struct perf_session *session);
517 int auxtrace_queues__add_sample(struct auxtrace_queues *queues,
518 				struct perf_session *session,
519 				struct perf_sample *sample, u64 data_offset,
520 				u64 reference);
521 void auxtrace_queues__free(struct auxtrace_queues *queues);
522 int auxtrace_queues__process_index(struct auxtrace_queues *queues,
523 				   struct perf_session *session);
524 int auxtrace_queue_data(struct perf_session *session, bool samples,
525 			bool events);
526 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
527 					      struct auxtrace_buffer *buffer);
528 void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd);
529 void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer);
530 void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer);
531 void auxtrace_buffer__free(struct auxtrace_buffer *buffer);
532 
533 int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
534 		       u64 ordinal);
535 void auxtrace_heap__pop(struct auxtrace_heap *heap);
536 void auxtrace_heap__free(struct auxtrace_heap *heap);
537 
538 struct auxtrace_cache_entry {
539 	struct hlist_node hash;
540 	u32 key;
541 };
542 
543 struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
544 					   unsigned int limit_percent);
545 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache);
546 void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c);
547 void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
548 int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
549 			struct auxtrace_cache_entry *entry);
550 void auxtrace_cache__remove(struct auxtrace_cache *c, u32 key);
551 void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key);
552 
553 struct auxtrace_record *auxtrace_record__init(struct evlist *evlist,
554 					      int *err);
555 
556 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
557 				    struct record_opts *opts,
558 				    const char *str);
559 int auxtrace_parse_sample_options(struct auxtrace_record *itr,
560 				  struct evlist *evlist,
561 				  struct record_opts *opts, const char *str);
562 void auxtrace_regroup_aux_output(struct evlist *evlist);
563 int auxtrace_record__options(struct auxtrace_record *itr,
564 			     struct evlist *evlist,
565 			     struct record_opts *opts);
566 size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr,
567 				       struct evlist *evlist);
568 int auxtrace_record__info_fill(struct auxtrace_record *itr,
569 			       struct perf_session *session,
570 			       struct perf_record_auxtrace_info *auxtrace_info,
571 			       size_t priv_size);
572 void auxtrace_record__free(struct auxtrace_record *itr);
573 int auxtrace_record__snapshot_start(struct auxtrace_record *itr);
574 int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit);
575 int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
576 				   struct auxtrace_mmap *mm,
577 				   unsigned char *data, u64 *head, u64 *old);
578 u64 auxtrace_record__reference(struct auxtrace_record *itr);
579 int auxtrace_record__read_finish(struct auxtrace_record *itr, int idx);
580 
581 int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
582 				   off_t file_offset);
583 int auxtrace_index__write(int fd, struct list_head *head);
584 int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
585 			    bool needs_swap);
586 void auxtrace_index__free(struct list_head *head);
587 
588 void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
589 			  int code, int cpu, pid_t pid, pid_t tid, u64 ip,
590 			  const char *msg, u64 timestamp);
591 
592 int perf_event__process_auxtrace_info(struct perf_session *session,
593 				      union perf_event *event);
594 s64 perf_event__process_auxtrace(struct perf_session *session,
595 				 union perf_event *event);
596 int perf_event__process_auxtrace_error(struct perf_session *session,
597 				       union perf_event *event);
598 int itrace_parse_synth_opts(const struct option *opt, const char *str,
599 			    int unset);
600 void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts,
601 				    bool no_sample);
602 
603 size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp);
604 void perf_session__auxtrace_error_inc(struct perf_session *session,
605 				      union perf_event *event);
606 void events_stats__auxtrace_error_warn(const struct events_stats *stats);
607 
608 void addr_filters__init(struct addr_filters *filts);
609 void addr_filters__exit(struct addr_filters *filts);
610 int addr_filters__parse_bare_filter(struct addr_filters *filts,
611 				    const char *filter);
612 int auxtrace_parse_filters(struct evlist *evlist);
613 
614 int auxtrace__process_event(struct perf_session *session, union perf_event *event,
615 			    struct perf_sample *sample, struct perf_tool *tool);
616 void auxtrace__dump_auxtrace_sample(struct perf_session *session,
617 				    struct perf_sample *sample);
618 int auxtrace__flush_events(struct perf_session *session, struct perf_tool *tool);
619 void auxtrace__free_events(struct perf_session *session);
620 void auxtrace__free(struct perf_session *session);
621 bool auxtrace__evsel_is_auxtrace(struct perf_session *session,
622 				 struct evsel *evsel);
623 
624 #define ITRACE_HELP \
625 "				i[period]:    		synthesize instructions events\n" \
626 "				b:	    		synthesize branches events (branch misses for Arm SPE)\n" \
627 "				c:	    		synthesize branches events (calls only)\n"	\
628 "				r:	    		synthesize branches events (returns only)\n" \
629 "				x:	    		synthesize transactions events\n"		\
630 "				w:	    		synthesize ptwrite events\n"		\
631 "				p:	    		synthesize power events\n"			\
632 "				o:			synthesize other events recorded due to the use\n" \
633 "							of aux-output (refer to perf record)\n"	\
634 "				e[flags]:		synthesize error events\n" \
635 "							each flag must be preceded by + or -\n" \
636 "							error flags are: o (overflow)\n" \
637 "									 l (data lost)\n" \
638 "				d[flags]:		create a debug log\n" \
639 "							each flag must be preceded by + or -\n" \
640 "							log flags are: a (all perf events)\n" \
641 "				f:	    		synthesize first level cache events\n" \
642 "				m:	    		synthesize last level cache events\n" \
643 "				t:	    		synthesize TLB events\n" \
644 "				a:	    		synthesize remote access events\n" \
645 "				g[len]:     		synthesize a call chain (use with i or x)\n" \
646 "				G[len]:			synthesize a call chain on existing event records\n" \
647 "				l[len]:     		synthesize last branch entries (use with i or x)\n" \
648 "				L[len]:			synthesize last branch entries on existing event records\n" \
649 "				sNUMBER:    		skip initial number of events\n"		\
650 "				q:			quicker (less detailed) decoding\n" \
651 "				PERIOD[ns|us|ms|i|t]:   specify period to sample stream\n" \
652 "				concatenate multiple options. Default is ibxwpe or cewp\n"
653 
654 static inline
655 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts,
656 				       struct perf_time_interval *ptime_range,
657 				       int range_num)
658 {
659 	opts->ptime_range = ptime_range;
660 	opts->range_num = range_num;
661 }
662 
663 static inline
664 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts)
665 {
666 	opts->ptime_range = NULL;
667 	opts->range_num = 0;
668 }
669 
670 #else
671 #include "debug.h"
672 
673 static inline struct auxtrace_record *
674 auxtrace_record__init(struct evlist *evlist __maybe_unused,
675 		      int *err)
676 {
677 	*err = 0;
678 	return NULL;
679 }
680 
681 static inline
682 void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused)
683 {
684 }
685 
686 static inline
687 int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused,
688 			     struct evlist *evlist __maybe_unused,
689 			     struct record_opts *opts __maybe_unused)
690 {
691 	return 0;
692 }
693 
694 #define perf_event__process_auxtrace_info		0
695 #define perf_event__process_auxtrace			0
696 #define perf_event__process_auxtrace_error		0
697 
698 static inline
699 void perf_session__auxtrace_error_inc(struct perf_session *session
700 				      __maybe_unused,
701 				      union perf_event *event
702 				      __maybe_unused)
703 {
704 }
705 
706 static inline
707 void events_stats__auxtrace_error_warn(const struct events_stats *stats
708 				       __maybe_unused)
709 {
710 }
711 
712 static inline
713 int itrace_parse_synth_opts(const struct option *opt __maybe_unused,
714 			    const char *str __maybe_unused,
715 			    int unset __maybe_unused)
716 {
717 	pr_err("AUX area tracing not supported\n");
718 	return -EINVAL;
719 }
720 
721 static inline
722 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused,
723 				    struct record_opts *opts __maybe_unused,
724 				    const char *str)
725 {
726 	if (!str)
727 		return 0;
728 	pr_err("AUX area tracing not supported\n");
729 	return -EINVAL;
730 }
731 
732 static inline
733 int auxtrace_parse_sample_options(struct auxtrace_record *itr __maybe_unused,
734 				  struct evlist *evlist __maybe_unused,
735 				  struct record_opts *opts __maybe_unused,
736 				  const char *str)
737 {
738 	if (!str)
739 		return 0;
740 	pr_err("AUX area tracing not supported\n");
741 	return -EINVAL;
742 }
743 
744 static inline
745 void auxtrace_regroup_aux_output(struct evlist *evlist __maybe_unused)
746 {
747 }
748 
749 static inline
750 int auxtrace__process_event(struct perf_session *session __maybe_unused,
751 			    union perf_event *event __maybe_unused,
752 			    struct perf_sample *sample __maybe_unused,
753 			    struct perf_tool *tool __maybe_unused)
754 {
755 	return 0;
756 }
757 
758 static inline
759 void auxtrace__dump_auxtrace_sample(struct perf_session *session __maybe_unused,
760 				    struct perf_sample *sample __maybe_unused)
761 {
762 }
763 
764 static inline
765 int auxtrace__flush_events(struct perf_session *session __maybe_unused,
766 			   struct perf_tool *tool __maybe_unused)
767 {
768 	return 0;
769 }
770 
771 static inline
772 void auxtrace__free_events(struct perf_session *session __maybe_unused)
773 {
774 }
775 
776 static inline
777 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused)
778 {
779 }
780 
781 static inline
782 void auxtrace__free(struct perf_session *session __maybe_unused)
783 {
784 }
785 
786 static inline
787 int auxtrace_index__write(int fd __maybe_unused,
788 			  struct list_head *head __maybe_unused)
789 {
790 	return -EINVAL;
791 }
792 
793 static inline
794 int auxtrace_index__process(int fd __maybe_unused,
795 			    u64 size __maybe_unused,
796 			    struct perf_session *session __maybe_unused,
797 			    bool needs_swap __maybe_unused)
798 {
799 	return -EINVAL;
800 }
801 
802 static inline
803 void auxtrace_index__free(struct list_head *head __maybe_unused)
804 {
805 }
806 
807 static inline
808 bool auxtrace__evsel_is_auxtrace(struct perf_session *session __maybe_unused,
809 				 struct evsel *evsel __maybe_unused)
810 {
811 	return false;
812 }
813 
814 static inline
815 int auxtrace_parse_filters(struct evlist *evlist __maybe_unused)
816 {
817 	return 0;
818 }
819 
820 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
821 			struct auxtrace_mmap_params *mp,
822 			void *userpg, int fd);
823 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
824 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
825 				off_t auxtrace_offset,
826 				unsigned int auxtrace_pages,
827 				bool auxtrace_overwrite);
828 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
829 				   struct evlist *evlist, int idx,
830 				   bool per_cpu);
831 
832 #define ITRACE_HELP ""
833 
834 static inline
835 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts
836 				       __maybe_unused,
837 				       struct perf_time_interval *ptime_range
838 				       __maybe_unused,
839 				       int range_num __maybe_unused)
840 {
841 }
842 
843 static inline
844 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts
845 					 __maybe_unused)
846 {
847 }
848 
849 #endif
850 
851 #endif
852