1 /* SPDX-License-Identifier: GPL-2.0-only */
2 /*
3 * auxtrace.h: AUX area trace support
4 * Copyright (c) 2013-2015, Intel Corporation.
5 */
6
7 #ifndef __PERF_AUXTRACE_H
8 #define __PERF_AUXTRACE_H
9
10 #include <sys/types.h>
11 #include <errno.h>
12 #include <stdbool.h>
13 #include <stddef.h>
14 #include <stdio.h> // FILE
15 #include <linux/list.h>
16 #include <linux/perf_event.h>
17 #include <linux/types.h>
18 #include <asm/bitsperlong.h>
19 #include <asm/barrier.h>
20
21 union perf_event;
22 struct perf_session;
23 struct evlist;
24 struct evsel;
25 struct perf_tool;
26 struct mmap;
27 struct perf_sample;
28 struct option;
29 struct record_opts;
30 struct perf_record_auxtrace_error;
31 struct perf_record_auxtrace_info;
32 struct events_stats;
33 struct perf_pmu;
34
35 enum auxtrace_error_type {
36 PERF_AUXTRACE_ERROR_ITRACE = 1,
37 PERF_AUXTRACE_ERROR_MAX
38 };
39
40 /* Auxtrace records must have the same alignment as perf event records */
41 #define PERF_AUXTRACE_RECORD_ALIGNMENT 8
42
43 enum auxtrace_type {
44 PERF_AUXTRACE_UNKNOWN,
45 PERF_AUXTRACE_INTEL_PT,
46 PERF_AUXTRACE_INTEL_BTS,
47 PERF_AUXTRACE_CS_ETM,
48 PERF_AUXTRACE_ARM_SPE,
49 PERF_AUXTRACE_S390_CPUMSF,
50 };
51
52 enum itrace_period_type {
53 PERF_ITRACE_PERIOD_INSTRUCTIONS,
54 PERF_ITRACE_PERIOD_TICKS,
55 PERF_ITRACE_PERIOD_NANOSECS,
56 };
57
58 #define AUXTRACE_ERR_FLG_OVERFLOW (1 << ('o' - 'a'))
59 #define AUXTRACE_ERR_FLG_DATA_LOST (1 << ('l' - 'a'))
60
61 #define AUXTRACE_LOG_FLG_ALL_PERF_EVTS (1 << ('a' - 'a'))
62
63 /**
64 * struct itrace_synth_opts - AUX area tracing synthesis options.
65 * @set: indicates whether or not options have been set
66 * @default_no_sample: Default to no sampling.
67 * @inject: indicates the event (not just the sample) must be fully synthesized
68 * because 'perf inject' will write it out
69 * @instructions: whether to synthesize 'instructions' events
70 * @branches: whether to synthesize 'branches' events
71 * (branch misses only for Arm SPE)
72 * @transactions: whether to synthesize events for transactions
73 * @ptwrites: whether to synthesize events for ptwrites
74 * @pwr_events: whether to synthesize power events
75 * @other_events: whether to synthesize other events recorded due to the use of
76 * aux_output
77 * @errors: whether to synthesize decoder error events
78 * @dont_decode: whether to skip decoding entirely
79 * @log: write a decoding log
80 * @calls: limit branch samples to calls (can be combined with @returns)
81 * @returns: limit branch samples to returns (can be combined with @calls)
82 * @callchain: add callchain to 'instructions' events
83 * @add_callchain: add callchain to existing event records
84 * @thread_stack: feed branches to the thread_stack
85 * @last_branch: add branch context to 'instruction' events
86 * @add_last_branch: add branch context to existing event records
87 * @flc: whether to synthesize first level cache events
88 * @llc: whether to synthesize last level cache events
89 * @tlb: whether to synthesize TLB events
90 * @remote_access: whether to synthesize remote access events
91 * @mem: whether to synthesize memory events
92 * @callchain_sz: maximum callchain size
93 * @last_branch_sz: branch context size
94 * @period: 'instructions' events period
95 * @period_type: 'instructions' events period type
96 * @initial_skip: skip N events at the beginning.
97 * @cpu_bitmap: CPUs for which to synthesize events, or NULL for all
98 * @ptime_range: time intervals to trace or NULL
99 * @range_num: number of time intervals to trace
100 * @error_plus_flags: flags to affect what errors are reported
101 * @error_minus_flags: flags to affect what errors are reported
102 * @log_plus_flags: flags to affect what is logged
103 * @log_minus_flags: flags to affect what is logged
104 * @quick: quicker (less detailed) decoding
105 */
106 struct itrace_synth_opts {
107 bool set;
108 bool default_no_sample;
109 bool inject;
110 bool instructions;
111 bool branches;
112 bool transactions;
113 bool ptwrites;
114 bool pwr_events;
115 bool other_events;
116 bool errors;
117 bool dont_decode;
118 bool log;
119 bool calls;
120 bool returns;
121 bool callchain;
122 bool add_callchain;
123 bool thread_stack;
124 bool last_branch;
125 bool add_last_branch;
126 bool flc;
127 bool llc;
128 bool tlb;
129 bool remote_access;
130 bool mem;
131 unsigned int callchain_sz;
132 unsigned int last_branch_sz;
133 unsigned long long period;
134 enum itrace_period_type period_type;
135 unsigned long initial_skip;
136 unsigned long *cpu_bitmap;
137 struct perf_time_interval *ptime_range;
138 int range_num;
139 unsigned int error_plus_flags;
140 unsigned int error_minus_flags;
141 unsigned int log_plus_flags;
142 unsigned int log_minus_flags;
143 unsigned int quick;
144 };
145
146 /**
147 * struct auxtrace_index_entry - indexes a AUX area tracing event within a
148 * perf.data file.
149 * @file_offset: offset within the perf.data file
150 * @sz: size of the event
151 */
152 struct auxtrace_index_entry {
153 u64 file_offset;
154 u64 sz;
155 };
156
157 #define PERF_AUXTRACE_INDEX_ENTRY_COUNT 256
158
159 /**
160 * struct auxtrace_index - index of AUX area tracing events within a perf.data
161 * file.
162 * @list: linking a number of arrays of entries
163 * @nr: number of entries
164 * @entries: array of entries
165 */
166 struct auxtrace_index {
167 struct list_head list;
168 size_t nr;
169 struct auxtrace_index_entry entries[PERF_AUXTRACE_INDEX_ENTRY_COUNT];
170 };
171
172 /**
173 * struct auxtrace - session callbacks to allow AUX area data decoding.
174 * @process_event: lets the decoder see all session events
175 * @process_auxtrace_event: process a PERF_RECORD_AUXTRACE event
176 * @queue_data: queue an AUX sample or PERF_RECORD_AUXTRACE event for later
177 * processing
178 * @dump_auxtrace_sample: dump AUX area sample data
179 * @flush_events: process any remaining data
180 * @free_events: free resources associated with event processing
181 * @free: free resources associated with the session
182 */
183 struct auxtrace {
184 int (*process_event)(struct perf_session *session,
185 union perf_event *event,
186 struct perf_sample *sample,
187 struct perf_tool *tool);
188 int (*process_auxtrace_event)(struct perf_session *session,
189 union perf_event *event,
190 struct perf_tool *tool);
191 int (*queue_data)(struct perf_session *session,
192 struct perf_sample *sample, union perf_event *event,
193 u64 data_offset);
194 void (*dump_auxtrace_sample)(struct perf_session *session,
195 struct perf_sample *sample);
196 int (*flush_events)(struct perf_session *session,
197 struct perf_tool *tool);
198 void (*free_events)(struct perf_session *session);
199 void (*free)(struct perf_session *session);
200 bool (*evsel_is_auxtrace)(struct perf_session *session,
201 struct evsel *evsel);
202 };
203
204 /**
205 * struct auxtrace_buffer - a buffer containing AUX area tracing data.
206 * @list: buffers are queued in a list held by struct auxtrace_queue
207 * @size: size of the buffer in bytes
208 * @pid: in per-thread mode, the pid this buffer is associated with
209 * @tid: in per-thread mode, the tid this buffer is associated with
210 * @cpu: in per-cpu mode, the cpu this buffer is associated with
211 * @data: actual buffer data (can be null if the data has not been loaded)
212 * @data_offset: file offset at which the buffer can be read
213 * @mmap_addr: mmap address at which the buffer can be read
214 * @mmap_size: size of the mmap at @mmap_addr
215 * @data_needs_freeing: @data was malloc'd so free it when it is no longer
216 * needed
217 * @consecutive: the original data was split up and this buffer is consecutive
218 * to the previous buffer
219 * @offset: offset as determined by aux_head / aux_tail members of struct
220 * perf_event_mmap_page
221 * @reference: an implementation-specific reference determined when the data is
222 * recorded
223 * @buffer_nr: used to number each buffer
224 * @use_size: implementation actually only uses this number of bytes
225 * @use_data: implementation actually only uses data starting at this address
226 */
227 struct auxtrace_buffer {
228 struct list_head list;
229 size_t size;
230 pid_t pid;
231 pid_t tid;
232 int cpu;
233 void *data;
234 off_t data_offset;
235 void *mmap_addr;
236 size_t mmap_size;
237 bool data_needs_freeing;
238 bool consecutive;
239 u64 offset;
240 u64 reference;
241 u64 buffer_nr;
242 size_t use_size;
243 void *use_data;
244 };
245
246 /**
247 * struct auxtrace_queue - a queue of AUX area tracing data buffers.
248 * @head: head of buffer list
249 * @tid: in per-thread mode, the tid this queue is associated with
250 * @cpu: in per-cpu mode, the cpu this queue is associated with
251 * @set: %true once this queue has been dedicated to a specific thread or cpu
252 * @priv: implementation-specific data
253 */
254 struct auxtrace_queue {
255 struct list_head head;
256 pid_t tid;
257 int cpu;
258 bool set;
259 void *priv;
260 };
261
262 /**
263 * struct auxtrace_queues - an array of AUX area tracing queues.
264 * @queue_array: array of queues
265 * @nr_queues: number of queues
266 * @new_data: set whenever new data is queued
267 * @populated: queues have been fully populated using the auxtrace_index
268 * @next_buffer_nr: used to number each buffer
269 */
270 struct auxtrace_queues {
271 struct auxtrace_queue *queue_array;
272 unsigned int nr_queues;
273 bool new_data;
274 bool populated;
275 u64 next_buffer_nr;
276 };
277
278 /**
279 * struct auxtrace_heap_item - element of struct auxtrace_heap.
280 * @queue_nr: queue number
281 * @ordinal: value used for sorting (lowest ordinal is top of the heap) expected
282 * to be a timestamp
283 */
284 struct auxtrace_heap_item {
285 unsigned int queue_nr;
286 u64 ordinal;
287 };
288
289 /**
290 * struct auxtrace_heap - a heap suitable for sorting AUX area tracing queues.
291 * @heap_array: the heap
292 * @heap_cnt: the number of elements in the heap
293 * @heap_sz: maximum number of elements (grows as needed)
294 */
295 struct auxtrace_heap {
296 struct auxtrace_heap_item *heap_array;
297 unsigned int heap_cnt;
298 unsigned int heap_sz;
299 };
300
301 /**
302 * struct auxtrace_mmap - records an mmap of the auxtrace buffer.
303 * @base: address of mapped area
304 * @userpg: pointer to buffer's perf_event_mmap_page
305 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
306 * @len: size of mapped area
307 * @prev: previous aux_head
308 * @idx: index of this mmap
309 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
310 * mmap) otherwise %0
311 * @cpu: cpu number for a per-cpu mmap otherwise %-1
312 */
313 struct auxtrace_mmap {
314 void *base;
315 void *userpg;
316 size_t mask;
317 size_t len;
318 u64 prev;
319 int idx;
320 pid_t tid;
321 int cpu;
322 };
323
324 /**
325 * struct auxtrace_mmap_params - parameters to set up struct auxtrace_mmap.
326 * @mask: %0 if @len is not a power of two, otherwise (@len - %1)
327 * @offset: file offset of mapped area
328 * @len: size of mapped area
329 * @prot: mmap memory protection
330 * @idx: index of this mmap
331 * @tid: tid for a per-thread mmap (also set if there is only 1 tid on a per-cpu
332 * mmap) otherwise %0
333 * @cpu: cpu number for a per-cpu mmap otherwise %-1
334 */
335 struct auxtrace_mmap_params {
336 size_t mask;
337 off_t offset;
338 size_t len;
339 int prot;
340 int idx;
341 pid_t tid;
342 int cpu;
343 };
344
345 /**
346 * struct auxtrace_record - callbacks for recording AUX area data.
347 * @recording_options: validate and process recording options
348 * @info_priv_size: return the size of the private data in auxtrace_info_event
349 * @info_fill: fill-in the private data in auxtrace_info_event
350 * @free: free this auxtrace record structure
351 * @snapshot_start: starting a snapshot
352 * @snapshot_finish: finishing a snapshot
353 * @find_snapshot: find data to snapshot within auxtrace mmap
354 * @parse_snapshot_options: parse snapshot options
355 * @reference: provide a 64-bit reference number for auxtrace_event
356 * @read_finish: called after reading from an auxtrace mmap
357 * @alignment: alignment (if any) for AUX area data
358 * @default_aux_sample_size: default sample size for --aux sample option
359 * @pmu: associated pmu
360 * @evlist: selected events list
361 */
362 struct auxtrace_record {
363 int (*recording_options)(struct auxtrace_record *itr,
364 struct evlist *evlist,
365 struct record_opts *opts);
366 size_t (*info_priv_size)(struct auxtrace_record *itr,
367 struct evlist *evlist);
368 int (*info_fill)(struct auxtrace_record *itr,
369 struct perf_session *session,
370 struct perf_record_auxtrace_info *auxtrace_info,
371 size_t priv_size);
372 void (*free)(struct auxtrace_record *itr);
373 int (*snapshot_start)(struct auxtrace_record *itr);
374 int (*snapshot_finish)(struct auxtrace_record *itr);
375 int (*find_snapshot)(struct auxtrace_record *itr, int idx,
376 struct auxtrace_mmap *mm, unsigned char *data,
377 u64 *head, u64 *old);
378 int (*parse_snapshot_options)(struct auxtrace_record *itr,
379 struct record_opts *opts,
380 const char *str);
381 u64 (*reference)(struct auxtrace_record *itr);
382 int (*read_finish)(struct auxtrace_record *itr, int idx);
383 unsigned int alignment;
384 unsigned int default_aux_sample_size;
385 struct perf_pmu *pmu;
386 struct evlist *evlist;
387 };
388
389 /**
390 * struct addr_filter - address filter.
391 * @list: list node
392 * @range: true if it is a range filter
393 * @start: true if action is 'filter' or 'start'
394 * @action: 'filter', 'start' or 'stop' ('tracestop' is accepted but converted
395 * to 'stop')
396 * @sym_from: symbol name for the filter address
397 * @sym_to: symbol name that determines the filter size
398 * @sym_from_idx: selects n'th from symbols with the same name (0 means global
399 * and less than 0 means symbol must be unique)
400 * @sym_to_idx: same as @sym_from_idx but for @sym_to
401 * @addr: filter address
402 * @size: filter region size (for range filters)
403 * @filename: DSO file name or NULL for the kernel
404 * @str: allocated string that contains the other string members
405 */
406 struct addr_filter {
407 struct list_head list;
408 bool range;
409 bool start;
410 const char *action;
411 const char *sym_from;
412 const char *sym_to;
413 int sym_from_idx;
414 int sym_to_idx;
415 u64 addr;
416 u64 size;
417 const char *filename;
418 char *str;
419 };
420
421 /**
422 * struct addr_filters - list of address filters.
423 * @head: list of address filters
424 * @cnt: number of address filters
425 */
426 struct addr_filters {
427 struct list_head head;
428 int cnt;
429 };
430
431 struct auxtrace_cache;
432
433 #ifdef HAVE_AUXTRACE_SUPPORT
434
435 /*
436 * In snapshot mode the mmapped page is read-only which makes using
437 * __sync_val_compare_and_swap() problematic. However, snapshot mode expects
438 * the buffer is not updated while the snapshot is made (e.g. Intel PT disables
439 * the event) so there is not a race anyway.
440 */
auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap * mm)441 static inline u64 auxtrace_mmap__read_snapshot_head(struct auxtrace_mmap *mm)
442 {
443 struct perf_event_mmap_page *pc = mm->userpg;
444 u64 head = READ_ONCE(pc->aux_head);
445
446 /* Ensure all reads are done after we read the head */
447 rmb();
448 return head;
449 }
450
auxtrace_mmap__read_head(struct auxtrace_mmap * mm)451 static inline u64 auxtrace_mmap__read_head(struct auxtrace_mmap *mm)
452 {
453 struct perf_event_mmap_page *pc = mm->userpg;
454 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
455 u64 head = READ_ONCE(pc->aux_head);
456 #else
457 u64 head = __sync_val_compare_and_swap(&pc->aux_head, 0, 0);
458 #endif
459
460 /* Ensure all reads are done after we read the head */
461 rmb();
462 return head;
463 }
464
auxtrace_mmap__write_tail(struct auxtrace_mmap * mm,u64 tail)465 static inline void auxtrace_mmap__write_tail(struct auxtrace_mmap *mm, u64 tail)
466 {
467 struct perf_event_mmap_page *pc = mm->userpg;
468 #if BITS_PER_LONG != 64 && defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
469 u64 old_tail;
470 #endif
471
472 /* Ensure all reads are done before we write the tail out */
473 mb();
474 #if BITS_PER_LONG == 64 || !defined(HAVE_SYNC_COMPARE_AND_SWAP_SUPPORT)
475 pc->aux_tail = tail;
476 #else
477 do {
478 old_tail = __sync_val_compare_and_swap(&pc->aux_tail, 0, 0);
479 } while (!__sync_bool_compare_and_swap(&pc->aux_tail, old_tail, tail));
480 #endif
481 }
482
483 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
484 struct auxtrace_mmap_params *mp,
485 void *userpg, int fd);
486 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
487 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
488 off_t auxtrace_offset,
489 unsigned int auxtrace_pages,
490 bool auxtrace_overwrite);
491 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
492 struct evlist *evlist, int idx,
493 bool per_cpu);
494
495 typedef int (*process_auxtrace_t)(struct perf_tool *tool,
496 struct mmap *map,
497 union perf_event *event, void *data1,
498 size_t len1, void *data2, size_t len2);
499
500 int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr,
501 struct perf_tool *tool, process_auxtrace_t fn);
502
503 int auxtrace_mmap__read_snapshot(struct mmap *map,
504 struct auxtrace_record *itr,
505 struct perf_tool *tool, process_auxtrace_t fn,
506 size_t snapshot_size);
507
508 int auxtrace_queues__init(struct auxtrace_queues *queues);
509 int auxtrace_queues__add_event(struct auxtrace_queues *queues,
510 struct perf_session *session,
511 union perf_event *event, off_t data_offset,
512 struct auxtrace_buffer **buffer_ptr);
513 struct auxtrace_queue *
514 auxtrace_queues__sample_queue(struct auxtrace_queues *queues,
515 struct perf_sample *sample,
516 struct perf_session *session);
517 int auxtrace_queues__add_sample(struct auxtrace_queues *queues,
518 struct perf_session *session,
519 struct perf_sample *sample, u64 data_offset,
520 u64 reference);
521 void auxtrace_queues__free(struct auxtrace_queues *queues);
522 int auxtrace_queues__process_index(struct auxtrace_queues *queues,
523 struct perf_session *session);
524 int auxtrace_queue_data(struct perf_session *session, bool samples,
525 bool events);
526 struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
527 struct auxtrace_buffer *buffer);
528 void *auxtrace_buffer__get_data(struct auxtrace_buffer *buffer, int fd);
529 void auxtrace_buffer__put_data(struct auxtrace_buffer *buffer);
530 void auxtrace_buffer__drop_data(struct auxtrace_buffer *buffer);
531 void auxtrace_buffer__free(struct auxtrace_buffer *buffer);
532
533 int auxtrace_heap__add(struct auxtrace_heap *heap, unsigned int queue_nr,
534 u64 ordinal);
535 void auxtrace_heap__pop(struct auxtrace_heap *heap);
536 void auxtrace_heap__free(struct auxtrace_heap *heap);
537
538 struct auxtrace_cache_entry {
539 struct hlist_node hash;
540 u32 key;
541 };
542
543 struct auxtrace_cache *auxtrace_cache__new(unsigned int bits, size_t entry_size,
544 unsigned int limit_percent);
545 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache);
546 void *auxtrace_cache__alloc_entry(struct auxtrace_cache *c);
547 void auxtrace_cache__free_entry(struct auxtrace_cache *c, void *entry);
548 int auxtrace_cache__add(struct auxtrace_cache *c, u32 key,
549 struct auxtrace_cache_entry *entry);
550 void auxtrace_cache__remove(struct auxtrace_cache *c, u32 key);
551 void *auxtrace_cache__lookup(struct auxtrace_cache *c, u32 key);
552
553 struct auxtrace_record *auxtrace_record__init(struct evlist *evlist,
554 int *err);
555
556 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr,
557 struct record_opts *opts,
558 const char *str);
559 int auxtrace_parse_sample_options(struct auxtrace_record *itr,
560 struct evlist *evlist,
561 struct record_opts *opts, const char *str);
562 void auxtrace_regroup_aux_output(struct evlist *evlist);
563 int auxtrace_record__options(struct auxtrace_record *itr,
564 struct evlist *evlist,
565 struct record_opts *opts);
566 size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr,
567 struct evlist *evlist);
568 int auxtrace_record__info_fill(struct auxtrace_record *itr,
569 struct perf_session *session,
570 struct perf_record_auxtrace_info *auxtrace_info,
571 size_t priv_size);
572 void auxtrace_record__free(struct auxtrace_record *itr);
573 int auxtrace_record__snapshot_start(struct auxtrace_record *itr);
574 int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit);
575 int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx,
576 struct auxtrace_mmap *mm,
577 unsigned char *data, u64 *head, u64 *old);
578 u64 auxtrace_record__reference(struct auxtrace_record *itr);
579 int auxtrace_record__read_finish(struct auxtrace_record *itr, int idx);
580
581 int auxtrace_index__auxtrace_event(struct list_head *head, union perf_event *event,
582 off_t file_offset);
583 int auxtrace_index__write(int fd, struct list_head *head);
584 int auxtrace_index__process(int fd, u64 size, struct perf_session *session,
585 bool needs_swap);
586 void auxtrace_index__free(struct list_head *head);
587
588 void auxtrace_synth_error(struct perf_record_auxtrace_error *auxtrace_error, int type,
589 int code, int cpu, pid_t pid, pid_t tid, u64 ip,
590 const char *msg, u64 timestamp);
591
592 int perf_event__process_auxtrace_info(struct perf_session *session,
593 union perf_event *event);
594 s64 perf_event__process_auxtrace(struct perf_session *session,
595 union perf_event *event);
596 int perf_event__process_auxtrace_error(struct perf_session *session,
597 union perf_event *event);
598 int itrace_parse_synth_opts(const struct option *opt, const char *str,
599 int unset);
600 void itrace_synth_opts__set_default(struct itrace_synth_opts *synth_opts,
601 bool no_sample);
602
603 size_t perf_event__fprintf_auxtrace_error(union perf_event *event, FILE *fp);
604 void perf_session__auxtrace_error_inc(struct perf_session *session,
605 union perf_event *event);
606 void events_stats__auxtrace_error_warn(const struct events_stats *stats);
607
608 void addr_filters__init(struct addr_filters *filts);
609 void addr_filters__exit(struct addr_filters *filts);
610 int addr_filters__parse_bare_filter(struct addr_filters *filts,
611 const char *filter);
612 int auxtrace_parse_filters(struct evlist *evlist);
613
614 int auxtrace__process_event(struct perf_session *session, union perf_event *event,
615 struct perf_sample *sample, struct perf_tool *tool);
616 void auxtrace__dump_auxtrace_sample(struct perf_session *session,
617 struct perf_sample *sample);
618 int auxtrace__flush_events(struct perf_session *session, struct perf_tool *tool);
619 void auxtrace__free_events(struct perf_session *session);
620 void auxtrace__free(struct perf_session *session);
621 bool auxtrace__evsel_is_auxtrace(struct perf_session *session,
622 struct evsel *evsel);
623
624 #define ITRACE_HELP \
625 " i[period]: synthesize instructions events\n" \
626 " b: synthesize branches events (branch misses for Arm SPE)\n" \
627 " c: synthesize branches events (calls only)\n" \
628 " r: synthesize branches events (returns only)\n" \
629 " x: synthesize transactions events\n" \
630 " w: synthesize ptwrite events\n" \
631 " p: synthesize power events\n" \
632 " o: synthesize other events recorded due to the use\n" \
633 " of aux-output (refer to perf record)\n" \
634 " e[flags]: synthesize error events\n" \
635 " each flag must be preceded by + or -\n" \
636 " error flags are: o (overflow)\n" \
637 " l (data lost)\n" \
638 " d[flags]: create a debug log\n" \
639 " each flag must be preceded by + or -\n" \
640 " log flags are: a (all perf events)\n" \
641 " f: synthesize first level cache events\n" \
642 " m: synthesize last level cache events\n" \
643 " t: synthesize TLB events\n" \
644 " a: synthesize remote access events\n" \
645 " g[len]: synthesize a call chain (use with i or x)\n" \
646 " G[len]: synthesize a call chain on existing event records\n" \
647 " l[len]: synthesize last branch entries (use with i or x)\n" \
648 " L[len]: synthesize last branch entries on existing event records\n" \
649 " sNUMBER: skip initial number of events\n" \
650 " q: quicker (less detailed) decoding\n" \
651 " PERIOD[ns|us|ms|i|t]: specify period to sample stream\n" \
652 " concatenate multiple options. Default is ibxwpe or cewp\n"
653
654 static inline
itrace_synth_opts__set_time_range(struct itrace_synth_opts * opts,struct perf_time_interval * ptime_range,int range_num)655 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts,
656 struct perf_time_interval *ptime_range,
657 int range_num)
658 {
659 opts->ptime_range = ptime_range;
660 opts->range_num = range_num;
661 }
662
663 static inline
itrace_synth_opts__clear_time_range(struct itrace_synth_opts * opts)664 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts)
665 {
666 opts->ptime_range = NULL;
667 opts->range_num = 0;
668 }
669
670 #else
671 #include "debug.h"
672
673 static inline struct auxtrace_record *
auxtrace_record__init(struct evlist * evlist __maybe_unused,int * err)674 auxtrace_record__init(struct evlist *evlist __maybe_unused,
675 int *err)
676 {
677 *err = 0;
678 return NULL;
679 }
680
681 static inline
auxtrace_record__free(struct auxtrace_record * itr __maybe_unused)682 void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused)
683 {
684 }
685
686 static inline
auxtrace_record__options(struct auxtrace_record * itr __maybe_unused,struct evlist * evlist __maybe_unused,struct record_opts * opts __maybe_unused)687 int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused,
688 struct evlist *evlist __maybe_unused,
689 struct record_opts *opts __maybe_unused)
690 {
691 return 0;
692 }
693
694 #define perf_event__process_auxtrace_info 0
695 #define perf_event__process_auxtrace 0
696 #define perf_event__process_auxtrace_error 0
697
698 static inline
perf_session__auxtrace_error_inc(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused)699 void perf_session__auxtrace_error_inc(struct perf_session *session
700 __maybe_unused,
701 union perf_event *event
702 __maybe_unused)
703 {
704 }
705
706 static inline
events_stats__auxtrace_error_warn(const struct events_stats * stats __maybe_unused)707 void events_stats__auxtrace_error_warn(const struct events_stats *stats
708 __maybe_unused)
709 {
710 }
711
712 static inline
itrace_parse_synth_opts(const struct option * opt __maybe_unused,const char * str __maybe_unused,int unset __maybe_unused)713 int itrace_parse_synth_opts(const struct option *opt __maybe_unused,
714 const char *str __maybe_unused,
715 int unset __maybe_unused)
716 {
717 pr_err("AUX area tracing not supported\n");
718 return -EINVAL;
719 }
720
721 static inline
auxtrace_parse_snapshot_options(struct auxtrace_record * itr __maybe_unused,struct record_opts * opts __maybe_unused,const char * str)722 int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused,
723 struct record_opts *opts __maybe_unused,
724 const char *str)
725 {
726 if (!str)
727 return 0;
728 pr_err("AUX area tracing not supported\n");
729 return -EINVAL;
730 }
731
732 static inline
auxtrace_parse_sample_options(struct auxtrace_record * itr __maybe_unused,struct evlist * evlist __maybe_unused,struct record_opts * opts __maybe_unused,const char * str)733 int auxtrace_parse_sample_options(struct auxtrace_record *itr __maybe_unused,
734 struct evlist *evlist __maybe_unused,
735 struct record_opts *opts __maybe_unused,
736 const char *str)
737 {
738 if (!str)
739 return 0;
740 pr_err("AUX area tracing not supported\n");
741 return -EINVAL;
742 }
743
744 static inline
auxtrace_regroup_aux_output(struct evlist * evlist __maybe_unused)745 void auxtrace_regroup_aux_output(struct evlist *evlist __maybe_unused)
746 {
747 }
748
749 static inline
auxtrace__process_event(struct perf_session * session __maybe_unused,union perf_event * event __maybe_unused,struct perf_sample * sample __maybe_unused,struct perf_tool * tool __maybe_unused)750 int auxtrace__process_event(struct perf_session *session __maybe_unused,
751 union perf_event *event __maybe_unused,
752 struct perf_sample *sample __maybe_unused,
753 struct perf_tool *tool __maybe_unused)
754 {
755 return 0;
756 }
757
758 static inline
auxtrace__dump_auxtrace_sample(struct perf_session * session __maybe_unused,struct perf_sample * sample __maybe_unused)759 void auxtrace__dump_auxtrace_sample(struct perf_session *session __maybe_unused,
760 struct perf_sample *sample __maybe_unused)
761 {
762 }
763
764 static inline
auxtrace__flush_events(struct perf_session * session __maybe_unused,struct perf_tool * tool __maybe_unused)765 int auxtrace__flush_events(struct perf_session *session __maybe_unused,
766 struct perf_tool *tool __maybe_unused)
767 {
768 return 0;
769 }
770
771 static inline
auxtrace__free_events(struct perf_session * session __maybe_unused)772 void auxtrace__free_events(struct perf_session *session __maybe_unused)
773 {
774 }
775
776 static inline
auxtrace_cache__free(struct auxtrace_cache * auxtrace_cache __maybe_unused)777 void auxtrace_cache__free(struct auxtrace_cache *auxtrace_cache __maybe_unused)
778 {
779 }
780
781 static inline
auxtrace__free(struct perf_session * session __maybe_unused)782 void auxtrace__free(struct perf_session *session __maybe_unused)
783 {
784 }
785
786 static inline
auxtrace_index__write(int fd __maybe_unused,struct list_head * head __maybe_unused)787 int auxtrace_index__write(int fd __maybe_unused,
788 struct list_head *head __maybe_unused)
789 {
790 return -EINVAL;
791 }
792
793 static inline
auxtrace_index__process(int fd __maybe_unused,u64 size __maybe_unused,struct perf_session * session __maybe_unused,bool needs_swap __maybe_unused)794 int auxtrace_index__process(int fd __maybe_unused,
795 u64 size __maybe_unused,
796 struct perf_session *session __maybe_unused,
797 bool needs_swap __maybe_unused)
798 {
799 return -EINVAL;
800 }
801
802 static inline
auxtrace_index__free(struct list_head * head __maybe_unused)803 void auxtrace_index__free(struct list_head *head __maybe_unused)
804 {
805 }
806
807 static inline
auxtrace__evsel_is_auxtrace(struct perf_session * session __maybe_unused,struct evsel * evsel __maybe_unused)808 bool auxtrace__evsel_is_auxtrace(struct perf_session *session __maybe_unused,
809 struct evsel *evsel __maybe_unused)
810 {
811 return false;
812 }
813
814 static inline
auxtrace_parse_filters(struct evlist * evlist __maybe_unused)815 int auxtrace_parse_filters(struct evlist *evlist __maybe_unused)
816 {
817 return 0;
818 }
819
820 int auxtrace_mmap__mmap(struct auxtrace_mmap *mm,
821 struct auxtrace_mmap_params *mp,
822 void *userpg, int fd);
823 void auxtrace_mmap__munmap(struct auxtrace_mmap *mm);
824 void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
825 off_t auxtrace_offset,
826 unsigned int auxtrace_pages,
827 bool auxtrace_overwrite);
828 void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
829 struct evlist *evlist, int idx,
830 bool per_cpu);
831
832 #define ITRACE_HELP ""
833
834 static inline
itrace_synth_opts__set_time_range(struct itrace_synth_opts * opts __maybe_unused,struct perf_time_interval * ptime_range __maybe_unused,int range_num __maybe_unused)835 void itrace_synth_opts__set_time_range(struct itrace_synth_opts *opts
836 __maybe_unused,
837 struct perf_time_interval *ptime_range
838 __maybe_unused,
839 int range_num __maybe_unused)
840 {
841 }
842
843 static inline
itrace_synth_opts__clear_time_range(struct itrace_synth_opts * opts __maybe_unused)844 void itrace_synth_opts__clear_time_range(struct itrace_synth_opts *opts
845 __maybe_unused)
846 {
847 }
848
849 #endif
850
851 #endif
852