xref: /linux/tools/perf/arch/x86/util/event.c (revision e91c37f1)
1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/types.h>
3 #include <linux/string.h>
4 #include <linux/zalloc.h>
5 #include <stdlib.h>
6 
7 #include "../../../util/event.h"
8 #include "../../../util/synthetic-events.h"
9 #include "../../../util/machine.h"
10 #include "../../../util/tool.h"
11 #include "../../../util/map.h"
12 #include "../../../util/debug.h"
13 #include "util/sample.h"
14 
15 #if defined(__x86_64__)
16 
17 struct perf_event__synthesize_extra_kmaps_cb_args {
18 	struct perf_tool *tool;
19 	perf_event__handler_t process;
20 	struct machine *machine;
21 	union perf_event *event;
22 };
23 
24 static int perf_event__synthesize_extra_kmaps_cb(struct map *map, void *data)
25 {
26 	struct perf_event__synthesize_extra_kmaps_cb_args *args = data;
27 	union perf_event *event = args->event;
28 	struct kmap *kmap;
29 	size_t size;
30 
31 	if (!__map__is_extra_kernel_map(map))
32 		return 0;
33 
34 	kmap = map__kmap(map);
35 
36 	size = sizeof(event->mmap) - sizeof(event->mmap.filename) +
37 		      PERF_ALIGN(strlen(kmap->name) + 1, sizeof(u64)) +
38 		      args->machine->id_hdr_size;
39 
40 	memset(event, 0, size);
41 
42 	event->mmap.header.type = PERF_RECORD_MMAP;
43 
44 	/*
45 	 * kernel uses 0 for user space maps, see kernel/perf_event.c
46 	 * __perf_event_mmap
47 	 */
48 	if (machine__is_host(args->machine))
49 		event->header.misc = PERF_RECORD_MISC_KERNEL;
50 	else
51 		event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
52 
53 	event->mmap.header.size = size;
54 
55 	event->mmap.start = map__start(map);
56 	event->mmap.len   = map__size(map);
57 	event->mmap.pgoff = map__pgoff(map);
58 	event->mmap.pid   = args->machine->pid;
59 
60 	strlcpy(event->mmap.filename, kmap->name, PATH_MAX);
61 
62 	if (perf_tool__process_synth_event(args->tool, event, args->machine, args->process) != 0)
63 		return -1;
64 
65 	return 0;
66 }
67 
68 int perf_event__synthesize_extra_kmaps(struct perf_tool *tool,
69 				       perf_event__handler_t process,
70 				       struct machine *machine)
71 {
72 	int rc;
73 	struct maps *kmaps = machine__kernel_maps(machine);
74 	struct perf_event__synthesize_extra_kmaps_cb_args args = {
75 		.tool = tool,
76 		.process = process,
77 		.machine = machine,
78 		.event = zalloc(sizeof(args.event->mmap) + machine->id_hdr_size),
79 	};
80 
81 	if (!args.event) {
82 		pr_debug("Not enough memory synthesizing mmap event "
83 			 "for extra kernel maps\n");
84 		return -1;
85 	}
86 
87 	rc = maps__for_each_map(kmaps, perf_event__synthesize_extra_kmaps_cb, &args);
88 
89 	free(args.event);
90 	return rc;
91 }
92 
93 #endif
94 
95 void arch_perf_parse_sample_weight(struct perf_sample *data,
96 				   const __u64 *array, u64 type)
97 {
98 	union perf_sample_weight weight;
99 
100 	weight.full = *array;
101 	if (type & PERF_SAMPLE_WEIGHT)
102 		data->weight = weight.full;
103 	else {
104 		data->weight = weight.var1_dw;
105 		data->ins_lat = weight.var2_w;
106 		data->retire_lat = weight.var3_w;
107 	}
108 }
109 
110 void arch_perf_synthesize_sample_weight(const struct perf_sample *data,
111 					__u64 *array, u64 type)
112 {
113 	*array = data->weight;
114 
115 	if (type & PERF_SAMPLE_WEIGHT_STRUCT) {
116 		*array &= 0xffffffff;
117 		*array |= ((u64)data->ins_lat << 32);
118 		*array |= ((u64)data->retire_lat << 48);
119 	}
120 }
121 
122 const char *arch_perf_header_entry(const char *se_header)
123 {
124 	if (!strcmp(se_header, "Local Pipeline Stage Cycle"))
125 		return "Local Retire Latency";
126 	else if (!strcmp(se_header, "Pipeline Stage Cycle"))
127 		return "Retire Latency";
128 
129 	return se_header;
130 }
131 
132 int arch_support_sort_key(const char *sort_key)
133 {
134 	if (!strcmp(sort_key, "p_stage_cyc"))
135 		return 1;
136 	if (!strcmp(sort_key, "local_p_stage_cyc"))
137 		return 1;
138 	return 0;
139 }
140