1 // SPDX-License-Identifier: GPL-2.0
2 #include <linux/types.h>
3 #include <linux/string.h>
4 #include <linux/zalloc.h>
5 #include <stdlib.h>
6
7 #include "../../../util/event.h"
8 #include "../../../util/synthetic-events.h"
9 #include "../../../util/machine.h"
10 #include "../../../util/tool.h"
11 #include "../../../util/map.h"
12 #include "../../../util/debug.h"
13 #include "util/sample.h"
14
15 #if defined(__x86_64__)
16
perf_event__synthesize_extra_kmaps(struct perf_tool * tool,perf_event__handler_t process,struct machine * machine)17 int perf_event__synthesize_extra_kmaps(struct perf_tool *tool,
18 perf_event__handler_t process,
19 struct machine *machine)
20 {
21 int rc = 0;
22 struct map_rb_node *pos;
23 struct maps *kmaps = machine__kernel_maps(machine);
24 union perf_event *event = zalloc(sizeof(event->mmap) +
25 machine->id_hdr_size);
26
27 if (!event) {
28 pr_debug("Not enough memory synthesizing mmap event "
29 "for extra kernel maps\n");
30 return -1;
31 }
32
33 maps__for_each_entry(kmaps, pos) {
34 struct kmap *kmap;
35 size_t size;
36 struct map *map = pos->map;
37
38 if (!__map__is_extra_kernel_map(map))
39 continue;
40
41 kmap = map__kmap(map);
42
43 size = sizeof(event->mmap) - sizeof(event->mmap.filename) +
44 PERF_ALIGN(strlen(kmap->name) + 1, sizeof(u64)) +
45 machine->id_hdr_size;
46
47 memset(event, 0, size);
48
49 event->mmap.header.type = PERF_RECORD_MMAP;
50
51 /*
52 * kernel uses 0 for user space maps, see kernel/perf_event.c
53 * __perf_event_mmap
54 */
55 if (machine__is_host(machine))
56 event->header.misc = PERF_RECORD_MISC_KERNEL;
57 else
58 event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL;
59
60 event->mmap.header.size = size;
61
62 event->mmap.start = map__start(map);
63 event->mmap.len = map__size(map);
64 event->mmap.pgoff = map__pgoff(map);
65 event->mmap.pid = machine->pid;
66
67 strlcpy(event->mmap.filename, kmap->name, PATH_MAX);
68
69 if (perf_tool__process_synth_event(tool, event, machine,
70 process) != 0) {
71 rc = -1;
72 break;
73 }
74 }
75
76 free(event);
77 return rc;
78 }
79
80 #endif
81
arch_perf_parse_sample_weight(struct perf_sample * data,const __u64 * array,u64 type)82 void arch_perf_parse_sample_weight(struct perf_sample *data,
83 const __u64 *array, u64 type)
84 {
85 union perf_sample_weight weight;
86
87 weight.full = *array;
88 if (type & PERF_SAMPLE_WEIGHT)
89 data->weight = weight.full;
90 else {
91 data->weight = weight.var1_dw;
92 data->ins_lat = weight.var2_w;
93 data->retire_lat = weight.var3_w;
94 }
95 }
96
arch_perf_synthesize_sample_weight(const struct perf_sample * data,__u64 * array,u64 type)97 void arch_perf_synthesize_sample_weight(const struct perf_sample *data,
98 __u64 *array, u64 type)
99 {
100 *array = data->weight;
101
102 if (type & PERF_SAMPLE_WEIGHT_STRUCT) {
103 *array &= 0xffffffff;
104 *array |= ((u64)data->ins_lat << 32);
105 *array |= ((u64)data->retire_lat << 48);
106 }
107 }
108
arch_perf_header_entry(const char * se_header)109 const char *arch_perf_header_entry(const char *se_header)
110 {
111 if (!strcmp(se_header, "Local Pipeline Stage Cycle"))
112 return "Local Retire Latency";
113 else if (!strcmp(se_header, "Pipeline Stage Cycle"))
114 return "Retire Latency";
115
116 return se_header;
117 }
118
arch_support_sort_key(const char * sort_key)119 int arch_support_sort_key(const char *sort_key)
120 {
121 if (!strcmp(sort_key, "p_stage_cyc"))
122 return 1;
123 if (!strcmp(sort_key, "local_p_stage_cyc"))
124 return 1;
125 return 0;
126 }
127