1 // SPDX-License-Identifier: GPL-2.0 2 #include <linux/types.h> 3 #include <linux/string.h> 4 #include <linux/zalloc.h> 5 #include <stdlib.h> 6 7 #include "../../../util/event.h" 8 #include "../../../util/synthetic-events.h" 9 #include "../../../util/machine.h" 10 #include "../../../util/tool.h" 11 #include "../../../util/map.h" 12 #include "../../../util/debug.h" 13 #include "util/sample.h" 14 15 #if defined(__x86_64__) 16 17 int perf_event__synthesize_extra_kmaps(struct perf_tool *tool, 18 perf_event__handler_t process, 19 struct machine *machine) 20 { 21 int rc = 0; 22 struct map_rb_node *pos; 23 struct maps *kmaps = machine__kernel_maps(machine); 24 union perf_event *event = zalloc(sizeof(event->mmap) + 25 machine->id_hdr_size); 26 27 if (!event) { 28 pr_debug("Not enough memory synthesizing mmap event " 29 "for extra kernel maps\n"); 30 return -1; 31 } 32 33 maps__for_each_entry(kmaps, pos) { 34 struct kmap *kmap; 35 size_t size; 36 struct map *map = pos->map; 37 38 if (!__map__is_extra_kernel_map(map)) 39 continue; 40 41 kmap = map__kmap(map); 42 43 size = sizeof(event->mmap) - sizeof(event->mmap.filename) + 44 PERF_ALIGN(strlen(kmap->name) + 1, sizeof(u64)) + 45 machine->id_hdr_size; 46 47 memset(event, 0, size); 48 49 event->mmap.header.type = PERF_RECORD_MMAP; 50 51 /* 52 * kernel uses 0 for user space maps, see kernel/perf_event.c 53 * __perf_event_mmap 54 */ 55 if (machine__is_host(machine)) 56 event->header.misc = PERF_RECORD_MISC_KERNEL; 57 else 58 event->header.misc = PERF_RECORD_MISC_GUEST_KERNEL; 59 60 event->mmap.header.size = size; 61 62 event->mmap.start = map__start(map); 63 event->mmap.len = map__size(map); 64 event->mmap.pgoff = map__pgoff(map); 65 event->mmap.pid = machine->pid; 66 67 strlcpy(event->mmap.filename, kmap->name, PATH_MAX); 68 69 if (perf_tool__process_synth_event(tool, event, machine, 70 process) != 0) { 71 rc = -1; 72 break; 73 } 74 } 75 76 free(event); 77 return rc; 78 } 79 80 #endif 81 82 void arch_perf_parse_sample_weight(struct perf_sample *data, 83 const __u64 *array, u64 type) 84 { 85 union perf_sample_weight weight; 86 87 weight.full = *array; 88 if (type & PERF_SAMPLE_WEIGHT) 89 data->weight = weight.full; 90 else { 91 data->weight = weight.var1_dw; 92 data->ins_lat = weight.var2_w; 93 data->retire_lat = weight.var3_w; 94 } 95 } 96 97 void arch_perf_synthesize_sample_weight(const struct perf_sample *data, 98 __u64 *array, u64 type) 99 { 100 *array = data->weight; 101 102 if (type & PERF_SAMPLE_WEIGHT_STRUCT) { 103 *array &= 0xffffffff; 104 *array |= ((u64)data->ins_lat << 32); 105 *array |= ((u64)data->retire_lat << 48); 106 } 107 } 108 109 const char *arch_perf_header_entry(const char *se_header) 110 { 111 if (!strcmp(se_header, "Local Pipeline Stage Cycle")) 112 return "Local Retire Latency"; 113 else if (!strcmp(se_header, "Pipeline Stage Cycle")) 114 return "Retire Latency"; 115 116 return se_header; 117 } 118 119 int arch_support_sort_key(const char *sort_key) 120 { 121 if (!strcmp(sort_key, "p_stage_cyc")) 122 return 1; 123 if (!strcmp(sort_key, "local_p_stage_cyc")) 124 return 1; 125 return 0; 126 } 127