1 #include <linux/kernel.h> 2 3 #include <unistd.h> 4 #include <sys/types.h> 5 6 #include "session.h" 7 #include "sort.h" 8 #include "util.h" 9 10 static int perf_session__open(struct perf_session *self, bool force) 11 { 12 struct stat input_stat; 13 14 self->fd = open(self->filename, O_RDONLY); 15 if (self->fd < 0) { 16 pr_err("failed to open file: %s", self->filename); 17 if (!strcmp(self->filename, "perf.data")) 18 pr_err(" (try 'perf record' first)"); 19 pr_err("\n"); 20 return -errno; 21 } 22 23 if (fstat(self->fd, &input_stat) < 0) 24 goto out_close; 25 26 if (!force && input_stat.st_uid && (input_stat.st_uid != geteuid())) { 27 pr_err("file %s not owned by current user or root\n", 28 self->filename); 29 goto out_close; 30 } 31 32 if (!input_stat.st_size) { 33 pr_info("zero-sized file (%s), nothing to do!\n", 34 self->filename); 35 goto out_close; 36 } 37 38 if (perf_header__read(&self->header, self->fd) < 0) { 39 pr_err("incompatible file format"); 40 goto out_close; 41 } 42 43 self->size = input_stat.st_size; 44 return 0; 45 46 out_close: 47 close(self->fd); 48 self->fd = -1; 49 return -1; 50 } 51 52 struct perf_session *perf_session__new(const char *filename, int mode, bool force) 53 { 54 size_t len = filename ? strlen(filename) + 1 : 0; 55 struct perf_session *self = zalloc(sizeof(*self) + len); 56 57 if (self == NULL) 58 goto out; 59 60 if (perf_header__init(&self->header) < 0) 61 goto out_free; 62 63 memcpy(self->filename, filename, len); 64 self->threads = RB_ROOT; 65 self->last_match = NULL; 66 self->mmap_window = 32; 67 self->cwd = NULL; 68 self->cwdlen = 0; 69 self->unknown_events = 0; 70 map_groups__init(&self->kmaps); 71 72 if (mode == O_RDONLY && perf_session__open(self, force) < 0) 73 goto out_delete; 74 75 self->sample_type = perf_header__sample_type(&self->header); 76 out: 77 return self; 78 out_free: 79 free(self); 80 return NULL; 81 out_delete: 82 perf_session__delete(self); 83 return NULL; 84 } 85 86 void perf_session__delete(struct perf_session *self) 87 { 88 perf_header__exit(&self->header); 89 close(self->fd); 90 free(self->cwd); 91 free(self); 92 } 93 94 static bool symbol__match_parent_regex(struct symbol *sym) 95 { 96 if (sym->name && !regexec(&parent_regex, sym->name, 0, NULL, 0)) 97 return 1; 98 99 return 0; 100 } 101 102 struct symbol **perf_session__resolve_callchain(struct perf_session *self, 103 struct thread *thread, 104 struct ip_callchain *chain, 105 struct symbol **parent) 106 { 107 u8 cpumode = PERF_RECORD_MISC_USER; 108 struct symbol **syms = NULL; 109 unsigned int i; 110 111 if (symbol_conf.use_callchain) { 112 syms = calloc(chain->nr, sizeof(*syms)); 113 if (!syms) { 114 fprintf(stderr, "Can't allocate memory for symbols\n"); 115 exit(-1); 116 } 117 } 118 119 for (i = 0; i < chain->nr; i++) { 120 u64 ip = chain->ips[i]; 121 struct addr_location al; 122 123 if (ip >= PERF_CONTEXT_MAX) { 124 switch (ip) { 125 case PERF_CONTEXT_HV: 126 cpumode = PERF_RECORD_MISC_HYPERVISOR; break; 127 case PERF_CONTEXT_KERNEL: 128 cpumode = PERF_RECORD_MISC_KERNEL; break; 129 case PERF_CONTEXT_USER: 130 cpumode = PERF_RECORD_MISC_USER; break; 131 default: 132 break; 133 } 134 continue; 135 } 136 137 thread__find_addr_location(thread, self, cpumode, 138 MAP__FUNCTION, ip, &al, NULL); 139 if (al.sym != NULL) { 140 if (sort__has_parent && !*parent && 141 symbol__match_parent_regex(al.sym)) 142 *parent = al.sym; 143 if (!symbol_conf.use_callchain) 144 break; 145 syms[i] = al.sym; 146 } 147 } 148 149 return syms; 150 } 151 152 static int process_event_stub(event_t *event __used, 153 struct perf_session *session __used) 154 { 155 dump_printf(": unhandled!\n"); 156 return 0; 157 } 158 159 static void perf_event_ops__fill_defaults(struct perf_event_ops *handler) 160 { 161 if (handler->sample == NULL) 162 handler->sample = process_event_stub; 163 if (handler->mmap == NULL) 164 handler->mmap = process_event_stub; 165 if (handler->comm == NULL) 166 handler->comm = process_event_stub; 167 if (handler->fork == NULL) 168 handler->fork = process_event_stub; 169 if (handler->exit == NULL) 170 handler->exit = process_event_stub; 171 if (handler->lost == NULL) 172 handler->lost = process_event_stub; 173 if (handler->read == NULL) 174 handler->read = process_event_stub; 175 if (handler->throttle == NULL) 176 handler->throttle = process_event_stub; 177 if (handler->unthrottle == NULL) 178 handler->unthrottle = process_event_stub; 179 } 180 181 static const char *event__name[] = { 182 [0] = "TOTAL", 183 [PERF_RECORD_MMAP] = "MMAP", 184 [PERF_RECORD_LOST] = "LOST", 185 [PERF_RECORD_COMM] = "COMM", 186 [PERF_RECORD_EXIT] = "EXIT", 187 [PERF_RECORD_THROTTLE] = "THROTTLE", 188 [PERF_RECORD_UNTHROTTLE] = "UNTHROTTLE", 189 [PERF_RECORD_FORK] = "FORK", 190 [PERF_RECORD_READ] = "READ", 191 [PERF_RECORD_SAMPLE] = "SAMPLE", 192 }; 193 194 unsigned long event__total[PERF_RECORD_MAX]; 195 196 void event__print_totals(void) 197 { 198 int i; 199 for (i = 0; i < PERF_RECORD_MAX; ++i) 200 pr_info("%10s events: %10ld\n", 201 event__name[i], event__total[i]); 202 } 203 204 static int perf_session__process_event(struct perf_session *self, 205 event_t *event, 206 struct perf_event_ops *ops, 207 unsigned long offset, unsigned long head) 208 { 209 trace_event(event); 210 211 if (event->header.type < PERF_RECORD_MAX) { 212 dump_printf("%#lx [%#x]: PERF_RECORD_%s", 213 offset + head, event->header.size, 214 event__name[event->header.type]); 215 ++event__total[0]; 216 ++event__total[event->header.type]; 217 } 218 219 switch (event->header.type) { 220 case PERF_RECORD_SAMPLE: 221 return ops->sample(event, self); 222 case PERF_RECORD_MMAP: 223 return ops->mmap(event, self); 224 case PERF_RECORD_COMM: 225 return ops->comm(event, self); 226 case PERF_RECORD_FORK: 227 return ops->fork(event, self); 228 case PERF_RECORD_EXIT: 229 return ops->exit(event, self); 230 case PERF_RECORD_LOST: 231 return ops->lost(event, self); 232 case PERF_RECORD_READ: 233 return ops->read(event, self); 234 case PERF_RECORD_THROTTLE: 235 return ops->throttle(event, self); 236 case PERF_RECORD_UNTHROTTLE: 237 return ops->unthrottle(event, self); 238 default: 239 self->unknown_events++; 240 return -1; 241 } 242 } 243 244 int perf_header__read_build_ids(int input, u64 offset, u64 size) 245 { 246 struct build_id_event bev; 247 char filename[PATH_MAX]; 248 u64 limit = offset + size; 249 int err = -1; 250 251 while (offset < limit) { 252 struct dso *dso; 253 ssize_t len; 254 struct list_head *head = &dsos__user; 255 256 if (read(input, &bev, sizeof(bev)) != sizeof(bev)) 257 goto out; 258 259 len = bev.header.size - sizeof(bev); 260 if (read(input, filename, len) != len) 261 goto out; 262 263 if (bev.header.misc & PERF_RECORD_MISC_KERNEL) 264 head = &dsos__kernel; 265 266 dso = __dsos__findnew(head, filename); 267 if (dso != NULL) { 268 dso__set_build_id(dso, &bev.build_id); 269 if (head == &dsos__kernel && filename[0] == '[') 270 dso->kernel = 1; 271 } 272 273 offset += bev.header.size; 274 } 275 err = 0; 276 out: 277 return err; 278 } 279 280 static struct thread *perf_session__register_idle_thread(struct perf_session *self) 281 { 282 struct thread *thread = perf_session__findnew(self, 0); 283 284 if (thread == NULL || thread__set_comm(thread, "swapper")) { 285 pr_err("problem inserting idle task.\n"); 286 thread = NULL; 287 } 288 289 return thread; 290 } 291 292 int perf_session__process_events(struct perf_session *self, 293 struct perf_event_ops *ops) 294 { 295 int err; 296 unsigned long head, shift; 297 unsigned long offset = 0; 298 size_t page_size; 299 event_t *event; 300 uint32_t size; 301 char *buf; 302 303 if (perf_session__register_idle_thread(self) == NULL) 304 return -ENOMEM; 305 306 perf_event_ops__fill_defaults(ops); 307 308 page_size = getpagesize(); 309 310 head = self->header.data_offset; 311 312 if (!symbol_conf.full_paths) { 313 char bf[PATH_MAX]; 314 315 if (getcwd(bf, sizeof(bf)) == NULL) { 316 err = -errno; 317 out_getcwd_err: 318 pr_err("failed to get the current directory\n"); 319 goto out_err; 320 } 321 self->cwd = strdup(bf); 322 if (self->cwd == NULL) { 323 err = -ENOMEM; 324 goto out_getcwd_err; 325 } 326 self->cwdlen = strlen(self->cwd); 327 } 328 329 shift = page_size * (head / page_size); 330 offset += shift; 331 head -= shift; 332 333 remap: 334 buf = mmap(NULL, page_size * self->mmap_window, PROT_READ, 335 MAP_SHARED, self->fd, offset); 336 if (buf == MAP_FAILED) { 337 pr_err("failed to mmap file\n"); 338 err = -errno; 339 goto out_err; 340 } 341 342 more: 343 event = (event_t *)(buf + head); 344 345 size = event->header.size; 346 if (size == 0) 347 size = 8; 348 349 if (head + event->header.size >= page_size * self->mmap_window) { 350 int munmap_ret; 351 352 shift = page_size * (head / page_size); 353 354 munmap_ret = munmap(buf, page_size * self->mmap_window); 355 assert(munmap_ret == 0); 356 357 offset += shift; 358 head -= shift; 359 goto remap; 360 } 361 362 size = event->header.size; 363 364 dump_printf("\n%#lx [%#x]: event: %d\n", 365 offset + head, event->header.size, event->header.type); 366 367 if (size == 0 || 368 perf_session__process_event(self, event, ops, offset, head) < 0) { 369 dump_printf("%#lx [%#x]: skipping unknown header type: %d\n", 370 offset + head, event->header.size, 371 event->header.type); 372 /* 373 * assume we lost track of the stream, check alignment, and 374 * increment a single u64 in the hope to catch on again 'soon'. 375 */ 376 if (unlikely(head & 7)) 377 head &= ~7ULL; 378 379 size = 8; 380 } 381 382 head += size; 383 384 if (offset + head >= self->header.data_offset + self->header.data_size) 385 goto done; 386 387 if (offset + head < self->size) 388 goto more; 389 done: 390 err = 0; 391 out_err: 392 return err; 393 } 394 395 bool perf_session__has_traces(struct perf_session *self, const char *msg) 396 { 397 if (!(self->sample_type & PERF_SAMPLE_RAW)) { 398 pr_err("No trace sample to read. Did you call 'perf %s'?\n", msg); 399 return false; 400 } 401 402 return true; 403 } 404 405 int perf_session__set_kallsyms_ref_reloc_sym(struct perf_session *self, 406 const char *symbol_name, 407 u64 addr) 408 { 409 char *bracket; 410 411 self->ref_reloc_sym.name = strdup(symbol_name); 412 if (self->ref_reloc_sym.name == NULL) 413 return -ENOMEM; 414 415 bracket = strchr(self->ref_reloc_sym.name, ']'); 416 if (bracket) 417 *bracket = '\0'; 418 419 self->ref_reloc_sym.addr = addr; 420 return 0; 421 } 422 423 static u64 map__reloc_map_ip(struct map *map, u64 ip) 424 { 425 return ip + (s64)map->pgoff; 426 } 427 428 static u64 map__reloc_unmap_ip(struct map *map, u64 ip) 429 { 430 return ip - (s64)map->pgoff; 431 } 432 433 void perf_session__reloc_vmlinux_maps(struct perf_session *self, 434 u64 unrelocated_addr) 435 { 436 enum map_type type; 437 s64 reloc = unrelocated_addr - self->ref_reloc_sym.addr; 438 439 if (!reloc) 440 return; 441 442 for (type = 0; type < MAP__NR_TYPES; ++type) { 443 struct map *map = self->vmlinux_maps[type]; 444 445 map->map_ip = map__reloc_map_ip; 446 map->unmap_ip = map__reloc_unmap_ip; 447 map->pgoff = reloc; 448 } 449 } 450