19251f904SBorislav Petkov /* 29251f904SBorislav Petkov * Performance events callchain code, extracted from core.c: 39251f904SBorislav Petkov * 49251f904SBorislav Petkov * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de> 59251f904SBorislav Petkov * Copyright (C) 2008-2011 Red Hat, Inc., Ingo Molnar 690eec103SPeter Zijlstra * Copyright (C) 2008-2011 Red Hat, Inc., Peter Zijlstra 79251f904SBorislav Petkov * Copyright � 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> 89251f904SBorislav Petkov * 99251f904SBorislav Petkov * For licensing details see kernel-base/COPYING 109251f904SBorislav Petkov */ 119251f904SBorislav Petkov 129251f904SBorislav Petkov #include <linux/perf_event.h> 139251f904SBorislav Petkov #include <linux/slab.h> 149251f904SBorislav Petkov #include "internal.h" 159251f904SBorislav Petkov 169251f904SBorislav Petkov struct callchain_cpus_entries { 179251f904SBorislav Petkov struct rcu_head rcu_head; 189251f904SBorislav Petkov struct perf_callchain_entry *cpu_entries[0]; 199251f904SBorislav Petkov }; 209251f904SBorislav Petkov 21c5dfd78eSArnaldo Carvalho de Melo int sysctl_perf_event_max_stack __read_mostly = PERF_MAX_STACK_DEPTH; 22c85b0334SArnaldo Carvalho de Melo int sysctl_perf_event_max_contexts_per_stack __read_mostly = PERF_MAX_CONTEXTS_PER_STACK; 23c5dfd78eSArnaldo Carvalho de Melo 24c5dfd78eSArnaldo Carvalho de Melo static inline size_t perf_callchain_entry__sizeof(void) 25c5dfd78eSArnaldo Carvalho de Melo { 26c5dfd78eSArnaldo Carvalho de Melo return (sizeof(struct perf_callchain_entry) + 27c85b0334SArnaldo Carvalho de Melo sizeof(__u64) * (sysctl_perf_event_max_stack + 28c85b0334SArnaldo Carvalho de Melo sysctl_perf_event_max_contexts_per_stack)); 29c5dfd78eSArnaldo Carvalho de Melo } 30c5dfd78eSArnaldo Carvalho de Melo 319251f904SBorislav Petkov static DEFINE_PER_CPU(int, callchain_recursion[PERF_NR_CONTEXTS]); 329251f904SBorislav Petkov static atomic_t nr_callchain_events; 339251f904SBorislav Petkov static DEFINE_MUTEX(callchain_mutex); 349251f904SBorislav Petkov static struct callchain_cpus_entries *callchain_cpus_entries; 359251f904SBorislav Petkov 369251f904SBorislav Petkov 37cfbcf468SArnaldo Carvalho de Melo __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, 389251f904SBorislav Petkov struct pt_regs *regs) 399251f904SBorislav Petkov { 409251f904SBorislav Petkov } 419251f904SBorislav Petkov 42cfbcf468SArnaldo Carvalho de Melo __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry, 439251f904SBorislav Petkov struct pt_regs *regs) 449251f904SBorislav Petkov { 459251f904SBorislav Petkov } 469251f904SBorislav Petkov 479251f904SBorislav Petkov static void release_callchain_buffers_rcu(struct rcu_head *head) 489251f904SBorislav Petkov { 499251f904SBorislav Petkov struct callchain_cpus_entries *entries; 509251f904SBorislav Petkov int cpu; 519251f904SBorislav Petkov 529251f904SBorislav Petkov entries = container_of(head, struct callchain_cpus_entries, rcu_head); 539251f904SBorislav Petkov 549251f904SBorislav Petkov for_each_possible_cpu(cpu) 559251f904SBorislav Petkov kfree(entries->cpu_entries[cpu]); 569251f904SBorislav Petkov 579251f904SBorislav Petkov kfree(entries); 589251f904SBorislav Petkov } 599251f904SBorislav Petkov 609251f904SBorislav Petkov static void release_callchain_buffers(void) 619251f904SBorislav Petkov { 629251f904SBorislav Petkov struct callchain_cpus_entries *entries; 639251f904SBorislav Petkov 649251f904SBorislav Petkov entries = callchain_cpus_entries; 65e0455e19SAndreea-Cristina Bernat RCU_INIT_POINTER(callchain_cpus_entries, NULL); 669251f904SBorislav Petkov call_rcu(&entries->rcu_head, release_callchain_buffers_rcu); 679251f904SBorislav Petkov } 689251f904SBorislav Petkov 699251f904SBorislav Petkov static int alloc_callchain_buffers(void) 709251f904SBorislav Petkov { 719251f904SBorislav Petkov int cpu; 729251f904SBorislav Petkov int size; 739251f904SBorislav Petkov struct callchain_cpus_entries *entries; 749251f904SBorislav Petkov 759251f904SBorislav Petkov /* 769251f904SBorislav Petkov * We can't use the percpu allocation API for data that can be 779251f904SBorislav Petkov * accessed from NMI. Use a temporary manual per cpu allocation 789251f904SBorislav Petkov * until that gets sorted out. 799251f904SBorislav Petkov */ 809251f904SBorislav Petkov size = offsetof(struct callchain_cpus_entries, cpu_entries[nr_cpu_ids]); 819251f904SBorislav Petkov 829251f904SBorislav Petkov entries = kzalloc(size, GFP_KERNEL); 839251f904SBorislav Petkov if (!entries) 849251f904SBorislav Petkov return -ENOMEM; 859251f904SBorislav Petkov 86c5dfd78eSArnaldo Carvalho de Melo size = perf_callchain_entry__sizeof() * PERF_NR_CONTEXTS; 879251f904SBorislav Petkov 889251f904SBorislav Petkov for_each_possible_cpu(cpu) { 899251f904SBorislav Petkov entries->cpu_entries[cpu] = kmalloc_node(size, GFP_KERNEL, 909251f904SBorislav Petkov cpu_to_node(cpu)); 919251f904SBorislav Petkov if (!entries->cpu_entries[cpu]) 929251f904SBorislav Petkov goto fail; 939251f904SBorislav Petkov } 949251f904SBorislav Petkov 959251f904SBorislav Petkov rcu_assign_pointer(callchain_cpus_entries, entries); 969251f904SBorislav Petkov 979251f904SBorislav Petkov return 0; 989251f904SBorislav Petkov 999251f904SBorislav Petkov fail: 1009251f904SBorislav Petkov for_each_possible_cpu(cpu) 1019251f904SBorislav Petkov kfree(entries->cpu_entries[cpu]); 1029251f904SBorislav Petkov kfree(entries); 1039251f904SBorislav Petkov 1049251f904SBorislav Petkov return -ENOMEM; 1059251f904SBorislav Petkov } 1069251f904SBorislav Petkov 107*97c79a38SArnaldo Carvalho de Melo int get_callchain_buffers(int event_max_stack) 1089251f904SBorislav Petkov { 1099251f904SBorislav Petkov int err = 0; 1109251f904SBorislav Petkov int count; 1119251f904SBorislav Petkov 1129251f904SBorislav Petkov mutex_lock(&callchain_mutex); 1139251f904SBorislav Petkov 1149251f904SBorislav Petkov count = atomic_inc_return(&nr_callchain_events); 1159251f904SBorislav Petkov if (WARN_ON_ONCE(count < 1)) { 1169251f904SBorislav Petkov err = -EINVAL; 1179251f904SBorislav Petkov goto exit; 1189251f904SBorislav Petkov } 1199251f904SBorislav Petkov 1209251f904SBorislav Petkov if (count > 1) { 1219251f904SBorislav Petkov /* If the allocation failed, give up */ 1229251f904SBorislav Petkov if (!callchain_cpus_entries) 1239251f904SBorislav Petkov err = -ENOMEM; 124*97c79a38SArnaldo Carvalho de Melo /* 125*97c79a38SArnaldo Carvalho de Melo * If requesting per event more than the global cap, 126*97c79a38SArnaldo Carvalho de Melo * return a different error to help userspace figure 127*97c79a38SArnaldo Carvalho de Melo * this out. 128*97c79a38SArnaldo Carvalho de Melo * 129*97c79a38SArnaldo Carvalho de Melo * And also do it here so that we have &callchain_mutex held. 130*97c79a38SArnaldo Carvalho de Melo */ 131*97c79a38SArnaldo Carvalho de Melo if (event_max_stack > sysctl_perf_event_max_stack) 132*97c79a38SArnaldo Carvalho de Melo err = -EOVERFLOW; 1339251f904SBorislav Petkov goto exit; 1349251f904SBorislav Petkov } 1359251f904SBorislav Petkov 1369251f904SBorislav Petkov err = alloc_callchain_buffers(); 1379251f904SBorislav Petkov exit: 13890983b16SFrederic Weisbecker if (err) 13990983b16SFrederic Weisbecker atomic_dec(&nr_callchain_events); 1409251f904SBorislav Petkov 141fc3b86d6SFrederic Weisbecker mutex_unlock(&callchain_mutex); 142fc3b86d6SFrederic Weisbecker 1439251f904SBorislav Petkov return err; 1449251f904SBorislav Petkov } 1459251f904SBorislav Petkov 1469251f904SBorislav Petkov void put_callchain_buffers(void) 1479251f904SBorislav Petkov { 1489251f904SBorislav Petkov if (atomic_dec_and_mutex_lock(&nr_callchain_events, &callchain_mutex)) { 1499251f904SBorislav Petkov release_callchain_buffers(); 1509251f904SBorislav Petkov mutex_unlock(&callchain_mutex); 1519251f904SBorislav Petkov } 1529251f904SBorislav Petkov } 1539251f904SBorislav Petkov 1549251f904SBorislav Petkov static struct perf_callchain_entry *get_callchain_entry(int *rctx) 1559251f904SBorislav Petkov { 1569251f904SBorislav Petkov int cpu; 1579251f904SBorislav Petkov struct callchain_cpus_entries *entries; 1589251f904SBorislav Petkov 1594a32fea9SChristoph Lameter *rctx = get_recursion_context(this_cpu_ptr(callchain_recursion)); 1609251f904SBorislav Petkov if (*rctx == -1) 1619251f904SBorislav Petkov return NULL; 1629251f904SBorislav Petkov 1639251f904SBorislav Petkov entries = rcu_dereference(callchain_cpus_entries); 1649251f904SBorislav Petkov if (!entries) 1659251f904SBorislav Petkov return NULL; 1669251f904SBorislav Petkov 1679251f904SBorislav Petkov cpu = smp_processor_id(); 1689251f904SBorislav Petkov 169c5dfd78eSArnaldo Carvalho de Melo return (((void *)entries->cpu_entries[cpu]) + 170c5dfd78eSArnaldo Carvalho de Melo (*rctx * perf_callchain_entry__sizeof())); 1719251f904SBorislav Petkov } 1729251f904SBorislav Petkov 1739251f904SBorislav Petkov static void 1749251f904SBorislav Petkov put_callchain_entry(int rctx) 1759251f904SBorislav Petkov { 1764a32fea9SChristoph Lameter put_recursion_context(this_cpu_ptr(callchain_recursion), rctx); 1779251f904SBorislav Petkov } 1789251f904SBorislav Petkov 179e6dab5ffSAndrew Vagin struct perf_callchain_entry * 180e6dab5ffSAndrew Vagin perf_callchain(struct perf_event *event, struct pt_regs *regs) 1819251f904SBorislav Petkov { 182568b329aSAlexei Starovoitov bool kernel = !event->attr.exclude_callchain_kernel; 183568b329aSAlexei Starovoitov bool user = !event->attr.exclude_callchain_user; 184568b329aSAlexei Starovoitov /* Disallow cross-task user callchains. */ 185568b329aSAlexei Starovoitov bool crosstask = event->ctx->task && event->ctx->task != current; 186*97c79a38SArnaldo Carvalho de Melo const u32 max_stack = event->attr.sample_max_stack; 187d0775264SFrederic Weisbecker 188d0775264SFrederic Weisbecker if (!kernel && !user) 189d0775264SFrederic Weisbecker return NULL; 1909251f904SBorislav Petkov 191*97c79a38SArnaldo Carvalho de Melo return get_perf_callchain(regs, 0, kernel, user, max_stack, crosstask, true); 192568b329aSAlexei Starovoitov } 193568b329aSAlexei Starovoitov 194568b329aSAlexei Starovoitov struct perf_callchain_entry * 195568b329aSAlexei Starovoitov get_perf_callchain(struct pt_regs *regs, u32 init_nr, bool kernel, bool user, 196cfbcf468SArnaldo Carvalho de Melo u32 max_stack, bool crosstask, bool add_mark) 197568b329aSAlexei Starovoitov { 198568b329aSAlexei Starovoitov struct perf_callchain_entry *entry; 199cfbcf468SArnaldo Carvalho de Melo struct perf_callchain_entry_ctx ctx; 200568b329aSAlexei Starovoitov int rctx; 201568b329aSAlexei Starovoitov 2029251f904SBorislav Petkov entry = get_callchain_entry(&rctx); 2039251f904SBorislav Petkov if (rctx == -1) 2049251f904SBorislav Petkov return NULL; 2059251f904SBorislav Petkov 2069251f904SBorislav Petkov if (!entry) 2079251f904SBorislav Petkov goto exit_put; 2089251f904SBorislav Petkov 209cfbcf468SArnaldo Carvalho de Melo ctx.entry = entry; 210cfbcf468SArnaldo Carvalho de Melo ctx.max_stack = max_stack; 2113b1fff08SArnaldo Carvalho de Melo ctx.nr = entry->nr = init_nr; 212c85b0334SArnaldo Carvalho de Melo ctx.contexts = 0; 213c85b0334SArnaldo Carvalho de Melo ctx.contexts_maxed = false; 2149251f904SBorislav Petkov 215d0775264SFrederic Weisbecker if (kernel && !user_mode(regs)) { 216568b329aSAlexei Starovoitov if (add_mark) 2173e4de4ecSArnaldo Carvalho de Melo perf_callchain_store_context(&ctx, PERF_CONTEXT_KERNEL); 218cfbcf468SArnaldo Carvalho de Melo perf_callchain_kernel(&ctx, regs); 219d0775264SFrederic Weisbecker } 220d0775264SFrederic Weisbecker 221d0775264SFrederic Weisbecker if (user) { 222d0775264SFrederic Weisbecker if (!user_mode(regs)) { 2239251f904SBorislav Petkov if (current->mm) 2249251f904SBorislav Petkov regs = task_pt_regs(current); 2259251f904SBorislav Petkov else 2269251f904SBorislav Petkov regs = NULL; 2279251f904SBorislav Petkov } 2289251f904SBorislav Petkov 2299251f904SBorislav Petkov if (regs) { 230568b329aSAlexei Starovoitov if (crosstask) 231e6dab5ffSAndrew Vagin goto exit_put; 232e6dab5ffSAndrew Vagin 233568b329aSAlexei Starovoitov if (add_mark) 2343e4de4ecSArnaldo Carvalho de Melo perf_callchain_store_context(&ctx, PERF_CONTEXT_USER); 235cfbcf468SArnaldo Carvalho de Melo perf_callchain_user(&ctx, regs); 2369251f904SBorislav Petkov } 237d0775264SFrederic Weisbecker } 2389251f904SBorislav Petkov 2399251f904SBorislav Petkov exit_put: 2409251f904SBorislav Petkov put_callchain_entry(rctx); 2419251f904SBorislav Petkov 2429251f904SBorislav Petkov return entry; 2439251f904SBorislav Petkov } 244c5dfd78eSArnaldo Carvalho de Melo 245c85b0334SArnaldo Carvalho de Melo /* 246c85b0334SArnaldo Carvalho de Melo * Used for sysctl_perf_event_max_stack and 247c85b0334SArnaldo Carvalho de Melo * sysctl_perf_event_max_contexts_per_stack. 248c85b0334SArnaldo Carvalho de Melo */ 249c5dfd78eSArnaldo Carvalho de Melo int perf_event_max_stack_handler(struct ctl_table *table, int write, 250c5dfd78eSArnaldo Carvalho de Melo void __user *buffer, size_t *lenp, loff_t *ppos) 251c5dfd78eSArnaldo Carvalho de Melo { 252a831100aSArnaldo Carvalho de Melo int *value = table->data; 253a831100aSArnaldo Carvalho de Melo int new_value = *value, ret; 254c5dfd78eSArnaldo Carvalho de Melo struct ctl_table new_table = *table; 255c5dfd78eSArnaldo Carvalho de Melo 256c5dfd78eSArnaldo Carvalho de Melo new_table.data = &new_value; 257c5dfd78eSArnaldo Carvalho de Melo ret = proc_dointvec_minmax(&new_table, write, buffer, lenp, ppos); 258c5dfd78eSArnaldo Carvalho de Melo if (ret || !write) 259c5dfd78eSArnaldo Carvalho de Melo return ret; 260c5dfd78eSArnaldo Carvalho de Melo 261c5dfd78eSArnaldo Carvalho de Melo mutex_lock(&callchain_mutex); 262c5dfd78eSArnaldo Carvalho de Melo if (atomic_read(&nr_callchain_events)) 263c5dfd78eSArnaldo Carvalho de Melo ret = -EBUSY; 264c5dfd78eSArnaldo Carvalho de Melo else 265a831100aSArnaldo Carvalho de Melo *value = new_value; 266c5dfd78eSArnaldo Carvalho de Melo 267c5dfd78eSArnaldo Carvalho de Melo mutex_unlock(&callchain_mutex); 268c5dfd78eSArnaldo Carvalho de Melo 269c5dfd78eSArnaldo Carvalho de Melo return ret; 270c5dfd78eSArnaldo Carvalho de Melo } 271