19251f904SBorislav Petkov /* 29251f904SBorislav Petkov * Performance events callchain code, extracted from core.c: 39251f904SBorislav Petkov * 49251f904SBorislav Petkov * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de> 59251f904SBorislav Petkov * Copyright (C) 2008-2011 Red Hat, Inc., Ingo Molnar 690eec103SPeter Zijlstra * Copyright (C) 2008-2011 Red Hat, Inc., Peter Zijlstra 79251f904SBorislav Petkov * Copyright � 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> 89251f904SBorislav Petkov * 99251f904SBorislav Petkov * For licensing details see kernel-base/COPYING 109251f904SBorislav Petkov */ 119251f904SBorislav Petkov 129251f904SBorislav Petkov #include <linux/perf_event.h> 139251f904SBorislav Petkov #include <linux/slab.h> 1468db0cf1SIngo Molnar #include <linux/sched/task_stack.h> 1568db0cf1SIngo Molnar 169251f904SBorislav Petkov #include "internal.h" 179251f904SBorislav Petkov 189251f904SBorislav Petkov struct callchain_cpus_entries { 199251f904SBorislav Petkov struct rcu_head rcu_head; 209251f904SBorislav Petkov struct perf_callchain_entry *cpu_entries[0]; 219251f904SBorislav Petkov }; 229251f904SBorislav Petkov 23c5dfd78eSArnaldo Carvalho de Melo int sysctl_perf_event_max_stack __read_mostly = PERF_MAX_STACK_DEPTH; 24c85b0334SArnaldo Carvalho de Melo int sysctl_perf_event_max_contexts_per_stack __read_mostly = PERF_MAX_CONTEXTS_PER_STACK; 25c5dfd78eSArnaldo Carvalho de Melo 26c5dfd78eSArnaldo Carvalho de Melo static inline size_t perf_callchain_entry__sizeof(void) 27c5dfd78eSArnaldo Carvalho de Melo { 28c5dfd78eSArnaldo Carvalho de Melo return (sizeof(struct perf_callchain_entry) + 29c85b0334SArnaldo Carvalho de Melo sizeof(__u64) * (sysctl_perf_event_max_stack + 30c85b0334SArnaldo Carvalho de Melo sysctl_perf_event_max_contexts_per_stack)); 31c5dfd78eSArnaldo Carvalho de Melo } 32c5dfd78eSArnaldo Carvalho de Melo 339251f904SBorislav Petkov static DEFINE_PER_CPU(int, callchain_recursion[PERF_NR_CONTEXTS]); 349251f904SBorislav Petkov static atomic_t nr_callchain_events; 359251f904SBorislav Petkov static DEFINE_MUTEX(callchain_mutex); 369251f904SBorislav Petkov static struct callchain_cpus_entries *callchain_cpus_entries; 379251f904SBorislav Petkov 389251f904SBorislav Petkov 39cfbcf468SArnaldo Carvalho de Melo __weak void perf_callchain_kernel(struct perf_callchain_entry_ctx *entry, 409251f904SBorislav Petkov struct pt_regs *regs) 419251f904SBorislav Petkov { 429251f904SBorislav Petkov } 439251f904SBorislav Petkov 44cfbcf468SArnaldo Carvalho de Melo __weak void perf_callchain_user(struct perf_callchain_entry_ctx *entry, 459251f904SBorislav Petkov struct pt_regs *regs) 469251f904SBorislav Petkov { 479251f904SBorislav Petkov } 489251f904SBorislav Petkov 499251f904SBorislav Petkov static void release_callchain_buffers_rcu(struct rcu_head *head) 509251f904SBorislav Petkov { 519251f904SBorislav Petkov struct callchain_cpus_entries *entries; 529251f904SBorislav Petkov int cpu; 539251f904SBorislav Petkov 549251f904SBorislav Petkov entries = container_of(head, struct callchain_cpus_entries, rcu_head); 559251f904SBorislav Petkov 569251f904SBorislav Petkov for_each_possible_cpu(cpu) 579251f904SBorislav Petkov kfree(entries->cpu_entries[cpu]); 589251f904SBorislav Petkov 599251f904SBorislav Petkov kfree(entries); 609251f904SBorislav Petkov } 619251f904SBorislav Petkov 629251f904SBorislav Petkov static void release_callchain_buffers(void) 639251f904SBorislav Petkov { 649251f904SBorislav Petkov struct callchain_cpus_entries *entries; 659251f904SBorislav Petkov 669251f904SBorislav Petkov entries = callchain_cpus_entries; 67e0455e19SAndreea-Cristina Bernat RCU_INIT_POINTER(callchain_cpus_entries, NULL); 689251f904SBorislav Petkov call_rcu(&entries->rcu_head, release_callchain_buffers_rcu); 699251f904SBorislav Petkov } 709251f904SBorislav Petkov 719251f904SBorislav Petkov static int alloc_callchain_buffers(void) 729251f904SBorislav Petkov { 739251f904SBorislav Petkov int cpu; 749251f904SBorislav Petkov int size; 759251f904SBorislav Petkov struct callchain_cpus_entries *entries; 769251f904SBorislav Petkov 779251f904SBorislav Petkov /* 789251f904SBorislav Petkov * We can't use the percpu allocation API for data that can be 799251f904SBorislav Petkov * accessed from NMI. Use a temporary manual per cpu allocation 809251f904SBorislav Petkov * until that gets sorted out. 819251f904SBorislav Petkov */ 829251f904SBorislav Petkov size = offsetof(struct callchain_cpus_entries, cpu_entries[nr_cpu_ids]); 839251f904SBorislav Petkov 849251f904SBorislav Petkov entries = kzalloc(size, GFP_KERNEL); 859251f904SBorislav Petkov if (!entries) 869251f904SBorislav Petkov return -ENOMEM; 879251f904SBorislav Petkov 88c5dfd78eSArnaldo Carvalho de Melo size = perf_callchain_entry__sizeof() * PERF_NR_CONTEXTS; 899251f904SBorislav Petkov 909251f904SBorislav Petkov for_each_possible_cpu(cpu) { 919251f904SBorislav Petkov entries->cpu_entries[cpu] = kmalloc_node(size, GFP_KERNEL, 929251f904SBorislav Petkov cpu_to_node(cpu)); 939251f904SBorislav Petkov if (!entries->cpu_entries[cpu]) 949251f904SBorislav Petkov goto fail; 959251f904SBorislav Petkov } 969251f904SBorislav Petkov 979251f904SBorislav Petkov rcu_assign_pointer(callchain_cpus_entries, entries); 989251f904SBorislav Petkov 999251f904SBorislav Petkov return 0; 1009251f904SBorislav Petkov 1019251f904SBorislav Petkov fail: 1029251f904SBorislav Petkov for_each_possible_cpu(cpu) 1039251f904SBorislav Petkov kfree(entries->cpu_entries[cpu]); 1049251f904SBorislav Petkov kfree(entries); 1059251f904SBorislav Petkov 1069251f904SBorislav Petkov return -ENOMEM; 1079251f904SBorislav Petkov } 1089251f904SBorislav Petkov 10997c79a38SArnaldo Carvalho de Melo int get_callchain_buffers(int event_max_stack) 1109251f904SBorislav Petkov { 1119251f904SBorislav Petkov int err = 0; 1129251f904SBorislav Petkov int count; 1139251f904SBorislav Petkov 1149251f904SBorislav Petkov mutex_lock(&callchain_mutex); 1159251f904SBorislav Petkov 1169251f904SBorislav Petkov count = atomic_inc_return(&nr_callchain_events); 1179251f904SBorislav Petkov if (WARN_ON_ONCE(count < 1)) { 1189251f904SBorislav Petkov err = -EINVAL; 1199251f904SBorislav Petkov goto exit; 1209251f904SBorislav Petkov } 1219251f904SBorislav Petkov 12297c79a38SArnaldo Carvalho de Melo /* 12397c79a38SArnaldo Carvalho de Melo * If requesting per event more than the global cap, 12497c79a38SArnaldo Carvalho de Melo * return a different error to help userspace figure 12597c79a38SArnaldo Carvalho de Melo * this out. 12697c79a38SArnaldo Carvalho de Melo * 12797c79a38SArnaldo Carvalho de Melo * And also do it here so that we have &callchain_mutex held. 12897c79a38SArnaldo Carvalho de Melo */ 129*5af44ca5SJiri Olsa if (event_max_stack > sysctl_perf_event_max_stack) { 13097c79a38SArnaldo Carvalho de Melo err = -EOVERFLOW; 1319251f904SBorislav Petkov goto exit; 1329251f904SBorislav Petkov } 1339251f904SBorislav Petkov 134*5af44ca5SJiri Olsa if (count > 1) { 135*5af44ca5SJiri Olsa /* If the allocation failed, give up */ 136*5af44ca5SJiri Olsa if (!callchain_cpus_entries) 137*5af44ca5SJiri Olsa err = -ENOMEM; 138*5af44ca5SJiri Olsa goto exit; 139*5af44ca5SJiri Olsa } 140*5af44ca5SJiri Olsa 1419251f904SBorislav Petkov err = alloc_callchain_buffers(); 1429251f904SBorislav Petkov exit: 14390983b16SFrederic Weisbecker if (err) 14490983b16SFrederic Weisbecker atomic_dec(&nr_callchain_events); 1459251f904SBorislav Petkov 146fc3b86d6SFrederic Weisbecker mutex_unlock(&callchain_mutex); 147fc3b86d6SFrederic Weisbecker 1489251f904SBorislav Petkov return err; 1499251f904SBorislav Petkov } 1509251f904SBorislav Petkov 1519251f904SBorislav Petkov void put_callchain_buffers(void) 1529251f904SBorislav Petkov { 1539251f904SBorislav Petkov if (atomic_dec_and_mutex_lock(&nr_callchain_events, &callchain_mutex)) { 1549251f904SBorislav Petkov release_callchain_buffers(); 1559251f904SBorislav Petkov mutex_unlock(&callchain_mutex); 1569251f904SBorislav Petkov } 1579251f904SBorislav Petkov } 1589251f904SBorislav Petkov 1599251f904SBorislav Petkov static struct perf_callchain_entry *get_callchain_entry(int *rctx) 1609251f904SBorislav Petkov { 1619251f904SBorislav Petkov int cpu; 1629251f904SBorislav Petkov struct callchain_cpus_entries *entries; 1639251f904SBorislav Petkov 1644a32fea9SChristoph Lameter *rctx = get_recursion_context(this_cpu_ptr(callchain_recursion)); 1659251f904SBorislav Petkov if (*rctx == -1) 1669251f904SBorislav Petkov return NULL; 1679251f904SBorislav Petkov 1689251f904SBorislav Petkov entries = rcu_dereference(callchain_cpus_entries); 1699251f904SBorislav Petkov if (!entries) 1709251f904SBorislav Petkov return NULL; 1719251f904SBorislav Petkov 1729251f904SBorislav Petkov cpu = smp_processor_id(); 1739251f904SBorislav Petkov 174c5dfd78eSArnaldo Carvalho de Melo return (((void *)entries->cpu_entries[cpu]) + 175c5dfd78eSArnaldo Carvalho de Melo (*rctx * perf_callchain_entry__sizeof())); 1769251f904SBorislav Petkov } 1779251f904SBorislav Petkov 1789251f904SBorislav Petkov static void 1799251f904SBorislav Petkov put_callchain_entry(int rctx) 1809251f904SBorislav Petkov { 1814a32fea9SChristoph Lameter put_recursion_context(this_cpu_ptr(callchain_recursion), rctx); 1829251f904SBorislav Petkov } 1839251f904SBorislav Petkov 184e6dab5ffSAndrew Vagin struct perf_callchain_entry * 185568b329aSAlexei Starovoitov get_perf_callchain(struct pt_regs *regs, u32 init_nr, bool kernel, bool user, 186cfbcf468SArnaldo Carvalho de Melo u32 max_stack, bool crosstask, bool add_mark) 187568b329aSAlexei Starovoitov { 188568b329aSAlexei Starovoitov struct perf_callchain_entry *entry; 189cfbcf468SArnaldo Carvalho de Melo struct perf_callchain_entry_ctx ctx; 190568b329aSAlexei Starovoitov int rctx; 191568b329aSAlexei Starovoitov 1929251f904SBorislav Petkov entry = get_callchain_entry(&rctx); 1939251f904SBorislav Petkov if (rctx == -1) 1949251f904SBorislav Petkov return NULL; 1959251f904SBorislav Petkov 1969251f904SBorislav Petkov if (!entry) 1979251f904SBorislav Petkov goto exit_put; 1989251f904SBorislav Petkov 199cfbcf468SArnaldo Carvalho de Melo ctx.entry = entry; 200cfbcf468SArnaldo Carvalho de Melo ctx.max_stack = max_stack; 2013b1fff08SArnaldo Carvalho de Melo ctx.nr = entry->nr = init_nr; 202c85b0334SArnaldo Carvalho de Melo ctx.contexts = 0; 203c85b0334SArnaldo Carvalho de Melo ctx.contexts_maxed = false; 2049251f904SBorislav Petkov 205d0775264SFrederic Weisbecker if (kernel && !user_mode(regs)) { 206568b329aSAlexei Starovoitov if (add_mark) 2073e4de4ecSArnaldo Carvalho de Melo perf_callchain_store_context(&ctx, PERF_CONTEXT_KERNEL); 208cfbcf468SArnaldo Carvalho de Melo perf_callchain_kernel(&ctx, regs); 209d0775264SFrederic Weisbecker } 210d0775264SFrederic Weisbecker 211d0775264SFrederic Weisbecker if (user) { 212d0775264SFrederic Weisbecker if (!user_mode(regs)) { 2139251f904SBorislav Petkov if (current->mm) 2149251f904SBorislav Petkov regs = task_pt_regs(current); 2159251f904SBorislav Petkov else 2169251f904SBorislav Petkov regs = NULL; 2179251f904SBorislav Petkov } 2189251f904SBorislav Petkov 2199251f904SBorislav Petkov if (regs) { 22088b0193dSWill Deacon mm_segment_t fs; 22188b0193dSWill Deacon 222568b329aSAlexei Starovoitov if (crosstask) 223e6dab5ffSAndrew Vagin goto exit_put; 224e6dab5ffSAndrew Vagin 225568b329aSAlexei Starovoitov if (add_mark) 2263e4de4ecSArnaldo Carvalho de Melo perf_callchain_store_context(&ctx, PERF_CONTEXT_USER); 22788b0193dSWill Deacon 22888b0193dSWill Deacon fs = get_fs(); 22988b0193dSWill Deacon set_fs(USER_DS); 230cfbcf468SArnaldo Carvalho de Melo perf_callchain_user(&ctx, regs); 23188b0193dSWill Deacon set_fs(fs); 2329251f904SBorislav Petkov } 233d0775264SFrederic Weisbecker } 2349251f904SBorislav Petkov 2359251f904SBorislav Petkov exit_put: 2369251f904SBorislav Petkov put_callchain_entry(rctx); 2379251f904SBorislav Petkov 2389251f904SBorislav Petkov return entry; 2399251f904SBorislav Petkov } 240c5dfd78eSArnaldo Carvalho de Melo 241c85b0334SArnaldo Carvalho de Melo /* 242c85b0334SArnaldo Carvalho de Melo * Used for sysctl_perf_event_max_stack and 243c85b0334SArnaldo Carvalho de Melo * sysctl_perf_event_max_contexts_per_stack. 244c85b0334SArnaldo Carvalho de Melo */ 245c5dfd78eSArnaldo Carvalho de Melo int perf_event_max_stack_handler(struct ctl_table *table, int write, 246c5dfd78eSArnaldo Carvalho de Melo void __user *buffer, size_t *lenp, loff_t *ppos) 247c5dfd78eSArnaldo Carvalho de Melo { 248a831100aSArnaldo Carvalho de Melo int *value = table->data; 249a831100aSArnaldo Carvalho de Melo int new_value = *value, ret; 250c5dfd78eSArnaldo Carvalho de Melo struct ctl_table new_table = *table; 251c5dfd78eSArnaldo Carvalho de Melo 252c5dfd78eSArnaldo Carvalho de Melo new_table.data = &new_value; 253c5dfd78eSArnaldo Carvalho de Melo ret = proc_dointvec_minmax(&new_table, write, buffer, lenp, ppos); 254c5dfd78eSArnaldo Carvalho de Melo if (ret || !write) 255c5dfd78eSArnaldo Carvalho de Melo return ret; 256c5dfd78eSArnaldo Carvalho de Melo 257c5dfd78eSArnaldo Carvalho de Melo mutex_lock(&callchain_mutex); 258c5dfd78eSArnaldo Carvalho de Melo if (atomic_read(&nr_callchain_events)) 259c5dfd78eSArnaldo Carvalho de Melo ret = -EBUSY; 260c5dfd78eSArnaldo Carvalho de Melo else 261a831100aSArnaldo Carvalho de Melo *value = new_value; 262c5dfd78eSArnaldo Carvalho de Melo 263c5dfd78eSArnaldo Carvalho de Melo mutex_unlock(&callchain_mutex); 264c5dfd78eSArnaldo Carvalho de Melo 265c5dfd78eSArnaldo Carvalho de Melo return ret; 266c5dfd78eSArnaldo Carvalho de Melo } 267