1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * Infrastructure for statistic tracing (histogram output). 4 * 5 * Copyright (C) 2008-2009 Frederic Weisbecker <fweisbec@gmail.com> 6 * 7 * Based on the code from trace_branch.c which is 8 * Copyright (C) 2008 Steven Rostedt <srostedt@redhat.com> 9 * 10 */ 11 12 #include <linux/security.h> 13 #include <linux/list.h> 14 #include <linux/slab.h> 15 #include <linux/rbtree.h> 16 #include <linux/tracefs.h> 17 #include "trace_stat.h" 18 #include "trace.h" 19 20 21 /* 22 * List of stat red-black nodes from a tracer 23 * We use a such tree to sort quickly the stat 24 * entries from the tracer. 25 */ 26 struct stat_node { 27 struct rb_node node; 28 void *stat; 29 }; 30 31 /* A stat session is the stats output in one file */ 32 struct stat_session { 33 struct list_head session_list; 34 struct tracer_stat *ts; 35 struct rb_root stat_root; 36 struct mutex stat_mutex; 37 struct dentry *file; 38 }; 39 40 /* All of the sessions currently in use. Each stat file embed one session */ 41 static LIST_HEAD(all_stat_sessions); 42 static DEFINE_MUTEX(all_stat_sessions_mutex); 43 44 /* The root directory for all stat files */ 45 static struct dentry *stat_dir; 46 47 static void __reset_stat_session(struct stat_session *session) 48 { 49 struct stat_node *snode, *n; 50 51 rbtree_postorder_for_each_entry_safe(snode, n, &session->stat_root, node) { 52 if (session->ts->stat_release) 53 session->ts->stat_release(snode->stat); 54 kfree(snode); 55 } 56 57 session->stat_root = RB_ROOT; 58 } 59 60 static void reset_stat_session(struct stat_session *session) 61 { 62 mutex_lock(&session->stat_mutex); 63 __reset_stat_session(session); 64 mutex_unlock(&session->stat_mutex); 65 } 66 67 static void destroy_session(struct stat_session *session) 68 { 69 tracefs_remove(session->file); 70 __reset_stat_session(session); 71 mutex_destroy(&session->stat_mutex); 72 kfree(session); 73 } 74 75 static int insert_stat(struct rb_root *root, void *stat, cmp_func_t cmp) 76 { 77 struct rb_node **new = &(root->rb_node), *parent = NULL; 78 struct stat_node *data; 79 80 data = kzalloc(sizeof(*data), GFP_KERNEL); 81 if (!data) 82 return -ENOMEM; 83 data->stat = stat; 84 85 /* 86 * Figure out where to put new node 87 * This is a descendent sorting 88 */ 89 while (*new) { 90 struct stat_node *this; 91 int result; 92 93 this = container_of(*new, struct stat_node, node); 94 result = cmp(data->stat, this->stat); 95 96 parent = *new; 97 if (result >= 0) 98 new = &((*new)->rb_left); 99 else 100 new = &((*new)->rb_right); 101 } 102 103 rb_link_node(&data->node, parent, new); 104 rb_insert_color(&data->node, root); 105 return 0; 106 } 107 108 /* 109 * For tracers that don't provide a stat_cmp callback. 110 * This one will force an insertion as right-most node 111 * in the rbtree. 112 */ 113 static int dummy_cmp(const void *p1, const void *p2) 114 { 115 return -1; 116 } 117 118 /* 119 * Initialize the stat rbtree at each trace_stat file opening. 120 * All of these copies and sorting are required on all opening 121 * since the stats could have changed between two file sessions. 122 */ 123 static int stat_seq_init(struct stat_session *session) 124 { 125 struct tracer_stat *ts = session->ts; 126 struct rb_root *root = &session->stat_root; 127 void *stat; 128 int ret = 0; 129 int i; 130 131 mutex_lock(&session->stat_mutex); 132 __reset_stat_session(session); 133 134 if (!ts->stat_cmp) 135 ts->stat_cmp = dummy_cmp; 136 137 stat = ts->stat_start(ts); 138 if (!stat) 139 goto exit; 140 141 ret = insert_stat(root, stat, ts->stat_cmp); 142 if (ret) 143 goto exit; 144 145 /* 146 * Iterate over the tracer stat entries and store them in an rbtree. 147 */ 148 for (i = 1; ; i++) { 149 stat = ts->stat_next(stat, i); 150 151 /* End of insertion */ 152 if (!stat) 153 break; 154 155 ret = insert_stat(root, stat, ts->stat_cmp); 156 if (ret) 157 goto exit_free_rbtree; 158 } 159 160 exit: 161 mutex_unlock(&session->stat_mutex); 162 return ret; 163 164 exit_free_rbtree: 165 __reset_stat_session(session); 166 mutex_unlock(&session->stat_mutex); 167 return ret; 168 } 169 170 171 static void *stat_seq_start(struct seq_file *s, loff_t *pos) 172 { 173 struct stat_session *session = s->private; 174 struct rb_node *node; 175 int n = *pos; 176 int i; 177 178 /* Prevent from tracer switch or rbtree modification */ 179 mutex_lock(&session->stat_mutex); 180 181 /* If we are in the beginning of the file, print the headers */ 182 if (session->ts->stat_headers) { 183 if (n == 0) 184 return SEQ_START_TOKEN; 185 n--; 186 } 187 188 node = rb_first(&session->stat_root); 189 for (i = 0; node && i < n; i++) 190 node = rb_next(node); 191 192 return node; 193 } 194 195 static void *stat_seq_next(struct seq_file *s, void *p, loff_t *pos) 196 { 197 struct stat_session *session = s->private; 198 struct rb_node *node = p; 199 200 (*pos)++; 201 202 if (p == SEQ_START_TOKEN) 203 return rb_first(&session->stat_root); 204 205 return rb_next(node); 206 } 207 208 static void stat_seq_stop(struct seq_file *s, void *p) 209 { 210 struct stat_session *session = s->private; 211 mutex_unlock(&session->stat_mutex); 212 } 213 214 static int stat_seq_show(struct seq_file *s, void *v) 215 { 216 struct stat_session *session = s->private; 217 struct stat_node *l = container_of(v, struct stat_node, node); 218 219 if (v == SEQ_START_TOKEN) 220 return session->ts->stat_headers(s); 221 222 return session->ts->stat_show(s, l->stat); 223 } 224 225 static const struct seq_operations trace_stat_seq_ops = { 226 .start = stat_seq_start, 227 .next = stat_seq_next, 228 .stop = stat_seq_stop, 229 .show = stat_seq_show 230 }; 231 232 /* The session stat is refilled and resorted at each stat file opening */ 233 static int tracing_stat_open(struct inode *inode, struct file *file) 234 { 235 int ret; 236 struct seq_file *m; 237 struct stat_session *session = inode->i_private; 238 239 ret = security_locked_down(LOCKDOWN_TRACEFS); 240 if (ret) 241 return ret; 242 243 ret = stat_seq_init(session); 244 if (ret) 245 return ret; 246 247 ret = seq_open(file, &trace_stat_seq_ops); 248 if (ret) { 249 reset_stat_session(session); 250 return ret; 251 } 252 253 m = file->private_data; 254 m->private = session; 255 return ret; 256 } 257 258 /* 259 * Avoid consuming memory with our now useless rbtree. 260 */ 261 static int tracing_stat_release(struct inode *i, struct file *f) 262 { 263 struct stat_session *session = i->i_private; 264 265 reset_stat_session(session); 266 267 return seq_release(i, f); 268 } 269 270 static const struct file_operations tracing_stat_fops = { 271 .open = tracing_stat_open, 272 .read = seq_read, 273 .llseek = seq_lseek, 274 .release = tracing_stat_release 275 }; 276 277 static int tracing_stat_init(void) 278 { 279 struct dentry *d_tracing; 280 281 d_tracing = tracing_init_dentry(); 282 if (IS_ERR(d_tracing)) 283 return 0; 284 285 stat_dir = tracefs_create_dir("trace_stat", d_tracing); 286 if (!stat_dir) 287 pr_warn("Could not create tracefs 'trace_stat' entry\n"); 288 return 0; 289 } 290 291 static int init_stat_file(struct stat_session *session) 292 { 293 if (!stat_dir && tracing_stat_init()) 294 return -ENODEV; 295 296 session->file = tracefs_create_file(session->ts->name, 0644, 297 stat_dir, 298 session, &tracing_stat_fops); 299 if (!session->file) 300 return -ENOMEM; 301 return 0; 302 } 303 304 int register_stat_tracer(struct tracer_stat *trace) 305 { 306 struct stat_session *session, *node; 307 int ret; 308 309 if (!trace) 310 return -EINVAL; 311 312 if (!trace->stat_start || !trace->stat_next || !trace->stat_show) 313 return -EINVAL; 314 315 /* Already registered? */ 316 mutex_lock(&all_stat_sessions_mutex); 317 list_for_each_entry(node, &all_stat_sessions, session_list) { 318 if (node->ts == trace) { 319 mutex_unlock(&all_stat_sessions_mutex); 320 return -EINVAL; 321 } 322 } 323 mutex_unlock(&all_stat_sessions_mutex); 324 325 /* Init the session */ 326 session = kzalloc(sizeof(*session), GFP_KERNEL); 327 if (!session) 328 return -ENOMEM; 329 330 session->ts = trace; 331 INIT_LIST_HEAD(&session->session_list); 332 mutex_init(&session->stat_mutex); 333 334 ret = init_stat_file(session); 335 if (ret) { 336 destroy_session(session); 337 return ret; 338 } 339 340 /* Register */ 341 mutex_lock(&all_stat_sessions_mutex); 342 list_add_tail(&session->session_list, &all_stat_sessions); 343 mutex_unlock(&all_stat_sessions_mutex); 344 345 return 0; 346 } 347 348 void unregister_stat_tracer(struct tracer_stat *trace) 349 { 350 struct stat_session *node, *tmp; 351 352 mutex_lock(&all_stat_sessions_mutex); 353 list_for_each_entry_safe(node, tmp, &all_stat_sessions, session_list) { 354 if (node->ts == trace) { 355 list_del(&node->session_list); 356 destroy_session(node); 357 break; 358 } 359 } 360 mutex_unlock(&all_stat_sessions_mutex); 361 } 362