xref: /openbmc/linux/kernel/trace/trace.c (revision a16be368)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * ring buffer based function tracer
4  *
5  * Copyright (C) 2007-2012 Steven Rostedt <srostedt@redhat.com>
6  * Copyright (C) 2008 Ingo Molnar <mingo@redhat.com>
7  *
8  * Originally taken from the RT patch by:
9  *    Arnaldo Carvalho de Melo <acme@redhat.com>
10  *
11  * Based on code from the latency_tracer, that is:
12  *  Copyright (C) 2004-2006 Ingo Molnar
13  *  Copyright (C) 2004 Nadia Yvette Chambers
14  */
15 #include <linux/ring_buffer.h>
16 #include <generated/utsrelease.h>
17 #include <linux/stacktrace.h>
18 #include <linux/writeback.h>
19 #include <linux/kallsyms.h>
20 #include <linux/security.h>
21 #include <linux/seq_file.h>
22 #include <linux/notifier.h>
23 #include <linux/irqflags.h>
24 #include <linux/debugfs.h>
25 #include <linux/tracefs.h>
26 #include <linux/pagemap.h>
27 #include <linux/hardirq.h>
28 #include <linux/linkage.h>
29 #include <linux/uaccess.h>
30 #include <linux/vmalloc.h>
31 #include <linux/ftrace.h>
32 #include <linux/module.h>
33 #include <linux/percpu.h>
34 #include <linux/splice.h>
35 #include <linux/kdebug.h>
36 #include <linux/string.h>
37 #include <linux/mount.h>
38 #include <linux/rwsem.h>
39 #include <linux/slab.h>
40 #include <linux/ctype.h>
41 #include <linux/init.h>
42 #include <linux/poll.h>
43 #include <linux/nmi.h>
44 #include <linux/fs.h>
45 #include <linux/trace.h>
46 #include <linux/sched/clock.h>
47 #include <linux/sched/rt.h>
48 #include <linux/fsnotify.h>
49 #include <linux/irq_work.h>
50 #include <linux/workqueue.h>
51 
52 #include "trace.h"
53 #include "trace_output.h"
54 
55 /*
56  * On boot up, the ring buffer is set to the minimum size, so that
57  * we do not waste memory on systems that are not using tracing.
58  */
59 bool ring_buffer_expanded;
60 
61 /*
62  * We need to change this state when a selftest is running.
63  * A selftest will lurk into the ring-buffer to count the
64  * entries inserted during the selftest although some concurrent
65  * insertions into the ring-buffer such as trace_printk could occurred
66  * at the same time, giving false positive or negative results.
67  */
68 static bool __read_mostly tracing_selftest_running;
69 
70 /*
71  * If a tracer is running, we do not want to run SELFTEST.
72  */
73 bool __read_mostly tracing_selftest_disabled;
74 
75 /* Pipe tracepoints to printk */
76 struct trace_iterator *tracepoint_print_iter;
77 int tracepoint_printk;
78 static DEFINE_STATIC_KEY_FALSE(tracepoint_printk_key);
79 
80 /* For tracers that don't implement custom flags */
81 static struct tracer_opt dummy_tracer_opt[] = {
82 	{ }
83 };
84 
85 static int
86 dummy_set_flag(struct trace_array *tr, u32 old_flags, u32 bit, int set)
87 {
88 	return 0;
89 }
90 
91 /*
92  * To prevent the comm cache from being overwritten when no
93  * tracing is active, only save the comm when a trace event
94  * occurred.
95  */
96 static DEFINE_PER_CPU(bool, trace_taskinfo_save);
97 
98 /*
99  * Kill all tracing for good (never come back).
100  * It is initialized to 1 but will turn to zero if the initialization
101  * of the tracer is successful. But that is the only place that sets
102  * this back to zero.
103  */
104 static int tracing_disabled = 1;
105 
106 cpumask_var_t __read_mostly	tracing_buffer_mask;
107 
108 /*
109  * ftrace_dump_on_oops - variable to dump ftrace buffer on oops
110  *
111  * If there is an oops (or kernel panic) and the ftrace_dump_on_oops
112  * is set, then ftrace_dump is called. This will output the contents
113  * of the ftrace buffers to the console.  This is very useful for
114  * capturing traces that lead to crashes and outputing it to a
115  * serial console.
116  *
117  * It is default off, but you can enable it with either specifying
118  * "ftrace_dump_on_oops" in the kernel command line, or setting
119  * /proc/sys/kernel/ftrace_dump_on_oops
120  * Set 1 if you want to dump buffers of all CPUs
121  * Set 2 if you want to dump the buffer of the CPU that triggered oops
122  */
123 
124 enum ftrace_dump_mode ftrace_dump_on_oops;
125 
126 /* When set, tracing will stop when a WARN*() is hit */
127 int __disable_trace_on_warning;
128 
129 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
130 /* Map of enums to their values, for "eval_map" file */
131 struct trace_eval_map_head {
132 	struct module			*mod;
133 	unsigned long			length;
134 };
135 
136 union trace_eval_map_item;
137 
138 struct trace_eval_map_tail {
139 	/*
140 	 * "end" is first and points to NULL as it must be different
141 	 * than "mod" or "eval_string"
142 	 */
143 	union trace_eval_map_item	*next;
144 	const char			*end;	/* points to NULL */
145 };
146 
147 static DEFINE_MUTEX(trace_eval_mutex);
148 
149 /*
150  * The trace_eval_maps are saved in an array with two extra elements,
151  * one at the beginning, and one at the end. The beginning item contains
152  * the count of the saved maps (head.length), and the module they
153  * belong to if not built in (head.mod). The ending item contains a
154  * pointer to the next array of saved eval_map items.
155  */
156 union trace_eval_map_item {
157 	struct trace_eval_map		map;
158 	struct trace_eval_map_head	head;
159 	struct trace_eval_map_tail	tail;
160 };
161 
162 static union trace_eval_map_item *trace_eval_maps;
163 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */
164 
165 int tracing_set_tracer(struct trace_array *tr, const char *buf);
166 static void ftrace_trace_userstack(struct trace_buffer *buffer,
167 				   unsigned long flags, int pc);
168 
169 #define MAX_TRACER_SIZE		100
170 static char bootup_tracer_buf[MAX_TRACER_SIZE] __initdata;
171 static char *default_bootup_tracer;
172 
173 static bool allocate_snapshot;
174 
175 static int __init set_cmdline_ftrace(char *str)
176 {
177 	strlcpy(bootup_tracer_buf, str, MAX_TRACER_SIZE);
178 	default_bootup_tracer = bootup_tracer_buf;
179 	/* We are using ftrace early, expand it */
180 	ring_buffer_expanded = true;
181 	return 1;
182 }
183 __setup("ftrace=", set_cmdline_ftrace);
184 
185 static int __init set_ftrace_dump_on_oops(char *str)
186 {
187 	if (*str++ != '=' || !*str) {
188 		ftrace_dump_on_oops = DUMP_ALL;
189 		return 1;
190 	}
191 
192 	if (!strcmp("orig_cpu", str)) {
193 		ftrace_dump_on_oops = DUMP_ORIG;
194                 return 1;
195         }
196 
197         return 0;
198 }
199 __setup("ftrace_dump_on_oops", set_ftrace_dump_on_oops);
200 
201 static int __init stop_trace_on_warning(char *str)
202 {
203 	if ((strcmp(str, "=0") != 0 && strcmp(str, "=off") != 0))
204 		__disable_trace_on_warning = 1;
205 	return 1;
206 }
207 __setup("traceoff_on_warning", stop_trace_on_warning);
208 
209 static int __init boot_alloc_snapshot(char *str)
210 {
211 	allocate_snapshot = true;
212 	/* We also need the main ring buffer expanded */
213 	ring_buffer_expanded = true;
214 	return 1;
215 }
216 __setup("alloc_snapshot", boot_alloc_snapshot);
217 
218 
219 static char trace_boot_options_buf[MAX_TRACER_SIZE] __initdata;
220 
221 static int __init set_trace_boot_options(char *str)
222 {
223 	strlcpy(trace_boot_options_buf, str, MAX_TRACER_SIZE);
224 	return 0;
225 }
226 __setup("trace_options=", set_trace_boot_options);
227 
228 static char trace_boot_clock_buf[MAX_TRACER_SIZE] __initdata;
229 static char *trace_boot_clock __initdata;
230 
231 static int __init set_trace_boot_clock(char *str)
232 {
233 	strlcpy(trace_boot_clock_buf, str, MAX_TRACER_SIZE);
234 	trace_boot_clock = trace_boot_clock_buf;
235 	return 0;
236 }
237 __setup("trace_clock=", set_trace_boot_clock);
238 
239 static int __init set_tracepoint_printk(char *str)
240 {
241 	if ((strcmp(str, "=0") != 0 && strcmp(str, "=off") != 0))
242 		tracepoint_printk = 1;
243 	return 1;
244 }
245 __setup("tp_printk", set_tracepoint_printk);
246 
247 unsigned long long ns2usecs(u64 nsec)
248 {
249 	nsec += 500;
250 	do_div(nsec, 1000);
251 	return nsec;
252 }
253 
254 /* trace_flags holds trace_options default values */
255 #define TRACE_DEFAULT_FLAGS						\
256 	(FUNCTION_DEFAULT_FLAGS |					\
257 	 TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK |			\
258 	 TRACE_ITER_ANNOTATE | TRACE_ITER_CONTEXT_INFO |		\
259 	 TRACE_ITER_RECORD_CMD | TRACE_ITER_OVERWRITE |			\
260 	 TRACE_ITER_IRQ_INFO | TRACE_ITER_MARKERS)
261 
262 /* trace_options that are only supported by global_trace */
263 #define TOP_LEVEL_TRACE_FLAGS (TRACE_ITER_PRINTK |			\
264 	       TRACE_ITER_PRINTK_MSGONLY | TRACE_ITER_RECORD_CMD)
265 
266 /* trace_flags that are default zero for instances */
267 #define ZEROED_TRACE_FLAGS \
268 	(TRACE_ITER_EVENT_FORK | TRACE_ITER_FUNC_FORK)
269 
270 /*
271  * The global_trace is the descriptor that holds the top-level tracing
272  * buffers for the live tracing.
273  */
274 static struct trace_array global_trace = {
275 	.trace_flags = TRACE_DEFAULT_FLAGS,
276 };
277 
278 LIST_HEAD(ftrace_trace_arrays);
279 
280 int trace_array_get(struct trace_array *this_tr)
281 {
282 	struct trace_array *tr;
283 	int ret = -ENODEV;
284 
285 	mutex_lock(&trace_types_lock);
286 	list_for_each_entry(tr, &ftrace_trace_arrays, list) {
287 		if (tr == this_tr) {
288 			tr->ref++;
289 			ret = 0;
290 			break;
291 		}
292 	}
293 	mutex_unlock(&trace_types_lock);
294 
295 	return ret;
296 }
297 
298 static void __trace_array_put(struct trace_array *this_tr)
299 {
300 	WARN_ON(!this_tr->ref);
301 	this_tr->ref--;
302 }
303 
304 /**
305  * trace_array_put - Decrement the reference counter for this trace array.
306  *
307  * NOTE: Use this when we no longer need the trace array returned by
308  * trace_array_get_by_name(). This ensures the trace array can be later
309  * destroyed.
310  *
311  */
312 void trace_array_put(struct trace_array *this_tr)
313 {
314 	if (!this_tr)
315 		return;
316 
317 	mutex_lock(&trace_types_lock);
318 	__trace_array_put(this_tr);
319 	mutex_unlock(&trace_types_lock);
320 }
321 EXPORT_SYMBOL_GPL(trace_array_put);
322 
323 int tracing_check_open_get_tr(struct trace_array *tr)
324 {
325 	int ret;
326 
327 	ret = security_locked_down(LOCKDOWN_TRACEFS);
328 	if (ret)
329 		return ret;
330 
331 	if (tracing_disabled)
332 		return -ENODEV;
333 
334 	if (tr && trace_array_get(tr) < 0)
335 		return -ENODEV;
336 
337 	return 0;
338 }
339 
340 int call_filter_check_discard(struct trace_event_call *call, void *rec,
341 			      struct trace_buffer *buffer,
342 			      struct ring_buffer_event *event)
343 {
344 	if (unlikely(call->flags & TRACE_EVENT_FL_FILTERED) &&
345 	    !filter_match_preds(call->filter, rec)) {
346 		__trace_event_discard_commit(buffer, event);
347 		return 1;
348 	}
349 
350 	return 0;
351 }
352 
353 void trace_free_pid_list(struct trace_pid_list *pid_list)
354 {
355 	vfree(pid_list->pids);
356 	kfree(pid_list);
357 }
358 
359 /**
360  * trace_find_filtered_pid - check if a pid exists in a filtered_pid list
361  * @filtered_pids: The list of pids to check
362  * @search_pid: The PID to find in @filtered_pids
363  *
364  * Returns true if @search_pid is fonud in @filtered_pids, and false otherwis.
365  */
366 bool
367 trace_find_filtered_pid(struct trace_pid_list *filtered_pids, pid_t search_pid)
368 {
369 	/*
370 	 * If pid_max changed after filtered_pids was created, we
371 	 * by default ignore all pids greater than the previous pid_max.
372 	 */
373 	if (search_pid >= filtered_pids->pid_max)
374 		return false;
375 
376 	return test_bit(search_pid, filtered_pids->pids);
377 }
378 
379 /**
380  * trace_ignore_this_task - should a task be ignored for tracing
381  * @filtered_pids: The list of pids to check
382  * @task: The task that should be ignored if not filtered
383  *
384  * Checks if @task should be traced or not from @filtered_pids.
385  * Returns true if @task should *NOT* be traced.
386  * Returns false if @task should be traced.
387  */
388 bool
389 trace_ignore_this_task(struct trace_pid_list *filtered_pids,
390 		       struct trace_pid_list *filtered_no_pids,
391 		       struct task_struct *task)
392 {
393 	/*
394 	 * If filterd_no_pids is not empty, and the task's pid is listed
395 	 * in filtered_no_pids, then return true.
396 	 * Otherwise, if filtered_pids is empty, that means we can
397 	 * trace all tasks. If it has content, then only trace pids
398 	 * within filtered_pids.
399 	 */
400 
401 	return (filtered_pids &&
402 		!trace_find_filtered_pid(filtered_pids, task->pid)) ||
403 		(filtered_no_pids &&
404 		 trace_find_filtered_pid(filtered_no_pids, task->pid));
405 }
406 
407 /**
408  * trace_filter_add_remove_task - Add or remove a task from a pid_list
409  * @pid_list: The list to modify
410  * @self: The current task for fork or NULL for exit
411  * @task: The task to add or remove
412  *
413  * If adding a task, if @self is defined, the task is only added if @self
414  * is also included in @pid_list. This happens on fork and tasks should
415  * only be added when the parent is listed. If @self is NULL, then the
416  * @task pid will be removed from the list, which would happen on exit
417  * of a task.
418  */
419 void trace_filter_add_remove_task(struct trace_pid_list *pid_list,
420 				  struct task_struct *self,
421 				  struct task_struct *task)
422 {
423 	if (!pid_list)
424 		return;
425 
426 	/* For forks, we only add if the forking task is listed */
427 	if (self) {
428 		if (!trace_find_filtered_pid(pid_list, self->pid))
429 			return;
430 	}
431 
432 	/* Sorry, but we don't support pid_max changing after setting */
433 	if (task->pid >= pid_list->pid_max)
434 		return;
435 
436 	/* "self" is set for forks, and NULL for exits */
437 	if (self)
438 		set_bit(task->pid, pid_list->pids);
439 	else
440 		clear_bit(task->pid, pid_list->pids);
441 }
442 
443 /**
444  * trace_pid_next - Used for seq_file to get to the next pid of a pid_list
445  * @pid_list: The pid list to show
446  * @v: The last pid that was shown (+1 the actual pid to let zero be displayed)
447  * @pos: The position of the file
448  *
449  * This is used by the seq_file "next" operation to iterate the pids
450  * listed in a trace_pid_list structure.
451  *
452  * Returns the pid+1 as we want to display pid of zero, but NULL would
453  * stop the iteration.
454  */
455 void *trace_pid_next(struct trace_pid_list *pid_list, void *v, loff_t *pos)
456 {
457 	unsigned long pid = (unsigned long)v;
458 
459 	(*pos)++;
460 
461 	/* pid already is +1 of the actual prevous bit */
462 	pid = find_next_bit(pid_list->pids, pid_list->pid_max, pid);
463 
464 	/* Return pid + 1 to allow zero to be represented */
465 	if (pid < pid_list->pid_max)
466 		return (void *)(pid + 1);
467 
468 	return NULL;
469 }
470 
471 /**
472  * trace_pid_start - Used for seq_file to start reading pid lists
473  * @pid_list: The pid list to show
474  * @pos: The position of the file
475  *
476  * This is used by seq_file "start" operation to start the iteration
477  * of listing pids.
478  *
479  * Returns the pid+1 as we want to display pid of zero, but NULL would
480  * stop the iteration.
481  */
482 void *trace_pid_start(struct trace_pid_list *pid_list, loff_t *pos)
483 {
484 	unsigned long pid;
485 	loff_t l = 0;
486 
487 	pid = find_first_bit(pid_list->pids, pid_list->pid_max);
488 	if (pid >= pid_list->pid_max)
489 		return NULL;
490 
491 	/* Return pid + 1 so that zero can be the exit value */
492 	for (pid++; pid && l < *pos;
493 	     pid = (unsigned long)trace_pid_next(pid_list, (void *)pid, &l))
494 		;
495 	return (void *)pid;
496 }
497 
498 /**
499  * trace_pid_show - show the current pid in seq_file processing
500  * @m: The seq_file structure to write into
501  * @v: A void pointer of the pid (+1) value to display
502  *
503  * Can be directly used by seq_file operations to display the current
504  * pid value.
505  */
506 int trace_pid_show(struct seq_file *m, void *v)
507 {
508 	unsigned long pid = (unsigned long)v - 1;
509 
510 	seq_printf(m, "%lu\n", pid);
511 	return 0;
512 }
513 
514 /* 128 should be much more than enough */
515 #define PID_BUF_SIZE		127
516 
517 int trace_pid_write(struct trace_pid_list *filtered_pids,
518 		    struct trace_pid_list **new_pid_list,
519 		    const char __user *ubuf, size_t cnt)
520 {
521 	struct trace_pid_list *pid_list;
522 	struct trace_parser parser;
523 	unsigned long val;
524 	int nr_pids = 0;
525 	ssize_t read = 0;
526 	ssize_t ret = 0;
527 	loff_t pos;
528 	pid_t pid;
529 
530 	if (trace_parser_get_init(&parser, PID_BUF_SIZE + 1))
531 		return -ENOMEM;
532 
533 	/*
534 	 * Always recreate a new array. The write is an all or nothing
535 	 * operation. Always create a new array when adding new pids by
536 	 * the user. If the operation fails, then the current list is
537 	 * not modified.
538 	 */
539 	pid_list = kmalloc(sizeof(*pid_list), GFP_KERNEL);
540 	if (!pid_list) {
541 		trace_parser_put(&parser);
542 		return -ENOMEM;
543 	}
544 
545 	pid_list->pid_max = READ_ONCE(pid_max);
546 
547 	/* Only truncating will shrink pid_max */
548 	if (filtered_pids && filtered_pids->pid_max > pid_list->pid_max)
549 		pid_list->pid_max = filtered_pids->pid_max;
550 
551 	pid_list->pids = vzalloc((pid_list->pid_max + 7) >> 3);
552 	if (!pid_list->pids) {
553 		trace_parser_put(&parser);
554 		kfree(pid_list);
555 		return -ENOMEM;
556 	}
557 
558 	if (filtered_pids) {
559 		/* copy the current bits to the new max */
560 		for_each_set_bit(pid, filtered_pids->pids,
561 				 filtered_pids->pid_max) {
562 			set_bit(pid, pid_list->pids);
563 			nr_pids++;
564 		}
565 	}
566 
567 	while (cnt > 0) {
568 
569 		pos = 0;
570 
571 		ret = trace_get_user(&parser, ubuf, cnt, &pos);
572 		if (ret < 0 || !trace_parser_loaded(&parser))
573 			break;
574 
575 		read += ret;
576 		ubuf += ret;
577 		cnt -= ret;
578 
579 		ret = -EINVAL;
580 		if (kstrtoul(parser.buffer, 0, &val))
581 			break;
582 		if (val >= pid_list->pid_max)
583 			break;
584 
585 		pid = (pid_t)val;
586 
587 		set_bit(pid, pid_list->pids);
588 		nr_pids++;
589 
590 		trace_parser_clear(&parser);
591 		ret = 0;
592 	}
593 	trace_parser_put(&parser);
594 
595 	if (ret < 0) {
596 		trace_free_pid_list(pid_list);
597 		return ret;
598 	}
599 
600 	if (!nr_pids) {
601 		/* Cleared the list of pids */
602 		trace_free_pid_list(pid_list);
603 		read = ret;
604 		pid_list = NULL;
605 	}
606 
607 	*new_pid_list = pid_list;
608 
609 	return read;
610 }
611 
612 static u64 buffer_ftrace_now(struct array_buffer *buf, int cpu)
613 {
614 	u64 ts;
615 
616 	/* Early boot up does not have a buffer yet */
617 	if (!buf->buffer)
618 		return trace_clock_local();
619 
620 	ts = ring_buffer_time_stamp(buf->buffer, cpu);
621 	ring_buffer_normalize_time_stamp(buf->buffer, cpu, &ts);
622 
623 	return ts;
624 }
625 
626 u64 ftrace_now(int cpu)
627 {
628 	return buffer_ftrace_now(&global_trace.array_buffer, cpu);
629 }
630 
631 /**
632  * tracing_is_enabled - Show if global_trace has been disabled
633  *
634  * Shows if the global trace has been enabled or not. It uses the
635  * mirror flag "buffer_disabled" to be used in fast paths such as for
636  * the irqsoff tracer. But it may be inaccurate due to races. If you
637  * need to know the accurate state, use tracing_is_on() which is a little
638  * slower, but accurate.
639  */
640 int tracing_is_enabled(void)
641 {
642 	/*
643 	 * For quick access (irqsoff uses this in fast path), just
644 	 * return the mirror variable of the state of the ring buffer.
645 	 * It's a little racy, but we don't really care.
646 	 */
647 	smp_rmb();
648 	return !global_trace.buffer_disabled;
649 }
650 
651 /*
652  * trace_buf_size is the size in bytes that is allocated
653  * for a buffer. Note, the number of bytes is always rounded
654  * to page size.
655  *
656  * This number is purposely set to a low number of 16384.
657  * If the dump on oops happens, it will be much appreciated
658  * to not have to wait for all that output. Anyway this can be
659  * boot time and run time configurable.
660  */
661 #define TRACE_BUF_SIZE_DEFAULT	1441792UL /* 16384 * 88 (sizeof(entry)) */
662 
663 static unsigned long		trace_buf_size = TRACE_BUF_SIZE_DEFAULT;
664 
665 /* trace_types holds a link list of available tracers. */
666 static struct tracer		*trace_types __read_mostly;
667 
668 /*
669  * trace_types_lock is used to protect the trace_types list.
670  */
671 DEFINE_MUTEX(trace_types_lock);
672 
673 /*
674  * serialize the access of the ring buffer
675  *
676  * ring buffer serializes readers, but it is low level protection.
677  * The validity of the events (which returns by ring_buffer_peek() ..etc)
678  * are not protected by ring buffer.
679  *
680  * The content of events may become garbage if we allow other process consumes
681  * these events concurrently:
682  *   A) the page of the consumed events may become a normal page
683  *      (not reader page) in ring buffer, and this page will be rewrited
684  *      by events producer.
685  *   B) The page of the consumed events may become a page for splice_read,
686  *      and this page will be returned to system.
687  *
688  * These primitives allow multi process access to different cpu ring buffer
689  * concurrently.
690  *
691  * These primitives don't distinguish read-only and read-consume access.
692  * Multi read-only access are also serialized.
693  */
694 
695 #ifdef CONFIG_SMP
696 static DECLARE_RWSEM(all_cpu_access_lock);
697 static DEFINE_PER_CPU(struct mutex, cpu_access_lock);
698 
699 static inline void trace_access_lock(int cpu)
700 {
701 	if (cpu == RING_BUFFER_ALL_CPUS) {
702 		/* gain it for accessing the whole ring buffer. */
703 		down_write(&all_cpu_access_lock);
704 	} else {
705 		/* gain it for accessing a cpu ring buffer. */
706 
707 		/* Firstly block other trace_access_lock(RING_BUFFER_ALL_CPUS). */
708 		down_read(&all_cpu_access_lock);
709 
710 		/* Secondly block other access to this @cpu ring buffer. */
711 		mutex_lock(&per_cpu(cpu_access_lock, cpu));
712 	}
713 }
714 
715 static inline void trace_access_unlock(int cpu)
716 {
717 	if (cpu == RING_BUFFER_ALL_CPUS) {
718 		up_write(&all_cpu_access_lock);
719 	} else {
720 		mutex_unlock(&per_cpu(cpu_access_lock, cpu));
721 		up_read(&all_cpu_access_lock);
722 	}
723 }
724 
725 static inline void trace_access_lock_init(void)
726 {
727 	int cpu;
728 
729 	for_each_possible_cpu(cpu)
730 		mutex_init(&per_cpu(cpu_access_lock, cpu));
731 }
732 
733 #else
734 
735 static DEFINE_MUTEX(access_lock);
736 
737 static inline void trace_access_lock(int cpu)
738 {
739 	(void)cpu;
740 	mutex_lock(&access_lock);
741 }
742 
743 static inline void trace_access_unlock(int cpu)
744 {
745 	(void)cpu;
746 	mutex_unlock(&access_lock);
747 }
748 
749 static inline void trace_access_lock_init(void)
750 {
751 }
752 
753 #endif
754 
755 #ifdef CONFIG_STACKTRACE
756 static void __ftrace_trace_stack(struct trace_buffer *buffer,
757 				 unsigned long flags,
758 				 int skip, int pc, struct pt_regs *regs);
759 static inline void ftrace_trace_stack(struct trace_array *tr,
760 				      struct trace_buffer *buffer,
761 				      unsigned long flags,
762 				      int skip, int pc, struct pt_regs *regs);
763 
764 #else
765 static inline void __ftrace_trace_stack(struct trace_buffer *buffer,
766 					unsigned long flags,
767 					int skip, int pc, struct pt_regs *regs)
768 {
769 }
770 static inline void ftrace_trace_stack(struct trace_array *tr,
771 				      struct trace_buffer *buffer,
772 				      unsigned long flags,
773 				      int skip, int pc, struct pt_regs *regs)
774 {
775 }
776 
777 #endif
778 
779 static __always_inline void
780 trace_event_setup(struct ring_buffer_event *event,
781 		  int type, unsigned long flags, int pc)
782 {
783 	struct trace_entry *ent = ring_buffer_event_data(event);
784 
785 	tracing_generic_entry_update(ent, type, flags, pc);
786 }
787 
788 static __always_inline struct ring_buffer_event *
789 __trace_buffer_lock_reserve(struct trace_buffer *buffer,
790 			  int type,
791 			  unsigned long len,
792 			  unsigned long flags, int pc)
793 {
794 	struct ring_buffer_event *event;
795 
796 	event = ring_buffer_lock_reserve(buffer, len);
797 	if (event != NULL)
798 		trace_event_setup(event, type, flags, pc);
799 
800 	return event;
801 }
802 
803 void tracer_tracing_on(struct trace_array *tr)
804 {
805 	if (tr->array_buffer.buffer)
806 		ring_buffer_record_on(tr->array_buffer.buffer);
807 	/*
808 	 * This flag is looked at when buffers haven't been allocated
809 	 * yet, or by some tracers (like irqsoff), that just want to
810 	 * know if the ring buffer has been disabled, but it can handle
811 	 * races of where it gets disabled but we still do a record.
812 	 * As the check is in the fast path of the tracers, it is more
813 	 * important to be fast than accurate.
814 	 */
815 	tr->buffer_disabled = 0;
816 	/* Make the flag seen by readers */
817 	smp_wmb();
818 }
819 
820 /**
821  * tracing_on - enable tracing buffers
822  *
823  * This function enables tracing buffers that may have been
824  * disabled with tracing_off.
825  */
826 void tracing_on(void)
827 {
828 	tracer_tracing_on(&global_trace);
829 }
830 EXPORT_SYMBOL_GPL(tracing_on);
831 
832 
833 static __always_inline void
834 __buffer_unlock_commit(struct trace_buffer *buffer, struct ring_buffer_event *event)
835 {
836 	__this_cpu_write(trace_taskinfo_save, true);
837 
838 	/* If this is the temp buffer, we need to commit fully */
839 	if (this_cpu_read(trace_buffered_event) == event) {
840 		/* Length is in event->array[0] */
841 		ring_buffer_write(buffer, event->array[0], &event->array[1]);
842 		/* Release the temp buffer */
843 		this_cpu_dec(trace_buffered_event_cnt);
844 	} else
845 		ring_buffer_unlock_commit(buffer, event);
846 }
847 
848 /**
849  * __trace_puts - write a constant string into the trace buffer.
850  * @ip:	   The address of the caller
851  * @str:   The constant string to write
852  * @size:  The size of the string.
853  */
854 int __trace_puts(unsigned long ip, const char *str, int size)
855 {
856 	struct ring_buffer_event *event;
857 	struct trace_buffer *buffer;
858 	struct print_entry *entry;
859 	unsigned long irq_flags;
860 	int alloc;
861 	int pc;
862 
863 	if (!(global_trace.trace_flags & TRACE_ITER_PRINTK))
864 		return 0;
865 
866 	pc = preempt_count();
867 
868 	if (unlikely(tracing_selftest_running || tracing_disabled))
869 		return 0;
870 
871 	alloc = sizeof(*entry) + size + 2; /* possible \n added */
872 
873 	local_save_flags(irq_flags);
874 	buffer = global_trace.array_buffer.buffer;
875 	ring_buffer_nest_start(buffer);
876 	event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, alloc,
877 					    irq_flags, pc);
878 	if (!event) {
879 		size = 0;
880 		goto out;
881 	}
882 
883 	entry = ring_buffer_event_data(event);
884 	entry->ip = ip;
885 
886 	memcpy(&entry->buf, str, size);
887 
888 	/* Add a newline if necessary */
889 	if (entry->buf[size - 1] != '\n') {
890 		entry->buf[size] = '\n';
891 		entry->buf[size + 1] = '\0';
892 	} else
893 		entry->buf[size] = '\0';
894 
895 	__buffer_unlock_commit(buffer, event);
896 	ftrace_trace_stack(&global_trace, buffer, irq_flags, 4, pc, NULL);
897  out:
898 	ring_buffer_nest_end(buffer);
899 	return size;
900 }
901 EXPORT_SYMBOL_GPL(__trace_puts);
902 
903 /**
904  * __trace_bputs - write the pointer to a constant string into trace buffer
905  * @ip:	   The address of the caller
906  * @str:   The constant string to write to the buffer to
907  */
908 int __trace_bputs(unsigned long ip, const char *str)
909 {
910 	struct ring_buffer_event *event;
911 	struct trace_buffer *buffer;
912 	struct bputs_entry *entry;
913 	unsigned long irq_flags;
914 	int size = sizeof(struct bputs_entry);
915 	int ret = 0;
916 	int pc;
917 
918 	if (!(global_trace.trace_flags & TRACE_ITER_PRINTK))
919 		return 0;
920 
921 	pc = preempt_count();
922 
923 	if (unlikely(tracing_selftest_running || tracing_disabled))
924 		return 0;
925 
926 	local_save_flags(irq_flags);
927 	buffer = global_trace.array_buffer.buffer;
928 
929 	ring_buffer_nest_start(buffer);
930 	event = __trace_buffer_lock_reserve(buffer, TRACE_BPUTS, size,
931 					    irq_flags, pc);
932 	if (!event)
933 		goto out;
934 
935 	entry = ring_buffer_event_data(event);
936 	entry->ip			= ip;
937 	entry->str			= str;
938 
939 	__buffer_unlock_commit(buffer, event);
940 	ftrace_trace_stack(&global_trace, buffer, irq_flags, 4, pc, NULL);
941 
942 	ret = 1;
943  out:
944 	ring_buffer_nest_end(buffer);
945 	return ret;
946 }
947 EXPORT_SYMBOL_GPL(__trace_bputs);
948 
949 #ifdef CONFIG_TRACER_SNAPSHOT
950 static void tracing_snapshot_instance_cond(struct trace_array *tr,
951 					   void *cond_data)
952 {
953 	struct tracer *tracer = tr->current_trace;
954 	unsigned long flags;
955 
956 	if (in_nmi()) {
957 		internal_trace_puts("*** SNAPSHOT CALLED FROM NMI CONTEXT ***\n");
958 		internal_trace_puts("*** snapshot is being ignored        ***\n");
959 		return;
960 	}
961 
962 	if (!tr->allocated_snapshot) {
963 		internal_trace_puts("*** SNAPSHOT NOT ALLOCATED ***\n");
964 		internal_trace_puts("*** stopping trace here!   ***\n");
965 		tracing_off();
966 		return;
967 	}
968 
969 	/* Note, snapshot can not be used when the tracer uses it */
970 	if (tracer->use_max_tr) {
971 		internal_trace_puts("*** LATENCY TRACER ACTIVE ***\n");
972 		internal_trace_puts("*** Can not use snapshot (sorry) ***\n");
973 		return;
974 	}
975 
976 	local_irq_save(flags);
977 	update_max_tr(tr, current, smp_processor_id(), cond_data);
978 	local_irq_restore(flags);
979 }
980 
981 void tracing_snapshot_instance(struct trace_array *tr)
982 {
983 	tracing_snapshot_instance_cond(tr, NULL);
984 }
985 
986 /**
987  * tracing_snapshot - take a snapshot of the current buffer.
988  *
989  * This causes a swap between the snapshot buffer and the current live
990  * tracing buffer. You can use this to take snapshots of the live
991  * trace when some condition is triggered, but continue to trace.
992  *
993  * Note, make sure to allocate the snapshot with either
994  * a tracing_snapshot_alloc(), or by doing it manually
995  * with: echo 1 > /sys/kernel/debug/tracing/snapshot
996  *
997  * If the snapshot buffer is not allocated, it will stop tracing.
998  * Basically making a permanent snapshot.
999  */
1000 void tracing_snapshot(void)
1001 {
1002 	struct trace_array *tr = &global_trace;
1003 
1004 	tracing_snapshot_instance(tr);
1005 }
1006 EXPORT_SYMBOL_GPL(tracing_snapshot);
1007 
1008 /**
1009  * tracing_snapshot_cond - conditionally take a snapshot of the current buffer.
1010  * @tr:		The tracing instance to snapshot
1011  * @cond_data:	The data to be tested conditionally, and possibly saved
1012  *
1013  * This is the same as tracing_snapshot() except that the snapshot is
1014  * conditional - the snapshot will only happen if the
1015  * cond_snapshot.update() implementation receiving the cond_data
1016  * returns true, which means that the trace array's cond_snapshot
1017  * update() operation used the cond_data to determine whether the
1018  * snapshot should be taken, and if it was, presumably saved it along
1019  * with the snapshot.
1020  */
1021 void tracing_snapshot_cond(struct trace_array *tr, void *cond_data)
1022 {
1023 	tracing_snapshot_instance_cond(tr, cond_data);
1024 }
1025 EXPORT_SYMBOL_GPL(tracing_snapshot_cond);
1026 
1027 /**
1028  * tracing_snapshot_cond_data - get the user data associated with a snapshot
1029  * @tr:		The tracing instance
1030  *
1031  * When the user enables a conditional snapshot using
1032  * tracing_snapshot_cond_enable(), the user-defined cond_data is saved
1033  * with the snapshot.  This accessor is used to retrieve it.
1034  *
1035  * Should not be called from cond_snapshot.update(), since it takes
1036  * the tr->max_lock lock, which the code calling
1037  * cond_snapshot.update() has already done.
1038  *
1039  * Returns the cond_data associated with the trace array's snapshot.
1040  */
1041 void *tracing_cond_snapshot_data(struct trace_array *tr)
1042 {
1043 	void *cond_data = NULL;
1044 
1045 	arch_spin_lock(&tr->max_lock);
1046 
1047 	if (tr->cond_snapshot)
1048 		cond_data = tr->cond_snapshot->cond_data;
1049 
1050 	arch_spin_unlock(&tr->max_lock);
1051 
1052 	return cond_data;
1053 }
1054 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data);
1055 
1056 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf,
1057 					struct array_buffer *size_buf, int cpu_id);
1058 static void set_buffer_entries(struct array_buffer *buf, unsigned long val);
1059 
1060 int tracing_alloc_snapshot_instance(struct trace_array *tr)
1061 {
1062 	int ret;
1063 
1064 	if (!tr->allocated_snapshot) {
1065 
1066 		/* allocate spare buffer */
1067 		ret = resize_buffer_duplicate_size(&tr->max_buffer,
1068 				   &tr->array_buffer, RING_BUFFER_ALL_CPUS);
1069 		if (ret < 0)
1070 			return ret;
1071 
1072 		tr->allocated_snapshot = true;
1073 	}
1074 
1075 	return 0;
1076 }
1077 
1078 static void free_snapshot(struct trace_array *tr)
1079 {
1080 	/*
1081 	 * We don't free the ring buffer. instead, resize it because
1082 	 * The max_tr ring buffer has some state (e.g. ring->clock) and
1083 	 * we want preserve it.
1084 	 */
1085 	ring_buffer_resize(tr->max_buffer.buffer, 1, RING_BUFFER_ALL_CPUS);
1086 	set_buffer_entries(&tr->max_buffer, 1);
1087 	tracing_reset_online_cpus(&tr->max_buffer);
1088 	tr->allocated_snapshot = false;
1089 }
1090 
1091 /**
1092  * tracing_alloc_snapshot - allocate snapshot buffer.
1093  *
1094  * This only allocates the snapshot buffer if it isn't already
1095  * allocated - it doesn't also take a snapshot.
1096  *
1097  * This is meant to be used in cases where the snapshot buffer needs
1098  * to be set up for events that can't sleep but need to be able to
1099  * trigger a snapshot.
1100  */
1101 int tracing_alloc_snapshot(void)
1102 {
1103 	struct trace_array *tr = &global_trace;
1104 	int ret;
1105 
1106 	ret = tracing_alloc_snapshot_instance(tr);
1107 	WARN_ON(ret < 0);
1108 
1109 	return ret;
1110 }
1111 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot);
1112 
1113 /**
1114  * tracing_snapshot_alloc - allocate and take a snapshot of the current buffer.
1115  *
1116  * This is similar to tracing_snapshot(), but it will allocate the
1117  * snapshot buffer if it isn't already allocated. Use this only
1118  * where it is safe to sleep, as the allocation may sleep.
1119  *
1120  * This causes a swap between the snapshot buffer and the current live
1121  * tracing buffer. You can use this to take snapshots of the live
1122  * trace when some condition is triggered, but continue to trace.
1123  */
1124 void tracing_snapshot_alloc(void)
1125 {
1126 	int ret;
1127 
1128 	ret = tracing_alloc_snapshot();
1129 	if (ret < 0)
1130 		return;
1131 
1132 	tracing_snapshot();
1133 }
1134 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc);
1135 
1136 /**
1137  * tracing_snapshot_cond_enable - enable conditional snapshot for an instance
1138  * @tr:		The tracing instance
1139  * @cond_data:	User data to associate with the snapshot
1140  * @update:	Implementation of the cond_snapshot update function
1141  *
1142  * Check whether the conditional snapshot for the given instance has
1143  * already been enabled, or if the current tracer is already using a
1144  * snapshot; if so, return -EBUSY, else create a cond_snapshot and
1145  * save the cond_data and update function inside.
1146  *
1147  * Returns 0 if successful, error otherwise.
1148  */
1149 int tracing_snapshot_cond_enable(struct trace_array *tr, void *cond_data,
1150 				 cond_update_fn_t update)
1151 {
1152 	struct cond_snapshot *cond_snapshot;
1153 	int ret = 0;
1154 
1155 	cond_snapshot = kzalloc(sizeof(*cond_snapshot), GFP_KERNEL);
1156 	if (!cond_snapshot)
1157 		return -ENOMEM;
1158 
1159 	cond_snapshot->cond_data = cond_data;
1160 	cond_snapshot->update = update;
1161 
1162 	mutex_lock(&trace_types_lock);
1163 
1164 	ret = tracing_alloc_snapshot_instance(tr);
1165 	if (ret)
1166 		goto fail_unlock;
1167 
1168 	if (tr->current_trace->use_max_tr) {
1169 		ret = -EBUSY;
1170 		goto fail_unlock;
1171 	}
1172 
1173 	/*
1174 	 * The cond_snapshot can only change to NULL without the
1175 	 * trace_types_lock. We don't care if we race with it going
1176 	 * to NULL, but we want to make sure that it's not set to
1177 	 * something other than NULL when we get here, which we can
1178 	 * do safely with only holding the trace_types_lock and not
1179 	 * having to take the max_lock.
1180 	 */
1181 	if (tr->cond_snapshot) {
1182 		ret = -EBUSY;
1183 		goto fail_unlock;
1184 	}
1185 
1186 	arch_spin_lock(&tr->max_lock);
1187 	tr->cond_snapshot = cond_snapshot;
1188 	arch_spin_unlock(&tr->max_lock);
1189 
1190 	mutex_unlock(&trace_types_lock);
1191 
1192 	return ret;
1193 
1194  fail_unlock:
1195 	mutex_unlock(&trace_types_lock);
1196 	kfree(cond_snapshot);
1197 	return ret;
1198 }
1199 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable);
1200 
1201 /**
1202  * tracing_snapshot_cond_disable - disable conditional snapshot for an instance
1203  * @tr:		The tracing instance
1204  *
1205  * Check whether the conditional snapshot for the given instance is
1206  * enabled; if so, free the cond_snapshot associated with it,
1207  * otherwise return -EINVAL.
1208  *
1209  * Returns 0 if successful, error otherwise.
1210  */
1211 int tracing_snapshot_cond_disable(struct trace_array *tr)
1212 {
1213 	int ret = 0;
1214 
1215 	arch_spin_lock(&tr->max_lock);
1216 
1217 	if (!tr->cond_snapshot)
1218 		ret = -EINVAL;
1219 	else {
1220 		kfree(tr->cond_snapshot);
1221 		tr->cond_snapshot = NULL;
1222 	}
1223 
1224 	arch_spin_unlock(&tr->max_lock);
1225 
1226 	return ret;
1227 }
1228 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable);
1229 #else
1230 void tracing_snapshot(void)
1231 {
1232 	WARN_ONCE(1, "Snapshot feature not enabled, but internal snapshot used");
1233 }
1234 EXPORT_SYMBOL_GPL(tracing_snapshot);
1235 void tracing_snapshot_cond(struct trace_array *tr, void *cond_data)
1236 {
1237 	WARN_ONCE(1, "Snapshot feature not enabled, but internal conditional snapshot used");
1238 }
1239 EXPORT_SYMBOL_GPL(tracing_snapshot_cond);
1240 int tracing_alloc_snapshot(void)
1241 {
1242 	WARN_ONCE(1, "Snapshot feature not enabled, but snapshot allocation used");
1243 	return -ENODEV;
1244 }
1245 EXPORT_SYMBOL_GPL(tracing_alloc_snapshot);
1246 void tracing_snapshot_alloc(void)
1247 {
1248 	/* Give warning */
1249 	tracing_snapshot();
1250 }
1251 EXPORT_SYMBOL_GPL(tracing_snapshot_alloc);
1252 void *tracing_cond_snapshot_data(struct trace_array *tr)
1253 {
1254 	return NULL;
1255 }
1256 EXPORT_SYMBOL_GPL(tracing_cond_snapshot_data);
1257 int tracing_snapshot_cond_enable(struct trace_array *tr, void *cond_data, cond_update_fn_t update)
1258 {
1259 	return -ENODEV;
1260 }
1261 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_enable);
1262 int tracing_snapshot_cond_disable(struct trace_array *tr)
1263 {
1264 	return false;
1265 }
1266 EXPORT_SYMBOL_GPL(tracing_snapshot_cond_disable);
1267 #endif /* CONFIG_TRACER_SNAPSHOT */
1268 
1269 void tracer_tracing_off(struct trace_array *tr)
1270 {
1271 	if (tr->array_buffer.buffer)
1272 		ring_buffer_record_off(tr->array_buffer.buffer);
1273 	/*
1274 	 * This flag is looked at when buffers haven't been allocated
1275 	 * yet, or by some tracers (like irqsoff), that just want to
1276 	 * know if the ring buffer has been disabled, but it can handle
1277 	 * races of where it gets disabled but we still do a record.
1278 	 * As the check is in the fast path of the tracers, it is more
1279 	 * important to be fast than accurate.
1280 	 */
1281 	tr->buffer_disabled = 1;
1282 	/* Make the flag seen by readers */
1283 	smp_wmb();
1284 }
1285 
1286 /**
1287  * tracing_off - turn off tracing buffers
1288  *
1289  * This function stops the tracing buffers from recording data.
1290  * It does not disable any overhead the tracers themselves may
1291  * be causing. This function simply causes all recording to
1292  * the ring buffers to fail.
1293  */
1294 void tracing_off(void)
1295 {
1296 	tracer_tracing_off(&global_trace);
1297 }
1298 EXPORT_SYMBOL_GPL(tracing_off);
1299 
1300 void disable_trace_on_warning(void)
1301 {
1302 	if (__disable_trace_on_warning) {
1303 		trace_array_printk_buf(global_trace.array_buffer.buffer, _THIS_IP_,
1304 			"Disabling tracing due to warning\n");
1305 		tracing_off();
1306 	}
1307 }
1308 
1309 /**
1310  * tracer_tracing_is_on - show real state of ring buffer enabled
1311  * @tr : the trace array to know if ring buffer is enabled
1312  *
1313  * Shows real state of the ring buffer if it is enabled or not.
1314  */
1315 bool tracer_tracing_is_on(struct trace_array *tr)
1316 {
1317 	if (tr->array_buffer.buffer)
1318 		return ring_buffer_record_is_on(tr->array_buffer.buffer);
1319 	return !tr->buffer_disabled;
1320 }
1321 
1322 /**
1323  * tracing_is_on - show state of ring buffers enabled
1324  */
1325 int tracing_is_on(void)
1326 {
1327 	return tracer_tracing_is_on(&global_trace);
1328 }
1329 EXPORT_SYMBOL_GPL(tracing_is_on);
1330 
1331 static int __init set_buf_size(char *str)
1332 {
1333 	unsigned long buf_size;
1334 
1335 	if (!str)
1336 		return 0;
1337 	buf_size = memparse(str, &str);
1338 	/* nr_entries can not be zero */
1339 	if (buf_size == 0)
1340 		return 0;
1341 	trace_buf_size = buf_size;
1342 	return 1;
1343 }
1344 __setup("trace_buf_size=", set_buf_size);
1345 
1346 static int __init set_tracing_thresh(char *str)
1347 {
1348 	unsigned long threshold;
1349 	int ret;
1350 
1351 	if (!str)
1352 		return 0;
1353 	ret = kstrtoul(str, 0, &threshold);
1354 	if (ret < 0)
1355 		return 0;
1356 	tracing_thresh = threshold * 1000;
1357 	return 1;
1358 }
1359 __setup("tracing_thresh=", set_tracing_thresh);
1360 
1361 unsigned long nsecs_to_usecs(unsigned long nsecs)
1362 {
1363 	return nsecs / 1000;
1364 }
1365 
1366 /*
1367  * TRACE_FLAGS is defined as a tuple matching bit masks with strings.
1368  * It uses C(a, b) where 'a' is the eval (enum) name and 'b' is the string that
1369  * matches it. By defining "C(a, b) b", TRACE_FLAGS becomes a list
1370  * of strings in the order that the evals (enum) were defined.
1371  */
1372 #undef C
1373 #define C(a, b) b
1374 
1375 /* These must match the bit postions in trace_iterator_flags */
1376 static const char *trace_options[] = {
1377 	TRACE_FLAGS
1378 	NULL
1379 };
1380 
1381 static struct {
1382 	u64 (*func)(void);
1383 	const char *name;
1384 	int in_ns;		/* is this clock in nanoseconds? */
1385 } trace_clocks[] = {
1386 	{ trace_clock_local,		"local",	1 },
1387 	{ trace_clock_global,		"global",	1 },
1388 	{ trace_clock_counter,		"counter",	0 },
1389 	{ trace_clock_jiffies,		"uptime",	0 },
1390 	{ trace_clock,			"perf",		1 },
1391 	{ ktime_get_mono_fast_ns,	"mono",		1 },
1392 	{ ktime_get_raw_fast_ns,	"mono_raw",	1 },
1393 	{ ktime_get_boot_fast_ns,	"boot",		1 },
1394 	ARCH_TRACE_CLOCKS
1395 };
1396 
1397 bool trace_clock_in_ns(struct trace_array *tr)
1398 {
1399 	if (trace_clocks[tr->clock_id].in_ns)
1400 		return true;
1401 
1402 	return false;
1403 }
1404 
1405 /*
1406  * trace_parser_get_init - gets the buffer for trace parser
1407  */
1408 int trace_parser_get_init(struct trace_parser *parser, int size)
1409 {
1410 	memset(parser, 0, sizeof(*parser));
1411 
1412 	parser->buffer = kmalloc(size, GFP_KERNEL);
1413 	if (!parser->buffer)
1414 		return 1;
1415 
1416 	parser->size = size;
1417 	return 0;
1418 }
1419 
1420 /*
1421  * trace_parser_put - frees the buffer for trace parser
1422  */
1423 void trace_parser_put(struct trace_parser *parser)
1424 {
1425 	kfree(parser->buffer);
1426 	parser->buffer = NULL;
1427 }
1428 
1429 /*
1430  * trace_get_user - reads the user input string separated by  space
1431  * (matched by isspace(ch))
1432  *
1433  * For each string found the 'struct trace_parser' is updated,
1434  * and the function returns.
1435  *
1436  * Returns number of bytes read.
1437  *
1438  * See kernel/trace/trace.h for 'struct trace_parser' details.
1439  */
1440 int trace_get_user(struct trace_parser *parser, const char __user *ubuf,
1441 	size_t cnt, loff_t *ppos)
1442 {
1443 	char ch;
1444 	size_t read = 0;
1445 	ssize_t ret;
1446 
1447 	if (!*ppos)
1448 		trace_parser_clear(parser);
1449 
1450 	ret = get_user(ch, ubuf++);
1451 	if (ret)
1452 		goto out;
1453 
1454 	read++;
1455 	cnt--;
1456 
1457 	/*
1458 	 * The parser is not finished with the last write,
1459 	 * continue reading the user input without skipping spaces.
1460 	 */
1461 	if (!parser->cont) {
1462 		/* skip white space */
1463 		while (cnt && isspace(ch)) {
1464 			ret = get_user(ch, ubuf++);
1465 			if (ret)
1466 				goto out;
1467 			read++;
1468 			cnt--;
1469 		}
1470 
1471 		parser->idx = 0;
1472 
1473 		/* only spaces were written */
1474 		if (isspace(ch) || !ch) {
1475 			*ppos += read;
1476 			ret = read;
1477 			goto out;
1478 		}
1479 	}
1480 
1481 	/* read the non-space input */
1482 	while (cnt && !isspace(ch) && ch) {
1483 		if (parser->idx < parser->size - 1)
1484 			parser->buffer[parser->idx++] = ch;
1485 		else {
1486 			ret = -EINVAL;
1487 			goto out;
1488 		}
1489 		ret = get_user(ch, ubuf++);
1490 		if (ret)
1491 			goto out;
1492 		read++;
1493 		cnt--;
1494 	}
1495 
1496 	/* We either got finished input or we have to wait for another call. */
1497 	if (isspace(ch) || !ch) {
1498 		parser->buffer[parser->idx] = 0;
1499 		parser->cont = false;
1500 	} else if (parser->idx < parser->size - 1) {
1501 		parser->cont = true;
1502 		parser->buffer[parser->idx++] = ch;
1503 		/* Make sure the parsed string always terminates with '\0'. */
1504 		parser->buffer[parser->idx] = 0;
1505 	} else {
1506 		ret = -EINVAL;
1507 		goto out;
1508 	}
1509 
1510 	*ppos += read;
1511 	ret = read;
1512 
1513 out:
1514 	return ret;
1515 }
1516 
1517 /* TODO add a seq_buf_to_buffer() */
1518 static ssize_t trace_seq_to_buffer(struct trace_seq *s, void *buf, size_t cnt)
1519 {
1520 	int len;
1521 
1522 	if (trace_seq_used(s) <= s->seq.readpos)
1523 		return -EBUSY;
1524 
1525 	len = trace_seq_used(s) - s->seq.readpos;
1526 	if (cnt > len)
1527 		cnt = len;
1528 	memcpy(buf, s->buffer + s->seq.readpos, cnt);
1529 
1530 	s->seq.readpos += cnt;
1531 	return cnt;
1532 }
1533 
1534 unsigned long __read_mostly	tracing_thresh;
1535 static const struct file_operations tracing_max_lat_fops;
1536 
1537 #if (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \
1538 	defined(CONFIG_FSNOTIFY)
1539 
1540 static struct workqueue_struct *fsnotify_wq;
1541 
1542 static void latency_fsnotify_workfn(struct work_struct *work)
1543 {
1544 	struct trace_array *tr = container_of(work, struct trace_array,
1545 					      fsnotify_work);
1546 	fsnotify(tr->d_max_latency->d_inode, FS_MODIFY,
1547 		 tr->d_max_latency->d_inode, FSNOTIFY_EVENT_INODE, NULL, 0);
1548 }
1549 
1550 static void latency_fsnotify_workfn_irq(struct irq_work *iwork)
1551 {
1552 	struct trace_array *tr = container_of(iwork, struct trace_array,
1553 					      fsnotify_irqwork);
1554 	queue_work(fsnotify_wq, &tr->fsnotify_work);
1555 }
1556 
1557 static void trace_create_maxlat_file(struct trace_array *tr,
1558 				     struct dentry *d_tracer)
1559 {
1560 	INIT_WORK(&tr->fsnotify_work, latency_fsnotify_workfn);
1561 	init_irq_work(&tr->fsnotify_irqwork, latency_fsnotify_workfn_irq);
1562 	tr->d_max_latency = trace_create_file("tracing_max_latency", 0644,
1563 					      d_tracer, &tr->max_latency,
1564 					      &tracing_max_lat_fops);
1565 }
1566 
1567 __init static int latency_fsnotify_init(void)
1568 {
1569 	fsnotify_wq = alloc_workqueue("tr_max_lat_wq",
1570 				      WQ_UNBOUND | WQ_HIGHPRI, 0);
1571 	if (!fsnotify_wq) {
1572 		pr_err("Unable to allocate tr_max_lat_wq\n");
1573 		return -ENOMEM;
1574 	}
1575 	return 0;
1576 }
1577 
1578 late_initcall_sync(latency_fsnotify_init);
1579 
1580 void latency_fsnotify(struct trace_array *tr)
1581 {
1582 	if (!fsnotify_wq)
1583 		return;
1584 	/*
1585 	 * We cannot call queue_work(&tr->fsnotify_work) from here because it's
1586 	 * possible that we are called from __schedule() or do_idle(), which
1587 	 * could cause a deadlock.
1588 	 */
1589 	irq_work_queue(&tr->fsnotify_irqwork);
1590 }
1591 
1592 /*
1593  * (defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)) && \
1594  *  defined(CONFIG_FSNOTIFY)
1595  */
1596 #else
1597 
1598 #define trace_create_maxlat_file(tr, d_tracer)				\
1599 	trace_create_file("tracing_max_latency", 0644, d_tracer,	\
1600 			  &tr->max_latency, &tracing_max_lat_fops)
1601 
1602 #endif
1603 
1604 #ifdef CONFIG_TRACER_MAX_TRACE
1605 /*
1606  * Copy the new maximum trace into the separate maximum-trace
1607  * structure. (this way the maximum trace is permanently saved,
1608  * for later retrieval via /sys/kernel/tracing/tracing_max_latency)
1609  */
1610 static void
1611 __update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu)
1612 {
1613 	struct array_buffer *trace_buf = &tr->array_buffer;
1614 	struct array_buffer *max_buf = &tr->max_buffer;
1615 	struct trace_array_cpu *data = per_cpu_ptr(trace_buf->data, cpu);
1616 	struct trace_array_cpu *max_data = per_cpu_ptr(max_buf->data, cpu);
1617 
1618 	max_buf->cpu = cpu;
1619 	max_buf->time_start = data->preempt_timestamp;
1620 
1621 	max_data->saved_latency = tr->max_latency;
1622 	max_data->critical_start = data->critical_start;
1623 	max_data->critical_end = data->critical_end;
1624 
1625 	strncpy(max_data->comm, tsk->comm, TASK_COMM_LEN);
1626 	max_data->pid = tsk->pid;
1627 	/*
1628 	 * If tsk == current, then use current_uid(), as that does not use
1629 	 * RCU. The irq tracer can be called out of RCU scope.
1630 	 */
1631 	if (tsk == current)
1632 		max_data->uid = current_uid();
1633 	else
1634 		max_data->uid = task_uid(tsk);
1635 
1636 	max_data->nice = tsk->static_prio - 20 - MAX_RT_PRIO;
1637 	max_data->policy = tsk->policy;
1638 	max_data->rt_priority = tsk->rt_priority;
1639 
1640 	/* record this tasks comm */
1641 	tracing_record_cmdline(tsk);
1642 	latency_fsnotify(tr);
1643 }
1644 
1645 /**
1646  * update_max_tr - snapshot all trace buffers from global_trace to max_tr
1647  * @tr: tracer
1648  * @tsk: the task with the latency
1649  * @cpu: The cpu that initiated the trace.
1650  * @cond_data: User data associated with a conditional snapshot
1651  *
1652  * Flip the buffers between the @tr and the max_tr and record information
1653  * about which task was the cause of this latency.
1654  */
1655 void
1656 update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu,
1657 	      void *cond_data)
1658 {
1659 	if (tr->stop_count)
1660 		return;
1661 
1662 	WARN_ON_ONCE(!irqs_disabled());
1663 
1664 	if (!tr->allocated_snapshot) {
1665 		/* Only the nop tracer should hit this when disabling */
1666 		WARN_ON_ONCE(tr->current_trace != &nop_trace);
1667 		return;
1668 	}
1669 
1670 	arch_spin_lock(&tr->max_lock);
1671 
1672 	/* Inherit the recordable setting from array_buffer */
1673 	if (ring_buffer_record_is_set_on(tr->array_buffer.buffer))
1674 		ring_buffer_record_on(tr->max_buffer.buffer);
1675 	else
1676 		ring_buffer_record_off(tr->max_buffer.buffer);
1677 
1678 #ifdef CONFIG_TRACER_SNAPSHOT
1679 	if (tr->cond_snapshot && !tr->cond_snapshot->update(tr, cond_data))
1680 		goto out_unlock;
1681 #endif
1682 	swap(tr->array_buffer.buffer, tr->max_buffer.buffer);
1683 
1684 	__update_max_tr(tr, tsk, cpu);
1685 
1686  out_unlock:
1687 	arch_spin_unlock(&tr->max_lock);
1688 }
1689 
1690 /**
1691  * update_max_tr_single - only copy one trace over, and reset the rest
1692  * @tr: tracer
1693  * @tsk: task with the latency
1694  * @cpu: the cpu of the buffer to copy.
1695  *
1696  * Flip the trace of a single CPU buffer between the @tr and the max_tr.
1697  */
1698 void
1699 update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu)
1700 {
1701 	int ret;
1702 
1703 	if (tr->stop_count)
1704 		return;
1705 
1706 	WARN_ON_ONCE(!irqs_disabled());
1707 	if (!tr->allocated_snapshot) {
1708 		/* Only the nop tracer should hit this when disabling */
1709 		WARN_ON_ONCE(tr->current_trace != &nop_trace);
1710 		return;
1711 	}
1712 
1713 	arch_spin_lock(&tr->max_lock);
1714 
1715 	ret = ring_buffer_swap_cpu(tr->max_buffer.buffer, tr->array_buffer.buffer, cpu);
1716 
1717 	if (ret == -EBUSY) {
1718 		/*
1719 		 * We failed to swap the buffer due to a commit taking
1720 		 * place on this CPU. We fail to record, but we reset
1721 		 * the max trace buffer (no one writes directly to it)
1722 		 * and flag that it failed.
1723 		 */
1724 		trace_array_printk_buf(tr->max_buffer.buffer, _THIS_IP_,
1725 			"Failed to swap buffers due to commit in progress\n");
1726 	}
1727 
1728 	WARN_ON_ONCE(ret && ret != -EAGAIN && ret != -EBUSY);
1729 
1730 	__update_max_tr(tr, tsk, cpu);
1731 	arch_spin_unlock(&tr->max_lock);
1732 }
1733 #endif /* CONFIG_TRACER_MAX_TRACE */
1734 
1735 static int wait_on_pipe(struct trace_iterator *iter, int full)
1736 {
1737 	/* Iterators are static, they should be filled or empty */
1738 	if (trace_buffer_iter(iter, iter->cpu_file))
1739 		return 0;
1740 
1741 	return ring_buffer_wait(iter->array_buffer->buffer, iter->cpu_file,
1742 				full);
1743 }
1744 
1745 #ifdef CONFIG_FTRACE_STARTUP_TEST
1746 static bool selftests_can_run;
1747 
1748 struct trace_selftests {
1749 	struct list_head		list;
1750 	struct tracer			*type;
1751 };
1752 
1753 static LIST_HEAD(postponed_selftests);
1754 
1755 static int save_selftest(struct tracer *type)
1756 {
1757 	struct trace_selftests *selftest;
1758 
1759 	selftest = kmalloc(sizeof(*selftest), GFP_KERNEL);
1760 	if (!selftest)
1761 		return -ENOMEM;
1762 
1763 	selftest->type = type;
1764 	list_add(&selftest->list, &postponed_selftests);
1765 	return 0;
1766 }
1767 
1768 static int run_tracer_selftest(struct tracer *type)
1769 {
1770 	struct trace_array *tr = &global_trace;
1771 	struct tracer *saved_tracer = tr->current_trace;
1772 	int ret;
1773 
1774 	if (!type->selftest || tracing_selftest_disabled)
1775 		return 0;
1776 
1777 	/*
1778 	 * If a tracer registers early in boot up (before scheduling is
1779 	 * initialized and such), then do not run its selftests yet.
1780 	 * Instead, run it a little later in the boot process.
1781 	 */
1782 	if (!selftests_can_run)
1783 		return save_selftest(type);
1784 
1785 	/*
1786 	 * Run a selftest on this tracer.
1787 	 * Here we reset the trace buffer, and set the current
1788 	 * tracer to be this tracer. The tracer can then run some
1789 	 * internal tracing to verify that everything is in order.
1790 	 * If we fail, we do not register this tracer.
1791 	 */
1792 	tracing_reset_online_cpus(&tr->array_buffer);
1793 
1794 	tr->current_trace = type;
1795 
1796 #ifdef CONFIG_TRACER_MAX_TRACE
1797 	if (type->use_max_tr) {
1798 		/* If we expanded the buffers, make sure the max is expanded too */
1799 		if (ring_buffer_expanded)
1800 			ring_buffer_resize(tr->max_buffer.buffer, trace_buf_size,
1801 					   RING_BUFFER_ALL_CPUS);
1802 		tr->allocated_snapshot = true;
1803 	}
1804 #endif
1805 
1806 	/* the test is responsible for initializing and enabling */
1807 	pr_info("Testing tracer %s: ", type->name);
1808 	ret = type->selftest(type, tr);
1809 	/* the test is responsible for resetting too */
1810 	tr->current_trace = saved_tracer;
1811 	if (ret) {
1812 		printk(KERN_CONT "FAILED!\n");
1813 		/* Add the warning after printing 'FAILED' */
1814 		WARN_ON(1);
1815 		return -1;
1816 	}
1817 	/* Only reset on passing, to avoid touching corrupted buffers */
1818 	tracing_reset_online_cpus(&tr->array_buffer);
1819 
1820 #ifdef CONFIG_TRACER_MAX_TRACE
1821 	if (type->use_max_tr) {
1822 		tr->allocated_snapshot = false;
1823 
1824 		/* Shrink the max buffer again */
1825 		if (ring_buffer_expanded)
1826 			ring_buffer_resize(tr->max_buffer.buffer, 1,
1827 					   RING_BUFFER_ALL_CPUS);
1828 	}
1829 #endif
1830 
1831 	printk(KERN_CONT "PASSED\n");
1832 	return 0;
1833 }
1834 
1835 static __init int init_trace_selftests(void)
1836 {
1837 	struct trace_selftests *p, *n;
1838 	struct tracer *t, **last;
1839 	int ret;
1840 
1841 	selftests_can_run = true;
1842 
1843 	mutex_lock(&trace_types_lock);
1844 
1845 	if (list_empty(&postponed_selftests))
1846 		goto out;
1847 
1848 	pr_info("Running postponed tracer tests:\n");
1849 
1850 	tracing_selftest_running = true;
1851 	list_for_each_entry_safe(p, n, &postponed_selftests, list) {
1852 		/* This loop can take minutes when sanitizers are enabled, so
1853 		 * lets make sure we allow RCU processing.
1854 		 */
1855 		cond_resched();
1856 		ret = run_tracer_selftest(p->type);
1857 		/* If the test fails, then warn and remove from available_tracers */
1858 		if (ret < 0) {
1859 			WARN(1, "tracer: %s failed selftest, disabling\n",
1860 			     p->type->name);
1861 			last = &trace_types;
1862 			for (t = trace_types; t; t = t->next) {
1863 				if (t == p->type) {
1864 					*last = t->next;
1865 					break;
1866 				}
1867 				last = &t->next;
1868 			}
1869 		}
1870 		list_del(&p->list);
1871 		kfree(p);
1872 	}
1873 	tracing_selftest_running = false;
1874 
1875  out:
1876 	mutex_unlock(&trace_types_lock);
1877 
1878 	return 0;
1879 }
1880 core_initcall(init_trace_selftests);
1881 #else
1882 static inline int run_tracer_selftest(struct tracer *type)
1883 {
1884 	return 0;
1885 }
1886 #endif /* CONFIG_FTRACE_STARTUP_TEST */
1887 
1888 static void add_tracer_options(struct trace_array *tr, struct tracer *t);
1889 
1890 static void __init apply_trace_boot_options(void);
1891 
1892 /**
1893  * register_tracer - register a tracer with the ftrace system.
1894  * @type: the plugin for the tracer
1895  *
1896  * Register a new plugin tracer.
1897  */
1898 int __init register_tracer(struct tracer *type)
1899 {
1900 	struct tracer *t;
1901 	int ret = 0;
1902 
1903 	if (!type->name) {
1904 		pr_info("Tracer must have a name\n");
1905 		return -1;
1906 	}
1907 
1908 	if (strlen(type->name) >= MAX_TRACER_SIZE) {
1909 		pr_info("Tracer has a name longer than %d\n", MAX_TRACER_SIZE);
1910 		return -1;
1911 	}
1912 
1913 	if (security_locked_down(LOCKDOWN_TRACEFS)) {
1914 		pr_warn("Can not register tracer %s due to lockdown\n",
1915 			   type->name);
1916 		return -EPERM;
1917 	}
1918 
1919 	mutex_lock(&trace_types_lock);
1920 
1921 	tracing_selftest_running = true;
1922 
1923 	for (t = trace_types; t; t = t->next) {
1924 		if (strcmp(type->name, t->name) == 0) {
1925 			/* already found */
1926 			pr_info("Tracer %s already registered\n",
1927 				type->name);
1928 			ret = -1;
1929 			goto out;
1930 		}
1931 	}
1932 
1933 	if (!type->set_flag)
1934 		type->set_flag = &dummy_set_flag;
1935 	if (!type->flags) {
1936 		/*allocate a dummy tracer_flags*/
1937 		type->flags = kmalloc(sizeof(*type->flags), GFP_KERNEL);
1938 		if (!type->flags) {
1939 			ret = -ENOMEM;
1940 			goto out;
1941 		}
1942 		type->flags->val = 0;
1943 		type->flags->opts = dummy_tracer_opt;
1944 	} else
1945 		if (!type->flags->opts)
1946 			type->flags->opts = dummy_tracer_opt;
1947 
1948 	/* store the tracer for __set_tracer_option */
1949 	type->flags->trace = type;
1950 
1951 	ret = run_tracer_selftest(type);
1952 	if (ret < 0)
1953 		goto out;
1954 
1955 	type->next = trace_types;
1956 	trace_types = type;
1957 	add_tracer_options(&global_trace, type);
1958 
1959  out:
1960 	tracing_selftest_running = false;
1961 	mutex_unlock(&trace_types_lock);
1962 
1963 	if (ret || !default_bootup_tracer)
1964 		goto out_unlock;
1965 
1966 	if (strncmp(default_bootup_tracer, type->name, MAX_TRACER_SIZE))
1967 		goto out_unlock;
1968 
1969 	printk(KERN_INFO "Starting tracer '%s'\n", type->name);
1970 	/* Do we want this tracer to start on bootup? */
1971 	tracing_set_tracer(&global_trace, type->name);
1972 	default_bootup_tracer = NULL;
1973 
1974 	apply_trace_boot_options();
1975 
1976 	/* disable other selftests, since this will break it. */
1977 	tracing_selftest_disabled = true;
1978 #ifdef CONFIG_FTRACE_STARTUP_TEST
1979 	printk(KERN_INFO "Disabling FTRACE selftests due to running tracer '%s'\n",
1980 	       type->name);
1981 #endif
1982 
1983  out_unlock:
1984 	return ret;
1985 }
1986 
1987 static void tracing_reset_cpu(struct array_buffer *buf, int cpu)
1988 {
1989 	struct trace_buffer *buffer = buf->buffer;
1990 
1991 	if (!buffer)
1992 		return;
1993 
1994 	ring_buffer_record_disable(buffer);
1995 
1996 	/* Make sure all commits have finished */
1997 	synchronize_rcu();
1998 	ring_buffer_reset_cpu(buffer, cpu);
1999 
2000 	ring_buffer_record_enable(buffer);
2001 }
2002 
2003 void tracing_reset_online_cpus(struct array_buffer *buf)
2004 {
2005 	struct trace_buffer *buffer = buf->buffer;
2006 	int cpu;
2007 
2008 	if (!buffer)
2009 		return;
2010 
2011 	ring_buffer_record_disable(buffer);
2012 
2013 	/* Make sure all commits have finished */
2014 	synchronize_rcu();
2015 
2016 	buf->time_start = buffer_ftrace_now(buf, buf->cpu);
2017 
2018 	for_each_online_cpu(cpu)
2019 		ring_buffer_reset_cpu(buffer, cpu);
2020 
2021 	ring_buffer_record_enable(buffer);
2022 }
2023 
2024 /* Must have trace_types_lock held */
2025 void tracing_reset_all_online_cpus(void)
2026 {
2027 	struct trace_array *tr;
2028 
2029 	list_for_each_entry(tr, &ftrace_trace_arrays, list) {
2030 		if (!tr->clear_trace)
2031 			continue;
2032 		tr->clear_trace = false;
2033 		tracing_reset_online_cpus(&tr->array_buffer);
2034 #ifdef CONFIG_TRACER_MAX_TRACE
2035 		tracing_reset_online_cpus(&tr->max_buffer);
2036 #endif
2037 	}
2038 }
2039 
2040 static int *tgid_map;
2041 
2042 #define SAVED_CMDLINES_DEFAULT 128
2043 #define NO_CMDLINE_MAP UINT_MAX
2044 static arch_spinlock_t trace_cmdline_lock = __ARCH_SPIN_LOCK_UNLOCKED;
2045 struct saved_cmdlines_buffer {
2046 	unsigned map_pid_to_cmdline[PID_MAX_DEFAULT+1];
2047 	unsigned *map_cmdline_to_pid;
2048 	unsigned cmdline_num;
2049 	int cmdline_idx;
2050 	char *saved_cmdlines;
2051 };
2052 static struct saved_cmdlines_buffer *savedcmd;
2053 
2054 /* temporary disable recording */
2055 static atomic_t trace_record_taskinfo_disabled __read_mostly;
2056 
2057 static inline char *get_saved_cmdlines(int idx)
2058 {
2059 	return &savedcmd->saved_cmdlines[idx * TASK_COMM_LEN];
2060 }
2061 
2062 static inline void set_cmdline(int idx, const char *cmdline)
2063 {
2064 	strncpy(get_saved_cmdlines(idx), cmdline, TASK_COMM_LEN);
2065 }
2066 
2067 static int allocate_cmdlines_buffer(unsigned int val,
2068 				    struct saved_cmdlines_buffer *s)
2069 {
2070 	s->map_cmdline_to_pid = kmalloc_array(val,
2071 					      sizeof(*s->map_cmdline_to_pid),
2072 					      GFP_KERNEL);
2073 	if (!s->map_cmdline_to_pid)
2074 		return -ENOMEM;
2075 
2076 	s->saved_cmdlines = kmalloc_array(TASK_COMM_LEN, val, GFP_KERNEL);
2077 	if (!s->saved_cmdlines) {
2078 		kfree(s->map_cmdline_to_pid);
2079 		return -ENOMEM;
2080 	}
2081 
2082 	s->cmdline_idx = 0;
2083 	s->cmdline_num = val;
2084 	memset(&s->map_pid_to_cmdline, NO_CMDLINE_MAP,
2085 	       sizeof(s->map_pid_to_cmdline));
2086 	memset(s->map_cmdline_to_pid, NO_CMDLINE_MAP,
2087 	       val * sizeof(*s->map_cmdline_to_pid));
2088 
2089 	return 0;
2090 }
2091 
2092 static int trace_create_savedcmd(void)
2093 {
2094 	int ret;
2095 
2096 	savedcmd = kmalloc(sizeof(*savedcmd), GFP_KERNEL);
2097 	if (!savedcmd)
2098 		return -ENOMEM;
2099 
2100 	ret = allocate_cmdlines_buffer(SAVED_CMDLINES_DEFAULT, savedcmd);
2101 	if (ret < 0) {
2102 		kfree(savedcmd);
2103 		savedcmd = NULL;
2104 		return -ENOMEM;
2105 	}
2106 
2107 	return 0;
2108 }
2109 
2110 int is_tracing_stopped(void)
2111 {
2112 	return global_trace.stop_count;
2113 }
2114 
2115 /**
2116  * tracing_start - quick start of the tracer
2117  *
2118  * If tracing is enabled but was stopped by tracing_stop,
2119  * this will start the tracer back up.
2120  */
2121 void tracing_start(void)
2122 {
2123 	struct trace_buffer *buffer;
2124 	unsigned long flags;
2125 
2126 	if (tracing_disabled)
2127 		return;
2128 
2129 	raw_spin_lock_irqsave(&global_trace.start_lock, flags);
2130 	if (--global_trace.stop_count) {
2131 		if (global_trace.stop_count < 0) {
2132 			/* Someone screwed up their debugging */
2133 			WARN_ON_ONCE(1);
2134 			global_trace.stop_count = 0;
2135 		}
2136 		goto out;
2137 	}
2138 
2139 	/* Prevent the buffers from switching */
2140 	arch_spin_lock(&global_trace.max_lock);
2141 
2142 	buffer = global_trace.array_buffer.buffer;
2143 	if (buffer)
2144 		ring_buffer_record_enable(buffer);
2145 
2146 #ifdef CONFIG_TRACER_MAX_TRACE
2147 	buffer = global_trace.max_buffer.buffer;
2148 	if (buffer)
2149 		ring_buffer_record_enable(buffer);
2150 #endif
2151 
2152 	arch_spin_unlock(&global_trace.max_lock);
2153 
2154  out:
2155 	raw_spin_unlock_irqrestore(&global_trace.start_lock, flags);
2156 }
2157 
2158 static void tracing_start_tr(struct trace_array *tr)
2159 {
2160 	struct trace_buffer *buffer;
2161 	unsigned long flags;
2162 
2163 	if (tracing_disabled)
2164 		return;
2165 
2166 	/* If global, we need to also start the max tracer */
2167 	if (tr->flags & TRACE_ARRAY_FL_GLOBAL)
2168 		return tracing_start();
2169 
2170 	raw_spin_lock_irqsave(&tr->start_lock, flags);
2171 
2172 	if (--tr->stop_count) {
2173 		if (tr->stop_count < 0) {
2174 			/* Someone screwed up their debugging */
2175 			WARN_ON_ONCE(1);
2176 			tr->stop_count = 0;
2177 		}
2178 		goto out;
2179 	}
2180 
2181 	buffer = tr->array_buffer.buffer;
2182 	if (buffer)
2183 		ring_buffer_record_enable(buffer);
2184 
2185  out:
2186 	raw_spin_unlock_irqrestore(&tr->start_lock, flags);
2187 }
2188 
2189 /**
2190  * tracing_stop - quick stop of the tracer
2191  *
2192  * Light weight way to stop tracing. Use in conjunction with
2193  * tracing_start.
2194  */
2195 void tracing_stop(void)
2196 {
2197 	struct trace_buffer *buffer;
2198 	unsigned long flags;
2199 
2200 	raw_spin_lock_irqsave(&global_trace.start_lock, flags);
2201 	if (global_trace.stop_count++)
2202 		goto out;
2203 
2204 	/* Prevent the buffers from switching */
2205 	arch_spin_lock(&global_trace.max_lock);
2206 
2207 	buffer = global_trace.array_buffer.buffer;
2208 	if (buffer)
2209 		ring_buffer_record_disable(buffer);
2210 
2211 #ifdef CONFIG_TRACER_MAX_TRACE
2212 	buffer = global_trace.max_buffer.buffer;
2213 	if (buffer)
2214 		ring_buffer_record_disable(buffer);
2215 #endif
2216 
2217 	arch_spin_unlock(&global_trace.max_lock);
2218 
2219  out:
2220 	raw_spin_unlock_irqrestore(&global_trace.start_lock, flags);
2221 }
2222 
2223 static void tracing_stop_tr(struct trace_array *tr)
2224 {
2225 	struct trace_buffer *buffer;
2226 	unsigned long flags;
2227 
2228 	/* If global, we need to also stop the max tracer */
2229 	if (tr->flags & TRACE_ARRAY_FL_GLOBAL)
2230 		return tracing_stop();
2231 
2232 	raw_spin_lock_irqsave(&tr->start_lock, flags);
2233 	if (tr->stop_count++)
2234 		goto out;
2235 
2236 	buffer = tr->array_buffer.buffer;
2237 	if (buffer)
2238 		ring_buffer_record_disable(buffer);
2239 
2240  out:
2241 	raw_spin_unlock_irqrestore(&tr->start_lock, flags);
2242 }
2243 
2244 static int trace_save_cmdline(struct task_struct *tsk)
2245 {
2246 	unsigned pid, idx;
2247 
2248 	/* treat recording of idle task as a success */
2249 	if (!tsk->pid)
2250 		return 1;
2251 
2252 	if (unlikely(tsk->pid > PID_MAX_DEFAULT))
2253 		return 0;
2254 
2255 	/*
2256 	 * It's not the end of the world if we don't get
2257 	 * the lock, but we also don't want to spin
2258 	 * nor do we want to disable interrupts,
2259 	 * so if we miss here, then better luck next time.
2260 	 */
2261 	if (!arch_spin_trylock(&trace_cmdline_lock))
2262 		return 0;
2263 
2264 	idx = savedcmd->map_pid_to_cmdline[tsk->pid];
2265 	if (idx == NO_CMDLINE_MAP) {
2266 		idx = (savedcmd->cmdline_idx + 1) % savedcmd->cmdline_num;
2267 
2268 		/*
2269 		 * Check whether the cmdline buffer at idx has a pid
2270 		 * mapped. We are going to overwrite that entry so we
2271 		 * need to clear the map_pid_to_cmdline. Otherwise we
2272 		 * would read the new comm for the old pid.
2273 		 */
2274 		pid = savedcmd->map_cmdline_to_pid[idx];
2275 		if (pid != NO_CMDLINE_MAP)
2276 			savedcmd->map_pid_to_cmdline[pid] = NO_CMDLINE_MAP;
2277 
2278 		savedcmd->map_cmdline_to_pid[idx] = tsk->pid;
2279 		savedcmd->map_pid_to_cmdline[tsk->pid] = idx;
2280 
2281 		savedcmd->cmdline_idx = idx;
2282 	}
2283 
2284 	set_cmdline(idx, tsk->comm);
2285 
2286 	arch_spin_unlock(&trace_cmdline_lock);
2287 
2288 	return 1;
2289 }
2290 
2291 static void __trace_find_cmdline(int pid, char comm[])
2292 {
2293 	unsigned map;
2294 
2295 	if (!pid) {
2296 		strcpy(comm, "<idle>");
2297 		return;
2298 	}
2299 
2300 	if (WARN_ON_ONCE(pid < 0)) {
2301 		strcpy(comm, "<XXX>");
2302 		return;
2303 	}
2304 
2305 	if (pid > PID_MAX_DEFAULT) {
2306 		strcpy(comm, "<...>");
2307 		return;
2308 	}
2309 
2310 	map = savedcmd->map_pid_to_cmdline[pid];
2311 	if (map != NO_CMDLINE_MAP)
2312 		strlcpy(comm, get_saved_cmdlines(map), TASK_COMM_LEN);
2313 	else
2314 		strcpy(comm, "<...>");
2315 }
2316 
2317 void trace_find_cmdline(int pid, char comm[])
2318 {
2319 	preempt_disable();
2320 	arch_spin_lock(&trace_cmdline_lock);
2321 
2322 	__trace_find_cmdline(pid, comm);
2323 
2324 	arch_spin_unlock(&trace_cmdline_lock);
2325 	preempt_enable();
2326 }
2327 
2328 int trace_find_tgid(int pid)
2329 {
2330 	if (unlikely(!tgid_map || !pid || pid > PID_MAX_DEFAULT))
2331 		return 0;
2332 
2333 	return tgid_map[pid];
2334 }
2335 
2336 static int trace_save_tgid(struct task_struct *tsk)
2337 {
2338 	/* treat recording of idle task as a success */
2339 	if (!tsk->pid)
2340 		return 1;
2341 
2342 	if (unlikely(!tgid_map || tsk->pid > PID_MAX_DEFAULT))
2343 		return 0;
2344 
2345 	tgid_map[tsk->pid] = tsk->tgid;
2346 	return 1;
2347 }
2348 
2349 static bool tracing_record_taskinfo_skip(int flags)
2350 {
2351 	if (unlikely(!(flags & (TRACE_RECORD_CMDLINE | TRACE_RECORD_TGID))))
2352 		return true;
2353 	if (atomic_read(&trace_record_taskinfo_disabled) || !tracing_is_on())
2354 		return true;
2355 	if (!__this_cpu_read(trace_taskinfo_save))
2356 		return true;
2357 	return false;
2358 }
2359 
2360 /**
2361  * tracing_record_taskinfo - record the task info of a task
2362  *
2363  * @task:  task to record
2364  * @flags: TRACE_RECORD_CMDLINE for recording comm
2365  *         TRACE_RECORD_TGID for recording tgid
2366  */
2367 void tracing_record_taskinfo(struct task_struct *task, int flags)
2368 {
2369 	bool done;
2370 
2371 	if (tracing_record_taskinfo_skip(flags))
2372 		return;
2373 
2374 	/*
2375 	 * Record as much task information as possible. If some fail, continue
2376 	 * to try to record the others.
2377 	 */
2378 	done = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(task);
2379 	done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(task);
2380 
2381 	/* If recording any information failed, retry again soon. */
2382 	if (!done)
2383 		return;
2384 
2385 	__this_cpu_write(trace_taskinfo_save, false);
2386 }
2387 
2388 /**
2389  * tracing_record_taskinfo_sched_switch - record task info for sched_switch
2390  *
2391  * @prev: previous task during sched_switch
2392  * @next: next task during sched_switch
2393  * @flags: TRACE_RECORD_CMDLINE for recording comm
2394  *         TRACE_RECORD_TGID for recording tgid
2395  */
2396 void tracing_record_taskinfo_sched_switch(struct task_struct *prev,
2397 					  struct task_struct *next, int flags)
2398 {
2399 	bool done;
2400 
2401 	if (tracing_record_taskinfo_skip(flags))
2402 		return;
2403 
2404 	/*
2405 	 * Record as much task information as possible. If some fail, continue
2406 	 * to try to record the others.
2407 	 */
2408 	done  = !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(prev);
2409 	done &= !(flags & TRACE_RECORD_CMDLINE) || trace_save_cmdline(next);
2410 	done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(prev);
2411 	done &= !(flags & TRACE_RECORD_TGID) || trace_save_tgid(next);
2412 
2413 	/* If recording any information failed, retry again soon. */
2414 	if (!done)
2415 		return;
2416 
2417 	__this_cpu_write(trace_taskinfo_save, false);
2418 }
2419 
2420 /* Helpers to record a specific task information */
2421 void tracing_record_cmdline(struct task_struct *task)
2422 {
2423 	tracing_record_taskinfo(task, TRACE_RECORD_CMDLINE);
2424 }
2425 
2426 void tracing_record_tgid(struct task_struct *task)
2427 {
2428 	tracing_record_taskinfo(task, TRACE_RECORD_TGID);
2429 }
2430 
2431 /*
2432  * Several functions return TRACE_TYPE_PARTIAL_LINE if the trace_seq
2433  * overflowed, and TRACE_TYPE_HANDLED otherwise. This helper function
2434  * simplifies those functions and keeps them in sync.
2435  */
2436 enum print_line_t trace_handle_return(struct trace_seq *s)
2437 {
2438 	return trace_seq_has_overflowed(s) ?
2439 		TRACE_TYPE_PARTIAL_LINE : TRACE_TYPE_HANDLED;
2440 }
2441 EXPORT_SYMBOL_GPL(trace_handle_return);
2442 
2443 void
2444 tracing_generic_entry_update(struct trace_entry *entry, unsigned short type,
2445 			     unsigned long flags, int pc)
2446 {
2447 	struct task_struct *tsk = current;
2448 
2449 	entry->preempt_count		= pc & 0xff;
2450 	entry->pid			= (tsk) ? tsk->pid : 0;
2451 	entry->type			= type;
2452 	entry->flags =
2453 #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT
2454 		(irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) |
2455 #else
2456 		TRACE_FLAG_IRQS_NOSUPPORT |
2457 #endif
2458 		((pc & NMI_MASK    ) ? TRACE_FLAG_NMI     : 0) |
2459 		((pc & HARDIRQ_MASK) ? TRACE_FLAG_HARDIRQ : 0) |
2460 		((pc & SOFTIRQ_OFFSET) ? TRACE_FLAG_SOFTIRQ : 0) |
2461 		(tif_need_resched() ? TRACE_FLAG_NEED_RESCHED : 0) |
2462 		(test_preempt_need_resched() ? TRACE_FLAG_PREEMPT_RESCHED : 0);
2463 }
2464 EXPORT_SYMBOL_GPL(tracing_generic_entry_update);
2465 
2466 struct ring_buffer_event *
2467 trace_buffer_lock_reserve(struct trace_buffer *buffer,
2468 			  int type,
2469 			  unsigned long len,
2470 			  unsigned long flags, int pc)
2471 {
2472 	return __trace_buffer_lock_reserve(buffer, type, len, flags, pc);
2473 }
2474 
2475 DEFINE_PER_CPU(struct ring_buffer_event *, trace_buffered_event);
2476 DEFINE_PER_CPU(int, trace_buffered_event_cnt);
2477 static int trace_buffered_event_ref;
2478 
2479 /**
2480  * trace_buffered_event_enable - enable buffering events
2481  *
2482  * When events are being filtered, it is quicker to use a temporary
2483  * buffer to write the event data into if there's a likely chance
2484  * that it will not be committed. The discard of the ring buffer
2485  * is not as fast as committing, and is much slower than copying
2486  * a commit.
2487  *
2488  * When an event is to be filtered, allocate per cpu buffers to
2489  * write the event data into, and if the event is filtered and discarded
2490  * it is simply dropped, otherwise, the entire data is to be committed
2491  * in one shot.
2492  */
2493 void trace_buffered_event_enable(void)
2494 {
2495 	struct ring_buffer_event *event;
2496 	struct page *page;
2497 	int cpu;
2498 
2499 	WARN_ON_ONCE(!mutex_is_locked(&event_mutex));
2500 
2501 	if (trace_buffered_event_ref++)
2502 		return;
2503 
2504 	for_each_tracing_cpu(cpu) {
2505 		page = alloc_pages_node(cpu_to_node(cpu),
2506 					GFP_KERNEL | __GFP_NORETRY, 0);
2507 		if (!page)
2508 			goto failed;
2509 
2510 		event = page_address(page);
2511 		memset(event, 0, sizeof(*event));
2512 
2513 		per_cpu(trace_buffered_event, cpu) = event;
2514 
2515 		preempt_disable();
2516 		if (cpu == smp_processor_id() &&
2517 		    this_cpu_read(trace_buffered_event) !=
2518 		    per_cpu(trace_buffered_event, cpu))
2519 			WARN_ON_ONCE(1);
2520 		preempt_enable();
2521 	}
2522 
2523 	return;
2524  failed:
2525 	trace_buffered_event_disable();
2526 }
2527 
2528 static void enable_trace_buffered_event(void *data)
2529 {
2530 	/* Probably not needed, but do it anyway */
2531 	smp_rmb();
2532 	this_cpu_dec(trace_buffered_event_cnt);
2533 }
2534 
2535 static void disable_trace_buffered_event(void *data)
2536 {
2537 	this_cpu_inc(trace_buffered_event_cnt);
2538 }
2539 
2540 /**
2541  * trace_buffered_event_disable - disable buffering events
2542  *
2543  * When a filter is removed, it is faster to not use the buffered
2544  * events, and to commit directly into the ring buffer. Free up
2545  * the temp buffers when there are no more users. This requires
2546  * special synchronization with current events.
2547  */
2548 void trace_buffered_event_disable(void)
2549 {
2550 	int cpu;
2551 
2552 	WARN_ON_ONCE(!mutex_is_locked(&event_mutex));
2553 
2554 	if (WARN_ON_ONCE(!trace_buffered_event_ref))
2555 		return;
2556 
2557 	if (--trace_buffered_event_ref)
2558 		return;
2559 
2560 	preempt_disable();
2561 	/* For each CPU, set the buffer as used. */
2562 	smp_call_function_many(tracing_buffer_mask,
2563 			       disable_trace_buffered_event, NULL, 1);
2564 	preempt_enable();
2565 
2566 	/* Wait for all current users to finish */
2567 	synchronize_rcu();
2568 
2569 	for_each_tracing_cpu(cpu) {
2570 		free_page((unsigned long)per_cpu(trace_buffered_event, cpu));
2571 		per_cpu(trace_buffered_event, cpu) = NULL;
2572 	}
2573 	/*
2574 	 * Make sure trace_buffered_event is NULL before clearing
2575 	 * trace_buffered_event_cnt.
2576 	 */
2577 	smp_wmb();
2578 
2579 	preempt_disable();
2580 	/* Do the work on each cpu */
2581 	smp_call_function_many(tracing_buffer_mask,
2582 			       enable_trace_buffered_event, NULL, 1);
2583 	preempt_enable();
2584 }
2585 
2586 static struct trace_buffer *temp_buffer;
2587 
2588 struct ring_buffer_event *
2589 trace_event_buffer_lock_reserve(struct trace_buffer **current_rb,
2590 			  struct trace_event_file *trace_file,
2591 			  int type, unsigned long len,
2592 			  unsigned long flags, int pc)
2593 {
2594 	struct ring_buffer_event *entry;
2595 	int val;
2596 
2597 	*current_rb = trace_file->tr->array_buffer.buffer;
2598 
2599 	if (!ring_buffer_time_stamp_abs(*current_rb) && (trace_file->flags &
2600 	     (EVENT_FILE_FL_SOFT_DISABLED | EVENT_FILE_FL_FILTERED)) &&
2601 	    (entry = this_cpu_read(trace_buffered_event))) {
2602 		/* Try to use the per cpu buffer first */
2603 		val = this_cpu_inc_return(trace_buffered_event_cnt);
2604 		if (val == 1) {
2605 			trace_event_setup(entry, type, flags, pc);
2606 			entry->array[0] = len;
2607 			return entry;
2608 		}
2609 		this_cpu_dec(trace_buffered_event_cnt);
2610 	}
2611 
2612 	entry = __trace_buffer_lock_reserve(*current_rb,
2613 					    type, len, flags, pc);
2614 	/*
2615 	 * If tracing is off, but we have triggers enabled
2616 	 * we still need to look at the event data. Use the temp_buffer
2617 	 * to store the trace event for the tigger to use. It's recusive
2618 	 * safe and will not be recorded anywhere.
2619 	 */
2620 	if (!entry && trace_file->flags & EVENT_FILE_FL_TRIGGER_COND) {
2621 		*current_rb = temp_buffer;
2622 		entry = __trace_buffer_lock_reserve(*current_rb,
2623 						    type, len, flags, pc);
2624 	}
2625 	return entry;
2626 }
2627 EXPORT_SYMBOL_GPL(trace_event_buffer_lock_reserve);
2628 
2629 static DEFINE_SPINLOCK(tracepoint_iter_lock);
2630 static DEFINE_MUTEX(tracepoint_printk_mutex);
2631 
2632 static void output_printk(struct trace_event_buffer *fbuffer)
2633 {
2634 	struct trace_event_call *event_call;
2635 	struct trace_event_file *file;
2636 	struct trace_event *event;
2637 	unsigned long flags;
2638 	struct trace_iterator *iter = tracepoint_print_iter;
2639 
2640 	/* We should never get here if iter is NULL */
2641 	if (WARN_ON_ONCE(!iter))
2642 		return;
2643 
2644 	event_call = fbuffer->trace_file->event_call;
2645 	if (!event_call || !event_call->event.funcs ||
2646 	    !event_call->event.funcs->trace)
2647 		return;
2648 
2649 	file = fbuffer->trace_file;
2650 	if (test_bit(EVENT_FILE_FL_SOFT_DISABLED_BIT, &file->flags) ||
2651 	    (unlikely(file->flags & EVENT_FILE_FL_FILTERED) &&
2652 	     !filter_match_preds(file->filter, fbuffer->entry)))
2653 		return;
2654 
2655 	event = &fbuffer->trace_file->event_call->event;
2656 
2657 	spin_lock_irqsave(&tracepoint_iter_lock, flags);
2658 	trace_seq_init(&iter->seq);
2659 	iter->ent = fbuffer->entry;
2660 	event_call->event.funcs->trace(iter, 0, event);
2661 	trace_seq_putc(&iter->seq, 0);
2662 	printk("%s", iter->seq.buffer);
2663 
2664 	spin_unlock_irqrestore(&tracepoint_iter_lock, flags);
2665 }
2666 
2667 int tracepoint_printk_sysctl(struct ctl_table *table, int write,
2668 			     void *buffer, size_t *lenp,
2669 			     loff_t *ppos)
2670 {
2671 	int save_tracepoint_printk;
2672 	int ret;
2673 
2674 	mutex_lock(&tracepoint_printk_mutex);
2675 	save_tracepoint_printk = tracepoint_printk;
2676 
2677 	ret = proc_dointvec(table, write, buffer, lenp, ppos);
2678 
2679 	/*
2680 	 * This will force exiting early, as tracepoint_printk
2681 	 * is always zero when tracepoint_printk_iter is not allocated
2682 	 */
2683 	if (!tracepoint_print_iter)
2684 		tracepoint_printk = 0;
2685 
2686 	if (save_tracepoint_printk == tracepoint_printk)
2687 		goto out;
2688 
2689 	if (tracepoint_printk)
2690 		static_key_enable(&tracepoint_printk_key.key);
2691 	else
2692 		static_key_disable(&tracepoint_printk_key.key);
2693 
2694  out:
2695 	mutex_unlock(&tracepoint_printk_mutex);
2696 
2697 	return ret;
2698 }
2699 
2700 void trace_event_buffer_commit(struct trace_event_buffer *fbuffer)
2701 {
2702 	if (static_key_false(&tracepoint_printk_key.key))
2703 		output_printk(fbuffer);
2704 
2705 	event_trigger_unlock_commit_regs(fbuffer->trace_file, fbuffer->buffer,
2706 				    fbuffer->event, fbuffer->entry,
2707 				    fbuffer->flags, fbuffer->pc, fbuffer->regs);
2708 }
2709 EXPORT_SYMBOL_GPL(trace_event_buffer_commit);
2710 
2711 /*
2712  * Skip 3:
2713  *
2714  *   trace_buffer_unlock_commit_regs()
2715  *   trace_event_buffer_commit()
2716  *   trace_event_raw_event_xxx()
2717  */
2718 # define STACK_SKIP 3
2719 
2720 void trace_buffer_unlock_commit_regs(struct trace_array *tr,
2721 				     struct trace_buffer *buffer,
2722 				     struct ring_buffer_event *event,
2723 				     unsigned long flags, int pc,
2724 				     struct pt_regs *regs)
2725 {
2726 	__buffer_unlock_commit(buffer, event);
2727 
2728 	/*
2729 	 * If regs is not set, then skip the necessary functions.
2730 	 * Note, we can still get here via blktrace, wakeup tracer
2731 	 * and mmiotrace, but that's ok if they lose a function or
2732 	 * two. They are not that meaningful.
2733 	 */
2734 	ftrace_trace_stack(tr, buffer, flags, regs ? 0 : STACK_SKIP, pc, regs);
2735 	ftrace_trace_userstack(buffer, flags, pc);
2736 }
2737 
2738 /*
2739  * Similar to trace_buffer_unlock_commit_regs() but do not dump stack.
2740  */
2741 void
2742 trace_buffer_unlock_commit_nostack(struct trace_buffer *buffer,
2743 				   struct ring_buffer_event *event)
2744 {
2745 	__buffer_unlock_commit(buffer, event);
2746 }
2747 
2748 static void
2749 trace_process_export(struct trace_export *export,
2750 	       struct ring_buffer_event *event)
2751 {
2752 	struct trace_entry *entry;
2753 	unsigned int size = 0;
2754 
2755 	entry = ring_buffer_event_data(event);
2756 	size = ring_buffer_event_length(event);
2757 	export->write(export, entry, size);
2758 }
2759 
2760 static DEFINE_MUTEX(ftrace_export_lock);
2761 
2762 static struct trace_export __rcu *ftrace_exports_list __read_mostly;
2763 
2764 static DEFINE_STATIC_KEY_FALSE(ftrace_exports_enabled);
2765 
2766 static inline void ftrace_exports_enable(void)
2767 {
2768 	static_branch_enable(&ftrace_exports_enabled);
2769 }
2770 
2771 static inline void ftrace_exports_disable(void)
2772 {
2773 	static_branch_disable(&ftrace_exports_enabled);
2774 }
2775 
2776 static void ftrace_exports(struct ring_buffer_event *event)
2777 {
2778 	struct trace_export *export;
2779 
2780 	preempt_disable_notrace();
2781 
2782 	export = rcu_dereference_raw_check(ftrace_exports_list);
2783 	while (export) {
2784 		trace_process_export(export, event);
2785 		export = rcu_dereference_raw_check(export->next);
2786 	}
2787 
2788 	preempt_enable_notrace();
2789 }
2790 
2791 static inline void
2792 add_trace_export(struct trace_export **list, struct trace_export *export)
2793 {
2794 	rcu_assign_pointer(export->next, *list);
2795 	/*
2796 	 * We are entering export into the list but another
2797 	 * CPU might be walking that list. We need to make sure
2798 	 * the export->next pointer is valid before another CPU sees
2799 	 * the export pointer included into the list.
2800 	 */
2801 	rcu_assign_pointer(*list, export);
2802 }
2803 
2804 static inline int
2805 rm_trace_export(struct trace_export **list, struct trace_export *export)
2806 {
2807 	struct trace_export **p;
2808 
2809 	for (p = list; *p != NULL; p = &(*p)->next)
2810 		if (*p == export)
2811 			break;
2812 
2813 	if (*p != export)
2814 		return -1;
2815 
2816 	rcu_assign_pointer(*p, (*p)->next);
2817 
2818 	return 0;
2819 }
2820 
2821 static inline void
2822 add_ftrace_export(struct trace_export **list, struct trace_export *export)
2823 {
2824 	if (*list == NULL)
2825 		ftrace_exports_enable();
2826 
2827 	add_trace_export(list, export);
2828 }
2829 
2830 static inline int
2831 rm_ftrace_export(struct trace_export **list, struct trace_export *export)
2832 {
2833 	int ret;
2834 
2835 	ret = rm_trace_export(list, export);
2836 	if (*list == NULL)
2837 		ftrace_exports_disable();
2838 
2839 	return ret;
2840 }
2841 
2842 int register_ftrace_export(struct trace_export *export)
2843 {
2844 	if (WARN_ON_ONCE(!export->write))
2845 		return -1;
2846 
2847 	mutex_lock(&ftrace_export_lock);
2848 
2849 	add_ftrace_export(&ftrace_exports_list, export);
2850 
2851 	mutex_unlock(&ftrace_export_lock);
2852 
2853 	return 0;
2854 }
2855 EXPORT_SYMBOL_GPL(register_ftrace_export);
2856 
2857 int unregister_ftrace_export(struct trace_export *export)
2858 {
2859 	int ret;
2860 
2861 	mutex_lock(&ftrace_export_lock);
2862 
2863 	ret = rm_ftrace_export(&ftrace_exports_list, export);
2864 
2865 	mutex_unlock(&ftrace_export_lock);
2866 
2867 	return ret;
2868 }
2869 EXPORT_SYMBOL_GPL(unregister_ftrace_export);
2870 
2871 void
2872 trace_function(struct trace_array *tr,
2873 	       unsigned long ip, unsigned long parent_ip, unsigned long flags,
2874 	       int pc)
2875 {
2876 	struct trace_event_call *call = &event_function;
2877 	struct trace_buffer *buffer = tr->array_buffer.buffer;
2878 	struct ring_buffer_event *event;
2879 	struct ftrace_entry *entry;
2880 
2881 	event = __trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry),
2882 					    flags, pc);
2883 	if (!event)
2884 		return;
2885 	entry	= ring_buffer_event_data(event);
2886 	entry->ip			= ip;
2887 	entry->parent_ip		= parent_ip;
2888 
2889 	if (!call_filter_check_discard(call, entry, buffer, event)) {
2890 		if (static_branch_unlikely(&ftrace_exports_enabled))
2891 			ftrace_exports(event);
2892 		__buffer_unlock_commit(buffer, event);
2893 	}
2894 }
2895 
2896 #ifdef CONFIG_STACKTRACE
2897 
2898 /* Allow 4 levels of nesting: normal, softirq, irq, NMI */
2899 #define FTRACE_KSTACK_NESTING	4
2900 
2901 #define FTRACE_KSTACK_ENTRIES	(PAGE_SIZE / FTRACE_KSTACK_NESTING)
2902 
2903 struct ftrace_stack {
2904 	unsigned long		calls[FTRACE_KSTACK_ENTRIES];
2905 };
2906 
2907 
2908 struct ftrace_stacks {
2909 	struct ftrace_stack	stacks[FTRACE_KSTACK_NESTING];
2910 };
2911 
2912 static DEFINE_PER_CPU(struct ftrace_stacks, ftrace_stacks);
2913 static DEFINE_PER_CPU(int, ftrace_stack_reserve);
2914 
2915 static void __ftrace_trace_stack(struct trace_buffer *buffer,
2916 				 unsigned long flags,
2917 				 int skip, int pc, struct pt_regs *regs)
2918 {
2919 	struct trace_event_call *call = &event_kernel_stack;
2920 	struct ring_buffer_event *event;
2921 	unsigned int size, nr_entries;
2922 	struct ftrace_stack *fstack;
2923 	struct stack_entry *entry;
2924 	int stackidx;
2925 
2926 	/*
2927 	 * Add one, for this function and the call to save_stack_trace()
2928 	 * If regs is set, then these functions will not be in the way.
2929 	 */
2930 #ifndef CONFIG_UNWINDER_ORC
2931 	if (!regs)
2932 		skip++;
2933 #endif
2934 
2935 	/*
2936 	 * Since events can happen in NMIs there's no safe way to
2937 	 * use the per cpu ftrace_stacks. We reserve it and if an interrupt
2938 	 * or NMI comes in, it will just have to use the default
2939 	 * FTRACE_STACK_SIZE.
2940 	 */
2941 	preempt_disable_notrace();
2942 
2943 	stackidx = __this_cpu_inc_return(ftrace_stack_reserve) - 1;
2944 
2945 	/* This should never happen. If it does, yell once and skip */
2946 	if (WARN_ON_ONCE(stackidx > FTRACE_KSTACK_NESTING))
2947 		goto out;
2948 
2949 	/*
2950 	 * The above __this_cpu_inc_return() is 'atomic' cpu local. An
2951 	 * interrupt will either see the value pre increment or post
2952 	 * increment. If the interrupt happens pre increment it will have
2953 	 * restored the counter when it returns.  We just need a barrier to
2954 	 * keep gcc from moving things around.
2955 	 */
2956 	barrier();
2957 
2958 	fstack = this_cpu_ptr(ftrace_stacks.stacks) + stackidx;
2959 	size = ARRAY_SIZE(fstack->calls);
2960 
2961 	if (regs) {
2962 		nr_entries = stack_trace_save_regs(regs, fstack->calls,
2963 						   size, skip);
2964 	} else {
2965 		nr_entries = stack_trace_save(fstack->calls, size, skip);
2966 	}
2967 
2968 	size = nr_entries * sizeof(unsigned long);
2969 	event = __trace_buffer_lock_reserve(buffer, TRACE_STACK,
2970 					    sizeof(*entry) + size, flags, pc);
2971 	if (!event)
2972 		goto out;
2973 	entry = ring_buffer_event_data(event);
2974 
2975 	memcpy(&entry->caller, fstack->calls, size);
2976 	entry->size = nr_entries;
2977 
2978 	if (!call_filter_check_discard(call, entry, buffer, event))
2979 		__buffer_unlock_commit(buffer, event);
2980 
2981  out:
2982 	/* Again, don't let gcc optimize things here */
2983 	barrier();
2984 	__this_cpu_dec(ftrace_stack_reserve);
2985 	preempt_enable_notrace();
2986 
2987 }
2988 
2989 static inline void ftrace_trace_stack(struct trace_array *tr,
2990 				      struct trace_buffer *buffer,
2991 				      unsigned long flags,
2992 				      int skip, int pc, struct pt_regs *regs)
2993 {
2994 	if (!(tr->trace_flags & TRACE_ITER_STACKTRACE))
2995 		return;
2996 
2997 	__ftrace_trace_stack(buffer, flags, skip, pc, regs);
2998 }
2999 
3000 void __trace_stack(struct trace_array *tr, unsigned long flags, int skip,
3001 		   int pc)
3002 {
3003 	struct trace_buffer *buffer = tr->array_buffer.buffer;
3004 
3005 	if (rcu_is_watching()) {
3006 		__ftrace_trace_stack(buffer, flags, skip, pc, NULL);
3007 		return;
3008 	}
3009 
3010 	/*
3011 	 * When an NMI triggers, RCU is enabled via rcu_nmi_enter(),
3012 	 * but if the above rcu_is_watching() failed, then the NMI
3013 	 * triggered someplace critical, and rcu_irq_enter() should
3014 	 * not be called from NMI.
3015 	 */
3016 	if (unlikely(in_nmi()))
3017 		return;
3018 
3019 	rcu_irq_enter_irqson();
3020 	__ftrace_trace_stack(buffer, flags, skip, pc, NULL);
3021 	rcu_irq_exit_irqson();
3022 }
3023 
3024 /**
3025  * trace_dump_stack - record a stack back trace in the trace buffer
3026  * @skip: Number of functions to skip (helper handlers)
3027  */
3028 void trace_dump_stack(int skip)
3029 {
3030 	unsigned long flags;
3031 
3032 	if (tracing_disabled || tracing_selftest_running)
3033 		return;
3034 
3035 	local_save_flags(flags);
3036 
3037 #ifndef CONFIG_UNWINDER_ORC
3038 	/* Skip 1 to skip this function. */
3039 	skip++;
3040 #endif
3041 	__ftrace_trace_stack(global_trace.array_buffer.buffer,
3042 			     flags, skip, preempt_count(), NULL);
3043 }
3044 EXPORT_SYMBOL_GPL(trace_dump_stack);
3045 
3046 #ifdef CONFIG_USER_STACKTRACE_SUPPORT
3047 static DEFINE_PER_CPU(int, user_stack_count);
3048 
3049 static void
3050 ftrace_trace_userstack(struct trace_buffer *buffer, unsigned long flags, int pc)
3051 {
3052 	struct trace_event_call *call = &event_user_stack;
3053 	struct ring_buffer_event *event;
3054 	struct userstack_entry *entry;
3055 
3056 	if (!(global_trace.trace_flags & TRACE_ITER_USERSTACKTRACE))
3057 		return;
3058 
3059 	/*
3060 	 * NMIs can not handle page faults, even with fix ups.
3061 	 * The save user stack can (and often does) fault.
3062 	 */
3063 	if (unlikely(in_nmi()))
3064 		return;
3065 
3066 	/*
3067 	 * prevent recursion, since the user stack tracing may
3068 	 * trigger other kernel events.
3069 	 */
3070 	preempt_disable();
3071 	if (__this_cpu_read(user_stack_count))
3072 		goto out;
3073 
3074 	__this_cpu_inc(user_stack_count);
3075 
3076 	event = __trace_buffer_lock_reserve(buffer, TRACE_USER_STACK,
3077 					    sizeof(*entry), flags, pc);
3078 	if (!event)
3079 		goto out_drop_count;
3080 	entry	= ring_buffer_event_data(event);
3081 
3082 	entry->tgid		= current->tgid;
3083 	memset(&entry->caller, 0, sizeof(entry->caller));
3084 
3085 	stack_trace_save_user(entry->caller, FTRACE_STACK_ENTRIES);
3086 	if (!call_filter_check_discard(call, entry, buffer, event))
3087 		__buffer_unlock_commit(buffer, event);
3088 
3089  out_drop_count:
3090 	__this_cpu_dec(user_stack_count);
3091  out:
3092 	preempt_enable();
3093 }
3094 #else /* CONFIG_USER_STACKTRACE_SUPPORT */
3095 static void ftrace_trace_userstack(struct trace_buffer *buffer,
3096 				   unsigned long flags, int pc)
3097 {
3098 }
3099 #endif /* !CONFIG_USER_STACKTRACE_SUPPORT */
3100 
3101 #endif /* CONFIG_STACKTRACE */
3102 
3103 /* created for use with alloc_percpu */
3104 struct trace_buffer_struct {
3105 	int nesting;
3106 	char buffer[4][TRACE_BUF_SIZE];
3107 };
3108 
3109 static struct trace_buffer_struct *trace_percpu_buffer;
3110 
3111 /*
3112  * Thise allows for lockless recording.  If we're nested too deeply, then
3113  * this returns NULL.
3114  */
3115 static char *get_trace_buf(void)
3116 {
3117 	struct trace_buffer_struct *buffer = this_cpu_ptr(trace_percpu_buffer);
3118 
3119 	if (!buffer || buffer->nesting >= 4)
3120 		return NULL;
3121 
3122 	buffer->nesting++;
3123 
3124 	/* Interrupts must see nesting incremented before we use the buffer */
3125 	barrier();
3126 	return &buffer->buffer[buffer->nesting][0];
3127 }
3128 
3129 static void put_trace_buf(void)
3130 {
3131 	/* Don't let the decrement of nesting leak before this */
3132 	barrier();
3133 	this_cpu_dec(trace_percpu_buffer->nesting);
3134 }
3135 
3136 static int alloc_percpu_trace_buffer(void)
3137 {
3138 	struct trace_buffer_struct *buffers;
3139 
3140 	buffers = alloc_percpu(struct trace_buffer_struct);
3141 	if (MEM_FAIL(!buffers, "Could not allocate percpu trace_printk buffer"))
3142 		return -ENOMEM;
3143 
3144 	trace_percpu_buffer = buffers;
3145 	return 0;
3146 }
3147 
3148 static int buffers_allocated;
3149 
3150 void trace_printk_init_buffers(void)
3151 {
3152 	if (buffers_allocated)
3153 		return;
3154 
3155 	if (alloc_percpu_trace_buffer())
3156 		return;
3157 
3158 	/* trace_printk() is for debug use only. Don't use it in production. */
3159 
3160 	pr_warn("\n");
3161 	pr_warn("**********************************************************\n");
3162 	pr_warn("**   NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE   **\n");
3163 	pr_warn("**                                                      **\n");
3164 	pr_warn("** trace_printk() being used. Allocating extra memory.  **\n");
3165 	pr_warn("**                                                      **\n");
3166 	pr_warn("** This means that this is a DEBUG kernel and it is     **\n");
3167 	pr_warn("** unsafe for production use.                           **\n");
3168 	pr_warn("**                                                      **\n");
3169 	pr_warn("** If you see this message and you are not debugging    **\n");
3170 	pr_warn("** the kernel, report this immediately to your vendor!  **\n");
3171 	pr_warn("**                                                      **\n");
3172 	pr_warn("**   NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE NOTICE   **\n");
3173 	pr_warn("**********************************************************\n");
3174 
3175 	/* Expand the buffers to set size */
3176 	tracing_update_buffers();
3177 
3178 	buffers_allocated = 1;
3179 
3180 	/*
3181 	 * trace_printk_init_buffers() can be called by modules.
3182 	 * If that happens, then we need to start cmdline recording
3183 	 * directly here. If the global_trace.buffer is already
3184 	 * allocated here, then this was called by module code.
3185 	 */
3186 	if (global_trace.array_buffer.buffer)
3187 		tracing_start_cmdline_record();
3188 }
3189 EXPORT_SYMBOL_GPL(trace_printk_init_buffers);
3190 
3191 void trace_printk_start_comm(void)
3192 {
3193 	/* Start tracing comms if trace printk is set */
3194 	if (!buffers_allocated)
3195 		return;
3196 	tracing_start_cmdline_record();
3197 }
3198 
3199 static void trace_printk_start_stop_comm(int enabled)
3200 {
3201 	if (!buffers_allocated)
3202 		return;
3203 
3204 	if (enabled)
3205 		tracing_start_cmdline_record();
3206 	else
3207 		tracing_stop_cmdline_record();
3208 }
3209 
3210 /**
3211  * trace_vbprintk - write binary msg to tracing buffer
3212  * @ip:    The address of the caller
3213  * @fmt:   The string format to write to the buffer
3214  * @args:  Arguments for @fmt
3215  */
3216 int trace_vbprintk(unsigned long ip, const char *fmt, va_list args)
3217 {
3218 	struct trace_event_call *call = &event_bprint;
3219 	struct ring_buffer_event *event;
3220 	struct trace_buffer *buffer;
3221 	struct trace_array *tr = &global_trace;
3222 	struct bprint_entry *entry;
3223 	unsigned long flags;
3224 	char *tbuffer;
3225 	int len = 0, size, pc;
3226 
3227 	if (unlikely(tracing_selftest_running || tracing_disabled))
3228 		return 0;
3229 
3230 	/* Don't pollute graph traces with trace_vprintk internals */
3231 	pause_graph_tracing();
3232 
3233 	pc = preempt_count();
3234 	preempt_disable_notrace();
3235 
3236 	tbuffer = get_trace_buf();
3237 	if (!tbuffer) {
3238 		len = 0;
3239 		goto out_nobuffer;
3240 	}
3241 
3242 	len = vbin_printf((u32 *)tbuffer, TRACE_BUF_SIZE/sizeof(int), fmt, args);
3243 
3244 	if (len > TRACE_BUF_SIZE/sizeof(int) || len < 0)
3245 		goto out_put;
3246 
3247 	local_save_flags(flags);
3248 	size = sizeof(*entry) + sizeof(u32) * len;
3249 	buffer = tr->array_buffer.buffer;
3250 	ring_buffer_nest_start(buffer);
3251 	event = __trace_buffer_lock_reserve(buffer, TRACE_BPRINT, size,
3252 					    flags, pc);
3253 	if (!event)
3254 		goto out;
3255 	entry = ring_buffer_event_data(event);
3256 	entry->ip			= ip;
3257 	entry->fmt			= fmt;
3258 
3259 	memcpy(entry->buf, tbuffer, sizeof(u32) * len);
3260 	if (!call_filter_check_discard(call, entry, buffer, event)) {
3261 		__buffer_unlock_commit(buffer, event);
3262 		ftrace_trace_stack(tr, buffer, flags, 6, pc, NULL);
3263 	}
3264 
3265 out:
3266 	ring_buffer_nest_end(buffer);
3267 out_put:
3268 	put_trace_buf();
3269 
3270 out_nobuffer:
3271 	preempt_enable_notrace();
3272 	unpause_graph_tracing();
3273 
3274 	return len;
3275 }
3276 EXPORT_SYMBOL_GPL(trace_vbprintk);
3277 
3278 __printf(3, 0)
3279 static int
3280 __trace_array_vprintk(struct trace_buffer *buffer,
3281 		      unsigned long ip, const char *fmt, va_list args)
3282 {
3283 	struct trace_event_call *call = &event_print;
3284 	struct ring_buffer_event *event;
3285 	int len = 0, size, pc;
3286 	struct print_entry *entry;
3287 	unsigned long flags;
3288 	char *tbuffer;
3289 
3290 	if (tracing_disabled || tracing_selftest_running)
3291 		return 0;
3292 
3293 	/* Don't pollute graph traces with trace_vprintk internals */
3294 	pause_graph_tracing();
3295 
3296 	pc = preempt_count();
3297 	preempt_disable_notrace();
3298 
3299 
3300 	tbuffer = get_trace_buf();
3301 	if (!tbuffer) {
3302 		len = 0;
3303 		goto out_nobuffer;
3304 	}
3305 
3306 	len = vscnprintf(tbuffer, TRACE_BUF_SIZE, fmt, args);
3307 
3308 	local_save_flags(flags);
3309 	size = sizeof(*entry) + len + 1;
3310 	ring_buffer_nest_start(buffer);
3311 	event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size,
3312 					    flags, pc);
3313 	if (!event)
3314 		goto out;
3315 	entry = ring_buffer_event_data(event);
3316 	entry->ip = ip;
3317 
3318 	memcpy(&entry->buf, tbuffer, len + 1);
3319 	if (!call_filter_check_discard(call, entry, buffer, event)) {
3320 		__buffer_unlock_commit(buffer, event);
3321 		ftrace_trace_stack(&global_trace, buffer, flags, 6, pc, NULL);
3322 	}
3323 
3324 out:
3325 	ring_buffer_nest_end(buffer);
3326 	put_trace_buf();
3327 
3328 out_nobuffer:
3329 	preempt_enable_notrace();
3330 	unpause_graph_tracing();
3331 
3332 	return len;
3333 }
3334 
3335 __printf(3, 0)
3336 int trace_array_vprintk(struct trace_array *tr,
3337 			unsigned long ip, const char *fmt, va_list args)
3338 {
3339 	return __trace_array_vprintk(tr->array_buffer.buffer, ip, fmt, args);
3340 }
3341 
3342 __printf(3, 0)
3343 int trace_array_printk(struct trace_array *tr,
3344 		       unsigned long ip, const char *fmt, ...)
3345 {
3346 	int ret;
3347 	va_list ap;
3348 
3349 	if (!(global_trace.trace_flags & TRACE_ITER_PRINTK))
3350 		return 0;
3351 
3352 	if (!tr)
3353 		return -ENOENT;
3354 
3355 	va_start(ap, fmt);
3356 	ret = trace_array_vprintk(tr, ip, fmt, ap);
3357 	va_end(ap);
3358 	return ret;
3359 }
3360 EXPORT_SYMBOL_GPL(trace_array_printk);
3361 
3362 __printf(3, 4)
3363 int trace_array_printk_buf(struct trace_buffer *buffer,
3364 			   unsigned long ip, const char *fmt, ...)
3365 {
3366 	int ret;
3367 	va_list ap;
3368 
3369 	if (!(global_trace.trace_flags & TRACE_ITER_PRINTK))
3370 		return 0;
3371 
3372 	va_start(ap, fmt);
3373 	ret = __trace_array_vprintk(buffer, ip, fmt, ap);
3374 	va_end(ap);
3375 	return ret;
3376 }
3377 
3378 __printf(2, 0)
3379 int trace_vprintk(unsigned long ip, const char *fmt, va_list args)
3380 {
3381 	return trace_array_vprintk(&global_trace, ip, fmt, args);
3382 }
3383 EXPORT_SYMBOL_GPL(trace_vprintk);
3384 
3385 static void trace_iterator_increment(struct trace_iterator *iter)
3386 {
3387 	struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, iter->cpu);
3388 
3389 	iter->idx++;
3390 	if (buf_iter)
3391 		ring_buffer_iter_advance(buf_iter);
3392 }
3393 
3394 static struct trace_entry *
3395 peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts,
3396 		unsigned long *lost_events)
3397 {
3398 	struct ring_buffer_event *event;
3399 	struct ring_buffer_iter *buf_iter = trace_buffer_iter(iter, cpu);
3400 
3401 	if (buf_iter) {
3402 		event = ring_buffer_iter_peek(buf_iter, ts);
3403 		if (lost_events)
3404 			*lost_events = ring_buffer_iter_dropped(buf_iter) ?
3405 				(unsigned long)-1 : 0;
3406 	} else {
3407 		event = ring_buffer_peek(iter->array_buffer->buffer, cpu, ts,
3408 					 lost_events);
3409 	}
3410 
3411 	if (event) {
3412 		iter->ent_size = ring_buffer_event_length(event);
3413 		return ring_buffer_event_data(event);
3414 	}
3415 	iter->ent_size = 0;
3416 	return NULL;
3417 }
3418 
3419 static struct trace_entry *
3420 __find_next_entry(struct trace_iterator *iter, int *ent_cpu,
3421 		  unsigned long *missing_events, u64 *ent_ts)
3422 {
3423 	struct trace_buffer *buffer = iter->array_buffer->buffer;
3424 	struct trace_entry *ent, *next = NULL;
3425 	unsigned long lost_events = 0, next_lost = 0;
3426 	int cpu_file = iter->cpu_file;
3427 	u64 next_ts = 0, ts;
3428 	int next_cpu = -1;
3429 	int next_size = 0;
3430 	int cpu;
3431 
3432 	/*
3433 	 * If we are in a per_cpu trace file, don't bother by iterating over
3434 	 * all cpu and peek directly.
3435 	 */
3436 	if (cpu_file > RING_BUFFER_ALL_CPUS) {
3437 		if (ring_buffer_empty_cpu(buffer, cpu_file))
3438 			return NULL;
3439 		ent = peek_next_entry(iter, cpu_file, ent_ts, missing_events);
3440 		if (ent_cpu)
3441 			*ent_cpu = cpu_file;
3442 
3443 		return ent;
3444 	}
3445 
3446 	for_each_tracing_cpu(cpu) {
3447 
3448 		if (ring_buffer_empty_cpu(buffer, cpu))
3449 			continue;
3450 
3451 		ent = peek_next_entry(iter, cpu, &ts, &lost_events);
3452 
3453 		/*
3454 		 * Pick the entry with the smallest timestamp:
3455 		 */
3456 		if (ent && (!next || ts < next_ts)) {
3457 			next = ent;
3458 			next_cpu = cpu;
3459 			next_ts = ts;
3460 			next_lost = lost_events;
3461 			next_size = iter->ent_size;
3462 		}
3463 	}
3464 
3465 	iter->ent_size = next_size;
3466 
3467 	if (ent_cpu)
3468 		*ent_cpu = next_cpu;
3469 
3470 	if (ent_ts)
3471 		*ent_ts = next_ts;
3472 
3473 	if (missing_events)
3474 		*missing_events = next_lost;
3475 
3476 	return next;
3477 }
3478 
3479 #define STATIC_TEMP_BUF_SIZE	128
3480 static char static_temp_buf[STATIC_TEMP_BUF_SIZE];
3481 
3482 /* Find the next real entry, without updating the iterator itself */
3483 struct trace_entry *trace_find_next_entry(struct trace_iterator *iter,
3484 					  int *ent_cpu, u64 *ent_ts)
3485 {
3486 	/* __find_next_entry will reset ent_size */
3487 	int ent_size = iter->ent_size;
3488 	struct trace_entry *entry;
3489 
3490 	/*
3491 	 * If called from ftrace_dump(), then the iter->temp buffer
3492 	 * will be the static_temp_buf and not created from kmalloc.
3493 	 * If the entry size is greater than the buffer, we can
3494 	 * not save it. Just return NULL in that case. This is only
3495 	 * used to add markers when two consecutive events' time
3496 	 * stamps have a large delta. See trace_print_lat_context()
3497 	 */
3498 	if (iter->temp == static_temp_buf &&
3499 	    STATIC_TEMP_BUF_SIZE < ent_size)
3500 		return NULL;
3501 
3502 	/*
3503 	 * The __find_next_entry() may call peek_next_entry(), which may
3504 	 * call ring_buffer_peek() that may make the contents of iter->ent
3505 	 * undefined. Need to copy iter->ent now.
3506 	 */
3507 	if (iter->ent && iter->ent != iter->temp) {
3508 		if ((!iter->temp || iter->temp_size < iter->ent_size) &&
3509 		    !WARN_ON_ONCE(iter->temp == static_temp_buf)) {
3510 			kfree(iter->temp);
3511 			iter->temp = kmalloc(iter->ent_size, GFP_KERNEL);
3512 			if (!iter->temp)
3513 				return NULL;
3514 		}
3515 		memcpy(iter->temp, iter->ent, iter->ent_size);
3516 		iter->temp_size = iter->ent_size;
3517 		iter->ent = iter->temp;
3518 	}
3519 	entry = __find_next_entry(iter, ent_cpu, NULL, ent_ts);
3520 	/* Put back the original ent_size */
3521 	iter->ent_size = ent_size;
3522 
3523 	return entry;
3524 }
3525 
3526 /* Find the next real entry, and increment the iterator to the next entry */
3527 void *trace_find_next_entry_inc(struct trace_iterator *iter)
3528 {
3529 	iter->ent = __find_next_entry(iter, &iter->cpu,
3530 				      &iter->lost_events, &iter->ts);
3531 
3532 	if (iter->ent)
3533 		trace_iterator_increment(iter);
3534 
3535 	return iter->ent ? iter : NULL;
3536 }
3537 
3538 static void trace_consume(struct trace_iterator *iter)
3539 {
3540 	ring_buffer_consume(iter->array_buffer->buffer, iter->cpu, &iter->ts,
3541 			    &iter->lost_events);
3542 }
3543 
3544 static void *s_next(struct seq_file *m, void *v, loff_t *pos)
3545 {
3546 	struct trace_iterator *iter = m->private;
3547 	int i = (int)*pos;
3548 	void *ent;
3549 
3550 	WARN_ON_ONCE(iter->leftover);
3551 
3552 	(*pos)++;
3553 
3554 	/* can't go backwards */
3555 	if (iter->idx > i)
3556 		return NULL;
3557 
3558 	if (iter->idx < 0)
3559 		ent = trace_find_next_entry_inc(iter);
3560 	else
3561 		ent = iter;
3562 
3563 	while (ent && iter->idx < i)
3564 		ent = trace_find_next_entry_inc(iter);
3565 
3566 	iter->pos = *pos;
3567 
3568 	return ent;
3569 }
3570 
3571 void tracing_iter_reset(struct trace_iterator *iter, int cpu)
3572 {
3573 	struct ring_buffer_event *event;
3574 	struct ring_buffer_iter *buf_iter;
3575 	unsigned long entries = 0;
3576 	u64 ts;
3577 
3578 	per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = 0;
3579 
3580 	buf_iter = trace_buffer_iter(iter, cpu);
3581 	if (!buf_iter)
3582 		return;
3583 
3584 	ring_buffer_iter_reset(buf_iter);
3585 
3586 	/*
3587 	 * We could have the case with the max latency tracers
3588 	 * that a reset never took place on a cpu. This is evident
3589 	 * by the timestamp being before the start of the buffer.
3590 	 */
3591 	while ((event = ring_buffer_iter_peek(buf_iter, &ts))) {
3592 		if (ts >= iter->array_buffer->time_start)
3593 			break;
3594 		entries++;
3595 		ring_buffer_iter_advance(buf_iter);
3596 	}
3597 
3598 	per_cpu_ptr(iter->array_buffer->data, cpu)->skipped_entries = entries;
3599 }
3600 
3601 /*
3602  * The current tracer is copied to avoid a global locking
3603  * all around.
3604  */
3605 static void *s_start(struct seq_file *m, loff_t *pos)
3606 {
3607 	struct trace_iterator *iter = m->private;
3608 	struct trace_array *tr = iter->tr;
3609 	int cpu_file = iter->cpu_file;
3610 	void *p = NULL;
3611 	loff_t l = 0;
3612 	int cpu;
3613 
3614 	/*
3615 	 * copy the tracer to avoid using a global lock all around.
3616 	 * iter->trace is a copy of current_trace, the pointer to the
3617 	 * name may be used instead of a strcmp(), as iter->trace->name
3618 	 * will point to the same string as current_trace->name.
3619 	 */
3620 	mutex_lock(&trace_types_lock);
3621 	if (unlikely(tr->current_trace && iter->trace->name != tr->current_trace->name))
3622 		*iter->trace = *tr->current_trace;
3623 	mutex_unlock(&trace_types_lock);
3624 
3625 #ifdef CONFIG_TRACER_MAX_TRACE
3626 	if (iter->snapshot && iter->trace->use_max_tr)
3627 		return ERR_PTR(-EBUSY);
3628 #endif
3629 
3630 	if (!iter->snapshot)
3631 		atomic_inc(&trace_record_taskinfo_disabled);
3632 
3633 	if (*pos != iter->pos) {
3634 		iter->ent = NULL;
3635 		iter->cpu = 0;
3636 		iter->idx = -1;
3637 
3638 		if (cpu_file == RING_BUFFER_ALL_CPUS) {
3639 			for_each_tracing_cpu(cpu)
3640 				tracing_iter_reset(iter, cpu);
3641 		} else
3642 			tracing_iter_reset(iter, cpu_file);
3643 
3644 		iter->leftover = 0;
3645 		for (p = iter; p && l < *pos; p = s_next(m, p, &l))
3646 			;
3647 
3648 	} else {
3649 		/*
3650 		 * If we overflowed the seq_file before, then we want
3651 		 * to just reuse the trace_seq buffer again.
3652 		 */
3653 		if (iter->leftover)
3654 			p = iter;
3655 		else {
3656 			l = *pos - 1;
3657 			p = s_next(m, p, &l);
3658 		}
3659 	}
3660 
3661 	trace_event_read_lock();
3662 	trace_access_lock(cpu_file);
3663 	return p;
3664 }
3665 
3666 static void s_stop(struct seq_file *m, void *p)
3667 {
3668 	struct trace_iterator *iter = m->private;
3669 
3670 #ifdef CONFIG_TRACER_MAX_TRACE
3671 	if (iter->snapshot && iter->trace->use_max_tr)
3672 		return;
3673 #endif
3674 
3675 	if (!iter->snapshot)
3676 		atomic_dec(&trace_record_taskinfo_disabled);
3677 
3678 	trace_access_unlock(iter->cpu_file);
3679 	trace_event_read_unlock();
3680 }
3681 
3682 static void
3683 get_total_entries_cpu(struct array_buffer *buf, unsigned long *total,
3684 		      unsigned long *entries, int cpu)
3685 {
3686 	unsigned long count;
3687 
3688 	count = ring_buffer_entries_cpu(buf->buffer, cpu);
3689 	/*
3690 	 * If this buffer has skipped entries, then we hold all
3691 	 * entries for the trace and we need to ignore the
3692 	 * ones before the time stamp.
3693 	 */
3694 	if (per_cpu_ptr(buf->data, cpu)->skipped_entries) {
3695 		count -= per_cpu_ptr(buf->data, cpu)->skipped_entries;
3696 		/* total is the same as the entries */
3697 		*total = count;
3698 	} else
3699 		*total = count +
3700 			ring_buffer_overrun_cpu(buf->buffer, cpu);
3701 	*entries = count;
3702 }
3703 
3704 static void
3705 get_total_entries(struct array_buffer *buf,
3706 		  unsigned long *total, unsigned long *entries)
3707 {
3708 	unsigned long t, e;
3709 	int cpu;
3710 
3711 	*total = 0;
3712 	*entries = 0;
3713 
3714 	for_each_tracing_cpu(cpu) {
3715 		get_total_entries_cpu(buf, &t, &e, cpu);
3716 		*total += t;
3717 		*entries += e;
3718 	}
3719 }
3720 
3721 unsigned long trace_total_entries_cpu(struct trace_array *tr, int cpu)
3722 {
3723 	unsigned long total, entries;
3724 
3725 	if (!tr)
3726 		tr = &global_trace;
3727 
3728 	get_total_entries_cpu(&tr->array_buffer, &total, &entries, cpu);
3729 
3730 	return entries;
3731 }
3732 
3733 unsigned long trace_total_entries(struct trace_array *tr)
3734 {
3735 	unsigned long total, entries;
3736 
3737 	if (!tr)
3738 		tr = &global_trace;
3739 
3740 	get_total_entries(&tr->array_buffer, &total, &entries);
3741 
3742 	return entries;
3743 }
3744 
3745 static void print_lat_help_header(struct seq_file *m)
3746 {
3747 	seq_puts(m, "#                  _------=> CPU#            \n"
3748 		    "#                 / _-----=> irqs-off        \n"
3749 		    "#                | / _----=> need-resched    \n"
3750 		    "#                || / _---=> hardirq/softirq \n"
3751 		    "#                ||| / _--=> preempt-depth   \n"
3752 		    "#                |||| /     delay            \n"
3753 		    "#  cmd     pid   ||||| time  |   caller      \n"
3754 		    "#     \\   /      |||||  \\    |   /         \n");
3755 }
3756 
3757 static void print_event_info(struct array_buffer *buf, struct seq_file *m)
3758 {
3759 	unsigned long total;
3760 	unsigned long entries;
3761 
3762 	get_total_entries(buf, &total, &entries);
3763 	seq_printf(m, "# entries-in-buffer/entries-written: %lu/%lu   #P:%d\n",
3764 		   entries, total, num_online_cpus());
3765 	seq_puts(m, "#\n");
3766 }
3767 
3768 static void print_func_help_header(struct array_buffer *buf, struct seq_file *m,
3769 				   unsigned int flags)
3770 {
3771 	bool tgid = flags & TRACE_ITER_RECORD_TGID;
3772 
3773 	print_event_info(buf, m);
3774 
3775 	seq_printf(m, "#           TASK-PID   %s  CPU#   TIMESTAMP  FUNCTION\n", tgid ? "TGID     " : "");
3776 	seq_printf(m, "#              | |     %s    |       |         |\n",	 tgid ? "  |      " : "");
3777 }
3778 
3779 static void print_func_help_header_irq(struct array_buffer *buf, struct seq_file *m,
3780 				       unsigned int flags)
3781 {
3782 	bool tgid = flags & TRACE_ITER_RECORD_TGID;
3783 	const char *space = "          ";
3784 	int prec = tgid ? 10 : 2;
3785 
3786 	print_event_info(buf, m);
3787 
3788 	seq_printf(m, "#                          %.*s  _-----=> irqs-off\n", prec, space);
3789 	seq_printf(m, "#                          %.*s / _----=> need-resched\n", prec, space);
3790 	seq_printf(m, "#                          %.*s| / _---=> hardirq/softirq\n", prec, space);
3791 	seq_printf(m, "#                          %.*s|| / _--=> preempt-depth\n", prec, space);
3792 	seq_printf(m, "#                          %.*s||| /     delay\n", prec, space);
3793 	seq_printf(m, "#           TASK-PID %.*sCPU#  ||||    TIMESTAMP  FUNCTION\n", prec, "   TGID   ");
3794 	seq_printf(m, "#              | |   %.*s  |   ||||       |         |\n", prec, "     |    ");
3795 }
3796 
3797 void
3798 print_trace_header(struct seq_file *m, struct trace_iterator *iter)
3799 {
3800 	unsigned long sym_flags = (global_trace.trace_flags & TRACE_ITER_SYM_MASK);
3801 	struct array_buffer *buf = iter->array_buffer;
3802 	struct trace_array_cpu *data = per_cpu_ptr(buf->data, buf->cpu);
3803 	struct tracer *type = iter->trace;
3804 	unsigned long entries;
3805 	unsigned long total;
3806 	const char *name = "preemption";
3807 
3808 	name = type->name;
3809 
3810 	get_total_entries(buf, &total, &entries);
3811 
3812 	seq_printf(m, "# %s latency trace v1.1.5 on %s\n",
3813 		   name, UTS_RELEASE);
3814 	seq_puts(m, "# -----------------------------------"
3815 		 "---------------------------------\n");
3816 	seq_printf(m, "# latency: %lu us, #%lu/%lu, CPU#%d |"
3817 		   " (M:%s VP:%d, KP:%d, SP:%d HP:%d",
3818 		   nsecs_to_usecs(data->saved_latency),
3819 		   entries,
3820 		   total,
3821 		   buf->cpu,
3822 #if defined(CONFIG_PREEMPT_NONE)
3823 		   "server",
3824 #elif defined(CONFIG_PREEMPT_VOLUNTARY)
3825 		   "desktop",
3826 #elif defined(CONFIG_PREEMPT)
3827 		   "preempt",
3828 #elif defined(CONFIG_PREEMPT_RT)
3829 		   "preempt_rt",
3830 #else
3831 		   "unknown",
3832 #endif
3833 		   /* These are reserved for later use */
3834 		   0, 0, 0, 0);
3835 #ifdef CONFIG_SMP
3836 	seq_printf(m, " #P:%d)\n", num_online_cpus());
3837 #else
3838 	seq_puts(m, ")\n");
3839 #endif
3840 	seq_puts(m, "#    -----------------\n");
3841 	seq_printf(m, "#    | task: %.16s-%d "
3842 		   "(uid:%d nice:%ld policy:%ld rt_prio:%ld)\n",
3843 		   data->comm, data->pid,
3844 		   from_kuid_munged(seq_user_ns(m), data->uid), data->nice,
3845 		   data->policy, data->rt_priority);
3846 	seq_puts(m, "#    -----------------\n");
3847 
3848 	if (data->critical_start) {
3849 		seq_puts(m, "#  => started at: ");
3850 		seq_print_ip_sym(&iter->seq, data->critical_start, sym_flags);
3851 		trace_print_seq(m, &iter->seq);
3852 		seq_puts(m, "\n#  => ended at:   ");
3853 		seq_print_ip_sym(&iter->seq, data->critical_end, sym_flags);
3854 		trace_print_seq(m, &iter->seq);
3855 		seq_puts(m, "\n#\n");
3856 	}
3857 
3858 	seq_puts(m, "#\n");
3859 }
3860 
3861 static void test_cpu_buff_start(struct trace_iterator *iter)
3862 {
3863 	struct trace_seq *s = &iter->seq;
3864 	struct trace_array *tr = iter->tr;
3865 
3866 	if (!(tr->trace_flags & TRACE_ITER_ANNOTATE))
3867 		return;
3868 
3869 	if (!(iter->iter_flags & TRACE_FILE_ANNOTATE))
3870 		return;
3871 
3872 	if (cpumask_available(iter->started) &&
3873 	    cpumask_test_cpu(iter->cpu, iter->started))
3874 		return;
3875 
3876 	if (per_cpu_ptr(iter->array_buffer->data, iter->cpu)->skipped_entries)
3877 		return;
3878 
3879 	if (cpumask_available(iter->started))
3880 		cpumask_set_cpu(iter->cpu, iter->started);
3881 
3882 	/* Don't print started cpu buffer for the first entry of the trace */
3883 	if (iter->idx > 1)
3884 		trace_seq_printf(s, "##### CPU %u buffer started ####\n",
3885 				iter->cpu);
3886 }
3887 
3888 static enum print_line_t print_trace_fmt(struct trace_iterator *iter)
3889 {
3890 	struct trace_array *tr = iter->tr;
3891 	struct trace_seq *s = &iter->seq;
3892 	unsigned long sym_flags = (tr->trace_flags & TRACE_ITER_SYM_MASK);
3893 	struct trace_entry *entry;
3894 	struct trace_event *event;
3895 
3896 	entry = iter->ent;
3897 
3898 	test_cpu_buff_start(iter);
3899 
3900 	event = ftrace_find_event(entry->type);
3901 
3902 	if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) {
3903 		if (iter->iter_flags & TRACE_FILE_LAT_FMT)
3904 			trace_print_lat_context(iter);
3905 		else
3906 			trace_print_context(iter);
3907 	}
3908 
3909 	if (trace_seq_has_overflowed(s))
3910 		return TRACE_TYPE_PARTIAL_LINE;
3911 
3912 	if (event)
3913 		return event->funcs->trace(iter, sym_flags, event);
3914 
3915 	trace_seq_printf(s, "Unknown type %d\n", entry->type);
3916 
3917 	return trace_handle_return(s);
3918 }
3919 
3920 static enum print_line_t print_raw_fmt(struct trace_iterator *iter)
3921 {
3922 	struct trace_array *tr = iter->tr;
3923 	struct trace_seq *s = &iter->seq;
3924 	struct trace_entry *entry;
3925 	struct trace_event *event;
3926 
3927 	entry = iter->ent;
3928 
3929 	if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO)
3930 		trace_seq_printf(s, "%d %d %llu ",
3931 				 entry->pid, iter->cpu, iter->ts);
3932 
3933 	if (trace_seq_has_overflowed(s))
3934 		return TRACE_TYPE_PARTIAL_LINE;
3935 
3936 	event = ftrace_find_event(entry->type);
3937 	if (event)
3938 		return event->funcs->raw(iter, 0, event);
3939 
3940 	trace_seq_printf(s, "%d ?\n", entry->type);
3941 
3942 	return trace_handle_return(s);
3943 }
3944 
3945 static enum print_line_t print_hex_fmt(struct trace_iterator *iter)
3946 {
3947 	struct trace_array *tr = iter->tr;
3948 	struct trace_seq *s = &iter->seq;
3949 	unsigned char newline = '\n';
3950 	struct trace_entry *entry;
3951 	struct trace_event *event;
3952 
3953 	entry = iter->ent;
3954 
3955 	if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) {
3956 		SEQ_PUT_HEX_FIELD(s, entry->pid);
3957 		SEQ_PUT_HEX_FIELD(s, iter->cpu);
3958 		SEQ_PUT_HEX_FIELD(s, iter->ts);
3959 		if (trace_seq_has_overflowed(s))
3960 			return TRACE_TYPE_PARTIAL_LINE;
3961 	}
3962 
3963 	event = ftrace_find_event(entry->type);
3964 	if (event) {
3965 		enum print_line_t ret = event->funcs->hex(iter, 0, event);
3966 		if (ret != TRACE_TYPE_HANDLED)
3967 			return ret;
3968 	}
3969 
3970 	SEQ_PUT_FIELD(s, newline);
3971 
3972 	return trace_handle_return(s);
3973 }
3974 
3975 static enum print_line_t print_bin_fmt(struct trace_iterator *iter)
3976 {
3977 	struct trace_array *tr = iter->tr;
3978 	struct trace_seq *s = &iter->seq;
3979 	struct trace_entry *entry;
3980 	struct trace_event *event;
3981 
3982 	entry = iter->ent;
3983 
3984 	if (tr->trace_flags & TRACE_ITER_CONTEXT_INFO) {
3985 		SEQ_PUT_FIELD(s, entry->pid);
3986 		SEQ_PUT_FIELD(s, iter->cpu);
3987 		SEQ_PUT_FIELD(s, iter->ts);
3988 		if (trace_seq_has_overflowed(s))
3989 			return TRACE_TYPE_PARTIAL_LINE;
3990 	}
3991 
3992 	event = ftrace_find_event(entry->type);
3993 	return event ? event->funcs->binary(iter, 0, event) :
3994 		TRACE_TYPE_HANDLED;
3995 }
3996 
3997 int trace_empty(struct trace_iterator *iter)
3998 {
3999 	struct ring_buffer_iter *buf_iter;
4000 	int cpu;
4001 
4002 	/* If we are looking at one CPU buffer, only check that one */
4003 	if (iter->cpu_file != RING_BUFFER_ALL_CPUS) {
4004 		cpu = iter->cpu_file;
4005 		buf_iter = trace_buffer_iter(iter, cpu);
4006 		if (buf_iter) {
4007 			if (!ring_buffer_iter_empty(buf_iter))
4008 				return 0;
4009 		} else {
4010 			if (!ring_buffer_empty_cpu(iter->array_buffer->buffer, cpu))
4011 				return 0;
4012 		}
4013 		return 1;
4014 	}
4015 
4016 	for_each_tracing_cpu(cpu) {
4017 		buf_iter = trace_buffer_iter(iter, cpu);
4018 		if (buf_iter) {
4019 			if (!ring_buffer_iter_empty(buf_iter))
4020 				return 0;
4021 		} else {
4022 			if (!ring_buffer_empty_cpu(iter->array_buffer->buffer, cpu))
4023 				return 0;
4024 		}
4025 	}
4026 
4027 	return 1;
4028 }
4029 
4030 /*  Called with trace_event_read_lock() held. */
4031 enum print_line_t print_trace_line(struct trace_iterator *iter)
4032 {
4033 	struct trace_array *tr = iter->tr;
4034 	unsigned long trace_flags = tr->trace_flags;
4035 	enum print_line_t ret;
4036 
4037 	if (iter->lost_events) {
4038 		if (iter->lost_events == (unsigned long)-1)
4039 			trace_seq_printf(&iter->seq, "CPU:%d [LOST EVENTS]\n",
4040 					 iter->cpu);
4041 		else
4042 			trace_seq_printf(&iter->seq, "CPU:%d [LOST %lu EVENTS]\n",
4043 					 iter->cpu, iter->lost_events);
4044 		if (trace_seq_has_overflowed(&iter->seq))
4045 			return TRACE_TYPE_PARTIAL_LINE;
4046 	}
4047 
4048 	if (iter->trace && iter->trace->print_line) {
4049 		ret = iter->trace->print_line(iter);
4050 		if (ret != TRACE_TYPE_UNHANDLED)
4051 			return ret;
4052 	}
4053 
4054 	if (iter->ent->type == TRACE_BPUTS &&
4055 			trace_flags & TRACE_ITER_PRINTK &&
4056 			trace_flags & TRACE_ITER_PRINTK_MSGONLY)
4057 		return trace_print_bputs_msg_only(iter);
4058 
4059 	if (iter->ent->type == TRACE_BPRINT &&
4060 			trace_flags & TRACE_ITER_PRINTK &&
4061 			trace_flags & TRACE_ITER_PRINTK_MSGONLY)
4062 		return trace_print_bprintk_msg_only(iter);
4063 
4064 	if (iter->ent->type == TRACE_PRINT &&
4065 			trace_flags & TRACE_ITER_PRINTK &&
4066 			trace_flags & TRACE_ITER_PRINTK_MSGONLY)
4067 		return trace_print_printk_msg_only(iter);
4068 
4069 	if (trace_flags & TRACE_ITER_BIN)
4070 		return print_bin_fmt(iter);
4071 
4072 	if (trace_flags & TRACE_ITER_HEX)
4073 		return print_hex_fmt(iter);
4074 
4075 	if (trace_flags & TRACE_ITER_RAW)
4076 		return print_raw_fmt(iter);
4077 
4078 	return print_trace_fmt(iter);
4079 }
4080 
4081 void trace_latency_header(struct seq_file *m)
4082 {
4083 	struct trace_iterator *iter = m->private;
4084 	struct trace_array *tr = iter->tr;
4085 
4086 	/* print nothing if the buffers are empty */
4087 	if (trace_empty(iter))
4088 		return;
4089 
4090 	if (iter->iter_flags & TRACE_FILE_LAT_FMT)
4091 		print_trace_header(m, iter);
4092 
4093 	if (!(tr->trace_flags & TRACE_ITER_VERBOSE))
4094 		print_lat_help_header(m);
4095 }
4096 
4097 void trace_default_header(struct seq_file *m)
4098 {
4099 	struct trace_iterator *iter = m->private;
4100 	struct trace_array *tr = iter->tr;
4101 	unsigned long trace_flags = tr->trace_flags;
4102 
4103 	if (!(trace_flags & TRACE_ITER_CONTEXT_INFO))
4104 		return;
4105 
4106 	if (iter->iter_flags & TRACE_FILE_LAT_FMT) {
4107 		/* print nothing if the buffers are empty */
4108 		if (trace_empty(iter))
4109 			return;
4110 		print_trace_header(m, iter);
4111 		if (!(trace_flags & TRACE_ITER_VERBOSE))
4112 			print_lat_help_header(m);
4113 	} else {
4114 		if (!(trace_flags & TRACE_ITER_VERBOSE)) {
4115 			if (trace_flags & TRACE_ITER_IRQ_INFO)
4116 				print_func_help_header_irq(iter->array_buffer,
4117 							   m, trace_flags);
4118 			else
4119 				print_func_help_header(iter->array_buffer, m,
4120 						       trace_flags);
4121 		}
4122 	}
4123 }
4124 
4125 static void test_ftrace_alive(struct seq_file *m)
4126 {
4127 	if (!ftrace_is_dead())
4128 		return;
4129 	seq_puts(m, "# WARNING: FUNCTION TRACING IS CORRUPTED\n"
4130 		    "#          MAY BE MISSING FUNCTION EVENTS\n");
4131 }
4132 
4133 #ifdef CONFIG_TRACER_MAX_TRACE
4134 static void show_snapshot_main_help(struct seq_file *m)
4135 {
4136 	seq_puts(m, "# echo 0 > snapshot : Clears and frees snapshot buffer\n"
4137 		    "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n"
4138 		    "#                      Takes a snapshot of the main buffer.\n"
4139 		    "# echo 2 > snapshot : Clears snapshot buffer (but does not allocate or free)\n"
4140 		    "#                      (Doesn't have to be '2' works with any number that\n"
4141 		    "#                       is not a '0' or '1')\n");
4142 }
4143 
4144 static void show_snapshot_percpu_help(struct seq_file *m)
4145 {
4146 	seq_puts(m, "# echo 0 > snapshot : Invalid for per_cpu snapshot file.\n");
4147 #ifdef CONFIG_RING_BUFFER_ALLOW_SWAP
4148 	seq_puts(m, "# echo 1 > snapshot : Allocates snapshot buffer, if not already allocated.\n"
4149 		    "#                      Takes a snapshot of the main buffer for this cpu.\n");
4150 #else
4151 	seq_puts(m, "# echo 1 > snapshot : Not supported with this kernel.\n"
4152 		    "#                     Must use main snapshot file to allocate.\n");
4153 #endif
4154 	seq_puts(m, "# echo 2 > snapshot : Clears this cpu's snapshot buffer (but does not allocate)\n"
4155 		    "#                      (Doesn't have to be '2' works with any number that\n"
4156 		    "#                       is not a '0' or '1')\n");
4157 }
4158 
4159 static void print_snapshot_help(struct seq_file *m, struct trace_iterator *iter)
4160 {
4161 	if (iter->tr->allocated_snapshot)
4162 		seq_puts(m, "#\n# * Snapshot is allocated *\n#\n");
4163 	else
4164 		seq_puts(m, "#\n# * Snapshot is freed *\n#\n");
4165 
4166 	seq_puts(m, "# Snapshot commands:\n");
4167 	if (iter->cpu_file == RING_BUFFER_ALL_CPUS)
4168 		show_snapshot_main_help(m);
4169 	else
4170 		show_snapshot_percpu_help(m);
4171 }
4172 #else
4173 /* Should never be called */
4174 static inline void print_snapshot_help(struct seq_file *m, struct trace_iterator *iter) { }
4175 #endif
4176 
4177 static int s_show(struct seq_file *m, void *v)
4178 {
4179 	struct trace_iterator *iter = v;
4180 	int ret;
4181 
4182 	if (iter->ent == NULL) {
4183 		if (iter->tr) {
4184 			seq_printf(m, "# tracer: %s\n", iter->trace->name);
4185 			seq_puts(m, "#\n");
4186 			test_ftrace_alive(m);
4187 		}
4188 		if (iter->snapshot && trace_empty(iter))
4189 			print_snapshot_help(m, iter);
4190 		else if (iter->trace && iter->trace->print_header)
4191 			iter->trace->print_header(m);
4192 		else
4193 			trace_default_header(m);
4194 
4195 	} else if (iter->leftover) {
4196 		/*
4197 		 * If we filled the seq_file buffer earlier, we
4198 		 * want to just show it now.
4199 		 */
4200 		ret = trace_print_seq(m, &iter->seq);
4201 
4202 		/* ret should this time be zero, but you never know */
4203 		iter->leftover = ret;
4204 
4205 	} else {
4206 		print_trace_line(iter);
4207 		ret = trace_print_seq(m, &iter->seq);
4208 		/*
4209 		 * If we overflow the seq_file buffer, then it will
4210 		 * ask us for this data again at start up.
4211 		 * Use that instead.
4212 		 *  ret is 0 if seq_file write succeeded.
4213 		 *        -1 otherwise.
4214 		 */
4215 		iter->leftover = ret;
4216 	}
4217 
4218 	return 0;
4219 }
4220 
4221 /*
4222  * Should be used after trace_array_get(), trace_types_lock
4223  * ensures that i_cdev was already initialized.
4224  */
4225 static inline int tracing_get_cpu(struct inode *inode)
4226 {
4227 	if (inode->i_cdev) /* See trace_create_cpu_file() */
4228 		return (long)inode->i_cdev - 1;
4229 	return RING_BUFFER_ALL_CPUS;
4230 }
4231 
4232 static const struct seq_operations tracer_seq_ops = {
4233 	.start		= s_start,
4234 	.next		= s_next,
4235 	.stop		= s_stop,
4236 	.show		= s_show,
4237 };
4238 
4239 static struct trace_iterator *
4240 __tracing_open(struct inode *inode, struct file *file, bool snapshot)
4241 {
4242 	struct trace_array *tr = inode->i_private;
4243 	struct trace_iterator *iter;
4244 	int cpu;
4245 
4246 	if (tracing_disabled)
4247 		return ERR_PTR(-ENODEV);
4248 
4249 	iter = __seq_open_private(file, &tracer_seq_ops, sizeof(*iter));
4250 	if (!iter)
4251 		return ERR_PTR(-ENOMEM);
4252 
4253 	iter->buffer_iter = kcalloc(nr_cpu_ids, sizeof(*iter->buffer_iter),
4254 				    GFP_KERNEL);
4255 	if (!iter->buffer_iter)
4256 		goto release;
4257 
4258 	/*
4259 	 * trace_find_next_entry() may need to save off iter->ent.
4260 	 * It will place it into the iter->temp buffer. As most
4261 	 * events are less than 128, allocate a buffer of that size.
4262 	 * If one is greater, then trace_find_next_entry() will
4263 	 * allocate a new buffer to adjust for the bigger iter->ent.
4264 	 * It's not critical if it fails to get allocated here.
4265 	 */
4266 	iter->temp = kmalloc(128, GFP_KERNEL);
4267 	if (iter->temp)
4268 		iter->temp_size = 128;
4269 
4270 	/*
4271 	 * We make a copy of the current tracer to avoid concurrent
4272 	 * changes on it while we are reading.
4273 	 */
4274 	mutex_lock(&trace_types_lock);
4275 	iter->trace = kzalloc(sizeof(*iter->trace), GFP_KERNEL);
4276 	if (!iter->trace)
4277 		goto fail;
4278 
4279 	*iter->trace = *tr->current_trace;
4280 
4281 	if (!zalloc_cpumask_var(&iter->started, GFP_KERNEL))
4282 		goto fail;
4283 
4284 	iter->tr = tr;
4285 
4286 #ifdef CONFIG_TRACER_MAX_TRACE
4287 	/* Currently only the top directory has a snapshot */
4288 	if (tr->current_trace->print_max || snapshot)
4289 		iter->array_buffer = &tr->max_buffer;
4290 	else
4291 #endif
4292 		iter->array_buffer = &tr->array_buffer;
4293 	iter->snapshot = snapshot;
4294 	iter->pos = -1;
4295 	iter->cpu_file = tracing_get_cpu(inode);
4296 	mutex_init(&iter->mutex);
4297 
4298 	/* Notify the tracer early; before we stop tracing. */
4299 	if (iter->trace->open)
4300 		iter->trace->open(iter);
4301 
4302 	/* Annotate start of buffers if we had overruns */
4303 	if (ring_buffer_overruns(iter->array_buffer->buffer))
4304 		iter->iter_flags |= TRACE_FILE_ANNOTATE;
4305 
4306 	/* Output in nanoseconds only if we are using a clock in nanoseconds. */
4307 	if (trace_clocks[tr->clock_id].in_ns)
4308 		iter->iter_flags |= TRACE_FILE_TIME_IN_NS;
4309 
4310 	/*
4311 	 * If pause-on-trace is enabled, then stop the trace while
4312 	 * dumping, unless this is the "snapshot" file
4313 	 */
4314 	if (!iter->snapshot && (tr->trace_flags & TRACE_ITER_PAUSE_ON_TRACE))
4315 		tracing_stop_tr(tr);
4316 
4317 	if (iter->cpu_file == RING_BUFFER_ALL_CPUS) {
4318 		for_each_tracing_cpu(cpu) {
4319 			iter->buffer_iter[cpu] =
4320 				ring_buffer_read_prepare(iter->array_buffer->buffer,
4321 							 cpu, GFP_KERNEL);
4322 		}
4323 		ring_buffer_read_prepare_sync();
4324 		for_each_tracing_cpu(cpu) {
4325 			ring_buffer_read_start(iter->buffer_iter[cpu]);
4326 			tracing_iter_reset(iter, cpu);
4327 		}
4328 	} else {
4329 		cpu = iter->cpu_file;
4330 		iter->buffer_iter[cpu] =
4331 			ring_buffer_read_prepare(iter->array_buffer->buffer,
4332 						 cpu, GFP_KERNEL);
4333 		ring_buffer_read_prepare_sync();
4334 		ring_buffer_read_start(iter->buffer_iter[cpu]);
4335 		tracing_iter_reset(iter, cpu);
4336 	}
4337 
4338 	mutex_unlock(&trace_types_lock);
4339 
4340 	return iter;
4341 
4342  fail:
4343 	mutex_unlock(&trace_types_lock);
4344 	kfree(iter->trace);
4345 	kfree(iter->temp);
4346 	kfree(iter->buffer_iter);
4347 release:
4348 	seq_release_private(inode, file);
4349 	return ERR_PTR(-ENOMEM);
4350 }
4351 
4352 int tracing_open_generic(struct inode *inode, struct file *filp)
4353 {
4354 	int ret;
4355 
4356 	ret = tracing_check_open_get_tr(NULL);
4357 	if (ret)
4358 		return ret;
4359 
4360 	filp->private_data = inode->i_private;
4361 	return 0;
4362 }
4363 
4364 bool tracing_is_disabled(void)
4365 {
4366 	return (tracing_disabled) ? true: false;
4367 }
4368 
4369 /*
4370  * Open and update trace_array ref count.
4371  * Must have the current trace_array passed to it.
4372  */
4373 int tracing_open_generic_tr(struct inode *inode, struct file *filp)
4374 {
4375 	struct trace_array *tr = inode->i_private;
4376 	int ret;
4377 
4378 	ret = tracing_check_open_get_tr(tr);
4379 	if (ret)
4380 		return ret;
4381 
4382 	filp->private_data = inode->i_private;
4383 
4384 	return 0;
4385 }
4386 
4387 static int tracing_release(struct inode *inode, struct file *file)
4388 {
4389 	struct trace_array *tr = inode->i_private;
4390 	struct seq_file *m = file->private_data;
4391 	struct trace_iterator *iter;
4392 	int cpu;
4393 
4394 	if (!(file->f_mode & FMODE_READ)) {
4395 		trace_array_put(tr);
4396 		return 0;
4397 	}
4398 
4399 	/* Writes do not use seq_file */
4400 	iter = m->private;
4401 	mutex_lock(&trace_types_lock);
4402 
4403 	for_each_tracing_cpu(cpu) {
4404 		if (iter->buffer_iter[cpu])
4405 			ring_buffer_read_finish(iter->buffer_iter[cpu]);
4406 	}
4407 
4408 	if (iter->trace && iter->trace->close)
4409 		iter->trace->close(iter);
4410 
4411 	if (!iter->snapshot && tr->stop_count)
4412 		/* reenable tracing if it was previously enabled */
4413 		tracing_start_tr(tr);
4414 
4415 	__trace_array_put(tr);
4416 
4417 	mutex_unlock(&trace_types_lock);
4418 
4419 	mutex_destroy(&iter->mutex);
4420 	free_cpumask_var(iter->started);
4421 	kfree(iter->temp);
4422 	kfree(iter->trace);
4423 	kfree(iter->buffer_iter);
4424 	seq_release_private(inode, file);
4425 
4426 	return 0;
4427 }
4428 
4429 static int tracing_release_generic_tr(struct inode *inode, struct file *file)
4430 {
4431 	struct trace_array *tr = inode->i_private;
4432 
4433 	trace_array_put(tr);
4434 	return 0;
4435 }
4436 
4437 static int tracing_single_release_tr(struct inode *inode, struct file *file)
4438 {
4439 	struct trace_array *tr = inode->i_private;
4440 
4441 	trace_array_put(tr);
4442 
4443 	return single_release(inode, file);
4444 }
4445 
4446 static int tracing_open(struct inode *inode, struct file *file)
4447 {
4448 	struct trace_array *tr = inode->i_private;
4449 	struct trace_iterator *iter;
4450 	int ret;
4451 
4452 	ret = tracing_check_open_get_tr(tr);
4453 	if (ret)
4454 		return ret;
4455 
4456 	/* If this file was open for write, then erase contents */
4457 	if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) {
4458 		int cpu = tracing_get_cpu(inode);
4459 		struct array_buffer *trace_buf = &tr->array_buffer;
4460 
4461 #ifdef CONFIG_TRACER_MAX_TRACE
4462 		if (tr->current_trace->print_max)
4463 			trace_buf = &tr->max_buffer;
4464 #endif
4465 
4466 		if (cpu == RING_BUFFER_ALL_CPUS)
4467 			tracing_reset_online_cpus(trace_buf);
4468 		else
4469 			tracing_reset_cpu(trace_buf, cpu);
4470 	}
4471 
4472 	if (file->f_mode & FMODE_READ) {
4473 		iter = __tracing_open(inode, file, false);
4474 		if (IS_ERR(iter))
4475 			ret = PTR_ERR(iter);
4476 		else if (tr->trace_flags & TRACE_ITER_LATENCY_FMT)
4477 			iter->iter_flags |= TRACE_FILE_LAT_FMT;
4478 	}
4479 
4480 	if (ret < 0)
4481 		trace_array_put(tr);
4482 
4483 	return ret;
4484 }
4485 
4486 /*
4487  * Some tracers are not suitable for instance buffers.
4488  * A tracer is always available for the global array (toplevel)
4489  * or if it explicitly states that it is.
4490  */
4491 static bool
4492 trace_ok_for_array(struct tracer *t, struct trace_array *tr)
4493 {
4494 	return (tr->flags & TRACE_ARRAY_FL_GLOBAL) || t->allow_instances;
4495 }
4496 
4497 /* Find the next tracer that this trace array may use */
4498 static struct tracer *
4499 get_tracer_for_array(struct trace_array *tr, struct tracer *t)
4500 {
4501 	while (t && !trace_ok_for_array(t, tr))
4502 		t = t->next;
4503 
4504 	return t;
4505 }
4506 
4507 static void *
4508 t_next(struct seq_file *m, void *v, loff_t *pos)
4509 {
4510 	struct trace_array *tr = m->private;
4511 	struct tracer *t = v;
4512 
4513 	(*pos)++;
4514 
4515 	if (t)
4516 		t = get_tracer_for_array(tr, t->next);
4517 
4518 	return t;
4519 }
4520 
4521 static void *t_start(struct seq_file *m, loff_t *pos)
4522 {
4523 	struct trace_array *tr = m->private;
4524 	struct tracer *t;
4525 	loff_t l = 0;
4526 
4527 	mutex_lock(&trace_types_lock);
4528 
4529 	t = get_tracer_for_array(tr, trace_types);
4530 	for (; t && l < *pos; t = t_next(m, t, &l))
4531 			;
4532 
4533 	return t;
4534 }
4535 
4536 static void t_stop(struct seq_file *m, void *p)
4537 {
4538 	mutex_unlock(&trace_types_lock);
4539 }
4540 
4541 static int t_show(struct seq_file *m, void *v)
4542 {
4543 	struct tracer *t = v;
4544 
4545 	if (!t)
4546 		return 0;
4547 
4548 	seq_puts(m, t->name);
4549 	if (t->next)
4550 		seq_putc(m, ' ');
4551 	else
4552 		seq_putc(m, '\n');
4553 
4554 	return 0;
4555 }
4556 
4557 static const struct seq_operations show_traces_seq_ops = {
4558 	.start		= t_start,
4559 	.next		= t_next,
4560 	.stop		= t_stop,
4561 	.show		= t_show,
4562 };
4563 
4564 static int show_traces_open(struct inode *inode, struct file *file)
4565 {
4566 	struct trace_array *tr = inode->i_private;
4567 	struct seq_file *m;
4568 	int ret;
4569 
4570 	ret = tracing_check_open_get_tr(tr);
4571 	if (ret)
4572 		return ret;
4573 
4574 	ret = seq_open(file, &show_traces_seq_ops);
4575 	if (ret) {
4576 		trace_array_put(tr);
4577 		return ret;
4578 	}
4579 
4580 	m = file->private_data;
4581 	m->private = tr;
4582 
4583 	return 0;
4584 }
4585 
4586 static int show_traces_release(struct inode *inode, struct file *file)
4587 {
4588 	struct trace_array *tr = inode->i_private;
4589 
4590 	trace_array_put(tr);
4591 	return seq_release(inode, file);
4592 }
4593 
4594 static ssize_t
4595 tracing_write_stub(struct file *filp, const char __user *ubuf,
4596 		   size_t count, loff_t *ppos)
4597 {
4598 	return count;
4599 }
4600 
4601 loff_t tracing_lseek(struct file *file, loff_t offset, int whence)
4602 {
4603 	int ret;
4604 
4605 	if (file->f_mode & FMODE_READ)
4606 		ret = seq_lseek(file, offset, whence);
4607 	else
4608 		file->f_pos = ret = 0;
4609 
4610 	return ret;
4611 }
4612 
4613 static const struct file_operations tracing_fops = {
4614 	.open		= tracing_open,
4615 	.read		= seq_read,
4616 	.write		= tracing_write_stub,
4617 	.llseek		= tracing_lseek,
4618 	.release	= tracing_release,
4619 };
4620 
4621 static const struct file_operations show_traces_fops = {
4622 	.open		= show_traces_open,
4623 	.read		= seq_read,
4624 	.llseek		= seq_lseek,
4625 	.release	= show_traces_release,
4626 };
4627 
4628 static ssize_t
4629 tracing_cpumask_read(struct file *filp, char __user *ubuf,
4630 		     size_t count, loff_t *ppos)
4631 {
4632 	struct trace_array *tr = file_inode(filp)->i_private;
4633 	char *mask_str;
4634 	int len;
4635 
4636 	len = snprintf(NULL, 0, "%*pb\n",
4637 		       cpumask_pr_args(tr->tracing_cpumask)) + 1;
4638 	mask_str = kmalloc(len, GFP_KERNEL);
4639 	if (!mask_str)
4640 		return -ENOMEM;
4641 
4642 	len = snprintf(mask_str, len, "%*pb\n",
4643 		       cpumask_pr_args(tr->tracing_cpumask));
4644 	if (len >= count) {
4645 		count = -EINVAL;
4646 		goto out_err;
4647 	}
4648 	count = simple_read_from_buffer(ubuf, count, ppos, mask_str, len);
4649 
4650 out_err:
4651 	kfree(mask_str);
4652 
4653 	return count;
4654 }
4655 
4656 int tracing_set_cpumask(struct trace_array *tr,
4657 			cpumask_var_t tracing_cpumask_new)
4658 {
4659 	int cpu;
4660 
4661 	if (!tr)
4662 		return -EINVAL;
4663 
4664 	local_irq_disable();
4665 	arch_spin_lock(&tr->max_lock);
4666 	for_each_tracing_cpu(cpu) {
4667 		/*
4668 		 * Increase/decrease the disabled counter if we are
4669 		 * about to flip a bit in the cpumask:
4670 		 */
4671 		if (cpumask_test_cpu(cpu, tr->tracing_cpumask) &&
4672 				!cpumask_test_cpu(cpu, tracing_cpumask_new)) {
4673 			atomic_inc(&per_cpu_ptr(tr->array_buffer.data, cpu)->disabled);
4674 			ring_buffer_record_disable_cpu(tr->array_buffer.buffer, cpu);
4675 		}
4676 		if (!cpumask_test_cpu(cpu, tr->tracing_cpumask) &&
4677 				cpumask_test_cpu(cpu, tracing_cpumask_new)) {
4678 			atomic_dec(&per_cpu_ptr(tr->array_buffer.data, cpu)->disabled);
4679 			ring_buffer_record_enable_cpu(tr->array_buffer.buffer, cpu);
4680 		}
4681 	}
4682 	arch_spin_unlock(&tr->max_lock);
4683 	local_irq_enable();
4684 
4685 	cpumask_copy(tr->tracing_cpumask, tracing_cpumask_new);
4686 
4687 	return 0;
4688 }
4689 
4690 static ssize_t
4691 tracing_cpumask_write(struct file *filp, const char __user *ubuf,
4692 		      size_t count, loff_t *ppos)
4693 {
4694 	struct trace_array *tr = file_inode(filp)->i_private;
4695 	cpumask_var_t tracing_cpumask_new;
4696 	int err;
4697 
4698 	if (!alloc_cpumask_var(&tracing_cpumask_new, GFP_KERNEL))
4699 		return -ENOMEM;
4700 
4701 	err = cpumask_parse_user(ubuf, count, tracing_cpumask_new);
4702 	if (err)
4703 		goto err_free;
4704 
4705 	err = tracing_set_cpumask(tr, tracing_cpumask_new);
4706 	if (err)
4707 		goto err_free;
4708 
4709 	free_cpumask_var(tracing_cpumask_new);
4710 
4711 	return count;
4712 
4713 err_free:
4714 	free_cpumask_var(tracing_cpumask_new);
4715 
4716 	return err;
4717 }
4718 
4719 static const struct file_operations tracing_cpumask_fops = {
4720 	.open		= tracing_open_generic_tr,
4721 	.read		= tracing_cpumask_read,
4722 	.write		= tracing_cpumask_write,
4723 	.release	= tracing_release_generic_tr,
4724 	.llseek		= generic_file_llseek,
4725 };
4726 
4727 static int tracing_trace_options_show(struct seq_file *m, void *v)
4728 {
4729 	struct tracer_opt *trace_opts;
4730 	struct trace_array *tr = m->private;
4731 	u32 tracer_flags;
4732 	int i;
4733 
4734 	mutex_lock(&trace_types_lock);
4735 	tracer_flags = tr->current_trace->flags->val;
4736 	trace_opts = tr->current_trace->flags->opts;
4737 
4738 	for (i = 0; trace_options[i]; i++) {
4739 		if (tr->trace_flags & (1 << i))
4740 			seq_printf(m, "%s\n", trace_options[i]);
4741 		else
4742 			seq_printf(m, "no%s\n", trace_options[i]);
4743 	}
4744 
4745 	for (i = 0; trace_opts[i].name; i++) {
4746 		if (tracer_flags & trace_opts[i].bit)
4747 			seq_printf(m, "%s\n", trace_opts[i].name);
4748 		else
4749 			seq_printf(m, "no%s\n", trace_opts[i].name);
4750 	}
4751 	mutex_unlock(&trace_types_lock);
4752 
4753 	return 0;
4754 }
4755 
4756 static int __set_tracer_option(struct trace_array *tr,
4757 			       struct tracer_flags *tracer_flags,
4758 			       struct tracer_opt *opts, int neg)
4759 {
4760 	struct tracer *trace = tracer_flags->trace;
4761 	int ret;
4762 
4763 	ret = trace->set_flag(tr, tracer_flags->val, opts->bit, !neg);
4764 	if (ret)
4765 		return ret;
4766 
4767 	if (neg)
4768 		tracer_flags->val &= ~opts->bit;
4769 	else
4770 		tracer_flags->val |= opts->bit;
4771 	return 0;
4772 }
4773 
4774 /* Try to assign a tracer specific option */
4775 static int set_tracer_option(struct trace_array *tr, char *cmp, int neg)
4776 {
4777 	struct tracer *trace = tr->current_trace;
4778 	struct tracer_flags *tracer_flags = trace->flags;
4779 	struct tracer_opt *opts = NULL;
4780 	int i;
4781 
4782 	for (i = 0; tracer_flags->opts[i].name; i++) {
4783 		opts = &tracer_flags->opts[i];
4784 
4785 		if (strcmp(cmp, opts->name) == 0)
4786 			return __set_tracer_option(tr, trace->flags, opts, neg);
4787 	}
4788 
4789 	return -EINVAL;
4790 }
4791 
4792 /* Some tracers require overwrite to stay enabled */
4793 int trace_keep_overwrite(struct tracer *tracer, u32 mask, int set)
4794 {
4795 	if (tracer->enabled && (mask & TRACE_ITER_OVERWRITE) && !set)
4796 		return -1;
4797 
4798 	return 0;
4799 }
4800 
4801 int set_tracer_flag(struct trace_array *tr, unsigned int mask, int enabled)
4802 {
4803 	if ((mask == TRACE_ITER_RECORD_TGID) ||
4804 	    (mask == TRACE_ITER_RECORD_CMD))
4805 		lockdep_assert_held(&event_mutex);
4806 
4807 	/* do nothing if flag is already set */
4808 	if (!!(tr->trace_flags & mask) == !!enabled)
4809 		return 0;
4810 
4811 	/* Give the tracer a chance to approve the change */
4812 	if (tr->current_trace->flag_changed)
4813 		if (tr->current_trace->flag_changed(tr, mask, !!enabled))
4814 			return -EINVAL;
4815 
4816 	if (enabled)
4817 		tr->trace_flags |= mask;
4818 	else
4819 		tr->trace_flags &= ~mask;
4820 
4821 	if (mask == TRACE_ITER_RECORD_CMD)
4822 		trace_event_enable_cmd_record(enabled);
4823 
4824 	if (mask == TRACE_ITER_RECORD_TGID) {
4825 		if (!tgid_map)
4826 			tgid_map = kvcalloc(PID_MAX_DEFAULT + 1,
4827 					   sizeof(*tgid_map),
4828 					   GFP_KERNEL);
4829 		if (!tgid_map) {
4830 			tr->trace_flags &= ~TRACE_ITER_RECORD_TGID;
4831 			return -ENOMEM;
4832 		}
4833 
4834 		trace_event_enable_tgid_record(enabled);
4835 	}
4836 
4837 	if (mask == TRACE_ITER_EVENT_FORK)
4838 		trace_event_follow_fork(tr, enabled);
4839 
4840 	if (mask == TRACE_ITER_FUNC_FORK)
4841 		ftrace_pid_follow_fork(tr, enabled);
4842 
4843 	if (mask == TRACE_ITER_OVERWRITE) {
4844 		ring_buffer_change_overwrite(tr->array_buffer.buffer, enabled);
4845 #ifdef CONFIG_TRACER_MAX_TRACE
4846 		ring_buffer_change_overwrite(tr->max_buffer.buffer, enabled);
4847 #endif
4848 	}
4849 
4850 	if (mask == TRACE_ITER_PRINTK) {
4851 		trace_printk_start_stop_comm(enabled);
4852 		trace_printk_control(enabled);
4853 	}
4854 
4855 	return 0;
4856 }
4857 
4858 int trace_set_options(struct trace_array *tr, char *option)
4859 {
4860 	char *cmp;
4861 	int neg = 0;
4862 	int ret;
4863 	size_t orig_len = strlen(option);
4864 	int len;
4865 
4866 	cmp = strstrip(option);
4867 
4868 	len = str_has_prefix(cmp, "no");
4869 	if (len)
4870 		neg = 1;
4871 
4872 	cmp += len;
4873 
4874 	mutex_lock(&event_mutex);
4875 	mutex_lock(&trace_types_lock);
4876 
4877 	ret = match_string(trace_options, -1, cmp);
4878 	/* If no option could be set, test the specific tracer options */
4879 	if (ret < 0)
4880 		ret = set_tracer_option(tr, cmp, neg);
4881 	else
4882 		ret = set_tracer_flag(tr, 1 << ret, !neg);
4883 
4884 	mutex_unlock(&trace_types_lock);
4885 	mutex_unlock(&event_mutex);
4886 
4887 	/*
4888 	 * If the first trailing whitespace is replaced with '\0' by strstrip,
4889 	 * turn it back into a space.
4890 	 */
4891 	if (orig_len > strlen(option))
4892 		option[strlen(option)] = ' ';
4893 
4894 	return ret;
4895 }
4896 
4897 static void __init apply_trace_boot_options(void)
4898 {
4899 	char *buf = trace_boot_options_buf;
4900 	char *option;
4901 
4902 	while (true) {
4903 		option = strsep(&buf, ",");
4904 
4905 		if (!option)
4906 			break;
4907 
4908 		if (*option)
4909 			trace_set_options(&global_trace, option);
4910 
4911 		/* Put back the comma to allow this to be called again */
4912 		if (buf)
4913 			*(buf - 1) = ',';
4914 	}
4915 }
4916 
4917 static ssize_t
4918 tracing_trace_options_write(struct file *filp, const char __user *ubuf,
4919 			size_t cnt, loff_t *ppos)
4920 {
4921 	struct seq_file *m = filp->private_data;
4922 	struct trace_array *tr = m->private;
4923 	char buf[64];
4924 	int ret;
4925 
4926 	if (cnt >= sizeof(buf))
4927 		return -EINVAL;
4928 
4929 	if (copy_from_user(buf, ubuf, cnt))
4930 		return -EFAULT;
4931 
4932 	buf[cnt] = 0;
4933 
4934 	ret = trace_set_options(tr, buf);
4935 	if (ret < 0)
4936 		return ret;
4937 
4938 	*ppos += cnt;
4939 
4940 	return cnt;
4941 }
4942 
4943 static int tracing_trace_options_open(struct inode *inode, struct file *file)
4944 {
4945 	struct trace_array *tr = inode->i_private;
4946 	int ret;
4947 
4948 	ret = tracing_check_open_get_tr(tr);
4949 	if (ret)
4950 		return ret;
4951 
4952 	ret = single_open(file, tracing_trace_options_show, inode->i_private);
4953 	if (ret < 0)
4954 		trace_array_put(tr);
4955 
4956 	return ret;
4957 }
4958 
4959 static const struct file_operations tracing_iter_fops = {
4960 	.open		= tracing_trace_options_open,
4961 	.read		= seq_read,
4962 	.llseek		= seq_lseek,
4963 	.release	= tracing_single_release_tr,
4964 	.write		= tracing_trace_options_write,
4965 };
4966 
4967 static const char readme_msg[] =
4968 	"tracing mini-HOWTO:\n\n"
4969 	"# echo 0 > tracing_on : quick way to disable tracing\n"
4970 	"# echo 1 > tracing_on : quick way to re-enable tracing\n\n"
4971 	" Important files:\n"
4972 	"  trace\t\t\t- The static contents of the buffer\n"
4973 	"\t\t\t  To clear the buffer write into this file: echo > trace\n"
4974 	"  trace_pipe\t\t- A consuming read to see the contents of the buffer\n"
4975 	"  current_tracer\t- function and latency tracers\n"
4976 	"  available_tracers\t- list of configured tracers for current_tracer\n"
4977 	"  error_log\t- error log for failed commands (that support it)\n"
4978 	"  buffer_size_kb\t- view and modify size of per cpu buffer\n"
4979 	"  buffer_total_size_kb  - view total size of all cpu buffers\n\n"
4980 	"  trace_clock\t\t-change the clock used to order events\n"
4981 	"       local:   Per cpu clock but may not be synced across CPUs\n"
4982 	"      global:   Synced across CPUs but slows tracing down.\n"
4983 	"     counter:   Not a clock, but just an increment\n"
4984 	"      uptime:   Jiffy counter from time of boot\n"
4985 	"        perf:   Same clock that perf events use\n"
4986 #ifdef CONFIG_X86_64
4987 	"     x86-tsc:   TSC cycle counter\n"
4988 #endif
4989 	"\n  timestamp_mode\t-view the mode used to timestamp events\n"
4990 	"       delta:   Delta difference against a buffer-wide timestamp\n"
4991 	"    absolute:   Absolute (standalone) timestamp\n"
4992 	"\n  trace_marker\t\t- Writes into this file writes into the kernel buffer\n"
4993 	"\n  trace_marker_raw\t\t- Writes into this file writes binary data into the kernel buffer\n"
4994 	"  tracing_cpumask\t- Limit which CPUs to trace\n"
4995 	"  instances\t\t- Make sub-buffers with: mkdir instances/foo\n"
4996 	"\t\t\t  Remove sub-buffer with rmdir\n"
4997 	"  trace_options\t\t- Set format or modify how tracing happens\n"
4998 	"\t\t\t  Disable an option by prefixing 'no' to the\n"
4999 	"\t\t\t  option name\n"
5000 	"  saved_cmdlines_size\t- echo command number in here to store comm-pid list\n"
5001 #ifdef CONFIG_DYNAMIC_FTRACE
5002 	"\n  available_filter_functions - list of functions that can be filtered on\n"
5003 	"  set_ftrace_filter\t- echo function name in here to only trace these\n"
5004 	"\t\t\t  functions\n"
5005 	"\t     accepts: func_full_name or glob-matching-pattern\n"
5006 	"\t     modules: Can select a group via module\n"
5007 	"\t      Format: :mod:<module-name>\n"
5008 	"\t     example: echo :mod:ext3 > set_ftrace_filter\n"
5009 	"\t    triggers: a command to perform when function is hit\n"
5010 	"\t      Format: <function>:<trigger>[:count]\n"
5011 	"\t     trigger: traceon, traceoff\n"
5012 	"\t\t      enable_event:<system>:<event>\n"
5013 	"\t\t      disable_event:<system>:<event>\n"
5014 #ifdef CONFIG_STACKTRACE
5015 	"\t\t      stacktrace\n"
5016 #endif
5017 #ifdef CONFIG_TRACER_SNAPSHOT
5018 	"\t\t      snapshot\n"
5019 #endif
5020 	"\t\t      dump\n"
5021 	"\t\t      cpudump\n"
5022 	"\t     example: echo do_fault:traceoff > set_ftrace_filter\n"
5023 	"\t              echo do_trap:traceoff:3 > set_ftrace_filter\n"
5024 	"\t     The first one will disable tracing every time do_fault is hit\n"
5025 	"\t     The second will disable tracing at most 3 times when do_trap is hit\n"
5026 	"\t       The first time do trap is hit and it disables tracing, the\n"
5027 	"\t       counter will decrement to 2. If tracing is already disabled,\n"
5028 	"\t       the counter will not decrement. It only decrements when the\n"
5029 	"\t       trigger did work\n"
5030 	"\t     To remove trigger without count:\n"
5031 	"\t       echo '!<function>:<trigger> > set_ftrace_filter\n"
5032 	"\t     To remove trigger with a count:\n"
5033 	"\t       echo '!<function>:<trigger>:0 > set_ftrace_filter\n"
5034 	"  set_ftrace_notrace\t- echo function name in here to never trace.\n"
5035 	"\t    accepts: func_full_name, *func_end, func_begin*, *func_middle*\n"
5036 	"\t    modules: Can select a group via module command :mod:\n"
5037 	"\t    Does not accept triggers\n"
5038 #endif /* CONFIG_DYNAMIC_FTRACE */
5039 #ifdef CONFIG_FUNCTION_TRACER
5040 	"  set_ftrace_pid\t- Write pid(s) to only function trace those pids\n"
5041 	"\t\t    (function)\n"
5042 	"  set_ftrace_notrace_pid\t- Write pid(s) to not function trace those pids\n"
5043 	"\t\t    (function)\n"
5044 #endif
5045 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
5046 	"  set_graph_function\t- Trace the nested calls of a function (function_graph)\n"
5047 	"  set_graph_notrace\t- Do not trace the nested calls of a function (function_graph)\n"
5048 	"  max_graph_depth\t- Trace a limited depth of nested calls (0 is unlimited)\n"
5049 #endif
5050 #ifdef CONFIG_TRACER_SNAPSHOT
5051 	"\n  snapshot\t\t- Like 'trace' but shows the content of the static\n"
5052 	"\t\t\t  snapshot buffer. Read the contents for more\n"
5053 	"\t\t\t  information\n"
5054 #endif
5055 #ifdef CONFIG_STACK_TRACER
5056 	"  stack_trace\t\t- Shows the max stack trace when active\n"
5057 	"  stack_max_size\t- Shows current max stack size that was traced\n"
5058 	"\t\t\t  Write into this file to reset the max size (trigger a\n"
5059 	"\t\t\t  new trace)\n"
5060 #ifdef CONFIG_DYNAMIC_FTRACE
5061 	"  stack_trace_filter\t- Like set_ftrace_filter but limits what stack_trace\n"
5062 	"\t\t\t  traces\n"
5063 #endif
5064 #endif /* CONFIG_STACK_TRACER */
5065 #ifdef CONFIG_DYNAMIC_EVENTS
5066 	"  dynamic_events\t\t- Create/append/remove/show the generic dynamic events\n"
5067 	"\t\t\t  Write into this file to define/undefine new trace events.\n"
5068 #endif
5069 #ifdef CONFIG_KPROBE_EVENTS
5070 	"  kprobe_events\t\t- Create/append/remove/show the kernel dynamic events\n"
5071 	"\t\t\t  Write into this file to define/undefine new trace events.\n"
5072 #endif
5073 #ifdef CONFIG_UPROBE_EVENTS
5074 	"  uprobe_events\t\t- Create/append/remove/show the userspace dynamic events\n"
5075 	"\t\t\t  Write into this file to define/undefine new trace events.\n"
5076 #endif
5077 #if defined(CONFIG_KPROBE_EVENTS) || defined(CONFIG_UPROBE_EVENTS)
5078 	"\t  accepts: event-definitions (one definition per line)\n"
5079 	"\t   Format: p[:[<group>/]<event>] <place> [<args>]\n"
5080 	"\t           r[maxactive][:[<group>/]<event>] <place> [<args>]\n"
5081 #ifdef CONFIG_HIST_TRIGGERS
5082 	"\t           s:[synthetic/]<event> <field> [<field>]\n"
5083 #endif
5084 	"\t           -:[<group>/]<event>\n"
5085 #ifdef CONFIG_KPROBE_EVENTS
5086 	"\t    place: [<module>:]<symbol>[+<offset>]|<memaddr>\n"
5087   "place (kretprobe): [<module>:]<symbol>[+<offset>]|<memaddr>\n"
5088 #endif
5089 #ifdef CONFIG_UPROBE_EVENTS
5090   "   place (uprobe): <path>:<offset>[(ref_ctr_offset)]\n"
5091 #endif
5092 	"\t     args: <name>=fetcharg[:type]\n"
5093 	"\t fetcharg: %<register>, @<address>, @<symbol>[+|-<offset>],\n"
5094 #ifdef CONFIG_HAVE_FUNCTION_ARG_ACCESS_API
5095 	"\t           $stack<index>, $stack, $retval, $comm, $arg<N>,\n"
5096 #else
5097 	"\t           $stack<index>, $stack, $retval, $comm,\n"
5098 #endif
5099 	"\t           +|-[u]<offset>(<fetcharg>), \\imm-value, \\\"imm-string\"\n"
5100 	"\t     type: s8/16/32/64, u8/16/32/64, x8/16/32/64, string, symbol,\n"
5101 	"\t           b<bit-width>@<bit-offset>/<container-size>, ustring,\n"
5102 	"\t           <type>\\[<array-size>\\]\n"
5103 #ifdef CONFIG_HIST_TRIGGERS
5104 	"\t    field: <stype> <name>;\n"
5105 	"\t    stype: u8/u16/u32/u64, s8/s16/s32/s64, pid_t,\n"
5106 	"\t           [unsigned] char/int/long\n"
5107 #endif
5108 #endif
5109 	"  events/\t\t- Directory containing all trace event subsystems:\n"
5110 	"      enable\t\t- Write 0/1 to enable/disable tracing of all events\n"
5111 	"  events/<system>/\t- Directory containing all trace events for <system>:\n"
5112 	"      enable\t\t- Write 0/1 to enable/disable tracing of all <system>\n"
5113 	"\t\t\t  events\n"
5114 	"      filter\t\t- If set, only events passing filter are traced\n"
5115 	"  events/<system>/<event>/\t- Directory containing control files for\n"
5116 	"\t\t\t  <event>:\n"
5117 	"      enable\t\t- Write 0/1 to enable/disable tracing of <event>\n"
5118 	"      filter\t\t- If set, only events passing filter are traced\n"
5119 	"      trigger\t\t- If set, a command to perform when event is hit\n"
5120 	"\t    Format: <trigger>[:count][if <filter>]\n"
5121 	"\t   trigger: traceon, traceoff\n"
5122 	"\t            enable_event:<system>:<event>\n"
5123 	"\t            disable_event:<system>:<event>\n"
5124 #ifdef CONFIG_HIST_TRIGGERS
5125 	"\t            enable_hist:<system>:<event>\n"
5126 	"\t            disable_hist:<system>:<event>\n"
5127 #endif
5128 #ifdef CONFIG_STACKTRACE
5129 	"\t\t    stacktrace\n"
5130 #endif
5131 #ifdef CONFIG_TRACER_SNAPSHOT
5132 	"\t\t    snapshot\n"
5133 #endif
5134 #ifdef CONFIG_HIST_TRIGGERS
5135 	"\t\t    hist (see below)\n"
5136 #endif
5137 	"\t   example: echo traceoff > events/block/block_unplug/trigger\n"
5138 	"\t            echo traceoff:3 > events/block/block_unplug/trigger\n"
5139 	"\t            echo 'enable_event:kmem:kmalloc:3 if nr_rq > 1' > \\\n"
5140 	"\t                  events/block/block_unplug/trigger\n"
5141 	"\t   The first disables tracing every time block_unplug is hit.\n"
5142 	"\t   The second disables tracing the first 3 times block_unplug is hit.\n"
5143 	"\t   The third enables the kmalloc event the first 3 times block_unplug\n"
5144 	"\t     is hit and has value of greater than 1 for the 'nr_rq' event field.\n"
5145 	"\t   Like function triggers, the counter is only decremented if it\n"
5146 	"\t    enabled or disabled tracing.\n"
5147 	"\t   To remove a trigger without a count:\n"
5148 	"\t     echo '!<trigger> > <system>/<event>/trigger\n"
5149 	"\t   To remove a trigger with a count:\n"
5150 	"\t     echo '!<trigger>:0 > <system>/<event>/trigger\n"
5151 	"\t   Filters can be ignored when removing a trigger.\n"
5152 #ifdef CONFIG_HIST_TRIGGERS
5153 	"      hist trigger\t- If set, event hits are aggregated into a hash table\n"
5154 	"\t    Format: hist:keys=<field1[,field2,...]>\n"
5155 	"\t            [:values=<field1[,field2,...]>]\n"
5156 	"\t            [:sort=<field1[,field2,...]>]\n"
5157 	"\t            [:size=#entries]\n"
5158 	"\t            [:pause][:continue][:clear]\n"
5159 	"\t            [:name=histname1]\n"
5160 	"\t            [:<handler>.<action>]\n"
5161 	"\t            [if <filter>]\n\n"
5162 	"\t    When a matching event is hit, an entry is added to a hash\n"
5163 	"\t    table using the key(s) and value(s) named, and the value of a\n"
5164 	"\t    sum called 'hitcount' is incremented.  Keys and values\n"
5165 	"\t    correspond to fields in the event's format description.  Keys\n"
5166 	"\t    can be any field, or the special string 'stacktrace'.\n"
5167 	"\t    Compound keys consisting of up to two fields can be specified\n"
5168 	"\t    by the 'keys' keyword.  Values must correspond to numeric\n"
5169 	"\t    fields.  Sort keys consisting of up to two fields can be\n"
5170 	"\t    specified using the 'sort' keyword.  The sort direction can\n"
5171 	"\t    be modified by appending '.descending' or '.ascending' to a\n"
5172 	"\t    sort field.  The 'size' parameter can be used to specify more\n"
5173 	"\t    or fewer than the default 2048 entries for the hashtable size.\n"
5174 	"\t    If a hist trigger is given a name using the 'name' parameter,\n"
5175 	"\t    its histogram data will be shared with other triggers of the\n"
5176 	"\t    same name, and trigger hits will update this common data.\n\n"
5177 	"\t    Reading the 'hist' file for the event will dump the hash\n"
5178 	"\t    table in its entirety to stdout.  If there are multiple hist\n"
5179 	"\t    triggers attached to an event, there will be a table for each\n"
5180 	"\t    trigger in the output.  The table displayed for a named\n"
5181 	"\t    trigger will be the same as any other instance having the\n"
5182 	"\t    same name.  The default format used to display a given field\n"
5183 	"\t    can be modified by appending any of the following modifiers\n"
5184 	"\t    to the field name, as applicable:\n\n"
5185 	"\t            .hex        display a number as a hex value\n"
5186 	"\t            .sym        display an address as a symbol\n"
5187 	"\t            .sym-offset display an address as a symbol and offset\n"
5188 	"\t            .execname   display a common_pid as a program name\n"
5189 	"\t            .syscall    display a syscall id as a syscall name\n"
5190 	"\t            .log2       display log2 value rather than raw number\n"
5191 	"\t            .usecs      display a common_timestamp in microseconds\n\n"
5192 	"\t    The 'pause' parameter can be used to pause an existing hist\n"
5193 	"\t    trigger or to start a hist trigger but not log any events\n"
5194 	"\t    until told to do so.  'continue' can be used to start or\n"
5195 	"\t    restart a paused hist trigger.\n\n"
5196 	"\t    The 'clear' parameter will clear the contents of a running\n"
5197 	"\t    hist trigger and leave its current paused/active state\n"
5198 	"\t    unchanged.\n\n"
5199 	"\t    The enable_hist and disable_hist triggers can be used to\n"
5200 	"\t    have one event conditionally start and stop another event's\n"
5201 	"\t    already-attached hist trigger.  The syntax is analogous to\n"
5202 	"\t    the enable_event and disable_event triggers.\n\n"
5203 	"\t    Hist trigger handlers and actions are executed whenever a\n"
5204 	"\t    a histogram entry is added or updated.  They take the form:\n\n"
5205 	"\t        <handler>.<action>\n\n"
5206 	"\t    The available handlers are:\n\n"
5207 	"\t        onmatch(matching.event)  - invoke on addition or update\n"
5208 	"\t        onmax(var)               - invoke if var exceeds current max\n"
5209 	"\t        onchange(var)            - invoke action if var changes\n\n"
5210 	"\t    The available actions are:\n\n"
5211 	"\t        trace(<synthetic_event>,param list)  - generate synthetic event\n"
5212 	"\t        save(field,...)                      - save current event fields\n"
5213 #ifdef CONFIG_TRACER_SNAPSHOT
5214 	"\t        snapshot()                           - snapshot the trace buffer\n"
5215 #endif
5216 #endif
5217 ;
5218 
5219 static ssize_t
5220 tracing_readme_read(struct file *filp, char __user *ubuf,
5221 		       size_t cnt, loff_t *ppos)
5222 {
5223 	return simple_read_from_buffer(ubuf, cnt, ppos,
5224 					readme_msg, strlen(readme_msg));
5225 }
5226 
5227 static const struct file_operations tracing_readme_fops = {
5228 	.open		= tracing_open_generic,
5229 	.read		= tracing_readme_read,
5230 	.llseek		= generic_file_llseek,
5231 };
5232 
5233 static void *saved_tgids_next(struct seq_file *m, void *v, loff_t *pos)
5234 {
5235 	int *ptr = v;
5236 
5237 	if (*pos || m->count)
5238 		ptr++;
5239 
5240 	(*pos)++;
5241 
5242 	for (; ptr <= &tgid_map[PID_MAX_DEFAULT]; ptr++) {
5243 		if (trace_find_tgid(*ptr))
5244 			return ptr;
5245 	}
5246 
5247 	return NULL;
5248 }
5249 
5250 static void *saved_tgids_start(struct seq_file *m, loff_t *pos)
5251 {
5252 	void *v;
5253 	loff_t l = 0;
5254 
5255 	if (!tgid_map)
5256 		return NULL;
5257 
5258 	v = &tgid_map[0];
5259 	while (l <= *pos) {
5260 		v = saved_tgids_next(m, v, &l);
5261 		if (!v)
5262 			return NULL;
5263 	}
5264 
5265 	return v;
5266 }
5267 
5268 static void saved_tgids_stop(struct seq_file *m, void *v)
5269 {
5270 }
5271 
5272 static int saved_tgids_show(struct seq_file *m, void *v)
5273 {
5274 	int pid = (int *)v - tgid_map;
5275 
5276 	seq_printf(m, "%d %d\n", pid, trace_find_tgid(pid));
5277 	return 0;
5278 }
5279 
5280 static const struct seq_operations tracing_saved_tgids_seq_ops = {
5281 	.start		= saved_tgids_start,
5282 	.stop		= saved_tgids_stop,
5283 	.next		= saved_tgids_next,
5284 	.show		= saved_tgids_show,
5285 };
5286 
5287 static int tracing_saved_tgids_open(struct inode *inode, struct file *filp)
5288 {
5289 	int ret;
5290 
5291 	ret = tracing_check_open_get_tr(NULL);
5292 	if (ret)
5293 		return ret;
5294 
5295 	return seq_open(filp, &tracing_saved_tgids_seq_ops);
5296 }
5297 
5298 
5299 static const struct file_operations tracing_saved_tgids_fops = {
5300 	.open		= tracing_saved_tgids_open,
5301 	.read		= seq_read,
5302 	.llseek		= seq_lseek,
5303 	.release	= seq_release,
5304 };
5305 
5306 static void *saved_cmdlines_next(struct seq_file *m, void *v, loff_t *pos)
5307 {
5308 	unsigned int *ptr = v;
5309 
5310 	if (*pos || m->count)
5311 		ptr++;
5312 
5313 	(*pos)++;
5314 
5315 	for (; ptr < &savedcmd->map_cmdline_to_pid[savedcmd->cmdline_num];
5316 	     ptr++) {
5317 		if (*ptr == -1 || *ptr == NO_CMDLINE_MAP)
5318 			continue;
5319 
5320 		return ptr;
5321 	}
5322 
5323 	return NULL;
5324 }
5325 
5326 static void *saved_cmdlines_start(struct seq_file *m, loff_t *pos)
5327 {
5328 	void *v;
5329 	loff_t l = 0;
5330 
5331 	preempt_disable();
5332 	arch_spin_lock(&trace_cmdline_lock);
5333 
5334 	v = &savedcmd->map_cmdline_to_pid[0];
5335 	while (l <= *pos) {
5336 		v = saved_cmdlines_next(m, v, &l);
5337 		if (!v)
5338 			return NULL;
5339 	}
5340 
5341 	return v;
5342 }
5343 
5344 static void saved_cmdlines_stop(struct seq_file *m, void *v)
5345 {
5346 	arch_spin_unlock(&trace_cmdline_lock);
5347 	preempt_enable();
5348 }
5349 
5350 static int saved_cmdlines_show(struct seq_file *m, void *v)
5351 {
5352 	char buf[TASK_COMM_LEN];
5353 	unsigned int *pid = v;
5354 
5355 	__trace_find_cmdline(*pid, buf);
5356 	seq_printf(m, "%d %s\n", *pid, buf);
5357 	return 0;
5358 }
5359 
5360 static const struct seq_operations tracing_saved_cmdlines_seq_ops = {
5361 	.start		= saved_cmdlines_start,
5362 	.next		= saved_cmdlines_next,
5363 	.stop		= saved_cmdlines_stop,
5364 	.show		= saved_cmdlines_show,
5365 };
5366 
5367 static int tracing_saved_cmdlines_open(struct inode *inode, struct file *filp)
5368 {
5369 	int ret;
5370 
5371 	ret = tracing_check_open_get_tr(NULL);
5372 	if (ret)
5373 		return ret;
5374 
5375 	return seq_open(filp, &tracing_saved_cmdlines_seq_ops);
5376 }
5377 
5378 static const struct file_operations tracing_saved_cmdlines_fops = {
5379 	.open		= tracing_saved_cmdlines_open,
5380 	.read		= seq_read,
5381 	.llseek		= seq_lseek,
5382 	.release	= seq_release,
5383 };
5384 
5385 static ssize_t
5386 tracing_saved_cmdlines_size_read(struct file *filp, char __user *ubuf,
5387 				 size_t cnt, loff_t *ppos)
5388 {
5389 	char buf[64];
5390 	int r;
5391 
5392 	arch_spin_lock(&trace_cmdline_lock);
5393 	r = scnprintf(buf, sizeof(buf), "%u\n", savedcmd->cmdline_num);
5394 	arch_spin_unlock(&trace_cmdline_lock);
5395 
5396 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
5397 }
5398 
5399 static void free_saved_cmdlines_buffer(struct saved_cmdlines_buffer *s)
5400 {
5401 	kfree(s->saved_cmdlines);
5402 	kfree(s->map_cmdline_to_pid);
5403 	kfree(s);
5404 }
5405 
5406 static int tracing_resize_saved_cmdlines(unsigned int val)
5407 {
5408 	struct saved_cmdlines_buffer *s, *savedcmd_temp;
5409 
5410 	s = kmalloc(sizeof(*s), GFP_KERNEL);
5411 	if (!s)
5412 		return -ENOMEM;
5413 
5414 	if (allocate_cmdlines_buffer(val, s) < 0) {
5415 		kfree(s);
5416 		return -ENOMEM;
5417 	}
5418 
5419 	arch_spin_lock(&trace_cmdline_lock);
5420 	savedcmd_temp = savedcmd;
5421 	savedcmd = s;
5422 	arch_spin_unlock(&trace_cmdline_lock);
5423 	free_saved_cmdlines_buffer(savedcmd_temp);
5424 
5425 	return 0;
5426 }
5427 
5428 static ssize_t
5429 tracing_saved_cmdlines_size_write(struct file *filp, const char __user *ubuf,
5430 				  size_t cnt, loff_t *ppos)
5431 {
5432 	unsigned long val;
5433 	int ret;
5434 
5435 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
5436 	if (ret)
5437 		return ret;
5438 
5439 	/* must have at least 1 entry or less than PID_MAX_DEFAULT */
5440 	if (!val || val > PID_MAX_DEFAULT)
5441 		return -EINVAL;
5442 
5443 	ret = tracing_resize_saved_cmdlines((unsigned int)val);
5444 	if (ret < 0)
5445 		return ret;
5446 
5447 	*ppos += cnt;
5448 
5449 	return cnt;
5450 }
5451 
5452 static const struct file_operations tracing_saved_cmdlines_size_fops = {
5453 	.open		= tracing_open_generic,
5454 	.read		= tracing_saved_cmdlines_size_read,
5455 	.write		= tracing_saved_cmdlines_size_write,
5456 };
5457 
5458 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
5459 static union trace_eval_map_item *
5460 update_eval_map(union trace_eval_map_item *ptr)
5461 {
5462 	if (!ptr->map.eval_string) {
5463 		if (ptr->tail.next) {
5464 			ptr = ptr->tail.next;
5465 			/* Set ptr to the next real item (skip head) */
5466 			ptr++;
5467 		} else
5468 			return NULL;
5469 	}
5470 	return ptr;
5471 }
5472 
5473 static void *eval_map_next(struct seq_file *m, void *v, loff_t *pos)
5474 {
5475 	union trace_eval_map_item *ptr = v;
5476 
5477 	/*
5478 	 * Paranoid! If ptr points to end, we don't want to increment past it.
5479 	 * This really should never happen.
5480 	 */
5481 	(*pos)++;
5482 	ptr = update_eval_map(ptr);
5483 	if (WARN_ON_ONCE(!ptr))
5484 		return NULL;
5485 
5486 	ptr++;
5487 	ptr = update_eval_map(ptr);
5488 
5489 	return ptr;
5490 }
5491 
5492 static void *eval_map_start(struct seq_file *m, loff_t *pos)
5493 {
5494 	union trace_eval_map_item *v;
5495 	loff_t l = 0;
5496 
5497 	mutex_lock(&trace_eval_mutex);
5498 
5499 	v = trace_eval_maps;
5500 	if (v)
5501 		v++;
5502 
5503 	while (v && l < *pos) {
5504 		v = eval_map_next(m, v, &l);
5505 	}
5506 
5507 	return v;
5508 }
5509 
5510 static void eval_map_stop(struct seq_file *m, void *v)
5511 {
5512 	mutex_unlock(&trace_eval_mutex);
5513 }
5514 
5515 static int eval_map_show(struct seq_file *m, void *v)
5516 {
5517 	union trace_eval_map_item *ptr = v;
5518 
5519 	seq_printf(m, "%s %ld (%s)\n",
5520 		   ptr->map.eval_string, ptr->map.eval_value,
5521 		   ptr->map.system);
5522 
5523 	return 0;
5524 }
5525 
5526 static const struct seq_operations tracing_eval_map_seq_ops = {
5527 	.start		= eval_map_start,
5528 	.next		= eval_map_next,
5529 	.stop		= eval_map_stop,
5530 	.show		= eval_map_show,
5531 };
5532 
5533 static int tracing_eval_map_open(struct inode *inode, struct file *filp)
5534 {
5535 	int ret;
5536 
5537 	ret = tracing_check_open_get_tr(NULL);
5538 	if (ret)
5539 		return ret;
5540 
5541 	return seq_open(filp, &tracing_eval_map_seq_ops);
5542 }
5543 
5544 static const struct file_operations tracing_eval_map_fops = {
5545 	.open		= tracing_eval_map_open,
5546 	.read		= seq_read,
5547 	.llseek		= seq_lseek,
5548 	.release	= seq_release,
5549 };
5550 
5551 static inline union trace_eval_map_item *
5552 trace_eval_jmp_to_tail(union trace_eval_map_item *ptr)
5553 {
5554 	/* Return tail of array given the head */
5555 	return ptr + ptr->head.length + 1;
5556 }
5557 
5558 static void
5559 trace_insert_eval_map_file(struct module *mod, struct trace_eval_map **start,
5560 			   int len)
5561 {
5562 	struct trace_eval_map **stop;
5563 	struct trace_eval_map **map;
5564 	union trace_eval_map_item *map_array;
5565 	union trace_eval_map_item *ptr;
5566 
5567 	stop = start + len;
5568 
5569 	/*
5570 	 * The trace_eval_maps contains the map plus a head and tail item,
5571 	 * where the head holds the module and length of array, and the
5572 	 * tail holds a pointer to the next list.
5573 	 */
5574 	map_array = kmalloc_array(len + 2, sizeof(*map_array), GFP_KERNEL);
5575 	if (!map_array) {
5576 		pr_warn("Unable to allocate trace eval mapping\n");
5577 		return;
5578 	}
5579 
5580 	mutex_lock(&trace_eval_mutex);
5581 
5582 	if (!trace_eval_maps)
5583 		trace_eval_maps = map_array;
5584 	else {
5585 		ptr = trace_eval_maps;
5586 		for (;;) {
5587 			ptr = trace_eval_jmp_to_tail(ptr);
5588 			if (!ptr->tail.next)
5589 				break;
5590 			ptr = ptr->tail.next;
5591 
5592 		}
5593 		ptr->tail.next = map_array;
5594 	}
5595 	map_array->head.mod = mod;
5596 	map_array->head.length = len;
5597 	map_array++;
5598 
5599 	for (map = start; (unsigned long)map < (unsigned long)stop; map++) {
5600 		map_array->map = **map;
5601 		map_array++;
5602 	}
5603 	memset(map_array, 0, sizeof(*map_array));
5604 
5605 	mutex_unlock(&trace_eval_mutex);
5606 }
5607 
5608 static void trace_create_eval_file(struct dentry *d_tracer)
5609 {
5610 	trace_create_file("eval_map", 0444, d_tracer,
5611 			  NULL, &tracing_eval_map_fops);
5612 }
5613 
5614 #else /* CONFIG_TRACE_EVAL_MAP_FILE */
5615 static inline void trace_create_eval_file(struct dentry *d_tracer) { }
5616 static inline void trace_insert_eval_map_file(struct module *mod,
5617 			      struct trace_eval_map **start, int len) { }
5618 #endif /* !CONFIG_TRACE_EVAL_MAP_FILE */
5619 
5620 static void trace_insert_eval_map(struct module *mod,
5621 				  struct trace_eval_map **start, int len)
5622 {
5623 	struct trace_eval_map **map;
5624 
5625 	if (len <= 0)
5626 		return;
5627 
5628 	map = start;
5629 
5630 	trace_event_eval_update(map, len);
5631 
5632 	trace_insert_eval_map_file(mod, start, len);
5633 }
5634 
5635 static ssize_t
5636 tracing_set_trace_read(struct file *filp, char __user *ubuf,
5637 		       size_t cnt, loff_t *ppos)
5638 {
5639 	struct trace_array *tr = filp->private_data;
5640 	char buf[MAX_TRACER_SIZE+2];
5641 	int r;
5642 
5643 	mutex_lock(&trace_types_lock);
5644 	r = sprintf(buf, "%s\n", tr->current_trace->name);
5645 	mutex_unlock(&trace_types_lock);
5646 
5647 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
5648 }
5649 
5650 int tracer_init(struct tracer *t, struct trace_array *tr)
5651 {
5652 	tracing_reset_online_cpus(&tr->array_buffer);
5653 	return t->init(tr);
5654 }
5655 
5656 static void set_buffer_entries(struct array_buffer *buf, unsigned long val)
5657 {
5658 	int cpu;
5659 
5660 	for_each_tracing_cpu(cpu)
5661 		per_cpu_ptr(buf->data, cpu)->entries = val;
5662 }
5663 
5664 #ifdef CONFIG_TRACER_MAX_TRACE
5665 /* resize @tr's buffer to the size of @size_tr's entries */
5666 static int resize_buffer_duplicate_size(struct array_buffer *trace_buf,
5667 					struct array_buffer *size_buf, int cpu_id)
5668 {
5669 	int cpu, ret = 0;
5670 
5671 	if (cpu_id == RING_BUFFER_ALL_CPUS) {
5672 		for_each_tracing_cpu(cpu) {
5673 			ret = ring_buffer_resize(trace_buf->buffer,
5674 				 per_cpu_ptr(size_buf->data, cpu)->entries, cpu);
5675 			if (ret < 0)
5676 				break;
5677 			per_cpu_ptr(trace_buf->data, cpu)->entries =
5678 				per_cpu_ptr(size_buf->data, cpu)->entries;
5679 		}
5680 	} else {
5681 		ret = ring_buffer_resize(trace_buf->buffer,
5682 				 per_cpu_ptr(size_buf->data, cpu_id)->entries, cpu_id);
5683 		if (ret == 0)
5684 			per_cpu_ptr(trace_buf->data, cpu_id)->entries =
5685 				per_cpu_ptr(size_buf->data, cpu_id)->entries;
5686 	}
5687 
5688 	return ret;
5689 }
5690 #endif /* CONFIG_TRACER_MAX_TRACE */
5691 
5692 static int __tracing_resize_ring_buffer(struct trace_array *tr,
5693 					unsigned long size, int cpu)
5694 {
5695 	int ret;
5696 
5697 	/*
5698 	 * If kernel or user changes the size of the ring buffer
5699 	 * we use the size that was given, and we can forget about
5700 	 * expanding it later.
5701 	 */
5702 	ring_buffer_expanded = true;
5703 
5704 	/* May be called before buffers are initialized */
5705 	if (!tr->array_buffer.buffer)
5706 		return 0;
5707 
5708 	ret = ring_buffer_resize(tr->array_buffer.buffer, size, cpu);
5709 	if (ret < 0)
5710 		return ret;
5711 
5712 #ifdef CONFIG_TRACER_MAX_TRACE
5713 	if (!(tr->flags & TRACE_ARRAY_FL_GLOBAL) ||
5714 	    !tr->current_trace->use_max_tr)
5715 		goto out;
5716 
5717 	ret = ring_buffer_resize(tr->max_buffer.buffer, size, cpu);
5718 	if (ret < 0) {
5719 		int r = resize_buffer_duplicate_size(&tr->array_buffer,
5720 						     &tr->array_buffer, cpu);
5721 		if (r < 0) {
5722 			/*
5723 			 * AARGH! We are left with different
5724 			 * size max buffer!!!!
5725 			 * The max buffer is our "snapshot" buffer.
5726 			 * When a tracer needs a snapshot (one of the
5727 			 * latency tracers), it swaps the max buffer
5728 			 * with the saved snap shot. We succeeded to
5729 			 * update the size of the main buffer, but failed to
5730 			 * update the size of the max buffer. But when we tried
5731 			 * to reset the main buffer to the original size, we
5732 			 * failed there too. This is very unlikely to
5733 			 * happen, but if it does, warn and kill all
5734 			 * tracing.
5735 			 */
5736 			WARN_ON(1);
5737 			tracing_disabled = 1;
5738 		}
5739 		return ret;
5740 	}
5741 
5742 	if (cpu == RING_BUFFER_ALL_CPUS)
5743 		set_buffer_entries(&tr->max_buffer, size);
5744 	else
5745 		per_cpu_ptr(tr->max_buffer.data, cpu)->entries = size;
5746 
5747  out:
5748 #endif /* CONFIG_TRACER_MAX_TRACE */
5749 
5750 	if (cpu == RING_BUFFER_ALL_CPUS)
5751 		set_buffer_entries(&tr->array_buffer, size);
5752 	else
5753 		per_cpu_ptr(tr->array_buffer.data, cpu)->entries = size;
5754 
5755 	return ret;
5756 }
5757 
5758 ssize_t tracing_resize_ring_buffer(struct trace_array *tr,
5759 				  unsigned long size, int cpu_id)
5760 {
5761 	int ret = size;
5762 
5763 	mutex_lock(&trace_types_lock);
5764 
5765 	if (cpu_id != RING_BUFFER_ALL_CPUS) {
5766 		/* make sure, this cpu is enabled in the mask */
5767 		if (!cpumask_test_cpu(cpu_id, tracing_buffer_mask)) {
5768 			ret = -EINVAL;
5769 			goto out;
5770 		}
5771 	}
5772 
5773 	ret = __tracing_resize_ring_buffer(tr, size, cpu_id);
5774 	if (ret < 0)
5775 		ret = -ENOMEM;
5776 
5777 out:
5778 	mutex_unlock(&trace_types_lock);
5779 
5780 	return ret;
5781 }
5782 
5783 
5784 /**
5785  * tracing_update_buffers - used by tracing facility to expand ring buffers
5786  *
5787  * To save on memory when the tracing is never used on a system with it
5788  * configured in. The ring buffers are set to a minimum size. But once
5789  * a user starts to use the tracing facility, then they need to grow
5790  * to their default size.
5791  *
5792  * This function is to be called when a tracer is about to be used.
5793  */
5794 int tracing_update_buffers(void)
5795 {
5796 	int ret = 0;
5797 
5798 	mutex_lock(&trace_types_lock);
5799 	if (!ring_buffer_expanded)
5800 		ret = __tracing_resize_ring_buffer(&global_trace, trace_buf_size,
5801 						RING_BUFFER_ALL_CPUS);
5802 	mutex_unlock(&trace_types_lock);
5803 
5804 	return ret;
5805 }
5806 
5807 struct trace_option_dentry;
5808 
5809 static void
5810 create_trace_option_files(struct trace_array *tr, struct tracer *tracer);
5811 
5812 /*
5813  * Used to clear out the tracer before deletion of an instance.
5814  * Must have trace_types_lock held.
5815  */
5816 static void tracing_set_nop(struct trace_array *tr)
5817 {
5818 	if (tr->current_trace == &nop_trace)
5819 		return;
5820 
5821 	tr->current_trace->enabled--;
5822 
5823 	if (tr->current_trace->reset)
5824 		tr->current_trace->reset(tr);
5825 
5826 	tr->current_trace = &nop_trace;
5827 }
5828 
5829 static void add_tracer_options(struct trace_array *tr, struct tracer *t)
5830 {
5831 	/* Only enable if the directory has been created already. */
5832 	if (!tr->dir)
5833 		return;
5834 
5835 	create_trace_option_files(tr, t);
5836 }
5837 
5838 int tracing_set_tracer(struct trace_array *tr, const char *buf)
5839 {
5840 	struct tracer *t;
5841 #ifdef CONFIG_TRACER_MAX_TRACE
5842 	bool had_max_tr;
5843 #endif
5844 	int ret = 0;
5845 
5846 	mutex_lock(&trace_types_lock);
5847 
5848 	if (!ring_buffer_expanded) {
5849 		ret = __tracing_resize_ring_buffer(tr, trace_buf_size,
5850 						RING_BUFFER_ALL_CPUS);
5851 		if (ret < 0)
5852 			goto out;
5853 		ret = 0;
5854 	}
5855 
5856 	for (t = trace_types; t; t = t->next) {
5857 		if (strcmp(t->name, buf) == 0)
5858 			break;
5859 	}
5860 	if (!t) {
5861 		ret = -EINVAL;
5862 		goto out;
5863 	}
5864 	if (t == tr->current_trace)
5865 		goto out;
5866 
5867 #ifdef CONFIG_TRACER_SNAPSHOT
5868 	if (t->use_max_tr) {
5869 		arch_spin_lock(&tr->max_lock);
5870 		if (tr->cond_snapshot)
5871 			ret = -EBUSY;
5872 		arch_spin_unlock(&tr->max_lock);
5873 		if (ret)
5874 			goto out;
5875 	}
5876 #endif
5877 	/* Some tracers won't work on kernel command line */
5878 	if (system_state < SYSTEM_RUNNING && t->noboot) {
5879 		pr_warn("Tracer '%s' is not allowed on command line, ignored\n",
5880 			t->name);
5881 		goto out;
5882 	}
5883 
5884 	/* Some tracers are only allowed for the top level buffer */
5885 	if (!trace_ok_for_array(t, tr)) {
5886 		ret = -EINVAL;
5887 		goto out;
5888 	}
5889 
5890 	/* If trace pipe files are being read, we can't change the tracer */
5891 	if (tr->current_trace->ref) {
5892 		ret = -EBUSY;
5893 		goto out;
5894 	}
5895 
5896 	trace_branch_disable();
5897 
5898 	tr->current_trace->enabled--;
5899 
5900 	if (tr->current_trace->reset)
5901 		tr->current_trace->reset(tr);
5902 
5903 	/* Current trace needs to be nop_trace before synchronize_rcu */
5904 	tr->current_trace = &nop_trace;
5905 
5906 #ifdef CONFIG_TRACER_MAX_TRACE
5907 	had_max_tr = tr->allocated_snapshot;
5908 
5909 	if (had_max_tr && !t->use_max_tr) {
5910 		/*
5911 		 * We need to make sure that the update_max_tr sees that
5912 		 * current_trace changed to nop_trace to keep it from
5913 		 * swapping the buffers after we resize it.
5914 		 * The update_max_tr is called from interrupts disabled
5915 		 * so a synchronized_sched() is sufficient.
5916 		 */
5917 		synchronize_rcu();
5918 		free_snapshot(tr);
5919 	}
5920 #endif
5921 
5922 #ifdef CONFIG_TRACER_MAX_TRACE
5923 	if (t->use_max_tr && !had_max_tr) {
5924 		ret = tracing_alloc_snapshot_instance(tr);
5925 		if (ret < 0)
5926 			goto out;
5927 	}
5928 #endif
5929 
5930 	if (t->init) {
5931 		ret = tracer_init(t, tr);
5932 		if (ret)
5933 			goto out;
5934 	}
5935 
5936 	tr->current_trace = t;
5937 	tr->current_trace->enabled++;
5938 	trace_branch_enable(tr);
5939  out:
5940 	mutex_unlock(&trace_types_lock);
5941 
5942 	return ret;
5943 }
5944 
5945 static ssize_t
5946 tracing_set_trace_write(struct file *filp, const char __user *ubuf,
5947 			size_t cnt, loff_t *ppos)
5948 {
5949 	struct trace_array *tr = filp->private_data;
5950 	char buf[MAX_TRACER_SIZE+1];
5951 	int i;
5952 	size_t ret;
5953 	int err;
5954 
5955 	ret = cnt;
5956 
5957 	if (cnt > MAX_TRACER_SIZE)
5958 		cnt = MAX_TRACER_SIZE;
5959 
5960 	if (copy_from_user(buf, ubuf, cnt))
5961 		return -EFAULT;
5962 
5963 	buf[cnt] = 0;
5964 
5965 	/* strip ending whitespace. */
5966 	for (i = cnt - 1; i > 0 && isspace(buf[i]); i--)
5967 		buf[i] = 0;
5968 
5969 	err = tracing_set_tracer(tr, buf);
5970 	if (err)
5971 		return err;
5972 
5973 	*ppos += ret;
5974 
5975 	return ret;
5976 }
5977 
5978 static ssize_t
5979 tracing_nsecs_read(unsigned long *ptr, char __user *ubuf,
5980 		   size_t cnt, loff_t *ppos)
5981 {
5982 	char buf[64];
5983 	int r;
5984 
5985 	r = snprintf(buf, sizeof(buf), "%ld\n",
5986 		     *ptr == (unsigned long)-1 ? -1 : nsecs_to_usecs(*ptr));
5987 	if (r > sizeof(buf))
5988 		r = sizeof(buf);
5989 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
5990 }
5991 
5992 static ssize_t
5993 tracing_nsecs_write(unsigned long *ptr, const char __user *ubuf,
5994 		    size_t cnt, loff_t *ppos)
5995 {
5996 	unsigned long val;
5997 	int ret;
5998 
5999 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
6000 	if (ret)
6001 		return ret;
6002 
6003 	*ptr = val * 1000;
6004 
6005 	return cnt;
6006 }
6007 
6008 static ssize_t
6009 tracing_thresh_read(struct file *filp, char __user *ubuf,
6010 		    size_t cnt, loff_t *ppos)
6011 {
6012 	return tracing_nsecs_read(&tracing_thresh, ubuf, cnt, ppos);
6013 }
6014 
6015 static ssize_t
6016 tracing_thresh_write(struct file *filp, const char __user *ubuf,
6017 		     size_t cnt, loff_t *ppos)
6018 {
6019 	struct trace_array *tr = filp->private_data;
6020 	int ret;
6021 
6022 	mutex_lock(&trace_types_lock);
6023 	ret = tracing_nsecs_write(&tracing_thresh, ubuf, cnt, ppos);
6024 	if (ret < 0)
6025 		goto out;
6026 
6027 	if (tr->current_trace->update_thresh) {
6028 		ret = tr->current_trace->update_thresh(tr);
6029 		if (ret < 0)
6030 			goto out;
6031 	}
6032 
6033 	ret = cnt;
6034 out:
6035 	mutex_unlock(&trace_types_lock);
6036 
6037 	return ret;
6038 }
6039 
6040 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
6041 
6042 static ssize_t
6043 tracing_max_lat_read(struct file *filp, char __user *ubuf,
6044 		     size_t cnt, loff_t *ppos)
6045 {
6046 	return tracing_nsecs_read(filp->private_data, ubuf, cnt, ppos);
6047 }
6048 
6049 static ssize_t
6050 tracing_max_lat_write(struct file *filp, const char __user *ubuf,
6051 		      size_t cnt, loff_t *ppos)
6052 {
6053 	return tracing_nsecs_write(filp->private_data, ubuf, cnt, ppos);
6054 }
6055 
6056 #endif
6057 
6058 static int tracing_open_pipe(struct inode *inode, struct file *filp)
6059 {
6060 	struct trace_array *tr = inode->i_private;
6061 	struct trace_iterator *iter;
6062 	int ret;
6063 
6064 	ret = tracing_check_open_get_tr(tr);
6065 	if (ret)
6066 		return ret;
6067 
6068 	mutex_lock(&trace_types_lock);
6069 
6070 	/* create a buffer to store the information to pass to userspace */
6071 	iter = kzalloc(sizeof(*iter), GFP_KERNEL);
6072 	if (!iter) {
6073 		ret = -ENOMEM;
6074 		__trace_array_put(tr);
6075 		goto out;
6076 	}
6077 
6078 	trace_seq_init(&iter->seq);
6079 	iter->trace = tr->current_trace;
6080 
6081 	if (!alloc_cpumask_var(&iter->started, GFP_KERNEL)) {
6082 		ret = -ENOMEM;
6083 		goto fail;
6084 	}
6085 
6086 	/* trace pipe does not show start of buffer */
6087 	cpumask_setall(iter->started);
6088 
6089 	if (tr->trace_flags & TRACE_ITER_LATENCY_FMT)
6090 		iter->iter_flags |= TRACE_FILE_LAT_FMT;
6091 
6092 	/* Output in nanoseconds only if we are using a clock in nanoseconds. */
6093 	if (trace_clocks[tr->clock_id].in_ns)
6094 		iter->iter_flags |= TRACE_FILE_TIME_IN_NS;
6095 
6096 	iter->tr = tr;
6097 	iter->array_buffer = &tr->array_buffer;
6098 	iter->cpu_file = tracing_get_cpu(inode);
6099 	mutex_init(&iter->mutex);
6100 	filp->private_data = iter;
6101 
6102 	if (iter->trace->pipe_open)
6103 		iter->trace->pipe_open(iter);
6104 
6105 	nonseekable_open(inode, filp);
6106 
6107 	tr->current_trace->ref++;
6108 out:
6109 	mutex_unlock(&trace_types_lock);
6110 	return ret;
6111 
6112 fail:
6113 	kfree(iter);
6114 	__trace_array_put(tr);
6115 	mutex_unlock(&trace_types_lock);
6116 	return ret;
6117 }
6118 
6119 static int tracing_release_pipe(struct inode *inode, struct file *file)
6120 {
6121 	struct trace_iterator *iter = file->private_data;
6122 	struct trace_array *tr = inode->i_private;
6123 
6124 	mutex_lock(&trace_types_lock);
6125 
6126 	tr->current_trace->ref--;
6127 
6128 	if (iter->trace->pipe_close)
6129 		iter->trace->pipe_close(iter);
6130 
6131 	mutex_unlock(&trace_types_lock);
6132 
6133 	free_cpumask_var(iter->started);
6134 	mutex_destroy(&iter->mutex);
6135 	kfree(iter);
6136 
6137 	trace_array_put(tr);
6138 
6139 	return 0;
6140 }
6141 
6142 static __poll_t
6143 trace_poll(struct trace_iterator *iter, struct file *filp, poll_table *poll_table)
6144 {
6145 	struct trace_array *tr = iter->tr;
6146 
6147 	/* Iterators are static, they should be filled or empty */
6148 	if (trace_buffer_iter(iter, iter->cpu_file))
6149 		return EPOLLIN | EPOLLRDNORM;
6150 
6151 	if (tr->trace_flags & TRACE_ITER_BLOCK)
6152 		/*
6153 		 * Always select as readable when in blocking mode
6154 		 */
6155 		return EPOLLIN | EPOLLRDNORM;
6156 	else
6157 		return ring_buffer_poll_wait(iter->array_buffer->buffer, iter->cpu_file,
6158 					     filp, poll_table);
6159 }
6160 
6161 static __poll_t
6162 tracing_poll_pipe(struct file *filp, poll_table *poll_table)
6163 {
6164 	struct trace_iterator *iter = filp->private_data;
6165 
6166 	return trace_poll(iter, filp, poll_table);
6167 }
6168 
6169 /* Must be called with iter->mutex held. */
6170 static int tracing_wait_pipe(struct file *filp)
6171 {
6172 	struct trace_iterator *iter = filp->private_data;
6173 	int ret;
6174 
6175 	while (trace_empty(iter)) {
6176 
6177 		if ((filp->f_flags & O_NONBLOCK)) {
6178 			return -EAGAIN;
6179 		}
6180 
6181 		/*
6182 		 * We block until we read something and tracing is disabled.
6183 		 * We still block if tracing is disabled, but we have never
6184 		 * read anything. This allows a user to cat this file, and
6185 		 * then enable tracing. But after we have read something,
6186 		 * we give an EOF when tracing is again disabled.
6187 		 *
6188 		 * iter->pos will be 0 if we haven't read anything.
6189 		 */
6190 		if (!tracer_tracing_is_on(iter->tr) && iter->pos)
6191 			break;
6192 
6193 		mutex_unlock(&iter->mutex);
6194 
6195 		ret = wait_on_pipe(iter, 0);
6196 
6197 		mutex_lock(&iter->mutex);
6198 
6199 		if (ret)
6200 			return ret;
6201 	}
6202 
6203 	return 1;
6204 }
6205 
6206 /*
6207  * Consumer reader.
6208  */
6209 static ssize_t
6210 tracing_read_pipe(struct file *filp, char __user *ubuf,
6211 		  size_t cnt, loff_t *ppos)
6212 {
6213 	struct trace_iterator *iter = filp->private_data;
6214 	ssize_t sret;
6215 
6216 	/*
6217 	 * Avoid more than one consumer on a single file descriptor
6218 	 * This is just a matter of traces coherency, the ring buffer itself
6219 	 * is protected.
6220 	 */
6221 	mutex_lock(&iter->mutex);
6222 
6223 	/* return any leftover data */
6224 	sret = trace_seq_to_user(&iter->seq, ubuf, cnt);
6225 	if (sret != -EBUSY)
6226 		goto out;
6227 
6228 	trace_seq_init(&iter->seq);
6229 
6230 	if (iter->trace->read) {
6231 		sret = iter->trace->read(iter, filp, ubuf, cnt, ppos);
6232 		if (sret)
6233 			goto out;
6234 	}
6235 
6236 waitagain:
6237 	sret = tracing_wait_pipe(filp);
6238 	if (sret <= 0)
6239 		goto out;
6240 
6241 	/* stop when tracing is finished */
6242 	if (trace_empty(iter)) {
6243 		sret = 0;
6244 		goto out;
6245 	}
6246 
6247 	if (cnt >= PAGE_SIZE)
6248 		cnt = PAGE_SIZE - 1;
6249 
6250 	/* reset all but tr, trace, and overruns */
6251 	memset(&iter->seq, 0,
6252 	       sizeof(struct trace_iterator) -
6253 	       offsetof(struct trace_iterator, seq));
6254 	cpumask_clear(iter->started);
6255 	trace_seq_init(&iter->seq);
6256 	iter->pos = -1;
6257 
6258 	trace_event_read_lock();
6259 	trace_access_lock(iter->cpu_file);
6260 	while (trace_find_next_entry_inc(iter) != NULL) {
6261 		enum print_line_t ret;
6262 		int save_len = iter->seq.seq.len;
6263 
6264 		ret = print_trace_line(iter);
6265 		if (ret == TRACE_TYPE_PARTIAL_LINE) {
6266 			/* don't print partial lines */
6267 			iter->seq.seq.len = save_len;
6268 			break;
6269 		}
6270 		if (ret != TRACE_TYPE_NO_CONSUME)
6271 			trace_consume(iter);
6272 
6273 		if (trace_seq_used(&iter->seq) >= cnt)
6274 			break;
6275 
6276 		/*
6277 		 * Setting the full flag means we reached the trace_seq buffer
6278 		 * size and we should leave by partial output condition above.
6279 		 * One of the trace_seq_* functions is not used properly.
6280 		 */
6281 		WARN_ONCE(iter->seq.full, "full flag set for trace type %d",
6282 			  iter->ent->type);
6283 	}
6284 	trace_access_unlock(iter->cpu_file);
6285 	trace_event_read_unlock();
6286 
6287 	/* Now copy what we have to the user */
6288 	sret = trace_seq_to_user(&iter->seq, ubuf, cnt);
6289 	if (iter->seq.seq.readpos >= trace_seq_used(&iter->seq))
6290 		trace_seq_init(&iter->seq);
6291 
6292 	/*
6293 	 * If there was nothing to send to user, in spite of consuming trace
6294 	 * entries, go back to wait for more entries.
6295 	 */
6296 	if (sret == -EBUSY)
6297 		goto waitagain;
6298 
6299 out:
6300 	mutex_unlock(&iter->mutex);
6301 
6302 	return sret;
6303 }
6304 
6305 static void tracing_spd_release_pipe(struct splice_pipe_desc *spd,
6306 				     unsigned int idx)
6307 {
6308 	__free_page(spd->pages[idx]);
6309 }
6310 
6311 static size_t
6312 tracing_fill_pipe_page(size_t rem, struct trace_iterator *iter)
6313 {
6314 	size_t count;
6315 	int save_len;
6316 	int ret;
6317 
6318 	/* Seq buffer is page-sized, exactly what we need. */
6319 	for (;;) {
6320 		save_len = iter->seq.seq.len;
6321 		ret = print_trace_line(iter);
6322 
6323 		if (trace_seq_has_overflowed(&iter->seq)) {
6324 			iter->seq.seq.len = save_len;
6325 			break;
6326 		}
6327 
6328 		/*
6329 		 * This should not be hit, because it should only
6330 		 * be set if the iter->seq overflowed. But check it
6331 		 * anyway to be safe.
6332 		 */
6333 		if (ret == TRACE_TYPE_PARTIAL_LINE) {
6334 			iter->seq.seq.len = save_len;
6335 			break;
6336 		}
6337 
6338 		count = trace_seq_used(&iter->seq) - save_len;
6339 		if (rem < count) {
6340 			rem = 0;
6341 			iter->seq.seq.len = save_len;
6342 			break;
6343 		}
6344 
6345 		if (ret != TRACE_TYPE_NO_CONSUME)
6346 			trace_consume(iter);
6347 		rem -= count;
6348 		if (!trace_find_next_entry_inc(iter))	{
6349 			rem = 0;
6350 			iter->ent = NULL;
6351 			break;
6352 		}
6353 	}
6354 
6355 	return rem;
6356 }
6357 
6358 static ssize_t tracing_splice_read_pipe(struct file *filp,
6359 					loff_t *ppos,
6360 					struct pipe_inode_info *pipe,
6361 					size_t len,
6362 					unsigned int flags)
6363 {
6364 	struct page *pages_def[PIPE_DEF_BUFFERS];
6365 	struct partial_page partial_def[PIPE_DEF_BUFFERS];
6366 	struct trace_iterator *iter = filp->private_data;
6367 	struct splice_pipe_desc spd = {
6368 		.pages		= pages_def,
6369 		.partial	= partial_def,
6370 		.nr_pages	= 0, /* This gets updated below. */
6371 		.nr_pages_max	= PIPE_DEF_BUFFERS,
6372 		.ops		= &default_pipe_buf_ops,
6373 		.spd_release	= tracing_spd_release_pipe,
6374 	};
6375 	ssize_t ret;
6376 	size_t rem;
6377 	unsigned int i;
6378 
6379 	if (splice_grow_spd(pipe, &spd))
6380 		return -ENOMEM;
6381 
6382 	mutex_lock(&iter->mutex);
6383 
6384 	if (iter->trace->splice_read) {
6385 		ret = iter->trace->splice_read(iter, filp,
6386 					       ppos, pipe, len, flags);
6387 		if (ret)
6388 			goto out_err;
6389 	}
6390 
6391 	ret = tracing_wait_pipe(filp);
6392 	if (ret <= 0)
6393 		goto out_err;
6394 
6395 	if (!iter->ent && !trace_find_next_entry_inc(iter)) {
6396 		ret = -EFAULT;
6397 		goto out_err;
6398 	}
6399 
6400 	trace_event_read_lock();
6401 	trace_access_lock(iter->cpu_file);
6402 
6403 	/* Fill as many pages as possible. */
6404 	for (i = 0, rem = len; i < spd.nr_pages_max && rem; i++) {
6405 		spd.pages[i] = alloc_page(GFP_KERNEL);
6406 		if (!spd.pages[i])
6407 			break;
6408 
6409 		rem = tracing_fill_pipe_page(rem, iter);
6410 
6411 		/* Copy the data into the page, so we can start over. */
6412 		ret = trace_seq_to_buffer(&iter->seq,
6413 					  page_address(spd.pages[i]),
6414 					  trace_seq_used(&iter->seq));
6415 		if (ret < 0) {
6416 			__free_page(spd.pages[i]);
6417 			break;
6418 		}
6419 		spd.partial[i].offset = 0;
6420 		spd.partial[i].len = trace_seq_used(&iter->seq);
6421 
6422 		trace_seq_init(&iter->seq);
6423 	}
6424 
6425 	trace_access_unlock(iter->cpu_file);
6426 	trace_event_read_unlock();
6427 	mutex_unlock(&iter->mutex);
6428 
6429 	spd.nr_pages = i;
6430 
6431 	if (i)
6432 		ret = splice_to_pipe(pipe, &spd);
6433 	else
6434 		ret = 0;
6435 out:
6436 	splice_shrink_spd(&spd);
6437 	return ret;
6438 
6439 out_err:
6440 	mutex_unlock(&iter->mutex);
6441 	goto out;
6442 }
6443 
6444 static ssize_t
6445 tracing_entries_read(struct file *filp, char __user *ubuf,
6446 		     size_t cnt, loff_t *ppos)
6447 {
6448 	struct inode *inode = file_inode(filp);
6449 	struct trace_array *tr = inode->i_private;
6450 	int cpu = tracing_get_cpu(inode);
6451 	char buf[64];
6452 	int r = 0;
6453 	ssize_t ret;
6454 
6455 	mutex_lock(&trace_types_lock);
6456 
6457 	if (cpu == RING_BUFFER_ALL_CPUS) {
6458 		int cpu, buf_size_same;
6459 		unsigned long size;
6460 
6461 		size = 0;
6462 		buf_size_same = 1;
6463 		/* check if all cpu sizes are same */
6464 		for_each_tracing_cpu(cpu) {
6465 			/* fill in the size from first enabled cpu */
6466 			if (size == 0)
6467 				size = per_cpu_ptr(tr->array_buffer.data, cpu)->entries;
6468 			if (size != per_cpu_ptr(tr->array_buffer.data, cpu)->entries) {
6469 				buf_size_same = 0;
6470 				break;
6471 			}
6472 		}
6473 
6474 		if (buf_size_same) {
6475 			if (!ring_buffer_expanded)
6476 				r = sprintf(buf, "%lu (expanded: %lu)\n",
6477 					    size >> 10,
6478 					    trace_buf_size >> 10);
6479 			else
6480 				r = sprintf(buf, "%lu\n", size >> 10);
6481 		} else
6482 			r = sprintf(buf, "X\n");
6483 	} else
6484 		r = sprintf(buf, "%lu\n", per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10);
6485 
6486 	mutex_unlock(&trace_types_lock);
6487 
6488 	ret = simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
6489 	return ret;
6490 }
6491 
6492 static ssize_t
6493 tracing_entries_write(struct file *filp, const char __user *ubuf,
6494 		      size_t cnt, loff_t *ppos)
6495 {
6496 	struct inode *inode = file_inode(filp);
6497 	struct trace_array *tr = inode->i_private;
6498 	unsigned long val;
6499 	int ret;
6500 
6501 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
6502 	if (ret)
6503 		return ret;
6504 
6505 	/* must have at least 1 entry */
6506 	if (!val)
6507 		return -EINVAL;
6508 
6509 	/* value is in KB */
6510 	val <<= 10;
6511 	ret = tracing_resize_ring_buffer(tr, val, tracing_get_cpu(inode));
6512 	if (ret < 0)
6513 		return ret;
6514 
6515 	*ppos += cnt;
6516 
6517 	return cnt;
6518 }
6519 
6520 static ssize_t
6521 tracing_total_entries_read(struct file *filp, char __user *ubuf,
6522 				size_t cnt, loff_t *ppos)
6523 {
6524 	struct trace_array *tr = filp->private_data;
6525 	char buf[64];
6526 	int r, cpu;
6527 	unsigned long size = 0, expanded_size = 0;
6528 
6529 	mutex_lock(&trace_types_lock);
6530 	for_each_tracing_cpu(cpu) {
6531 		size += per_cpu_ptr(tr->array_buffer.data, cpu)->entries >> 10;
6532 		if (!ring_buffer_expanded)
6533 			expanded_size += trace_buf_size >> 10;
6534 	}
6535 	if (ring_buffer_expanded)
6536 		r = sprintf(buf, "%lu\n", size);
6537 	else
6538 		r = sprintf(buf, "%lu (expanded: %lu)\n", size, expanded_size);
6539 	mutex_unlock(&trace_types_lock);
6540 
6541 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
6542 }
6543 
6544 static ssize_t
6545 tracing_free_buffer_write(struct file *filp, const char __user *ubuf,
6546 			  size_t cnt, loff_t *ppos)
6547 {
6548 	/*
6549 	 * There is no need to read what the user has written, this function
6550 	 * is just to make sure that there is no error when "echo" is used
6551 	 */
6552 
6553 	*ppos += cnt;
6554 
6555 	return cnt;
6556 }
6557 
6558 static int
6559 tracing_free_buffer_release(struct inode *inode, struct file *filp)
6560 {
6561 	struct trace_array *tr = inode->i_private;
6562 
6563 	/* disable tracing ? */
6564 	if (tr->trace_flags & TRACE_ITER_STOP_ON_FREE)
6565 		tracer_tracing_off(tr);
6566 	/* resize the ring buffer to 0 */
6567 	tracing_resize_ring_buffer(tr, 0, RING_BUFFER_ALL_CPUS);
6568 
6569 	trace_array_put(tr);
6570 
6571 	return 0;
6572 }
6573 
6574 static ssize_t
6575 tracing_mark_write(struct file *filp, const char __user *ubuf,
6576 					size_t cnt, loff_t *fpos)
6577 {
6578 	struct trace_array *tr = filp->private_data;
6579 	struct ring_buffer_event *event;
6580 	enum event_trigger_type tt = ETT_NONE;
6581 	struct trace_buffer *buffer;
6582 	struct print_entry *entry;
6583 	unsigned long irq_flags;
6584 	ssize_t written;
6585 	int size;
6586 	int len;
6587 
6588 /* Used in tracing_mark_raw_write() as well */
6589 #define FAULTED_STR "<faulted>"
6590 #define FAULTED_SIZE (sizeof(FAULTED_STR) - 1) /* '\0' is already accounted for */
6591 
6592 	if (tracing_disabled)
6593 		return -EINVAL;
6594 
6595 	if (!(tr->trace_flags & TRACE_ITER_MARKERS))
6596 		return -EINVAL;
6597 
6598 	if (cnt > TRACE_BUF_SIZE)
6599 		cnt = TRACE_BUF_SIZE;
6600 
6601 	BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE);
6602 
6603 	local_save_flags(irq_flags);
6604 	size = sizeof(*entry) + cnt + 2; /* add '\0' and possible '\n' */
6605 
6606 	/* If less than "<faulted>", then make sure we can still add that */
6607 	if (cnt < FAULTED_SIZE)
6608 		size += FAULTED_SIZE - cnt;
6609 
6610 	buffer = tr->array_buffer.buffer;
6611 	event = __trace_buffer_lock_reserve(buffer, TRACE_PRINT, size,
6612 					    irq_flags, preempt_count());
6613 	if (unlikely(!event))
6614 		/* Ring buffer disabled, return as if not open for write */
6615 		return -EBADF;
6616 
6617 	entry = ring_buffer_event_data(event);
6618 	entry->ip = _THIS_IP_;
6619 
6620 	len = __copy_from_user_inatomic(&entry->buf, ubuf, cnt);
6621 	if (len) {
6622 		memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE);
6623 		cnt = FAULTED_SIZE;
6624 		written = -EFAULT;
6625 	} else
6626 		written = cnt;
6627 	len = cnt;
6628 
6629 	if (tr->trace_marker_file && !list_empty(&tr->trace_marker_file->triggers)) {
6630 		/* do not add \n before testing triggers, but add \0 */
6631 		entry->buf[cnt] = '\0';
6632 		tt = event_triggers_call(tr->trace_marker_file, entry, event);
6633 	}
6634 
6635 	if (entry->buf[cnt - 1] != '\n') {
6636 		entry->buf[cnt] = '\n';
6637 		entry->buf[cnt + 1] = '\0';
6638 	} else
6639 		entry->buf[cnt] = '\0';
6640 
6641 	__buffer_unlock_commit(buffer, event);
6642 
6643 	if (tt)
6644 		event_triggers_post_call(tr->trace_marker_file, tt);
6645 
6646 	if (written > 0)
6647 		*fpos += written;
6648 
6649 	return written;
6650 }
6651 
6652 /* Limit it for now to 3K (including tag) */
6653 #define RAW_DATA_MAX_SIZE (1024*3)
6654 
6655 static ssize_t
6656 tracing_mark_raw_write(struct file *filp, const char __user *ubuf,
6657 					size_t cnt, loff_t *fpos)
6658 {
6659 	struct trace_array *tr = filp->private_data;
6660 	struct ring_buffer_event *event;
6661 	struct trace_buffer *buffer;
6662 	struct raw_data_entry *entry;
6663 	unsigned long irq_flags;
6664 	ssize_t written;
6665 	int size;
6666 	int len;
6667 
6668 #define FAULT_SIZE_ID (FAULTED_SIZE + sizeof(int))
6669 
6670 	if (tracing_disabled)
6671 		return -EINVAL;
6672 
6673 	if (!(tr->trace_flags & TRACE_ITER_MARKERS))
6674 		return -EINVAL;
6675 
6676 	/* The marker must at least have a tag id */
6677 	if (cnt < sizeof(unsigned int) || cnt > RAW_DATA_MAX_SIZE)
6678 		return -EINVAL;
6679 
6680 	if (cnt > TRACE_BUF_SIZE)
6681 		cnt = TRACE_BUF_SIZE;
6682 
6683 	BUILD_BUG_ON(TRACE_BUF_SIZE >= PAGE_SIZE);
6684 
6685 	local_save_flags(irq_flags);
6686 	size = sizeof(*entry) + cnt;
6687 	if (cnt < FAULT_SIZE_ID)
6688 		size += FAULT_SIZE_ID - cnt;
6689 
6690 	buffer = tr->array_buffer.buffer;
6691 	event = __trace_buffer_lock_reserve(buffer, TRACE_RAW_DATA, size,
6692 					    irq_flags, preempt_count());
6693 	if (!event)
6694 		/* Ring buffer disabled, return as if not open for write */
6695 		return -EBADF;
6696 
6697 	entry = ring_buffer_event_data(event);
6698 
6699 	len = __copy_from_user_inatomic(&entry->id, ubuf, cnt);
6700 	if (len) {
6701 		entry->id = -1;
6702 		memcpy(&entry->buf, FAULTED_STR, FAULTED_SIZE);
6703 		written = -EFAULT;
6704 	} else
6705 		written = cnt;
6706 
6707 	__buffer_unlock_commit(buffer, event);
6708 
6709 	if (written > 0)
6710 		*fpos += written;
6711 
6712 	return written;
6713 }
6714 
6715 static int tracing_clock_show(struct seq_file *m, void *v)
6716 {
6717 	struct trace_array *tr = m->private;
6718 	int i;
6719 
6720 	for (i = 0; i < ARRAY_SIZE(trace_clocks); i++)
6721 		seq_printf(m,
6722 			"%s%s%s%s", i ? " " : "",
6723 			i == tr->clock_id ? "[" : "", trace_clocks[i].name,
6724 			i == tr->clock_id ? "]" : "");
6725 	seq_putc(m, '\n');
6726 
6727 	return 0;
6728 }
6729 
6730 int tracing_set_clock(struct trace_array *tr, const char *clockstr)
6731 {
6732 	int i;
6733 
6734 	for (i = 0; i < ARRAY_SIZE(trace_clocks); i++) {
6735 		if (strcmp(trace_clocks[i].name, clockstr) == 0)
6736 			break;
6737 	}
6738 	if (i == ARRAY_SIZE(trace_clocks))
6739 		return -EINVAL;
6740 
6741 	mutex_lock(&trace_types_lock);
6742 
6743 	tr->clock_id = i;
6744 
6745 	ring_buffer_set_clock(tr->array_buffer.buffer, trace_clocks[i].func);
6746 
6747 	/*
6748 	 * New clock may not be consistent with the previous clock.
6749 	 * Reset the buffer so that it doesn't have incomparable timestamps.
6750 	 */
6751 	tracing_reset_online_cpus(&tr->array_buffer);
6752 
6753 #ifdef CONFIG_TRACER_MAX_TRACE
6754 	if (tr->max_buffer.buffer)
6755 		ring_buffer_set_clock(tr->max_buffer.buffer, trace_clocks[i].func);
6756 	tracing_reset_online_cpus(&tr->max_buffer);
6757 #endif
6758 
6759 	mutex_unlock(&trace_types_lock);
6760 
6761 	return 0;
6762 }
6763 
6764 static ssize_t tracing_clock_write(struct file *filp, const char __user *ubuf,
6765 				   size_t cnt, loff_t *fpos)
6766 {
6767 	struct seq_file *m = filp->private_data;
6768 	struct trace_array *tr = m->private;
6769 	char buf[64];
6770 	const char *clockstr;
6771 	int ret;
6772 
6773 	if (cnt >= sizeof(buf))
6774 		return -EINVAL;
6775 
6776 	if (copy_from_user(buf, ubuf, cnt))
6777 		return -EFAULT;
6778 
6779 	buf[cnt] = 0;
6780 
6781 	clockstr = strstrip(buf);
6782 
6783 	ret = tracing_set_clock(tr, clockstr);
6784 	if (ret)
6785 		return ret;
6786 
6787 	*fpos += cnt;
6788 
6789 	return cnt;
6790 }
6791 
6792 static int tracing_clock_open(struct inode *inode, struct file *file)
6793 {
6794 	struct trace_array *tr = inode->i_private;
6795 	int ret;
6796 
6797 	ret = tracing_check_open_get_tr(tr);
6798 	if (ret)
6799 		return ret;
6800 
6801 	ret = single_open(file, tracing_clock_show, inode->i_private);
6802 	if (ret < 0)
6803 		trace_array_put(tr);
6804 
6805 	return ret;
6806 }
6807 
6808 static int tracing_time_stamp_mode_show(struct seq_file *m, void *v)
6809 {
6810 	struct trace_array *tr = m->private;
6811 
6812 	mutex_lock(&trace_types_lock);
6813 
6814 	if (ring_buffer_time_stamp_abs(tr->array_buffer.buffer))
6815 		seq_puts(m, "delta [absolute]\n");
6816 	else
6817 		seq_puts(m, "[delta] absolute\n");
6818 
6819 	mutex_unlock(&trace_types_lock);
6820 
6821 	return 0;
6822 }
6823 
6824 static int tracing_time_stamp_mode_open(struct inode *inode, struct file *file)
6825 {
6826 	struct trace_array *tr = inode->i_private;
6827 	int ret;
6828 
6829 	ret = tracing_check_open_get_tr(tr);
6830 	if (ret)
6831 		return ret;
6832 
6833 	ret = single_open(file, tracing_time_stamp_mode_show, inode->i_private);
6834 	if (ret < 0)
6835 		trace_array_put(tr);
6836 
6837 	return ret;
6838 }
6839 
6840 int tracing_set_time_stamp_abs(struct trace_array *tr, bool abs)
6841 {
6842 	int ret = 0;
6843 
6844 	mutex_lock(&trace_types_lock);
6845 
6846 	if (abs && tr->time_stamp_abs_ref++)
6847 		goto out;
6848 
6849 	if (!abs) {
6850 		if (WARN_ON_ONCE(!tr->time_stamp_abs_ref)) {
6851 			ret = -EINVAL;
6852 			goto out;
6853 		}
6854 
6855 		if (--tr->time_stamp_abs_ref)
6856 			goto out;
6857 	}
6858 
6859 	ring_buffer_set_time_stamp_abs(tr->array_buffer.buffer, abs);
6860 
6861 #ifdef CONFIG_TRACER_MAX_TRACE
6862 	if (tr->max_buffer.buffer)
6863 		ring_buffer_set_time_stamp_abs(tr->max_buffer.buffer, abs);
6864 #endif
6865  out:
6866 	mutex_unlock(&trace_types_lock);
6867 
6868 	return ret;
6869 }
6870 
6871 struct ftrace_buffer_info {
6872 	struct trace_iterator	iter;
6873 	void			*spare;
6874 	unsigned int		spare_cpu;
6875 	unsigned int		read;
6876 };
6877 
6878 #ifdef CONFIG_TRACER_SNAPSHOT
6879 static int tracing_snapshot_open(struct inode *inode, struct file *file)
6880 {
6881 	struct trace_array *tr = inode->i_private;
6882 	struct trace_iterator *iter;
6883 	struct seq_file *m;
6884 	int ret;
6885 
6886 	ret = tracing_check_open_get_tr(tr);
6887 	if (ret)
6888 		return ret;
6889 
6890 	if (file->f_mode & FMODE_READ) {
6891 		iter = __tracing_open(inode, file, true);
6892 		if (IS_ERR(iter))
6893 			ret = PTR_ERR(iter);
6894 	} else {
6895 		/* Writes still need the seq_file to hold the private data */
6896 		ret = -ENOMEM;
6897 		m = kzalloc(sizeof(*m), GFP_KERNEL);
6898 		if (!m)
6899 			goto out;
6900 		iter = kzalloc(sizeof(*iter), GFP_KERNEL);
6901 		if (!iter) {
6902 			kfree(m);
6903 			goto out;
6904 		}
6905 		ret = 0;
6906 
6907 		iter->tr = tr;
6908 		iter->array_buffer = &tr->max_buffer;
6909 		iter->cpu_file = tracing_get_cpu(inode);
6910 		m->private = iter;
6911 		file->private_data = m;
6912 	}
6913 out:
6914 	if (ret < 0)
6915 		trace_array_put(tr);
6916 
6917 	return ret;
6918 }
6919 
6920 static ssize_t
6921 tracing_snapshot_write(struct file *filp, const char __user *ubuf, size_t cnt,
6922 		       loff_t *ppos)
6923 {
6924 	struct seq_file *m = filp->private_data;
6925 	struct trace_iterator *iter = m->private;
6926 	struct trace_array *tr = iter->tr;
6927 	unsigned long val;
6928 	int ret;
6929 
6930 	ret = tracing_update_buffers();
6931 	if (ret < 0)
6932 		return ret;
6933 
6934 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
6935 	if (ret)
6936 		return ret;
6937 
6938 	mutex_lock(&trace_types_lock);
6939 
6940 	if (tr->current_trace->use_max_tr) {
6941 		ret = -EBUSY;
6942 		goto out;
6943 	}
6944 
6945 	arch_spin_lock(&tr->max_lock);
6946 	if (tr->cond_snapshot)
6947 		ret = -EBUSY;
6948 	arch_spin_unlock(&tr->max_lock);
6949 	if (ret)
6950 		goto out;
6951 
6952 	switch (val) {
6953 	case 0:
6954 		if (iter->cpu_file != RING_BUFFER_ALL_CPUS) {
6955 			ret = -EINVAL;
6956 			break;
6957 		}
6958 		if (tr->allocated_snapshot)
6959 			free_snapshot(tr);
6960 		break;
6961 	case 1:
6962 /* Only allow per-cpu swap if the ring buffer supports it */
6963 #ifndef CONFIG_RING_BUFFER_ALLOW_SWAP
6964 		if (iter->cpu_file != RING_BUFFER_ALL_CPUS) {
6965 			ret = -EINVAL;
6966 			break;
6967 		}
6968 #endif
6969 		if (tr->allocated_snapshot)
6970 			ret = resize_buffer_duplicate_size(&tr->max_buffer,
6971 					&tr->array_buffer, iter->cpu_file);
6972 		else
6973 			ret = tracing_alloc_snapshot_instance(tr);
6974 		if (ret < 0)
6975 			break;
6976 		local_irq_disable();
6977 		/* Now, we're going to swap */
6978 		if (iter->cpu_file == RING_BUFFER_ALL_CPUS)
6979 			update_max_tr(tr, current, smp_processor_id(), NULL);
6980 		else
6981 			update_max_tr_single(tr, current, iter->cpu_file);
6982 		local_irq_enable();
6983 		break;
6984 	default:
6985 		if (tr->allocated_snapshot) {
6986 			if (iter->cpu_file == RING_BUFFER_ALL_CPUS)
6987 				tracing_reset_online_cpus(&tr->max_buffer);
6988 			else
6989 				tracing_reset_cpu(&tr->max_buffer, iter->cpu_file);
6990 		}
6991 		break;
6992 	}
6993 
6994 	if (ret >= 0) {
6995 		*ppos += cnt;
6996 		ret = cnt;
6997 	}
6998 out:
6999 	mutex_unlock(&trace_types_lock);
7000 	return ret;
7001 }
7002 
7003 static int tracing_snapshot_release(struct inode *inode, struct file *file)
7004 {
7005 	struct seq_file *m = file->private_data;
7006 	int ret;
7007 
7008 	ret = tracing_release(inode, file);
7009 
7010 	if (file->f_mode & FMODE_READ)
7011 		return ret;
7012 
7013 	/* If write only, the seq_file is just a stub */
7014 	if (m)
7015 		kfree(m->private);
7016 	kfree(m);
7017 
7018 	return 0;
7019 }
7020 
7021 static int tracing_buffers_open(struct inode *inode, struct file *filp);
7022 static ssize_t tracing_buffers_read(struct file *filp, char __user *ubuf,
7023 				    size_t count, loff_t *ppos);
7024 static int tracing_buffers_release(struct inode *inode, struct file *file);
7025 static ssize_t tracing_buffers_splice_read(struct file *file, loff_t *ppos,
7026 		   struct pipe_inode_info *pipe, size_t len, unsigned int flags);
7027 
7028 static int snapshot_raw_open(struct inode *inode, struct file *filp)
7029 {
7030 	struct ftrace_buffer_info *info;
7031 	int ret;
7032 
7033 	/* The following checks for tracefs lockdown */
7034 	ret = tracing_buffers_open(inode, filp);
7035 	if (ret < 0)
7036 		return ret;
7037 
7038 	info = filp->private_data;
7039 
7040 	if (info->iter.trace->use_max_tr) {
7041 		tracing_buffers_release(inode, filp);
7042 		return -EBUSY;
7043 	}
7044 
7045 	info->iter.snapshot = true;
7046 	info->iter.array_buffer = &info->iter.tr->max_buffer;
7047 
7048 	return ret;
7049 }
7050 
7051 #endif /* CONFIG_TRACER_SNAPSHOT */
7052 
7053 
7054 static const struct file_operations tracing_thresh_fops = {
7055 	.open		= tracing_open_generic,
7056 	.read		= tracing_thresh_read,
7057 	.write		= tracing_thresh_write,
7058 	.llseek		= generic_file_llseek,
7059 };
7060 
7061 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
7062 static const struct file_operations tracing_max_lat_fops = {
7063 	.open		= tracing_open_generic,
7064 	.read		= tracing_max_lat_read,
7065 	.write		= tracing_max_lat_write,
7066 	.llseek		= generic_file_llseek,
7067 };
7068 #endif
7069 
7070 static const struct file_operations set_tracer_fops = {
7071 	.open		= tracing_open_generic,
7072 	.read		= tracing_set_trace_read,
7073 	.write		= tracing_set_trace_write,
7074 	.llseek		= generic_file_llseek,
7075 };
7076 
7077 static const struct file_operations tracing_pipe_fops = {
7078 	.open		= tracing_open_pipe,
7079 	.poll		= tracing_poll_pipe,
7080 	.read		= tracing_read_pipe,
7081 	.splice_read	= tracing_splice_read_pipe,
7082 	.release	= tracing_release_pipe,
7083 	.llseek		= no_llseek,
7084 };
7085 
7086 static const struct file_operations tracing_entries_fops = {
7087 	.open		= tracing_open_generic_tr,
7088 	.read		= tracing_entries_read,
7089 	.write		= tracing_entries_write,
7090 	.llseek		= generic_file_llseek,
7091 	.release	= tracing_release_generic_tr,
7092 };
7093 
7094 static const struct file_operations tracing_total_entries_fops = {
7095 	.open		= tracing_open_generic_tr,
7096 	.read		= tracing_total_entries_read,
7097 	.llseek		= generic_file_llseek,
7098 	.release	= tracing_release_generic_tr,
7099 };
7100 
7101 static const struct file_operations tracing_free_buffer_fops = {
7102 	.open		= tracing_open_generic_tr,
7103 	.write		= tracing_free_buffer_write,
7104 	.release	= tracing_free_buffer_release,
7105 };
7106 
7107 static const struct file_operations tracing_mark_fops = {
7108 	.open		= tracing_open_generic_tr,
7109 	.write		= tracing_mark_write,
7110 	.llseek		= generic_file_llseek,
7111 	.release	= tracing_release_generic_tr,
7112 };
7113 
7114 static const struct file_operations tracing_mark_raw_fops = {
7115 	.open		= tracing_open_generic_tr,
7116 	.write		= tracing_mark_raw_write,
7117 	.llseek		= generic_file_llseek,
7118 	.release	= tracing_release_generic_tr,
7119 };
7120 
7121 static const struct file_operations trace_clock_fops = {
7122 	.open		= tracing_clock_open,
7123 	.read		= seq_read,
7124 	.llseek		= seq_lseek,
7125 	.release	= tracing_single_release_tr,
7126 	.write		= tracing_clock_write,
7127 };
7128 
7129 static const struct file_operations trace_time_stamp_mode_fops = {
7130 	.open		= tracing_time_stamp_mode_open,
7131 	.read		= seq_read,
7132 	.llseek		= seq_lseek,
7133 	.release	= tracing_single_release_tr,
7134 };
7135 
7136 #ifdef CONFIG_TRACER_SNAPSHOT
7137 static const struct file_operations snapshot_fops = {
7138 	.open		= tracing_snapshot_open,
7139 	.read		= seq_read,
7140 	.write		= tracing_snapshot_write,
7141 	.llseek		= tracing_lseek,
7142 	.release	= tracing_snapshot_release,
7143 };
7144 
7145 static const struct file_operations snapshot_raw_fops = {
7146 	.open		= snapshot_raw_open,
7147 	.read		= tracing_buffers_read,
7148 	.release	= tracing_buffers_release,
7149 	.splice_read	= tracing_buffers_splice_read,
7150 	.llseek		= no_llseek,
7151 };
7152 
7153 #endif /* CONFIG_TRACER_SNAPSHOT */
7154 
7155 #define TRACING_LOG_ERRS_MAX	8
7156 #define TRACING_LOG_LOC_MAX	128
7157 
7158 #define CMD_PREFIX "  Command: "
7159 
7160 struct err_info {
7161 	const char	**errs;	/* ptr to loc-specific array of err strings */
7162 	u8		type;	/* index into errs -> specific err string */
7163 	u8		pos;	/* MAX_FILTER_STR_VAL = 256 */
7164 	u64		ts;
7165 };
7166 
7167 struct tracing_log_err {
7168 	struct list_head	list;
7169 	struct err_info		info;
7170 	char			loc[TRACING_LOG_LOC_MAX]; /* err location */
7171 	char			cmd[MAX_FILTER_STR_VAL]; /* what caused err */
7172 };
7173 
7174 static DEFINE_MUTEX(tracing_err_log_lock);
7175 
7176 static struct tracing_log_err *get_tracing_log_err(struct trace_array *tr)
7177 {
7178 	struct tracing_log_err *err;
7179 
7180 	if (tr->n_err_log_entries < TRACING_LOG_ERRS_MAX) {
7181 		err = kzalloc(sizeof(*err), GFP_KERNEL);
7182 		if (!err)
7183 			err = ERR_PTR(-ENOMEM);
7184 		tr->n_err_log_entries++;
7185 
7186 		return err;
7187 	}
7188 
7189 	err = list_first_entry(&tr->err_log, struct tracing_log_err, list);
7190 	list_del(&err->list);
7191 
7192 	return err;
7193 }
7194 
7195 /**
7196  * err_pos - find the position of a string within a command for error careting
7197  * @cmd: The tracing command that caused the error
7198  * @str: The string to position the caret at within @cmd
7199  *
7200  * Finds the position of the first occurence of @str within @cmd.  The
7201  * return value can be passed to tracing_log_err() for caret placement
7202  * within @cmd.
7203  *
7204  * Returns the index within @cmd of the first occurence of @str or 0
7205  * if @str was not found.
7206  */
7207 unsigned int err_pos(char *cmd, const char *str)
7208 {
7209 	char *found;
7210 
7211 	if (WARN_ON(!strlen(cmd)))
7212 		return 0;
7213 
7214 	found = strstr(cmd, str);
7215 	if (found)
7216 		return found - cmd;
7217 
7218 	return 0;
7219 }
7220 
7221 /**
7222  * tracing_log_err - write an error to the tracing error log
7223  * @tr: The associated trace array for the error (NULL for top level array)
7224  * @loc: A string describing where the error occurred
7225  * @cmd: The tracing command that caused the error
7226  * @errs: The array of loc-specific static error strings
7227  * @type: The index into errs[], which produces the specific static err string
7228  * @pos: The position the caret should be placed in the cmd
7229  *
7230  * Writes an error into tracing/error_log of the form:
7231  *
7232  * <loc>: error: <text>
7233  *   Command: <cmd>
7234  *              ^
7235  *
7236  * tracing/error_log is a small log file containing the last
7237  * TRACING_LOG_ERRS_MAX errors (8).  Memory for errors isn't allocated
7238  * unless there has been a tracing error, and the error log can be
7239  * cleared and have its memory freed by writing the empty string in
7240  * truncation mode to it i.e. echo > tracing/error_log.
7241  *
7242  * NOTE: the @errs array along with the @type param are used to
7243  * produce a static error string - this string is not copied and saved
7244  * when the error is logged - only a pointer to it is saved.  See
7245  * existing callers for examples of how static strings are typically
7246  * defined for use with tracing_log_err().
7247  */
7248 void tracing_log_err(struct trace_array *tr,
7249 		     const char *loc, const char *cmd,
7250 		     const char **errs, u8 type, u8 pos)
7251 {
7252 	struct tracing_log_err *err;
7253 
7254 	if (!tr)
7255 		tr = &global_trace;
7256 
7257 	mutex_lock(&tracing_err_log_lock);
7258 	err = get_tracing_log_err(tr);
7259 	if (PTR_ERR(err) == -ENOMEM) {
7260 		mutex_unlock(&tracing_err_log_lock);
7261 		return;
7262 	}
7263 
7264 	snprintf(err->loc, TRACING_LOG_LOC_MAX, "%s: error: ", loc);
7265 	snprintf(err->cmd, MAX_FILTER_STR_VAL,"\n" CMD_PREFIX "%s\n", cmd);
7266 
7267 	err->info.errs = errs;
7268 	err->info.type = type;
7269 	err->info.pos = pos;
7270 	err->info.ts = local_clock();
7271 
7272 	list_add_tail(&err->list, &tr->err_log);
7273 	mutex_unlock(&tracing_err_log_lock);
7274 }
7275 
7276 static void clear_tracing_err_log(struct trace_array *tr)
7277 {
7278 	struct tracing_log_err *err, *next;
7279 
7280 	mutex_lock(&tracing_err_log_lock);
7281 	list_for_each_entry_safe(err, next, &tr->err_log, list) {
7282 		list_del(&err->list);
7283 		kfree(err);
7284 	}
7285 
7286 	tr->n_err_log_entries = 0;
7287 	mutex_unlock(&tracing_err_log_lock);
7288 }
7289 
7290 static void *tracing_err_log_seq_start(struct seq_file *m, loff_t *pos)
7291 {
7292 	struct trace_array *tr = m->private;
7293 
7294 	mutex_lock(&tracing_err_log_lock);
7295 
7296 	return seq_list_start(&tr->err_log, *pos);
7297 }
7298 
7299 static void *tracing_err_log_seq_next(struct seq_file *m, void *v, loff_t *pos)
7300 {
7301 	struct trace_array *tr = m->private;
7302 
7303 	return seq_list_next(v, &tr->err_log, pos);
7304 }
7305 
7306 static void tracing_err_log_seq_stop(struct seq_file *m, void *v)
7307 {
7308 	mutex_unlock(&tracing_err_log_lock);
7309 }
7310 
7311 static void tracing_err_log_show_pos(struct seq_file *m, u8 pos)
7312 {
7313 	u8 i;
7314 
7315 	for (i = 0; i < sizeof(CMD_PREFIX) - 1; i++)
7316 		seq_putc(m, ' ');
7317 	for (i = 0; i < pos; i++)
7318 		seq_putc(m, ' ');
7319 	seq_puts(m, "^\n");
7320 }
7321 
7322 static int tracing_err_log_seq_show(struct seq_file *m, void *v)
7323 {
7324 	struct tracing_log_err *err = v;
7325 
7326 	if (err) {
7327 		const char *err_text = err->info.errs[err->info.type];
7328 		u64 sec = err->info.ts;
7329 		u32 nsec;
7330 
7331 		nsec = do_div(sec, NSEC_PER_SEC);
7332 		seq_printf(m, "[%5llu.%06u] %s%s", sec, nsec / 1000,
7333 			   err->loc, err_text);
7334 		seq_printf(m, "%s", err->cmd);
7335 		tracing_err_log_show_pos(m, err->info.pos);
7336 	}
7337 
7338 	return 0;
7339 }
7340 
7341 static const struct seq_operations tracing_err_log_seq_ops = {
7342 	.start  = tracing_err_log_seq_start,
7343 	.next   = tracing_err_log_seq_next,
7344 	.stop   = tracing_err_log_seq_stop,
7345 	.show   = tracing_err_log_seq_show
7346 };
7347 
7348 static int tracing_err_log_open(struct inode *inode, struct file *file)
7349 {
7350 	struct trace_array *tr = inode->i_private;
7351 	int ret = 0;
7352 
7353 	ret = tracing_check_open_get_tr(tr);
7354 	if (ret)
7355 		return ret;
7356 
7357 	/* If this file was opened for write, then erase contents */
7358 	if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC))
7359 		clear_tracing_err_log(tr);
7360 
7361 	if (file->f_mode & FMODE_READ) {
7362 		ret = seq_open(file, &tracing_err_log_seq_ops);
7363 		if (!ret) {
7364 			struct seq_file *m = file->private_data;
7365 			m->private = tr;
7366 		} else {
7367 			trace_array_put(tr);
7368 		}
7369 	}
7370 	return ret;
7371 }
7372 
7373 static ssize_t tracing_err_log_write(struct file *file,
7374 				     const char __user *buffer,
7375 				     size_t count, loff_t *ppos)
7376 {
7377 	return count;
7378 }
7379 
7380 static int tracing_err_log_release(struct inode *inode, struct file *file)
7381 {
7382 	struct trace_array *tr = inode->i_private;
7383 
7384 	trace_array_put(tr);
7385 
7386 	if (file->f_mode & FMODE_READ)
7387 		seq_release(inode, file);
7388 
7389 	return 0;
7390 }
7391 
7392 static const struct file_operations tracing_err_log_fops = {
7393 	.open           = tracing_err_log_open,
7394 	.write		= tracing_err_log_write,
7395 	.read           = seq_read,
7396 	.llseek         = seq_lseek,
7397 	.release        = tracing_err_log_release,
7398 };
7399 
7400 static int tracing_buffers_open(struct inode *inode, struct file *filp)
7401 {
7402 	struct trace_array *tr = inode->i_private;
7403 	struct ftrace_buffer_info *info;
7404 	int ret;
7405 
7406 	ret = tracing_check_open_get_tr(tr);
7407 	if (ret)
7408 		return ret;
7409 
7410 	info = kzalloc(sizeof(*info), GFP_KERNEL);
7411 	if (!info) {
7412 		trace_array_put(tr);
7413 		return -ENOMEM;
7414 	}
7415 
7416 	mutex_lock(&trace_types_lock);
7417 
7418 	info->iter.tr		= tr;
7419 	info->iter.cpu_file	= tracing_get_cpu(inode);
7420 	info->iter.trace	= tr->current_trace;
7421 	info->iter.array_buffer = &tr->array_buffer;
7422 	info->spare		= NULL;
7423 	/* Force reading ring buffer for first read */
7424 	info->read		= (unsigned int)-1;
7425 
7426 	filp->private_data = info;
7427 
7428 	tr->current_trace->ref++;
7429 
7430 	mutex_unlock(&trace_types_lock);
7431 
7432 	ret = nonseekable_open(inode, filp);
7433 	if (ret < 0)
7434 		trace_array_put(tr);
7435 
7436 	return ret;
7437 }
7438 
7439 static __poll_t
7440 tracing_buffers_poll(struct file *filp, poll_table *poll_table)
7441 {
7442 	struct ftrace_buffer_info *info = filp->private_data;
7443 	struct trace_iterator *iter = &info->iter;
7444 
7445 	return trace_poll(iter, filp, poll_table);
7446 }
7447 
7448 static ssize_t
7449 tracing_buffers_read(struct file *filp, char __user *ubuf,
7450 		     size_t count, loff_t *ppos)
7451 {
7452 	struct ftrace_buffer_info *info = filp->private_data;
7453 	struct trace_iterator *iter = &info->iter;
7454 	ssize_t ret = 0;
7455 	ssize_t size;
7456 
7457 	if (!count)
7458 		return 0;
7459 
7460 #ifdef CONFIG_TRACER_MAX_TRACE
7461 	if (iter->snapshot && iter->tr->current_trace->use_max_tr)
7462 		return -EBUSY;
7463 #endif
7464 
7465 	if (!info->spare) {
7466 		info->spare = ring_buffer_alloc_read_page(iter->array_buffer->buffer,
7467 							  iter->cpu_file);
7468 		if (IS_ERR(info->spare)) {
7469 			ret = PTR_ERR(info->spare);
7470 			info->spare = NULL;
7471 		} else {
7472 			info->spare_cpu = iter->cpu_file;
7473 		}
7474 	}
7475 	if (!info->spare)
7476 		return ret;
7477 
7478 	/* Do we have previous read data to read? */
7479 	if (info->read < PAGE_SIZE)
7480 		goto read;
7481 
7482  again:
7483 	trace_access_lock(iter->cpu_file);
7484 	ret = ring_buffer_read_page(iter->array_buffer->buffer,
7485 				    &info->spare,
7486 				    count,
7487 				    iter->cpu_file, 0);
7488 	trace_access_unlock(iter->cpu_file);
7489 
7490 	if (ret < 0) {
7491 		if (trace_empty(iter)) {
7492 			if ((filp->f_flags & O_NONBLOCK))
7493 				return -EAGAIN;
7494 
7495 			ret = wait_on_pipe(iter, 0);
7496 			if (ret)
7497 				return ret;
7498 
7499 			goto again;
7500 		}
7501 		return 0;
7502 	}
7503 
7504 	info->read = 0;
7505  read:
7506 	size = PAGE_SIZE - info->read;
7507 	if (size > count)
7508 		size = count;
7509 
7510 	ret = copy_to_user(ubuf, info->spare + info->read, size);
7511 	if (ret == size)
7512 		return -EFAULT;
7513 
7514 	size -= ret;
7515 
7516 	*ppos += size;
7517 	info->read += size;
7518 
7519 	return size;
7520 }
7521 
7522 static int tracing_buffers_release(struct inode *inode, struct file *file)
7523 {
7524 	struct ftrace_buffer_info *info = file->private_data;
7525 	struct trace_iterator *iter = &info->iter;
7526 
7527 	mutex_lock(&trace_types_lock);
7528 
7529 	iter->tr->current_trace->ref--;
7530 
7531 	__trace_array_put(iter->tr);
7532 
7533 	if (info->spare)
7534 		ring_buffer_free_read_page(iter->array_buffer->buffer,
7535 					   info->spare_cpu, info->spare);
7536 	kfree(info);
7537 
7538 	mutex_unlock(&trace_types_lock);
7539 
7540 	return 0;
7541 }
7542 
7543 struct buffer_ref {
7544 	struct trace_buffer	*buffer;
7545 	void			*page;
7546 	int			cpu;
7547 	refcount_t		refcount;
7548 };
7549 
7550 static void buffer_ref_release(struct buffer_ref *ref)
7551 {
7552 	if (!refcount_dec_and_test(&ref->refcount))
7553 		return;
7554 	ring_buffer_free_read_page(ref->buffer, ref->cpu, ref->page);
7555 	kfree(ref);
7556 }
7557 
7558 static void buffer_pipe_buf_release(struct pipe_inode_info *pipe,
7559 				    struct pipe_buffer *buf)
7560 {
7561 	struct buffer_ref *ref = (struct buffer_ref *)buf->private;
7562 
7563 	buffer_ref_release(ref);
7564 	buf->private = 0;
7565 }
7566 
7567 static bool buffer_pipe_buf_get(struct pipe_inode_info *pipe,
7568 				struct pipe_buffer *buf)
7569 {
7570 	struct buffer_ref *ref = (struct buffer_ref *)buf->private;
7571 
7572 	if (refcount_read(&ref->refcount) > INT_MAX/2)
7573 		return false;
7574 
7575 	refcount_inc(&ref->refcount);
7576 	return true;
7577 }
7578 
7579 /* Pipe buffer operations for a buffer. */
7580 static const struct pipe_buf_operations buffer_pipe_buf_ops = {
7581 	.release		= buffer_pipe_buf_release,
7582 	.get			= buffer_pipe_buf_get,
7583 };
7584 
7585 /*
7586  * Callback from splice_to_pipe(), if we need to release some pages
7587  * at the end of the spd in case we error'ed out in filling the pipe.
7588  */
7589 static void buffer_spd_release(struct splice_pipe_desc *spd, unsigned int i)
7590 {
7591 	struct buffer_ref *ref =
7592 		(struct buffer_ref *)spd->partial[i].private;
7593 
7594 	buffer_ref_release(ref);
7595 	spd->partial[i].private = 0;
7596 }
7597 
7598 static ssize_t
7599 tracing_buffers_splice_read(struct file *file, loff_t *ppos,
7600 			    struct pipe_inode_info *pipe, size_t len,
7601 			    unsigned int flags)
7602 {
7603 	struct ftrace_buffer_info *info = file->private_data;
7604 	struct trace_iterator *iter = &info->iter;
7605 	struct partial_page partial_def[PIPE_DEF_BUFFERS];
7606 	struct page *pages_def[PIPE_DEF_BUFFERS];
7607 	struct splice_pipe_desc spd = {
7608 		.pages		= pages_def,
7609 		.partial	= partial_def,
7610 		.nr_pages_max	= PIPE_DEF_BUFFERS,
7611 		.ops		= &buffer_pipe_buf_ops,
7612 		.spd_release	= buffer_spd_release,
7613 	};
7614 	struct buffer_ref *ref;
7615 	int entries, i;
7616 	ssize_t ret = 0;
7617 
7618 #ifdef CONFIG_TRACER_MAX_TRACE
7619 	if (iter->snapshot && iter->tr->current_trace->use_max_tr)
7620 		return -EBUSY;
7621 #endif
7622 
7623 	if (*ppos & (PAGE_SIZE - 1))
7624 		return -EINVAL;
7625 
7626 	if (len & (PAGE_SIZE - 1)) {
7627 		if (len < PAGE_SIZE)
7628 			return -EINVAL;
7629 		len &= PAGE_MASK;
7630 	}
7631 
7632 	if (splice_grow_spd(pipe, &spd))
7633 		return -ENOMEM;
7634 
7635  again:
7636 	trace_access_lock(iter->cpu_file);
7637 	entries = ring_buffer_entries_cpu(iter->array_buffer->buffer, iter->cpu_file);
7638 
7639 	for (i = 0; i < spd.nr_pages_max && len && entries; i++, len -= PAGE_SIZE) {
7640 		struct page *page;
7641 		int r;
7642 
7643 		ref = kzalloc(sizeof(*ref), GFP_KERNEL);
7644 		if (!ref) {
7645 			ret = -ENOMEM;
7646 			break;
7647 		}
7648 
7649 		refcount_set(&ref->refcount, 1);
7650 		ref->buffer = iter->array_buffer->buffer;
7651 		ref->page = ring_buffer_alloc_read_page(ref->buffer, iter->cpu_file);
7652 		if (IS_ERR(ref->page)) {
7653 			ret = PTR_ERR(ref->page);
7654 			ref->page = NULL;
7655 			kfree(ref);
7656 			break;
7657 		}
7658 		ref->cpu = iter->cpu_file;
7659 
7660 		r = ring_buffer_read_page(ref->buffer, &ref->page,
7661 					  len, iter->cpu_file, 1);
7662 		if (r < 0) {
7663 			ring_buffer_free_read_page(ref->buffer, ref->cpu,
7664 						   ref->page);
7665 			kfree(ref);
7666 			break;
7667 		}
7668 
7669 		page = virt_to_page(ref->page);
7670 
7671 		spd.pages[i] = page;
7672 		spd.partial[i].len = PAGE_SIZE;
7673 		spd.partial[i].offset = 0;
7674 		spd.partial[i].private = (unsigned long)ref;
7675 		spd.nr_pages++;
7676 		*ppos += PAGE_SIZE;
7677 
7678 		entries = ring_buffer_entries_cpu(iter->array_buffer->buffer, iter->cpu_file);
7679 	}
7680 
7681 	trace_access_unlock(iter->cpu_file);
7682 	spd.nr_pages = i;
7683 
7684 	/* did we read anything? */
7685 	if (!spd.nr_pages) {
7686 		if (ret)
7687 			goto out;
7688 
7689 		ret = -EAGAIN;
7690 		if ((file->f_flags & O_NONBLOCK) || (flags & SPLICE_F_NONBLOCK))
7691 			goto out;
7692 
7693 		ret = wait_on_pipe(iter, iter->tr->buffer_percent);
7694 		if (ret)
7695 			goto out;
7696 
7697 		goto again;
7698 	}
7699 
7700 	ret = splice_to_pipe(pipe, &spd);
7701 out:
7702 	splice_shrink_spd(&spd);
7703 
7704 	return ret;
7705 }
7706 
7707 static const struct file_operations tracing_buffers_fops = {
7708 	.open		= tracing_buffers_open,
7709 	.read		= tracing_buffers_read,
7710 	.poll		= tracing_buffers_poll,
7711 	.release	= tracing_buffers_release,
7712 	.splice_read	= tracing_buffers_splice_read,
7713 	.llseek		= no_llseek,
7714 };
7715 
7716 static ssize_t
7717 tracing_stats_read(struct file *filp, char __user *ubuf,
7718 		   size_t count, loff_t *ppos)
7719 {
7720 	struct inode *inode = file_inode(filp);
7721 	struct trace_array *tr = inode->i_private;
7722 	struct array_buffer *trace_buf = &tr->array_buffer;
7723 	int cpu = tracing_get_cpu(inode);
7724 	struct trace_seq *s;
7725 	unsigned long cnt;
7726 	unsigned long long t;
7727 	unsigned long usec_rem;
7728 
7729 	s = kmalloc(sizeof(*s), GFP_KERNEL);
7730 	if (!s)
7731 		return -ENOMEM;
7732 
7733 	trace_seq_init(s);
7734 
7735 	cnt = ring_buffer_entries_cpu(trace_buf->buffer, cpu);
7736 	trace_seq_printf(s, "entries: %ld\n", cnt);
7737 
7738 	cnt = ring_buffer_overrun_cpu(trace_buf->buffer, cpu);
7739 	trace_seq_printf(s, "overrun: %ld\n", cnt);
7740 
7741 	cnt = ring_buffer_commit_overrun_cpu(trace_buf->buffer, cpu);
7742 	trace_seq_printf(s, "commit overrun: %ld\n", cnt);
7743 
7744 	cnt = ring_buffer_bytes_cpu(trace_buf->buffer, cpu);
7745 	trace_seq_printf(s, "bytes: %ld\n", cnt);
7746 
7747 	if (trace_clocks[tr->clock_id].in_ns) {
7748 		/* local or global for trace_clock */
7749 		t = ns2usecs(ring_buffer_oldest_event_ts(trace_buf->buffer, cpu));
7750 		usec_rem = do_div(t, USEC_PER_SEC);
7751 		trace_seq_printf(s, "oldest event ts: %5llu.%06lu\n",
7752 								t, usec_rem);
7753 
7754 		t = ns2usecs(ring_buffer_time_stamp(trace_buf->buffer, cpu));
7755 		usec_rem = do_div(t, USEC_PER_SEC);
7756 		trace_seq_printf(s, "now ts: %5llu.%06lu\n", t, usec_rem);
7757 	} else {
7758 		/* counter or tsc mode for trace_clock */
7759 		trace_seq_printf(s, "oldest event ts: %llu\n",
7760 				ring_buffer_oldest_event_ts(trace_buf->buffer, cpu));
7761 
7762 		trace_seq_printf(s, "now ts: %llu\n",
7763 				ring_buffer_time_stamp(trace_buf->buffer, cpu));
7764 	}
7765 
7766 	cnt = ring_buffer_dropped_events_cpu(trace_buf->buffer, cpu);
7767 	trace_seq_printf(s, "dropped events: %ld\n", cnt);
7768 
7769 	cnt = ring_buffer_read_events_cpu(trace_buf->buffer, cpu);
7770 	trace_seq_printf(s, "read events: %ld\n", cnt);
7771 
7772 	count = simple_read_from_buffer(ubuf, count, ppos,
7773 					s->buffer, trace_seq_used(s));
7774 
7775 	kfree(s);
7776 
7777 	return count;
7778 }
7779 
7780 static const struct file_operations tracing_stats_fops = {
7781 	.open		= tracing_open_generic_tr,
7782 	.read		= tracing_stats_read,
7783 	.llseek		= generic_file_llseek,
7784 	.release	= tracing_release_generic_tr,
7785 };
7786 
7787 #ifdef CONFIG_DYNAMIC_FTRACE
7788 
7789 static ssize_t
7790 tracing_read_dyn_info(struct file *filp, char __user *ubuf,
7791 		  size_t cnt, loff_t *ppos)
7792 {
7793 	ssize_t ret;
7794 	char *buf;
7795 	int r;
7796 
7797 	/* 256 should be plenty to hold the amount needed */
7798 	buf = kmalloc(256, GFP_KERNEL);
7799 	if (!buf)
7800 		return -ENOMEM;
7801 
7802 	r = scnprintf(buf, 256, "%ld pages:%ld groups: %ld\n",
7803 		      ftrace_update_tot_cnt,
7804 		      ftrace_number_of_pages,
7805 		      ftrace_number_of_groups);
7806 
7807 	ret = simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
7808 	kfree(buf);
7809 	return ret;
7810 }
7811 
7812 static const struct file_operations tracing_dyn_info_fops = {
7813 	.open		= tracing_open_generic,
7814 	.read		= tracing_read_dyn_info,
7815 	.llseek		= generic_file_llseek,
7816 };
7817 #endif /* CONFIG_DYNAMIC_FTRACE */
7818 
7819 #if defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE)
7820 static void
7821 ftrace_snapshot(unsigned long ip, unsigned long parent_ip,
7822 		struct trace_array *tr, struct ftrace_probe_ops *ops,
7823 		void *data)
7824 {
7825 	tracing_snapshot_instance(tr);
7826 }
7827 
7828 static void
7829 ftrace_count_snapshot(unsigned long ip, unsigned long parent_ip,
7830 		      struct trace_array *tr, struct ftrace_probe_ops *ops,
7831 		      void *data)
7832 {
7833 	struct ftrace_func_mapper *mapper = data;
7834 	long *count = NULL;
7835 
7836 	if (mapper)
7837 		count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
7838 
7839 	if (count) {
7840 
7841 		if (*count <= 0)
7842 			return;
7843 
7844 		(*count)--;
7845 	}
7846 
7847 	tracing_snapshot_instance(tr);
7848 }
7849 
7850 static int
7851 ftrace_snapshot_print(struct seq_file *m, unsigned long ip,
7852 		      struct ftrace_probe_ops *ops, void *data)
7853 {
7854 	struct ftrace_func_mapper *mapper = data;
7855 	long *count = NULL;
7856 
7857 	seq_printf(m, "%ps:", (void *)ip);
7858 
7859 	seq_puts(m, "snapshot");
7860 
7861 	if (mapper)
7862 		count = (long *)ftrace_func_mapper_find_ip(mapper, ip);
7863 
7864 	if (count)
7865 		seq_printf(m, ":count=%ld\n", *count);
7866 	else
7867 		seq_puts(m, ":unlimited\n");
7868 
7869 	return 0;
7870 }
7871 
7872 static int
7873 ftrace_snapshot_init(struct ftrace_probe_ops *ops, struct trace_array *tr,
7874 		     unsigned long ip, void *init_data, void **data)
7875 {
7876 	struct ftrace_func_mapper *mapper = *data;
7877 
7878 	if (!mapper) {
7879 		mapper = allocate_ftrace_func_mapper();
7880 		if (!mapper)
7881 			return -ENOMEM;
7882 		*data = mapper;
7883 	}
7884 
7885 	return ftrace_func_mapper_add_ip(mapper, ip, init_data);
7886 }
7887 
7888 static void
7889 ftrace_snapshot_free(struct ftrace_probe_ops *ops, struct trace_array *tr,
7890 		     unsigned long ip, void *data)
7891 {
7892 	struct ftrace_func_mapper *mapper = data;
7893 
7894 	if (!ip) {
7895 		if (!mapper)
7896 			return;
7897 		free_ftrace_func_mapper(mapper, NULL);
7898 		return;
7899 	}
7900 
7901 	ftrace_func_mapper_remove_ip(mapper, ip);
7902 }
7903 
7904 static struct ftrace_probe_ops snapshot_probe_ops = {
7905 	.func			= ftrace_snapshot,
7906 	.print			= ftrace_snapshot_print,
7907 };
7908 
7909 static struct ftrace_probe_ops snapshot_count_probe_ops = {
7910 	.func			= ftrace_count_snapshot,
7911 	.print			= ftrace_snapshot_print,
7912 	.init			= ftrace_snapshot_init,
7913 	.free			= ftrace_snapshot_free,
7914 };
7915 
7916 static int
7917 ftrace_trace_snapshot_callback(struct trace_array *tr, struct ftrace_hash *hash,
7918 			       char *glob, char *cmd, char *param, int enable)
7919 {
7920 	struct ftrace_probe_ops *ops;
7921 	void *count = (void *)-1;
7922 	char *number;
7923 	int ret;
7924 
7925 	if (!tr)
7926 		return -ENODEV;
7927 
7928 	/* hash funcs only work with set_ftrace_filter */
7929 	if (!enable)
7930 		return -EINVAL;
7931 
7932 	ops = param ? &snapshot_count_probe_ops :  &snapshot_probe_ops;
7933 
7934 	if (glob[0] == '!')
7935 		return unregister_ftrace_function_probe_func(glob+1, tr, ops);
7936 
7937 	if (!param)
7938 		goto out_reg;
7939 
7940 	number = strsep(&param, ":");
7941 
7942 	if (!strlen(number))
7943 		goto out_reg;
7944 
7945 	/*
7946 	 * We use the callback data field (which is a pointer)
7947 	 * as our counter.
7948 	 */
7949 	ret = kstrtoul(number, 0, (unsigned long *)&count);
7950 	if (ret)
7951 		return ret;
7952 
7953  out_reg:
7954 	ret = tracing_alloc_snapshot_instance(tr);
7955 	if (ret < 0)
7956 		goto out;
7957 
7958 	ret = register_ftrace_function_probe(glob, tr, ops, count);
7959 
7960  out:
7961 	return ret < 0 ? ret : 0;
7962 }
7963 
7964 static struct ftrace_func_command ftrace_snapshot_cmd = {
7965 	.name			= "snapshot",
7966 	.func			= ftrace_trace_snapshot_callback,
7967 };
7968 
7969 static __init int register_snapshot_cmd(void)
7970 {
7971 	return register_ftrace_command(&ftrace_snapshot_cmd);
7972 }
7973 #else
7974 static inline __init int register_snapshot_cmd(void) { return 0; }
7975 #endif /* defined(CONFIG_TRACER_SNAPSHOT) && defined(CONFIG_DYNAMIC_FTRACE) */
7976 
7977 static struct dentry *tracing_get_dentry(struct trace_array *tr)
7978 {
7979 	if (WARN_ON(!tr->dir))
7980 		return ERR_PTR(-ENODEV);
7981 
7982 	/* Top directory uses NULL as the parent */
7983 	if (tr->flags & TRACE_ARRAY_FL_GLOBAL)
7984 		return NULL;
7985 
7986 	/* All sub buffers have a descriptor */
7987 	return tr->dir;
7988 }
7989 
7990 static struct dentry *tracing_dentry_percpu(struct trace_array *tr, int cpu)
7991 {
7992 	struct dentry *d_tracer;
7993 
7994 	if (tr->percpu_dir)
7995 		return tr->percpu_dir;
7996 
7997 	d_tracer = tracing_get_dentry(tr);
7998 	if (IS_ERR(d_tracer))
7999 		return NULL;
8000 
8001 	tr->percpu_dir = tracefs_create_dir("per_cpu", d_tracer);
8002 
8003 	MEM_FAIL(!tr->percpu_dir,
8004 		  "Could not create tracefs directory 'per_cpu/%d'\n", cpu);
8005 
8006 	return tr->percpu_dir;
8007 }
8008 
8009 static struct dentry *
8010 trace_create_cpu_file(const char *name, umode_t mode, struct dentry *parent,
8011 		      void *data, long cpu, const struct file_operations *fops)
8012 {
8013 	struct dentry *ret = trace_create_file(name, mode, parent, data, fops);
8014 
8015 	if (ret) /* See tracing_get_cpu() */
8016 		d_inode(ret)->i_cdev = (void *)(cpu + 1);
8017 	return ret;
8018 }
8019 
8020 static void
8021 tracing_init_tracefs_percpu(struct trace_array *tr, long cpu)
8022 {
8023 	struct dentry *d_percpu = tracing_dentry_percpu(tr, cpu);
8024 	struct dentry *d_cpu;
8025 	char cpu_dir[30]; /* 30 characters should be more than enough */
8026 
8027 	if (!d_percpu)
8028 		return;
8029 
8030 	snprintf(cpu_dir, 30, "cpu%ld", cpu);
8031 	d_cpu = tracefs_create_dir(cpu_dir, d_percpu);
8032 	if (!d_cpu) {
8033 		pr_warn("Could not create tracefs '%s' entry\n", cpu_dir);
8034 		return;
8035 	}
8036 
8037 	/* per cpu trace_pipe */
8038 	trace_create_cpu_file("trace_pipe", 0444, d_cpu,
8039 				tr, cpu, &tracing_pipe_fops);
8040 
8041 	/* per cpu trace */
8042 	trace_create_cpu_file("trace", 0644, d_cpu,
8043 				tr, cpu, &tracing_fops);
8044 
8045 	trace_create_cpu_file("trace_pipe_raw", 0444, d_cpu,
8046 				tr, cpu, &tracing_buffers_fops);
8047 
8048 	trace_create_cpu_file("stats", 0444, d_cpu,
8049 				tr, cpu, &tracing_stats_fops);
8050 
8051 	trace_create_cpu_file("buffer_size_kb", 0444, d_cpu,
8052 				tr, cpu, &tracing_entries_fops);
8053 
8054 #ifdef CONFIG_TRACER_SNAPSHOT
8055 	trace_create_cpu_file("snapshot", 0644, d_cpu,
8056 				tr, cpu, &snapshot_fops);
8057 
8058 	trace_create_cpu_file("snapshot_raw", 0444, d_cpu,
8059 				tr, cpu, &snapshot_raw_fops);
8060 #endif
8061 }
8062 
8063 #ifdef CONFIG_FTRACE_SELFTEST
8064 /* Let selftest have access to static functions in this file */
8065 #include "trace_selftest.c"
8066 #endif
8067 
8068 static ssize_t
8069 trace_options_read(struct file *filp, char __user *ubuf, size_t cnt,
8070 			loff_t *ppos)
8071 {
8072 	struct trace_option_dentry *topt = filp->private_data;
8073 	char *buf;
8074 
8075 	if (topt->flags->val & topt->opt->bit)
8076 		buf = "1\n";
8077 	else
8078 		buf = "0\n";
8079 
8080 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, 2);
8081 }
8082 
8083 static ssize_t
8084 trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt,
8085 			 loff_t *ppos)
8086 {
8087 	struct trace_option_dentry *topt = filp->private_data;
8088 	unsigned long val;
8089 	int ret;
8090 
8091 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
8092 	if (ret)
8093 		return ret;
8094 
8095 	if (val != 0 && val != 1)
8096 		return -EINVAL;
8097 
8098 	if (!!(topt->flags->val & topt->opt->bit) != val) {
8099 		mutex_lock(&trace_types_lock);
8100 		ret = __set_tracer_option(topt->tr, topt->flags,
8101 					  topt->opt, !val);
8102 		mutex_unlock(&trace_types_lock);
8103 		if (ret)
8104 			return ret;
8105 	}
8106 
8107 	*ppos += cnt;
8108 
8109 	return cnt;
8110 }
8111 
8112 
8113 static const struct file_operations trace_options_fops = {
8114 	.open = tracing_open_generic,
8115 	.read = trace_options_read,
8116 	.write = trace_options_write,
8117 	.llseek	= generic_file_llseek,
8118 };
8119 
8120 /*
8121  * In order to pass in both the trace_array descriptor as well as the index
8122  * to the flag that the trace option file represents, the trace_array
8123  * has a character array of trace_flags_index[], which holds the index
8124  * of the bit for the flag it represents. index[0] == 0, index[1] == 1, etc.
8125  * The address of this character array is passed to the flag option file
8126  * read/write callbacks.
8127  *
8128  * In order to extract both the index and the trace_array descriptor,
8129  * get_tr_index() uses the following algorithm.
8130  *
8131  *   idx = *ptr;
8132  *
8133  * As the pointer itself contains the address of the index (remember
8134  * index[1] == 1).
8135  *
8136  * Then to get the trace_array descriptor, by subtracting that index
8137  * from the ptr, we get to the start of the index itself.
8138  *
8139  *   ptr - idx == &index[0]
8140  *
8141  * Then a simple container_of() from that pointer gets us to the
8142  * trace_array descriptor.
8143  */
8144 static void get_tr_index(void *data, struct trace_array **ptr,
8145 			 unsigned int *pindex)
8146 {
8147 	*pindex = *(unsigned char *)data;
8148 
8149 	*ptr = container_of(data - *pindex, struct trace_array,
8150 			    trace_flags_index);
8151 }
8152 
8153 static ssize_t
8154 trace_options_core_read(struct file *filp, char __user *ubuf, size_t cnt,
8155 			loff_t *ppos)
8156 {
8157 	void *tr_index = filp->private_data;
8158 	struct trace_array *tr;
8159 	unsigned int index;
8160 	char *buf;
8161 
8162 	get_tr_index(tr_index, &tr, &index);
8163 
8164 	if (tr->trace_flags & (1 << index))
8165 		buf = "1\n";
8166 	else
8167 		buf = "0\n";
8168 
8169 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, 2);
8170 }
8171 
8172 static ssize_t
8173 trace_options_core_write(struct file *filp, const char __user *ubuf, size_t cnt,
8174 			 loff_t *ppos)
8175 {
8176 	void *tr_index = filp->private_data;
8177 	struct trace_array *tr;
8178 	unsigned int index;
8179 	unsigned long val;
8180 	int ret;
8181 
8182 	get_tr_index(tr_index, &tr, &index);
8183 
8184 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
8185 	if (ret)
8186 		return ret;
8187 
8188 	if (val != 0 && val != 1)
8189 		return -EINVAL;
8190 
8191 	mutex_lock(&event_mutex);
8192 	mutex_lock(&trace_types_lock);
8193 	ret = set_tracer_flag(tr, 1 << index, val);
8194 	mutex_unlock(&trace_types_lock);
8195 	mutex_unlock(&event_mutex);
8196 
8197 	if (ret < 0)
8198 		return ret;
8199 
8200 	*ppos += cnt;
8201 
8202 	return cnt;
8203 }
8204 
8205 static const struct file_operations trace_options_core_fops = {
8206 	.open = tracing_open_generic,
8207 	.read = trace_options_core_read,
8208 	.write = trace_options_core_write,
8209 	.llseek = generic_file_llseek,
8210 };
8211 
8212 struct dentry *trace_create_file(const char *name,
8213 				 umode_t mode,
8214 				 struct dentry *parent,
8215 				 void *data,
8216 				 const struct file_operations *fops)
8217 {
8218 	struct dentry *ret;
8219 
8220 	ret = tracefs_create_file(name, mode, parent, data, fops);
8221 	if (!ret)
8222 		pr_warn("Could not create tracefs '%s' entry\n", name);
8223 
8224 	return ret;
8225 }
8226 
8227 
8228 static struct dentry *trace_options_init_dentry(struct trace_array *tr)
8229 {
8230 	struct dentry *d_tracer;
8231 
8232 	if (tr->options)
8233 		return tr->options;
8234 
8235 	d_tracer = tracing_get_dentry(tr);
8236 	if (IS_ERR(d_tracer))
8237 		return NULL;
8238 
8239 	tr->options = tracefs_create_dir("options", d_tracer);
8240 	if (!tr->options) {
8241 		pr_warn("Could not create tracefs directory 'options'\n");
8242 		return NULL;
8243 	}
8244 
8245 	return tr->options;
8246 }
8247 
8248 static void
8249 create_trace_option_file(struct trace_array *tr,
8250 			 struct trace_option_dentry *topt,
8251 			 struct tracer_flags *flags,
8252 			 struct tracer_opt *opt)
8253 {
8254 	struct dentry *t_options;
8255 
8256 	t_options = trace_options_init_dentry(tr);
8257 	if (!t_options)
8258 		return;
8259 
8260 	topt->flags = flags;
8261 	topt->opt = opt;
8262 	topt->tr = tr;
8263 
8264 	topt->entry = trace_create_file(opt->name, 0644, t_options, topt,
8265 				    &trace_options_fops);
8266 
8267 }
8268 
8269 static void
8270 create_trace_option_files(struct trace_array *tr, struct tracer *tracer)
8271 {
8272 	struct trace_option_dentry *topts;
8273 	struct trace_options *tr_topts;
8274 	struct tracer_flags *flags;
8275 	struct tracer_opt *opts;
8276 	int cnt;
8277 	int i;
8278 
8279 	if (!tracer)
8280 		return;
8281 
8282 	flags = tracer->flags;
8283 
8284 	if (!flags || !flags->opts)
8285 		return;
8286 
8287 	/*
8288 	 * If this is an instance, only create flags for tracers
8289 	 * the instance may have.
8290 	 */
8291 	if (!trace_ok_for_array(tracer, tr))
8292 		return;
8293 
8294 	for (i = 0; i < tr->nr_topts; i++) {
8295 		/* Make sure there's no duplicate flags. */
8296 		if (WARN_ON_ONCE(tr->topts[i].tracer->flags == tracer->flags))
8297 			return;
8298 	}
8299 
8300 	opts = flags->opts;
8301 
8302 	for (cnt = 0; opts[cnt].name; cnt++)
8303 		;
8304 
8305 	topts = kcalloc(cnt + 1, sizeof(*topts), GFP_KERNEL);
8306 	if (!topts)
8307 		return;
8308 
8309 	tr_topts = krealloc(tr->topts, sizeof(*tr->topts) * (tr->nr_topts + 1),
8310 			    GFP_KERNEL);
8311 	if (!tr_topts) {
8312 		kfree(topts);
8313 		return;
8314 	}
8315 
8316 	tr->topts = tr_topts;
8317 	tr->topts[tr->nr_topts].tracer = tracer;
8318 	tr->topts[tr->nr_topts].topts = topts;
8319 	tr->nr_topts++;
8320 
8321 	for (cnt = 0; opts[cnt].name; cnt++) {
8322 		create_trace_option_file(tr, &topts[cnt], flags,
8323 					 &opts[cnt]);
8324 		MEM_FAIL(topts[cnt].entry == NULL,
8325 			  "Failed to create trace option: %s",
8326 			  opts[cnt].name);
8327 	}
8328 }
8329 
8330 static struct dentry *
8331 create_trace_option_core_file(struct trace_array *tr,
8332 			      const char *option, long index)
8333 {
8334 	struct dentry *t_options;
8335 
8336 	t_options = trace_options_init_dentry(tr);
8337 	if (!t_options)
8338 		return NULL;
8339 
8340 	return trace_create_file(option, 0644, t_options,
8341 				 (void *)&tr->trace_flags_index[index],
8342 				 &trace_options_core_fops);
8343 }
8344 
8345 static void create_trace_options_dir(struct trace_array *tr)
8346 {
8347 	struct dentry *t_options;
8348 	bool top_level = tr == &global_trace;
8349 	int i;
8350 
8351 	t_options = trace_options_init_dentry(tr);
8352 	if (!t_options)
8353 		return;
8354 
8355 	for (i = 0; trace_options[i]; i++) {
8356 		if (top_level ||
8357 		    !((1 << i) & TOP_LEVEL_TRACE_FLAGS))
8358 			create_trace_option_core_file(tr, trace_options[i], i);
8359 	}
8360 }
8361 
8362 static ssize_t
8363 rb_simple_read(struct file *filp, char __user *ubuf,
8364 	       size_t cnt, loff_t *ppos)
8365 {
8366 	struct trace_array *tr = filp->private_data;
8367 	char buf[64];
8368 	int r;
8369 
8370 	r = tracer_tracing_is_on(tr);
8371 	r = sprintf(buf, "%d\n", r);
8372 
8373 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
8374 }
8375 
8376 static ssize_t
8377 rb_simple_write(struct file *filp, const char __user *ubuf,
8378 		size_t cnt, loff_t *ppos)
8379 {
8380 	struct trace_array *tr = filp->private_data;
8381 	struct trace_buffer *buffer = tr->array_buffer.buffer;
8382 	unsigned long val;
8383 	int ret;
8384 
8385 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
8386 	if (ret)
8387 		return ret;
8388 
8389 	if (buffer) {
8390 		mutex_lock(&trace_types_lock);
8391 		if (!!val == tracer_tracing_is_on(tr)) {
8392 			val = 0; /* do nothing */
8393 		} else if (val) {
8394 			tracer_tracing_on(tr);
8395 			if (tr->current_trace->start)
8396 				tr->current_trace->start(tr);
8397 		} else {
8398 			tracer_tracing_off(tr);
8399 			if (tr->current_trace->stop)
8400 				tr->current_trace->stop(tr);
8401 		}
8402 		mutex_unlock(&trace_types_lock);
8403 	}
8404 
8405 	(*ppos)++;
8406 
8407 	return cnt;
8408 }
8409 
8410 static const struct file_operations rb_simple_fops = {
8411 	.open		= tracing_open_generic_tr,
8412 	.read		= rb_simple_read,
8413 	.write		= rb_simple_write,
8414 	.release	= tracing_release_generic_tr,
8415 	.llseek		= default_llseek,
8416 };
8417 
8418 static ssize_t
8419 buffer_percent_read(struct file *filp, char __user *ubuf,
8420 		    size_t cnt, loff_t *ppos)
8421 {
8422 	struct trace_array *tr = filp->private_data;
8423 	char buf[64];
8424 	int r;
8425 
8426 	r = tr->buffer_percent;
8427 	r = sprintf(buf, "%d\n", r);
8428 
8429 	return simple_read_from_buffer(ubuf, cnt, ppos, buf, r);
8430 }
8431 
8432 static ssize_t
8433 buffer_percent_write(struct file *filp, const char __user *ubuf,
8434 		     size_t cnt, loff_t *ppos)
8435 {
8436 	struct trace_array *tr = filp->private_data;
8437 	unsigned long val;
8438 	int ret;
8439 
8440 	ret = kstrtoul_from_user(ubuf, cnt, 10, &val);
8441 	if (ret)
8442 		return ret;
8443 
8444 	if (val > 100)
8445 		return -EINVAL;
8446 
8447 	if (!val)
8448 		val = 1;
8449 
8450 	tr->buffer_percent = val;
8451 
8452 	(*ppos)++;
8453 
8454 	return cnt;
8455 }
8456 
8457 static const struct file_operations buffer_percent_fops = {
8458 	.open		= tracing_open_generic_tr,
8459 	.read		= buffer_percent_read,
8460 	.write		= buffer_percent_write,
8461 	.release	= tracing_release_generic_tr,
8462 	.llseek		= default_llseek,
8463 };
8464 
8465 static struct dentry *trace_instance_dir;
8466 
8467 static void
8468 init_tracer_tracefs(struct trace_array *tr, struct dentry *d_tracer);
8469 
8470 static int
8471 allocate_trace_buffer(struct trace_array *tr, struct array_buffer *buf, int size)
8472 {
8473 	enum ring_buffer_flags rb_flags;
8474 
8475 	rb_flags = tr->trace_flags & TRACE_ITER_OVERWRITE ? RB_FL_OVERWRITE : 0;
8476 
8477 	buf->tr = tr;
8478 
8479 	buf->buffer = ring_buffer_alloc(size, rb_flags);
8480 	if (!buf->buffer)
8481 		return -ENOMEM;
8482 
8483 	buf->data = alloc_percpu(struct trace_array_cpu);
8484 	if (!buf->data) {
8485 		ring_buffer_free(buf->buffer);
8486 		buf->buffer = NULL;
8487 		return -ENOMEM;
8488 	}
8489 
8490 	/* Allocate the first page for all buffers */
8491 	set_buffer_entries(&tr->array_buffer,
8492 			   ring_buffer_size(tr->array_buffer.buffer, 0));
8493 
8494 	return 0;
8495 }
8496 
8497 static int allocate_trace_buffers(struct trace_array *tr, int size)
8498 {
8499 	int ret;
8500 
8501 	ret = allocate_trace_buffer(tr, &tr->array_buffer, size);
8502 	if (ret)
8503 		return ret;
8504 
8505 #ifdef CONFIG_TRACER_MAX_TRACE
8506 	ret = allocate_trace_buffer(tr, &tr->max_buffer,
8507 				    allocate_snapshot ? size : 1);
8508 	if (MEM_FAIL(ret, "Failed to allocate trace buffer\n")) {
8509 		ring_buffer_free(tr->array_buffer.buffer);
8510 		tr->array_buffer.buffer = NULL;
8511 		free_percpu(tr->array_buffer.data);
8512 		tr->array_buffer.data = NULL;
8513 		return -ENOMEM;
8514 	}
8515 	tr->allocated_snapshot = allocate_snapshot;
8516 
8517 	/*
8518 	 * Only the top level trace array gets its snapshot allocated
8519 	 * from the kernel command line.
8520 	 */
8521 	allocate_snapshot = false;
8522 #endif
8523 
8524 	return 0;
8525 }
8526 
8527 static void free_trace_buffer(struct array_buffer *buf)
8528 {
8529 	if (buf->buffer) {
8530 		ring_buffer_free(buf->buffer);
8531 		buf->buffer = NULL;
8532 		free_percpu(buf->data);
8533 		buf->data = NULL;
8534 	}
8535 }
8536 
8537 static void free_trace_buffers(struct trace_array *tr)
8538 {
8539 	if (!tr)
8540 		return;
8541 
8542 	free_trace_buffer(&tr->array_buffer);
8543 
8544 #ifdef CONFIG_TRACER_MAX_TRACE
8545 	free_trace_buffer(&tr->max_buffer);
8546 #endif
8547 }
8548 
8549 static void init_trace_flags_index(struct trace_array *tr)
8550 {
8551 	int i;
8552 
8553 	/* Used by the trace options files */
8554 	for (i = 0; i < TRACE_FLAGS_MAX_SIZE; i++)
8555 		tr->trace_flags_index[i] = i;
8556 }
8557 
8558 static void __update_tracer_options(struct trace_array *tr)
8559 {
8560 	struct tracer *t;
8561 
8562 	for (t = trace_types; t; t = t->next)
8563 		add_tracer_options(tr, t);
8564 }
8565 
8566 static void update_tracer_options(struct trace_array *tr)
8567 {
8568 	mutex_lock(&trace_types_lock);
8569 	__update_tracer_options(tr);
8570 	mutex_unlock(&trace_types_lock);
8571 }
8572 
8573 /* Must have trace_types_lock held */
8574 struct trace_array *trace_array_find(const char *instance)
8575 {
8576 	struct trace_array *tr, *found = NULL;
8577 
8578 	list_for_each_entry(tr, &ftrace_trace_arrays, list) {
8579 		if (tr->name && strcmp(tr->name, instance) == 0) {
8580 			found = tr;
8581 			break;
8582 		}
8583 	}
8584 
8585 	return found;
8586 }
8587 
8588 struct trace_array *trace_array_find_get(const char *instance)
8589 {
8590 	struct trace_array *tr;
8591 
8592 	mutex_lock(&trace_types_lock);
8593 	tr = trace_array_find(instance);
8594 	if (tr)
8595 		tr->ref++;
8596 	mutex_unlock(&trace_types_lock);
8597 
8598 	return tr;
8599 }
8600 
8601 static struct trace_array *trace_array_create(const char *name)
8602 {
8603 	struct trace_array *tr;
8604 	int ret;
8605 
8606 	ret = -ENOMEM;
8607 	tr = kzalloc(sizeof(*tr), GFP_KERNEL);
8608 	if (!tr)
8609 		return ERR_PTR(ret);
8610 
8611 	tr->name = kstrdup(name, GFP_KERNEL);
8612 	if (!tr->name)
8613 		goto out_free_tr;
8614 
8615 	if (!alloc_cpumask_var(&tr->tracing_cpumask, GFP_KERNEL))
8616 		goto out_free_tr;
8617 
8618 	tr->trace_flags = global_trace.trace_flags & ~ZEROED_TRACE_FLAGS;
8619 
8620 	cpumask_copy(tr->tracing_cpumask, cpu_all_mask);
8621 
8622 	raw_spin_lock_init(&tr->start_lock);
8623 
8624 	tr->max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
8625 
8626 	tr->current_trace = &nop_trace;
8627 
8628 	INIT_LIST_HEAD(&tr->systems);
8629 	INIT_LIST_HEAD(&tr->events);
8630 	INIT_LIST_HEAD(&tr->hist_vars);
8631 	INIT_LIST_HEAD(&tr->err_log);
8632 
8633 	if (allocate_trace_buffers(tr, trace_buf_size) < 0)
8634 		goto out_free_tr;
8635 
8636 	tr->dir = tracefs_create_dir(name, trace_instance_dir);
8637 	if (!tr->dir)
8638 		goto out_free_tr;
8639 
8640 	ret = event_trace_add_tracer(tr->dir, tr);
8641 	if (ret) {
8642 		tracefs_remove(tr->dir);
8643 		goto out_free_tr;
8644 	}
8645 
8646 	ftrace_init_trace_array(tr);
8647 
8648 	init_tracer_tracefs(tr, tr->dir);
8649 	init_trace_flags_index(tr);
8650 	__update_tracer_options(tr);
8651 
8652 	list_add(&tr->list, &ftrace_trace_arrays);
8653 
8654 	tr->ref++;
8655 
8656 
8657 	return tr;
8658 
8659  out_free_tr:
8660 	free_trace_buffers(tr);
8661 	free_cpumask_var(tr->tracing_cpumask);
8662 	kfree(tr->name);
8663 	kfree(tr);
8664 
8665 	return ERR_PTR(ret);
8666 }
8667 
8668 static int instance_mkdir(const char *name)
8669 {
8670 	struct trace_array *tr;
8671 	int ret;
8672 
8673 	mutex_lock(&event_mutex);
8674 	mutex_lock(&trace_types_lock);
8675 
8676 	ret = -EEXIST;
8677 	if (trace_array_find(name))
8678 		goto out_unlock;
8679 
8680 	tr = trace_array_create(name);
8681 
8682 	ret = PTR_ERR_OR_ZERO(tr);
8683 
8684 out_unlock:
8685 	mutex_unlock(&trace_types_lock);
8686 	mutex_unlock(&event_mutex);
8687 	return ret;
8688 }
8689 
8690 /**
8691  * trace_array_get_by_name - Create/Lookup a trace array, given its name.
8692  * @name: The name of the trace array to be looked up/created.
8693  *
8694  * Returns pointer to trace array with given name.
8695  * NULL, if it cannot be created.
8696  *
8697  * NOTE: This function increments the reference counter associated with the
8698  * trace array returned. This makes sure it cannot be freed while in use.
8699  * Use trace_array_put() once the trace array is no longer needed.
8700  * If the trace_array is to be freed, trace_array_destroy() needs to
8701  * be called after the trace_array_put(), or simply let user space delete
8702  * it from the tracefs instances directory. But until the
8703  * trace_array_put() is called, user space can not delete it.
8704  *
8705  */
8706 struct trace_array *trace_array_get_by_name(const char *name)
8707 {
8708 	struct trace_array *tr;
8709 
8710 	mutex_lock(&event_mutex);
8711 	mutex_lock(&trace_types_lock);
8712 
8713 	list_for_each_entry(tr, &ftrace_trace_arrays, list) {
8714 		if (tr->name && strcmp(tr->name, name) == 0)
8715 			goto out_unlock;
8716 	}
8717 
8718 	tr = trace_array_create(name);
8719 
8720 	if (IS_ERR(tr))
8721 		tr = NULL;
8722 out_unlock:
8723 	if (tr)
8724 		tr->ref++;
8725 
8726 	mutex_unlock(&trace_types_lock);
8727 	mutex_unlock(&event_mutex);
8728 	return tr;
8729 }
8730 EXPORT_SYMBOL_GPL(trace_array_get_by_name);
8731 
8732 static int __remove_instance(struct trace_array *tr)
8733 {
8734 	int i;
8735 
8736 	/* Reference counter for a newly created trace array = 1. */
8737 	if (tr->ref > 1 || (tr->current_trace && tr->current_trace->ref))
8738 		return -EBUSY;
8739 
8740 	list_del(&tr->list);
8741 
8742 	/* Disable all the flags that were enabled coming in */
8743 	for (i = 0; i < TRACE_FLAGS_MAX_SIZE; i++) {
8744 		if ((1 << i) & ZEROED_TRACE_FLAGS)
8745 			set_tracer_flag(tr, 1 << i, 0);
8746 	}
8747 
8748 	tracing_set_nop(tr);
8749 	clear_ftrace_function_probes(tr);
8750 	event_trace_del_tracer(tr);
8751 	ftrace_clear_pids(tr);
8752 	ftrace_destroy_function_files(tr);
8753 	tracefs_remove(tr->dir);
8754 	free_trace_buffers(tr);
8755 
8756 	for (i = 0; i < tr->nr_topts; i++) {
8757 		kfree(tr->topts[i].topts);
8758 	}
8759 	kfree(tr->topts);
8760 
8761 	free_cpumask_var(tr->tracing_cpumask);
8762 	kfree(tr->name);
8763 	kfree(tr);
8764 	tr = NULL;
8765 
8766 	return 0;
8767 }
8768 
8769 int trace_array_destroy(struct trace_array *this_tr)
8770 {
8771 	struct trace_array *tr;
8772 	int ret;
8773 
8774 	if (!this_tr)
8775 		return -EINVAL;
8776 
8777 	mutex_lock(&event_mutex);
8778 	mutex_lock(&trace_types_lock);
8779 
8780 	ret = -ENODEV;
8781 
8782 	/* Making sure trace array exists before destroying it. */
8783 	list_for_each_entry(tr, &ftrace_trace_arrays, list) {
8784 		if (tr == this_tr) {
8785 			ret = __remove_instance(tr);
8786 			break;
8787 		}
8788 	}
8789 
8790 	mutex_unlock(&trace_types_lock);
8791 	mutex_unlock(&event_mutex);
8792 
8793 	return ret;
8794 }
8795 EXPORT_SYMBOL_GPL(trace_array_destroy);
8796 
8797 static int instance_rmdir(const char *name)
8798 {
8799 	struct trace_array *tr;
8800 	int ret;
8801 
8802 	mutex_lock(&event_mutex);
8803 	mutex_lock(&trace_types_lock);
8804 
8805 	ret = -ENODEV;
8806 	tr = trace_array_find(name);
8807 	if (tr)
8808 		ret = __remove_instance(tr);
8809 
8810 	mutex_unlock(&trace_types_lock);
8811 	mutex_unlock(&event_mutex);
8812 
8813 	return ret;
8814 }
8815 
8816 static __init void create_trace_instances(struct dentry *d_tracer)
8817 {
8818 	trace_instance_dir = tracefs_create_instance_dir("instances", d_tracer,
8819 							 instance_mkdir,
8820 							 instance_rmdir);
8821 	if (MEM_FAIL(!trace_instance_dir, "Failed to create instances directory\n"))
8822 		return;
8823 }
8824 
8825 static void
8826 init_tracer_tracefs(struct trace_array *tr, struct dentry *d_tracer)
8827 {
8828 	struct trace_event_file *file;
8829 	int cpu;
8830 
8831 	trace_create_file("available_tracers", 0444, d_tracer,
8832 			tr, &show_traces_fops);
8833 
8834 	trace_create_file("current_tracer", 0644, d_tracer,
8835 			tr, &set_tracer_fops);
8836 
8837 	trace_create_file("tracing_cpumask", 0644, d_tracer,
8838 			  tr, &tracing_cpumask_fops);
8839 
8840 	trace_create_file("trace_options", 0644, d_tracer,
8841 			  tr, &tracing_iter_fops);
8842 
8843 	trace_create_file("trace", 0644, d_tracer,
8844 			  tr, &tracing_fops);
8845 
8846 	trace_create_file("trace_pipe", 0444, d_tracer,
8847 			  tr, &tracing_pipe_fops);
8848 
8849 	trace_create_file("buffer_size_kb", 0644, d_tracer,
8850 			  tr, &tracing_entries_fops);
8851 
8852 	trace_create_file("buffer_total_size_kb", 0444, d_tracer,
8853 			  tr, &tracing_total_entries_fops);
8854 
8855 	trace_create_file("free_buffer", 0200, d_tracer,
8856 			  tr, &tracing_free_buffer_fops);
8857 
8858 	trace_create_file("trace_marker", 0220, d_tracer,
8859 			  tr, &tracing_mark_fops);
8860 
8861 	file = __find_event_file(tr, "ftrace", "print");
8862 	if (file && file->dir)
8863 		trace_create_file("trigger", 0644, file->dir, file,
8864 				  &event_trigger_fops);
8865 	tr->trace_marker_file = file;
8866 
8867 	trace_create_file("trace_marker_raw", 0220, d_tracer,
8868 			  tr, &tracing_mark_raw_fops);
8869 
8870 	trace_create_file("trace_clock", 0644, d_tracer, tr,
8871 			  &trace_clock_fops);
8872 
8873 	trace_create_file("tracing_on", 0644, d_tracer,
8874 			  tr, &rb_simple_fops);
8875 
8876 	trace_create_file("timestamp_mode", 0444, d_tracer, tr,
8877 			  &trace_time_stamp_mode_fops);
8878 
8879 	tr->buffer_percent = 50;
8880 
8881 	trace_create_file("buffer_percent", 0444, d_tracer,
8882 			tr, &buffer_percent_fops);
8883 
8884 	create_trace_options_dir(tr);
8885 
8886 #if defined(CONFIG_TRACER_MAX_TRACE) || defined(CONFIG_HWLAT_TRACER)
8887 	trace_create_maxlat_file(tr, d_tracer);
8888 #endif
8889 
8890 	if (ftrace_create_function_files(tr, d_tracer))
8891 		MEM_FAIL(1, "Could not allocate function filter files");
8892 
8893 #ifdef CONFIG_TRACER_SNAPSHOT
8894 	trace_create_file("snapshot", 0644, d_tracer,
8895 			  tr, &snapshot_fops);
8896 #endif
8897 
8898 	trace_create_file("error_log", 0644, d_tracer,
8899 			  tr, &tracing_err_log_fops);
8900 
8901 	for_each_tracing_cpu(cpu)
8902 		tracing_init_tracefs_percpu(tr, cpu);
8903 
8904 	ftrace_init_tracefs(tr, d_tracer);
8905 }
8906 
8907 static struct vfsmount *trace_automount(struct dentry *mntpt, void *ingore)
8908 {
8909 	struct vfsmount *mnt;
8910 	struct file_system_type *type;
8911 
8912 	/*
8913 	 * To maintain backward compatibility for tools that mount
8914 	 * debugfs to get to the tracing facility, tracefs is automatically
8915 	 * mounted to the debugfs/tracing directory.
8916 	 */
8917 	type = get_fs_type("tracefs");
8918 	if (!type)
8919 		return NULL;
8920 	mnt = vfs_submount(mntpt, type, "tracefs", NULL);
8921 	put_filesystem(type);
8922 	if (IS_ERR(mnt))
8923 		return NULL;
8924 	mntget(mnt);
8925 
8926 	return mnt;
8927 }
8928 
8929 /**
8930  * tracing_init_dentry - initialize top level trace array
8931  *
8932  * This is called when creating files or directories in the tracing
8933  * directory. It is called via fs_initcall() by any of the boot up code
8934  * and expects to return the dentry of the top level tracing directory.
8935  */
8936 struct dentry *tracing_init_dentry(void)
8937 {
8938 	struct trace_array *tr = &global_trace;
8939 
8940 	if (security_locked_down(LOCKDOWN_TRACEFS)) {
8941 		pr_warn("Tracing disabled due to lockdown\n");
8942 		return ERR_PTR(-EPERM);
8943 	}
8944 
8945 	/* The top level trace array uses  NULL as parent */
8946 	if (tr->dir)
8947 		return NULL;
8948 
8949 	if (WARN_ON(!tracefs_initialized()) ||
8950 		(IS_ENABLED(CONFIG_DEBUG_FS) &&
8951 		 WARN_ON(!debugfs_initialized())))
8952 		return ERR_PTR(-ENODEV);
8953 
8954 	/*
8955 	 * As there may still be users that expect the tracing
8956 	 * files to exist in debugfs/tracing, we must automount
8957 	 * the tracefs file system there, so older tools still
8958 	 * work with the newer kerenl.
8959 	 */
8960 	tr->dir = debugfs_create_automount("tracing", NULL,
8961 					   trace_automount, NULL);
8962 
8963 	return NULL;
8964 }
8965 
8966 extern struct trace_eval_map *__start_ftrace_eval_maps[];
8967 extern struct trace_eval_map *__stop_ftrace_eval_maps[];
8968 
8969 static void __init trace_eval_init(void)
8970 {
8971 	int len;
8972 
8973 	len = __stop_ftrace_eval_maps - __start_ftrace_eval_maps;
8974 	trace_insert_eval_map(NULL, __start_ftrace_eval_maps, len);
8975 }
8976 
8977 #ifdef CONFIG_MODULES
8978 static void trace_module_add_evals(struct module *mod)
8979 {
8980 	if (!mod->num_trace_evals)
8981 		return;
8982 
8983 	/*
8984 	 * Modules with bad taint do not have events created, do
8985 	 * not bother with enums either.
8986 	 */
8987 	if (trace_module_has_bad_taint(mod))
8988 		return;
8989 
8990 	trace_insert_eval_map(mod, mod->trace_evals, mod->num_trace_evals);
8991 }
8992 
8993 #ifdef CONFIG_TRACE_EVAL_MAP_FILE
8994 static void trace_module_remove_evals(struct module *mod)
8995 {
8996 	union trace_eval_map_item *map;
8997 	union trace_eval_map_item **last = &trace_eval_maps;
8998 
8999 	if (!mod->num_trace_evals)
9000 		return;
9001 
9002 	mutex_lock(&trace_eval_mutex);
9003 
9004 	map = trace_eval_maps;
9005 
9006 	while (map) {
9007 		if (map->head.mod == mod)
9008 			break;
9009 		map = trace_eval_jmp_to_tail(map);
9010 		last = &map->tail.next;
9011 		map = map->tail.next;
9012 	}
9013 	if (!map)
9014 		goto out;
9015 
9016 	*last = trace_eval_jmp_to_tail(map)->tail.next;
9017 	kfree(map);
9018  out:
9019 	mutex_unlock(&trace_eval_mutex);
9020 }
9021 #else
9022 static inline void trace_module_remove_evals(struct module *mod) { }
9023 #endif /* CONFIG_TRACE_EVAL_MAP_FILE */
9024 
9025 static int trace_module_notify(struct notifier_block *self,
9026 			       unsigned long val, void *data)
9027 {
9028 	struct module *mod = data;
9029 
9030 	switch (val) {
9031 	case MODULE_STATE_COMING:
9032 		trace_module_add_evals(mod);
9033 		break;
9034 	case MODULE_STATE_GOING:
9035 		trace_module_remove_evals(mod);
9036 		break;
9037 	}
9038 
9039 	return 0;
9040 }
9041 
9042 static struct notifier_block trace_module_nb = {
9043 	.notifier_call = trace_module_notify,
9044 	.priority = 0,
9045 };
9046 #endif /* CONFIG_MODULES */
9047 
9048 static __init int tracer_init_tracefs(void)
9049 {
9050 	struct dentry *d_tracer;
9051 
9052 	trace_access_lock_init();
9053 
9054 	d_tracer = tracing_init_dentry();
9055 	if (IS_ERR(d_tracer))
9056 		return 0;
9057 
9058 	event_trace_init();
9059 
9060 	init_tracer_tracefs(&global_trace, d_tracer);
9061 	ftrace_init_tracefs_toplevel(&global_trace, d_tracer);
9062 
9063 	trace_create_file("tracing_thresh", 0644, d_tracer,
9064 			&global_trace, &tracing_thresh_fops);
9065 
9066 	trace_create_file("README", 0444, d_tracer,
9067 			NULL, &tracing_readme_fops);
9068 
9069 	trace_create_file("saved_cmdlines", 0444, d_tracer,
9070 			NULL, &tracing_saved_cmdlines_fops);
9071 
9072 	trace_create_file("saved_cmdlines_size", 0644, d_tracer,
9073 			  NULL, &tracing_saved_cmdlines_size_fops);
9074 
9075 	trace_create_file("saved_tgids", 0444, d_tracer,
9076 			NULL, &tracing_saved_tgids_fops);
9077 
9078 	trace_eval_init();
9079 
9080 	trace_create_eval_file(d_tracer);
9081 
9082 #ifdef CONFIG_MODULES
9083 	register_module_notifier(&trace_module_nb);
9084 #endif
9085 
9086 #ifdef CONFIG_DYNAMIC_FTRACE
9087 	trace_create_file("dyn_ftrace_total_info", 0444, d_tracer,
9088 			NULL, &tracing_dyn_info_fops);
9089 #endif
9090 
9091 	create_trace_instances(d_tracer);
9092 
9093 	update_tracer_options(&global_trace);
9094 
9095 	return 0;
9096 }
9097 
9098 static int trace_panic_handler(struct notifier_block *this,
9099 			       unsigned long event, void *unused)
9100 {
9101 	if (ftrace_dump_on_oops)
9102 		ftrace_dump(ftrace_dump_on_oops);
9103 	return NOTIFY_OK;
9104 }
9105 
9106 static struct notifier_block trace_panic_notifier = {
9107 	.notifier_call  = trace_panic_handler,
9108 	.next           = NULL,
9109 	.priority       = 150   /* priority: INT_MAX >= x >= 0 */
9110 };
9111 
9112 static int trace_die_handler(struct notifier_block *self,
9113 			     unsigned long val,
9114 			     void *data)
9115 {
9116 	switch (val) {
9117 	case DIE_OOPS:
9118 		if (ftrace_dump_on_oops)
9119 			ftrace_dump(ftrace_dump_on_oops);
9120 		break;
9121 	default:
9122 		break;
9123 	}
9124 	return NOTIFY_OK;
9125 }
9126 
9127 static struct notifier_block trace_die_notifier = {
9128 	.notifier_call = trace_die_handler,
9129 	.priority = 200
9130 };
9131 
9132 /*
9133  * printk is set to max of 1024, we really don't need it that big.
9134  * Nothing should be printing 1000 characters anyway.
9135  */
9136 #define TRACE_MAX_PRINT		1000
9137 
9138 /*
9139  * Define here KERN_TRACE so that we have one place to modify
9140  * it if we decide to change what log level the ftrace dump
9141  * should be at.
9142  */
9143 #define KERN_TRACE		KERN_EMERG
9144 
9145 void
9146 trace_printk_seq(struct trace_seq *s)
9147 {
9148 	/* Probably should print a warning here. */
9149 	if (s->seq.len >= TRACE_MAX_PRINT)
9150 		s->seq.len = TRACE_MAX_PRINT;
9151 
9152 	/*
9153 	 * More paranoid code. Although the buffer size is set to
9154 	 * PAGE_SIZE, and TRACE_MAX_PRINT is 1000, this is just
9155 	 * an extra layer of protection.
9156 	 */
9157 	if (WARN_ON_ONCE(s->seq.len >= s->seq.size))
9158 		s->seq.len = s->seq.size - 1;
9159 
9160 	/* should be zero ended, but we are paranoid. */
9161 	s->buffer[s->seq.len] = 0;
9162 
9163 	printk(KERN_TRACE "%s", s->buffer);
9164 
9165 	trace_seq_init(s);
9166 }
9167 
9168 void trace_init_global_iter(struct trace_iterator *iter)
9169 {
9170 	iter->tr = &global_trace;
9171 	iter->trace = iter->tr->current_trace;
9172 	iter->cpu_file = RING_BUFFER_ALL_CPUS;
9173 	iter->array_buffer = &global_trace.array_buffer;
9174 
9175 	if (iter->trace && iter->trace->open)
9176 		iter->trace->open(iter);
9177 
9178 	/* Annotate start of buffers if we had overruns */
9179 	if (ring_buffer_overruns(iter->array_buffer->buffer))
9180 		iter->iter_flags |= TRACE_FILE_ANNOTATE;
9181 
9182 	/* Output in nanoseconds only if we are using a clock in nanoseconds. */
9183 	if (trace_clocks[iter->tr->clock_id].in_ns)
9184 		iter->iter_flags |= TRACE_FILE_TIME_IN_NS;
9185 }
9186 
9187 void ftrace_dump(enum ftrace_dump_mode oops_dump_mode)
9188 {
9189 	/* use static because iter can be a bit big for the stack */
9190 	static struct trace_iterator iter;
9191 	static atomic_t dump_running;
9192 	struct trace_array *tr = &global_trace;
9193 	unsigned int old_userobj;
9194 	unsigned long flags;
9195 	int cnt = 0, cpu;
9196 
9197 	/* Only allow one dump user at a time. */
9198 	if (atomic_inc_return(&dump_running) != 1) {
9199 		atomic_dec(&dump_running);
9200 		return;
9201 	}
9202 
9203 	/*
9204 	 * Always turn off tracing when we dump.
9205 	 * We don't need to show trace output of what happens
9206 	 * between multiple crashes.
9207 	 *
9208 	 * If the user does a sysrq-z, then they can re-enable
9209 	 * tracing with echo 1 > tracing_on.
9210 	 */
9211 	tracing_off();
9212 
9213 	local_irq_save(flags);
9214 	printk_nmi_direct_enter();
9215 
9216 	/* Simulate the iterator */
9217 	trace_init_global_iter(&iter);
9218 	/* Can not use kmalloc for iter.temp */
9219 	iter.temp = static_temp_buf;
9220 	iter.temp_size = STATIC_TEMP_BUF_SIZE;
9221 
9222 	for_each_tracing_cpu(cpu) {
9223 		atomic_inc(&per_cpu_ptr(iter.array_buffer->data, cpu)->disabled);
9224 	}
9225 
9226 	old_userobj = tr->trace_flags & TRACE_ITER_SYM_USEROBJ;
9227 
9228 	/* don't look at user memory in panic mode */
9229 	tr->trace_flags &= ~TRACE_ITER_SYM_USEROBJ;
9230 
9231 	switch (oops_dump_mode) {
9232 	case DUMP_ALL:
9233 		iter.cpu_file = RING_BUFFER_ALL_CPUS;
9234 		break;
9235 	case DUMP_ORIG:
9236 		iter.cpu_file = raw_smp_processor_id();
9237 		break;
9238 	case DUMP_NONE:
9239 		goto out_enable;
9240 	default:
9241 		printk(KERN_TRACE "Bad dumping mode, switching to all CPUs dump\n");
9242 		iter.cpu_file = RING_BUFFER_ALL_CPUS;
9243 	}
9244 
9245 	printk(KERN_TRACE "Dumping ftrace buffer:\n");
9246 
9247 	/* Did function tracer already get disabled? */
9248 	if (ftrace_is_dead()) {
9249 		printk("# WARNING: FUNCTION TRACING IS CORRUPTED\n");
9250 		printk("#          MAY BE MISSING FUNCTION EVENTS\n");
9251 	}
9252 
9253 	/*
9254 	 * We need to stop all tracing on all CPUS to read the
9255 	 * the next buffer. This is a bit expensive, but is
9256 	 * not done often. We fill all what we can read,
9257 	 * and then release the locks again.
9258 	 */
9259 
9260 	while (!trace_empty(&iter)) {
9261 
9262 		if (!cnt)
9263 			printk(KERN_TRACE "---------------------------------\n");
9264 
9265 		cnt++;
9266 
9267 		trace_iterator_reset(&iter);
9268 		iter.iter_flags |= TRACE_FILE_LAT_FMT;
9269 
9270 		if (trace_find_next_entry_inc(&iter) != NULL) {
9271 			int ret;
9272 
9273 			ret = print_trace_line(&iter);
9274 			if (ret != TRACE_TYPE_NO_CONSUME)
9275 				trace_consume(&iter);
9276 		}
9277 		touch_nmi_watchdog();
9278 
9279 		trace_printk_seq(&iter.seq);
9280 	}
9281 
9282 	if (!cnt)
9283 		printk(KERN_TRACE "   (ftrace buffer empty)\n");
9284 	else
9285 		printk(KERN_TRACE "---------------------------------\n");
9286 
9287  out_enable:
9288 	tr->trace_flags |= old_userobj;
9289 
9290 	for_each_tracing_cpu(cpu) {
9291 		atomic_dec(&per_cpu_ptr(iter.array_buffer->data, cpu)->disabled);
9292 	}
9293 	atomic_dec(&dump_running);
9294 	printk_nmi_direct_exit();
9295 	local_irq_restore(flags);
9296 }
9297 EXPORT_SYMBOL_GPL(ftrace_dump);
9298 
9299 int trace_run_command(const char *buf, int (*createfn)(int, char **))
9300 {
9301 	char **argv;
9302 	int argc, ret;
9303 
9304 	argc = 0;
9305 	ret = 0;
9306 	argv = argv_split(GFP_KERNEL, buf, &argc);
9307 	if (!argv)
9308 		return -ENOMEM;
9309 
9310 	if (argc)
9311 		ret = createfn(argc, argv);
9312 
9313 	argv_free(argv);
9314 
9315 	return ret;
9316 }
9317 
9318 #define WRITE_BUFSIZE  4096
9319 
9320 ssize_t trace_parse_run_command(struct file *file, const char __user *buffer,
9321 				size_t count, loff_t *ppos,
9322 				int (*createfn)(int, char **))
9323 {
9324 	char *kbuf, *buf, *tmp;
9325 	int ret = 0;
9326 	size_t done = 0;
9327 	size_t size;
9328 
9329 	kbuf = kmalloc(WRITE_BUFSIZE, GFP_KERNEL);
9330 	if (!kbuf)
9331 		return -ENOMEM;
9332 
9333 	while (done < count) {
9334 		size = count - done;
9335 
9336 		if (size >= WRITE_BUFSIZE)
9337 			size = WRITE_BUFSIZE - 1;
9338 
9339 		if (copy_from_user(kbuf, buffer + done, size)) {
9340 			ret = -EFAULT;
9341 			goto out;
9342 		}
9343 		kbuf[size] = '\0';
9344 		buf = kbuf;
9345 		do {
9346 			tmp = strchr(buf, '\n');
9347 			if (tmp) {
9348 				*tmp = '\0';
9349 				size = tmp - buf + 1;
9350 			} else {
9351 				size = strlen(buf);
9352 				if (done + size < count) {
9353 					if (buf != kbuf)
9354 						break;
9355 					/* This can accept WRITE_BUFSIZE - 2 ('\n' + '\0') */
9356 					pr_warn("Line length is too long: Should be less than %d\n",
9357 						WRITE_BUFSIZE - 2);
9358 					ret = -EINVAL;
9359 					goto out;
9360 				}
9361 			}
9362 			done += size;
9363 
9364 			/* Remove comments */
9365 			tmp = strchr(buf, '#');
9366 
9367 			if (tmp)
9368 				*tmp = '\0';
9369 
9370 			ret = trace_run_command(buf, createfn);
9371 			if (ret)
9372 				goto out;
9373 			buf += size;
9374 
9375 		} while (done < count);
9376 	}
9377 	ret = done;
9378 
9379 out:
9380 	kfree(kbuf);
9381 
9382 	return ret;
9383 }
9384 
9385 __init static int tracer_alloc_buffers(void)
9386 {
9387 	int ring_buf_size;
9388 	int ret = -ENOMEM;
9389 
9390 
9391 	if (security_locked_down(LOCKDOWN_TRACEFS)) {
9392 		pr_warn("Tracing disabled due to lockdown\n");
9393 		return -EPERM;
9394 	}
9395 
9396 	/*
9397 	 * Make sure we don't accidently add more trace options
9398 	 * than we have bits for.
9399 	 */
9400 	BUILD_BUG_ON(TRACE_ITER_LAST_BIT > TRACE_FLAGS_MAX_SIZE);
9401 
9402 	if (!alloc_cpumask_var(&tracing_buffer_mask, GFP_KERNEL))
9403 		goto out;
9404 
9405 	if (!alloc_cpumask_var(&global_trace.tracing_cpumask, GFP_KERNEL))
9406 		goto out_free_buffer_mask;
9407 
9408 	/* Only allocate trace_printk buffers if a trace_printk exists */
9409 	if (&__stop___trace_bprintk_fmt != &__start___trace_bprintk_fmt)
9410 		/* Must be called before global_trace.buffer is allocated */
9411 		trace_printk_init_buffers();
9412 
9413 	/* To save memory, keep the ring buffer size to its minimum */
9414 	if (ring_buffer_expanded)
9415 		ring_buf_size = trace_buf_size;
9416 	else
9417 		ring_buf_size = 1;
9418 
9419 	cpumask_copy(tracing_buffer_mask, cpu_possible_mask);
9420 	cpumask_copy(global_trace.tracing_cpumask, cpu_all_mask);
9421 
9422 	raw_spin_lock_init(&global_trace.start_lock);
9423 
9424 	/*
9425 	 * The prepare callbacks allocates some memory for the ring buffer. We
9426 	 * don't free the buffer if the if the CPU goes down. If we were to free
9427 	 * the buffer, then the user would lose any trace that was in the
9428 	 * buffer. The memory will be removed once the "instance" is removed.
9429 	 */
9430 	ret = cpuhp_setup_state_multi(CPUHP_TRACE_RB_PREPARE,
9431 				      "trace/RB:preapre", trace_rb_cpu_prepare,
9432 				      NULL);
9433 	if (ret < 0)
9434 		goto out_free_cpumask;
9435 	/* Used for event triggers */
9436 	ret = -ENOMEM;
9437 	temp_buffer = ring_buffer_alloc(PAGE_SIZE, RB_FL_OVERWRITE);
9438 	if (!temp_buffer)
9439 		goto out_rm_hp_state;
9440 
9441 	if (trace_create_savedcmd() < 0)
9442 		goto out_free_temp_buffer;
9443 
9444 	/* TODO: make the number of buffers hot pluggable with CPUS */
9445 	if (allocate_trace_buffers(&global_trace, ring_buf_size) < 0) {
9446 		MEM_FAIL(1, "tracer: failed to allocate ring buffer!\n");
9447 		goto out_free_savedcmd;
9448 	}
9449 
9450 	if (global_trace.buffer_disabled)
9451 		tracing_off();
9452 
9453 	if (trace_boot_clock) {
9454 		ret = tracing_set_clock(&global_trace, trace_boot_clock);
9455 		if (ret < 0)
9456 			pr_warn("Trace clock %s not defined, going back to default\n",
9457 				trace_boot_clock);
9458 	}
9459 
9460 	/*
9461 	 * register_tracer() might reference current_trace, so it
9462 	 * needs to be set before we register anything. This is
9463 	 * just a bootstrap of current_trace anyway.
9464 	 */
9465 	global_trace.current_trace = &nop_trace;
9466 
9467 	global_trace.max_lock = (arch_spinlock_t)__ARCH_SPIN_LOCK_UNLOCKED;
9468 
9469 	ftrace_init_global_array_ops(&global_trace);
9470 
9471 	init_trace_flags_index(&global_trace);
9472 
9473 	register_tracer(&nop_trace);
9474 
9475 	/* Function tracing may start here (via kernel command line) */
9476 	init_function_trace();
9477 
9478 	/* All seems OK, enable tracing */
9479 	tracing_disabled = 0;
9480 
9481 	atomic_notifier_chain_register(&panic_notifier_list,
9482 				       &trace_panic_notifier);
9483 
9484 	register_die_notifier(&trace_die_notifier);
9485 
9486 	global_trace.flags = TRACE_ARRAY_FL_GLOBAL;
9487 
9488 	INIT_LIST_HEAD(&global_trace.systems);
9489 	INIT_LIST_HEAD(&global_trace.events);
9490 	INIT_LIST_HEAD(&global_trace.hist_vars);
9491 	INIT_LIST_HEAD(&global_trace.err_log);
9492 	list_add(&global_trace.list, &ftrace_trace_arrays);
9493 
9494 	apply_trace_boot_options();
9495 
9496 	register_snapshot_cmd();
9497 
9498 	return 0;
9499 
9500 out_free_savedcmd:
9501 	free_saved_cmdlines_buffer(savedcmd);
9502 out_free_temp_buffer:
9503 	ring_buffer_free(temp_buffer);
9504 out_rm_hp_state:
9505 	cpuhp_remove_multi_state(CPUHP_TRACE_RB_PREPARE);
9506 out_free_cpumask:
9507 	free_cpumask_var(global_trace.tracing_cpumask);
9508 out_free_buffer_mask:
9509 	free_cpumask_var(tracing_buffer_mask);
9510 out:
9511 	return ret;
9512 }
9513 
9514 void __init early_trace_init(void)
9515 {
9516 	if (tracepoint_printk) {
9517 		tracepoint_print_iter =
9518 			kmalloc(sizeof(*tracepoint_print_iter), GFP_KERNEL);
9519 		if (MEM_FAIL(!tracepoint_print_iter,
9520 			     "Failed to allocate trace iterator\n"))
9521 			tracepoint_printk = 0;
9522 		else
9523 			static_key_enable(&tracepoint_printk_key.key);
9524 	}
9525 	tracer_alloc_buffers();
9526 }
9527 
9528 void __init trace_init(void)
9529 {
9530 	trace_event_init();
9531 }
9532 
9533 __init static int clear_boot_tracer(void)
9534 {
9535 	/*
9536 	 * The default tracer at boot buffer is an init section.
9537 	 * This function is called in lateinit. If we did not
9538 	 * find the boot tracer, then clear it out, to prevent
9539 	 * later registration from accessing the buffer that is
9540 	 * about to be freed.
9541 	 */
9542 	if (!default_bootup_tracer)
9543 		return 0;
9544 
9545 	printk(KERN_INFO "ftrace bootup tracer '%s' not registered.\n",
9546 	       default_bootup_tracer);
9547 	default_bootup_tracer = NULL;
9548 
9549 	return 0;
9550 }
9551 
9552 fs_initcall(tracer_init_tracefs);
9553 late_initcall_sync(clear_boot_tracer);
9554 
9555 #ifdef CONFIG_HAVE_UNSTABLE_SCHED_CLOCK
9556 __init static int tracing_set_default_clock(void)
9557 {
9558 	/* sched_clock_stable() is determined in late_initcall */
9559 	if (!trace_boot_clock && !sched_clock_stable()) {
9560 		if (security_locked_down(LOCKDOWN_TRACEFS)) {
9561 			pr_warn("Can not set tracing clock due to lockdown\n");
9562 			return -EPERM;
9563 		}
9564 
9565 		printk(KERN_WARNING
9566 		       "Unstable clock detected, switching default tracing clock to \"global\"\n"
9567 		       "If you want to keep using the local clock, then add:\n"
9568 		       "  \"trace_clock=local\"\n"
9569 		       "on the kernel command line\n");
9570 		tracing_set_clock(&global_trace, "global");
9571 	}
9572 
9573 	return 0;
9574 }
9575 late_initcall_sync(tracing_set_default_clock);
9576 #endif
9577