1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * trace_events_synth - synthetic trace events
4  *
5  * Copyright (C) 2015, 2020 Tom Zanussi <tom.zanussi@linux.intel.com>
6  */
7 
8 #include <linux/module.h>
9 #include <linux/kallsyms.h>
10 #include <linux/security.h>
11 #include <linux/mutex.h>
12 #include <linux/slab.h>
13 #include <linux/stacktrace.h>
14 #include <linux/rculist.h>
15 #include <linux/tracefs.h>
16 
17 /* for gfp flag names */
18 #include <linux/trace_events.h>
19 #include <trace/events/mmflags.h>
20 #include "trace_probe.h"
21 #include "trace_probe_kernel.h"
22 
23 #include "trace_synth.h"
24 
25 #undef ERRORS
26 #define ERRORS	\
27 	C(BAD_NAME,		"Illegal name"),		\
28 	C(INVALID_CMD,		"Command must be of the form: <name> field[;field] ..."),\
29 	C(INVALID_DYN_CMD,	"Command must be of the form: s or -:[synthetic/]<name> field[;field] ..."),\
30 	C(EVENT_EXISTS,		"Event already exists"),	\
31 	C(TOO_MANY_FIELDS,	"Too many fields"),		\
32 	C(INCOMPLETE_TYPE,	"Incomplete type"),		\
33 	C(INVALID_TYPE,		"Invalid type"),		\
34 	C(INVALID_FIELD,        "Invalid field"),		\
35 	C(INVALID_ARRAY_SPEC,	"Invalid array specification"),
36 
37 #undef C
38 #define C(a, b)		SYNTH_ERR_##a
39 
40 enum { ERRORS };
41 
42 #undef C
43 #define C(a, b)		b
44 
45 static const char *err_text[] = { ERRORS };
46 
47 static DEFINE_MUTEX(lastcmd_mutex);
48 static char *last_cmd;
49 
50 static int errpos(const char *str)
51 {
52 	int ret = 0;
53 
54 	mutex_lock(&lastcmd_mutex);
55 	if (!str || !last_cmd)
56 		goto out;
57 
58 	ret = err_pos(last_cmd, str);
59  out:
60 	mutex_unlock(&lastcmd_mutex);
61 	return ret;
62 }
63 
64 static void last_cmd_set(const char *str)
65 {
66 	if (!str)
67 		return;
68 
69 	mutex_lock(&lastcmd_mutex);
70 	kfree(last_cmd);
71 	last_cmd = kstrdup(str, GFP_KERNEL);
72 	mutex_unlock(&lastcmd_mutex);
73 }
74 
75 static void synth_err(u8 err_type, u16 err_pos)
76 {
77 	mutex_lock(&lastcmd_mutex);
78 	if (!last_cmd)
79 		goto out;
80 
81 	tracing_log_err(NULL, "synthetic_events", last_cmd, err_text,
82 			err_type, err_pos);
83  out:
84 	mutex_unlock(&lastcmd_mutex);
85 }
86 
87 static int create_synth_event(const char *raw_command);
88 static int synth_event_show(struct seq_file *m, struct dyn_event *ev);
89 static int synth_event_release(struct dyn_event *ev);
90 static bool synth_event_is_busy(struct dyn_event *ev);
91 static bool synth_event_match(const char *system, const char *event,
92 			int argc, const char **argv, struct dyn_event *ev);
93 
94 static struct dyn_event_operations synth_event_ops = {
95 	.create = create_synth_event,
96 	.show = synth_event_show,
97 	.is_busy = synth_event_is_busy,
98 	.free = synth_event_release,
99 	.match = synth_event_match,
100 };
101 
102 static bool is_synth_event(struct dyn_event *ev)
103 {
104 	return ev->ops == &synth_event_ops;
105 }
106 
107 static struct synth_event *to_synth_event(struct dyn_event *ev)
108 {
109 	return container_of(ev, struct synth_event, devent);
110 }
111 
112 static bool synth_event_is_busy(struct dyn_event *ev)
113 {
114 	struct synth_event *event = to_synth_event(ev);
115 
116 	return event->ref != 0;
117 }
118 
119 static bool synth_event_match(const char *system, const char *event,
120 			int argc, const char **argv, struct dyn_event *ev)
121 {
122 	struct synth_event *sev = to_synth_event(ev);
123 
124 	return strcmp(sev->name, event) == 0 &&
125 		(!system || strcmp(system, SYNTH_SYSTEM) == 0);
126 }
127 
128 struct synth_trace_event {
129 	struct trace_entry	ent;
130 	u64			fields[];
131 };
132 
133 static int synth_event_define_fields(struct trace_event_call *call)
134 {
135 	struct synth_trace_event trace;
136 	int offset = offsetof(typeof(trace), fields);
137 	struct synth_event *event = call->data;
138 	unsigned int i, size, n_u64;
139 	char *name, *type;
140 	bool is_signed;
141 	int ret = 0;
142 
143 	for (i = 0, n_u64 = 0; i < event->n_fields; i++) {
144 		size = event->fields[i]->size;
145 		is_signed = event->fields[i]->is_signed;
146 		type = event->fields[i]->type;
147 		name = event->fields[i]->name;
148 		ret = trace_define_field(call, type, name, offset, size,
149 					 is_signed, FILTER_OTHER);
150 		if (ret)
151 			break;
152 
153 		event->fields[i]->offset = n_u64;
154 
155 		if (event->fields[i]->is_string && !event->fields[i]->is_dynamic) {
156 			offset += STR_VAR_LEN_MAX;
157 			n_u64 += STR_VAR_LEN_MAX / sizeof(u64);
158 		} else {
159 			offset += sizeof(u64);
160 			n_u64++;
161 		}
162 	}
163 
164 	event->n_u64 = n_u64;
165 
166 	return ret;
167 }
168 
169 static bool synth_field_signed(char *type)
170 {
171 	if (str_has_prefix(type, "u"))
172 		return false;
173 	if (strcmp(type, "gfp_t") == 0)
174 		return false;
175 
176 	return true;
177 }
178 
179 static int synth_field_is_string(char *type)
180 {
181 	if (strstr(type, "char[") != NULL)
182 		return true;
183 
184 	return false;
185 }
186 
187 static int synth_field_is_stack(char *type)
188 {
189 	if (strstr(type, "long[") != NULL)
190 		return true;
191 
192 	return false;
193 }
194 
195 static int synth_field_string_size(char *type)
196 {
197 	char buf[4], *end, *start;
198 	unsigned int len;
199 	int size, err;
200 
201 	start = strstr(type, "char[");
202 	if (start == NULL)
203 		return -EINVAL;
204 	start += sizeof("char[") - 1;
205 
206 	end = strchr(type, ']');
207 	if (!end || end < start || type + strlen(type) > end + 1)
208 		return -EINVAL;
209 
210 	len = end - start;
211 	if (len > 3)
212 		return -EINVAL;
213 
214 	if (len == 0)
215 		return 0; /* variable-length string */
216 
217 	strncpy(buf, start, len);
218 	buf[len] = '\0';
219 
220 	err = kstrtouint(buf, 0, &size);
221 	if (err)
222 		return err;
223 
224 	if (size > STR_VAR_LEN_MAX)
225 		return -EINVAL;
226 
227 	return size;
228 }
229 
230 static int synth_field_size(char *type)
231 {
232 	int size = 0;
233 
234 	if (strcmp(type, "s64") == 0)
235 		size = sizeof(s64);
236 	else if (strcmp(type, "u64") == 0)
237 		size = sizeof(u64);
238 	else if (strcmp(type, "s32") == 0)
239 		size = sizeof(s32);
240 	else if (strcmp(type, "u32") == 0)
241 		size = sizeof(u32);
242 	else if (strcmp(type, "s16") == 0)
243 		size = sizeof(s16);
244 	else if (strcmp(type, "u16") == 0)
245 		size = sizeof(u16);
246 	else if (strcmp(type, "s8") == 0)
247 		size = sizeof(s8);
248 	else if (strcmp(type, "u8") == 0)
249 		size = sizeof(u8);
250 	else if (strcmp(type, "char") == 0)
251 		size = sizeof(char);
252 	else if (strcmp(type, "unsigned char") == 0)
253 		size = sizeof(unsigned char);
254 	else if (strcmp(type, "int") == 0)
255 		size = sizeof(int);
256 	else if (strcmp(type, "unsigned int") == 0)
257 		size = sizeof(unsigned int);
258 	else if (strcmp(type, "long") == 0)
259 		size = sizeof(long);
260 	else if (strcmp(type, "unsigned long") == 0)
261 		size = sizeof(unsigned long);
262 	else if (strcmp(type, "bool") == 0)
263 		size = sizeof(bool);
264 	else if (strcmp(type, "pid_t") == 0)
265 		size = sizeof(pid_t);
266 	else if (strcmp(type, "gfp_t") == 0)
267 		size = sizeof(gfp_t);
268 	else if (synth_field_is_string(type))
269 		size = synth_field_string_size(type);
270 	else if (synth_field_is_stack(type))
271 		size = 0;
272 
273 	return size;
274 }
275 
276 static const char *synth_field_fmt(char *type)
277 {
278 	const char *fmt = "%llu";
279 
280 	if (strcmp(type, "s64") == 0)
281 		fmt = "%lld";
282 	else if (strcmp(type, "u64") == 0)
283 		fmt = "%llu";
284 	else if (strcmp(type, "s32") == 0)
285 		fmt = "%d";
286 	else if (strcmp(type, "u32") == 0)
287 		fmt = "%u";
288 	else if (strcmp(type, "s16") == 0)
289 		fmt = "%d";
290 	else if (strcmp(type, "u16") == 0)
291 		fmt = "%u";
292 	else if (strcmp(type, "s8") == 0)
293 		fmt = "%d";
294 	else if (strcmp(type, "u8") == 0)
295 		fmt = "%u";
296 	else if (strcmp(type, "char") == 0)
297 		fmt = "%d";
298 	else if (strcmp(type, "unsigned char") == 0)
299 		fmt = "%u";
300 	else if (strcmp(type, "int") == 0)
301 		fmt = "%d";
302 	else if (strcmp(type, "unsigned int") == 0)
303 		fmt = "%u";
304 	else if (strcmp(type, "long") == 0)
305 		fmt = "%ld";
306 	else if (strcmp(type, "unsigned long") == 0)
307 		fmt = "%lu";
308 	else if (strcmp(type, "bool") == 0)
309 		fmt = "%d";
310 	else if (strcmp(type, "pid_t") == 0)
311 		fmt = "%d";
312 	else if (strcmp(type, "gfp_t") == 0)
313 		fmt = "%x";
314 	else if (synth_field_is_string(type))
315 		fmt = "%.*s";
316 	else if (synth_field_is_stack(type))
317 		fmt = "%s";
318 
319 	return fmt;
320 }
321 
322 static void print_synth_event_num_val(struct trace_seq *s,
323 				      char *print_fmt, char *name,
324 				      int size, u64 val, char *space)
325 {
326 	switch (size) {
327 	case 1:
328 		trace_seq_printf(s, print_fmt, name, (u8)val, space);
329 		break;
330 
331 	case 2:
332 		trace_seq_printf(s, print_fmt, name, (u16)val, space);
333 		break;
334 
335 	case 4:
336 		trace_seq_printf(s, print_fmt, name, (u32)val, space);
337 		break;
338 
339 	default:
340 		trace_seq_printf(s, print_fmt, name, val, space);
341 		break;
342 	}
343 }
344 
345 static enum print_line_t print_synth_event(struct trace_iterator *iter,
346 					   int flags,
347 					   struct trace_event *event)
348 {
349 	struct trace_array *tr = iter->tr;
350 	struct trace_seq *s = &iter->seq;
351 	struct synth_trace_event *entry;
352 	struct synth_event *se;
353 	unsigned int i, n_u64;
354 	char print_fmt[32];
355 	const char *fmt;
356 
357 	entry = (struct synth_trace_event *)iter->ent;
358 	se = container_of(event, struct synth_event, call.event);
359 
360 	trace_seq_printf(s, "%s: ", se->name);
361 
362 	for (i = 0, n_u64 = 0; i < se->n_fields; i++) {
363 		if (trace_seq_has_overflowed(s))
364 			goto end;
365 
366 		fmt = synth_field_fmt(se->fields[i]->type);
367 
368 		/* parameter types */
369 		if (tr && tr->trace_flags & TRACE_ITER_VERBOSE)
370 			trace_seq_printf(s, "%s ", fmt);
371 
372 		snprintf(print_fmt, sizeof(print_fmt), "%%s=%s%%s", fmt);
373 
374 		/* parameter values */
375 		if (se->fields[i]->is_string) {
376 			if (se->fields[i]->is_dynamic) {
377 				u32 offset, data_offset;
378 				char *str_field;
379 
380 				offset = (u32)entry->fields[n_u64];
381 				data_offset = offset & 0xffff;
382 
383 				str_field = (char *)entry + data_offset;
384 
385 				trace_seq_printf(s, print_fmt, se->fields[i]->name,
386 						 STR_VAR_LEN_MAX,
387 						 str_field,
388 						 i == se->n_fields - 1 ? "" : " ");
389 				n_u64++;
390 			} else {
391 				trace_seq_printf(s, print_fmt, se->fields[i]->name,
392 						 STR_VAR_LEN_MAX,
393 						 (char *)&entry->fields[n_u64],
394 						 i == se->n_fields - 1 ? "" : " ");
395 				n_u64 += STR_VAR_LEN_MAX / sizeof(u64);
396 			}
397 		} else if (se->fields[i]->is_stack) {
398 			u32 offset, data_offset, len;
399 			unsigned long *p, *end;
400 
401 			offset = (u32)entry->fields[n_u64];
402 			data_offset = offset & 0xffff;
403 			len = offset >> 16;
404 
405 			p = (void *)entry + data_offset;
406 			end = (void *)p + len - (sizeof(long) - 1);
407 
408 			trace_seq_printf(s, "%s=STACK:\n", se->fields[i]->name);
409 
410 			for (; *p && p < end; p++)
411 				trace_seq_printf(s, "=> %pS\n", (void *)*p);
412 			n_u64++;
413 
414 		} else {
415 			struct trace_print_flags __flags[] = {
416 			    __def_gfpflag_names, {-1, NULL} };
417 			char *space = (i == se->n_fields - 1 ? "" : " ");
418 
419 			print_synth_event_num_val(s, print_fmt,
420 						  se->fields[i]->name,
421 						  se->fields[i]->size,
422 						  entry->fields[n_u64],
423 						  space);
424 
425 			if (strcmp(se->fields[i]->type, "gfp_t") == 0) {
426 				trace_seq_puts(s, " (");
427 				trace_print_flags_seq(s, "|",
428 						      entry->fields[n_u64],
429 						      __flags);
430 				trace_seq_putc(s, ')');
431 			}
432 			n_u64++;
433 		}
434 	}
435 end:
436 	trace_seq_putc(s, '\n');
437 
438 	return trace_handle_return(s);
439 }
440 
441 static struct trace_event_functions synth_event_funcs = {
442 	.trace		= print_synth_event
443 };
444 
445 static unsigned int trace_string(struct synth_trace_event *entry,
446 				 struct synth_event *event,
447 				 char *str_val,
448 				 bool is_dynamic,
449 				 unsigned int data_size,
450 				 unsigned int *n_u64)
451 {
452 	unsigned int len = 0;
453 	char *str_field;
454 	int ret;
455 
456 	if (is_dynamic) {
457 		u32 data_offset;
458 
459 		data_offset = struct_size(entry, fields, event->n_u64);
460 		data_offset += data_size;
461 
462 		len = fetch_store_strlen((unsigned long)str_val);
463 
464 		data_offset |= len << 16;
465 		*(u32 *)&entry->fields[*n_u64] = data_offset;
466 
467 		ret = fetch_store_string((unsigned long)str_val, &entry->fields[*n_u64], entry);
468 
469 		(*n_u64)++;
470 	} else {
471 		str_field = (char *)&entry->fields[*n_u64];
472 
473 #ifdef CONFIG_ARCH_HAS_NON_OVERLAPPING_ADDRESS_SPACE
474 		if ((unsigned long)str_val < TASK_SIZE)
475 			ret = strncpy_from_user_nofault(str_field, str_val, STR_VAR_LEN_MAX);
476 		else
477 #endif
478 			ret = strncpy_from_kernel_nofault(str_field, str_val, STR_VAR_LEN_MAX);
479 
480 		if (ret < 0)
481 			strcpy(str_field, FAULT_STRING);
482 
483 		(*n_u64) += STR_VAR_LEN_MAX / sizeof(u64);
484 	}
485 
486 	return len;
487 }
488 
489 static unsigned int trace_stack(struct synth_trace_event *entry,
490 				 struct synth_event *event,
491 				 long *stack,
492 				 unsigned int data_size,
493 				 unsigned int *n_u64)
494 {
495 	unsigned int len;
496 	u32 data_offset;
497 	void *data_loc;
498 
499 	data_offset = struct_size(entry, fields, event->n_u64);
500 	data_offset += data_size;
501 
502 	for (len = 0; len < HIST_STACKTRACE_DEPTH; len++) {
503 		if (!stack[len])
504 			break;
505 	}
506 
507 	/* Include the zero'd element if it fits */
508 	if (len < HIST_STACKTRACE_DEPTH)
509 		len++;
510 
511 	len *= sizeof(long);
512 
513 	/* Find the dynamic section to copy the stack into. */
514 	data_loc = (void *)entry + data_offset;
515 	memcpy(data_loc, stack, len);
516 
517 	/* Fill in the field that holds the offset/len combo */
518 	data_offset |= len << 16;
519 	*(u32 *)&entry->fields[*n_u64] = data_offset;
520 
521 	(*n_u64)++;
522 
523 	return len;
524 }
525 
526 static notrace void trace_event_raw_event_synth(void *__data,
527 						u64 *var_ref_vals,
528 						unsigned int *var_ref_idx)
529 {
530 	unsigned int i, n_u64, val_idx, len, data_size = 0;
531 	struct trace_event_file *trace_file = __data;
532 	struct synth_trace_event *entry;
533 	struct trace_event_buffer fbuffer;
534 	struct trace_buffer *buffer;
535 	struct synth_event *event;
536 	int fields_size = 0;
537 
538 	event = trace_file->event_call->data;
539 
540 	if (trace_trigger_soft_disabled(trace_file))
541 		return;
542 
543 	fields_size = event->n_u64 * sizeof(u64);
544 
545 	for (i = 0; i < event->n_dynamic_fields; i++) {
546 		unsigned int field_pos = event->dynamic_fields[i]->field_pos;
547 		char *str_val;
548 
549 		val_idx = var_ref_idx[field_pos];
550 		str_val = (char *)(long)var_ref_vals[val_idx];
551 
552 		if (event->dynamic_fields[i]->is_stack) {
553 			len = *((unsigned long *)str_val);
554 			len *= sizeof(unsigned long);
555 		} else {
556 			len = fetch_store_strlen((unsigned long)str_val);
557 		}
558 
559 		fields_size += len;
560 	}
561 
562 	/*
563 	 * Avoid ring buffer recursion detection, as this event
564 	 * is being performed within another event.
565 	 */
566 	buffer = trace_file->tr->array_buffer.buffer;
567 	ring_buffer_nest_start(buffer);
568 
569 	entry = trace_event_buffer_reserve(&fbuffer, trace_file,
570 					   sizeof(*entry) + fields_size);
571 	if (!entry)
572 		goto out;
573 
574 	for (i = 0, n_u64 = 0; i < event->n_fields; i++) {
575 		val_idx = var_ref_idx[i];
576 		if (event->fields[i]->is_string) {
577 			char *str_val = (char *)(long)var_ref_vals[val_idx];
578 
579 			len = trace_string(entry, event, str_val,
580 					   event->fields[i]->is_dynamic,
581 					   data_size, &n_u64);
582 			data_size += len; /* only dynamic string increments */
583 		} else if (event->fields[i]->is_stack) {
584 			long *stack = (long *)(long)var_ref_vals[val_idx];
585 
586 			len = trace_stack(entry, event, stack,
587 					   data_size, &n_u64);
588 			data_size += len;
589 		} else {
590 			struct synth_field *field = event->fields[i];
591 			u64 val = var_ref_vals[val_idx];
592 
593 			switch (field->size) {
594 			case 1:
595 				*(u8 *)&entry->fields[n_u64] = (u8)val;
596 				break;
597 
598 			case 2:
599 				*(u16 *)&entry->fields[n_u64] = (u16)val;
600 				break;
601 
602 			case 4:
603 				*(u32 *)&entry->fields[n_u64] = (u32)val;
604 				break;
605 
606 			default:
607 				entry->fields[n_u64] = val;
608 				break;
609 			}
610 			n_u64++;
611 		}
612 	}
613 
614 	trace_event_buffer_commit(&fbuffer);
615 out:
616 	ring_buffer_nest_end(buffer);
617 }
618 
619 static void free_synth_event_print_fmt(struct trace_event_call *call)
620 {
621 	if (call) {
622 		kfree(call->print_fmt);
623 		call->print_fmt = NULL;
624 	}
625 }
626 
627 static int __set_synth_event_print_fmt(struct synth_event *event,
628 				       char *buf, int len)
629 {
630 	const char *fmt;
631 	int pos = 0;
632 	int i;
633 
634 	/* When len=0, we just calculate the needed length */
635 #define LEN_OR_ZERO (len ? len - pos : 0)
636 
637 	pos += snprintf(buf + pos, LEN_OR_ZERO, "\"");
638 	for (i = 0; i < event->n_fields; i++) {
639 		fmt = synth_field_fmt(event->fields[i]->type);
640 		pos += snprintf(buf + pos, LEN_OR_ZERO, "%s=%s%s",
641 				event->fields[i]->name, fmt,
642 				i == event->n_fields - 1 ? "" : ", ");
643 	}
644 	pos += snprintf(buf + pos, LEN_OR_ZERO, "\"");
645 
646 	for (i = 0; i < event->n_fields; i++) {
647 		if (event->fields[i]->is_string &&
648 		    event->fields[i]->is_dynamic)
649 			pos += snprintf(buf + pos, LEN_OR_ZERO,
650 				", __get_str(%s)", event->fields[i]->name);
651 		else if (event->fields[i]->is_stack)
652 			pos += snprintf(buf + pos, LEN_OR_ZERO,
653 				", __get_stacktrace(%s)", event->fields[i]->name);
654 		else
655 			pos += snprintf(buf + pos, LEN_OR_ZERO,
656 					", REC->%s", event->fields[i]->name);
657 	}
658 
659 #undef LEN_OR_ZERO
660 
661 	/* return the length of print_fmt */
662 	return pos;
663 }
664 
665 static int set_synth_event_print_fmt(struct trace_event_call *call)
666 {
667 	struct synth_event *event = call->data;
668 	char *print_fmt;
669 	int len;
670 
671 	/* First: called with 0 length to calculate the needed length */
672 	len = __set_synth_event_print_fmt(event, NULL, 0);
673 
674 	print_fmt = kmalloc(len + 1, GFP_KERNEL);
675 	if (!print_fmt)
676 		return -ENOMEM;
677 
678 	/* Second: actually write the @print_fmt */
679 	__set_synth_event_print_fmt(event, print_fmt, len + 1);
680 	call->print_fmt = print_fmt;
681 
682 	return 0;
683 }
684 
685 static void free_synth_field(struct synth_field *field)
686 {
687 	kfree(field->type);
688 	kfree(field->name);
689 	kfree(field);
690 }
691 
692 static int check_field_version(const char *prefix, const char *field_type,
693 			       const char *field_name)
694 {
695 	/*
696 	 * For backward compatibility, the old synthetic event command
697 	 * format did not require semicolons, and in order to not
698 	 * break user space, that old format must still work. If a new
699 	 * feature is added, then the format that uses the new feature
700 	 * will be required to have semicolons, as nothing that uses
701 	 * the old format would be using the new, yet to be created,
702 	 * feature. When a new feature is added, this will detect it,
703 	 * and return a number greater than 1, and require the format
704 	 * to use semicolons.
705 	 */
706 	return 1;
707 }
708 
709 static struct synth_field *parse_synth_field(int argc, char **argv,
710 					     int *consumed, int *field_version)
711 {
712 	const char *prefix = NULL, *field_type = argv[0], *field_name, *array;
713 	struct synth_field *field;
714 	int len, ret = -ENOMEM;
715 	struct seq_buf s;
716 	ssize_t size;
717 
718 	if (!strcmp(field_type, "unsigned")) {
719 		if (argc < 3) {
720 			synth_err(SYNTH_ERR_INCOMPLETE_TYPE, errpos(field_type));
721 			return ERR_PTR(-EINVAL);
722 		}
723 		prefix = "unsigned ";
724 		field_type = argv[1];
725 		field_name = argv[2];
726 		*consumed += 3;
727 	} else {
728 		field_name = argv[1];
729 		*consumed += 2;
730 	}
731 
732 	if (!field_name) {
733 		synth_err(SYNTH_ERR_INVALID_FIELD, errpos(field_type));
734 		return ERR_PTR(-EINVAL);
735 	}
736 
737 	*field_version = check_field_version(prefix, field_type, field_name);
738 
739 	field = kzalloc(sizeof(*field), GFP_KERNEL);
740 	if (!field)
741 		return ERR_PTR(-ENOMEM);
742 
743 	len = strlen(field_name);
744 	array = strchr(field_name, '[');
745 	if (array)
746 		len -= strlen(array);
747 
748 	field->name = kmemdup_nul(field_name, len, GFP_KERNEL);
749 	if (!field->name)
750 		goto free;
751 
752 	if (!is_good_name(field->name)) {
753 		synth_err(SYNTH_ERR_BAD_NAME, errpos(field_name));
754 		ret = -EINVAL;
755 		goto free;
756 	}
757 
758 	len = strlen(field_type) + 1;
759 
760 	if (array)
761 		len += strlen(array);
762 
763 	if (prefix)
764 		len += strlen(prefix);
765 
766 	field->type = kzalloc(len, GFP_KERNEL);
767 	if (!field->type)
768 		goto free;
769 
770 	seq_buf_init(&s, field->type, len);
771 	if (prefix)
772 		seq_buf_puts(&s, prefix);
773 	seq_buf_puts(&s, field_type);
774 	if (array)
775 		seq_buf_puts(&s, array);
776 	if (WARN_ON_ONCE(!seq_buf_buffer_left(&s)))
777 		goto free;
778 
779 	s.buffer[s.len] = '\0';
780 
781 	size = synth_field_size(field->type);
782 	if (size < 0) {
783 		if (array)
784 			synth_err(SYNTH_ERR_INVALID_ARRAY_SPEC, errpos(field_name));
785 		else
786 			synth_err(SYNTH_ERR_INVALID_TYPE, errpos(field_type));
787 		ret = -EINVAL;
788 		goto free;
789 	} else if (size == 0) {
790 		if (synth_field_is_string(field->type) ||
791 		    synth_field_is_stack(field->type)) {
792 			char *type;
793 
794 			len = sizeof("__data_loc ") + strlen(field->type) + 1;
795 			type = kzalloc(len, GFP_KERNEL);
796 			if (!type)
797 				goto free;
798 
799 			seq_buf_init(&s, type, len);
800 			seq_buf_puts(&s, "__data_loc ");
801 			seq_buf_puts(&s, field->type);
802 
803 			if (WARN_ON_ONCE(!seq_buf_buffer_left(&s)))
804 				goto free;
805 			s.buffer[s.len] = '\0';
806 
807 			kfree(field->type);
808 			field->type = type;
809 
810 			field->is_dynamic = true;
811 			size = sizeof(u64);
812 		} else {
813 			synth_err(SYNTH_ERR_INVALID_TYPE, errpos(field_type));
814 			ret = -EINVAL;
815 			goto free;
816 		}
817 	}
818 	field->size = size;
819 
820 	if (synth_field_is_string(field->type))
821 		field->is_string = true;
822 	else if (synth_field_is_stack(field->type))
823 		field->is_stack = true;
824 
825 	field->is_signed = synth_field_signed(field->type);
826  out:
827 	return field;
828  free:
829 	free_synth_field(field);
830 	field = ERR_PTR(ret);
831 	goto out;
832 }
833 
834 static void free_synth_tracepoint(struct tracepoint *tp)
835 {
836 	if (!tp)
837 		return;
838 
839 	kfree(tp->name);
840 	kfree(tp);
841 }
842 
843 static struct tracepoint *alloc_synth_tracepoint(char *name)
844 {
845 	struct tracepoint *tp;
846 
847 	tp = kzalloc(sizeof(*tp), GFP_KERNEL);
848 	if (!tp)
849 		return ERR_PTR(-ENOMEM);
850 
851 	tp->name = kstrdup(name, GFP_KERNEL);
852 	if (!tp->name) {
853 		kfree(tp);
854 		return ERR_PTR(-ENOMEM);
855 	}
856 
857 	return tp;
858 }
859 
860 struct synth_event *find_synth_event(const char *name)
861 {
862 	struct dyn_event *pos;
863 	struct synth_event *event;
864 
865 	for_each_dyn_event(pos) {
866 		if (!is_synth_event(pos))
867 			continue;
868 		event = to_synth_event(pos);
869 		if (strcmp(event->name, name) == 0)
870 			return event;
871 	}
872 
873 	return NULL;
874 }
875 
876 static struct trace_event_fields synth_event_fields_array[] = {
877 	{ .type = TRACE_FUNCTION_TYPE,
878 	  .define_fields = synth_event_define_fields },
879 	{}
880 };
881 
882 static int register_synth_event(struct synth_event *event)
883 {
884 	struct trace_event_call *call = &event->call;
885 	int ret = 0;
886 
887 	event->call.class = &event->class;
888 	event->class.system = kstrdup(SYNTH_SYSTEM, GFP_KERNEL);
889 	if (!event->class.system) {
890 		ret = -ENOMEM;
891 		goto out;
892 	}
893 
894 	event->tp = alloc_synth_tracepoint(event->name);
895 	if (IS_ERR(event->tp)) {
896 		ret = PTR_ERR(event->tp);
897 		event->tp = NULL;
898 		goto out;
899 	}
900 
901 	INIT_LIST_HEAD(&call->class->fields);
902 	call->event.funcs = &synth_event_funcs;
903 	call->class->fields_array = synth_event_fields_array;
904 
905 	ret = register_trace_event(&call->event);
906 	if (!ret) {
907 		ret = -ENODEV;
908 		goto out;
909 	}
910 	call->flags = TRACE_EVENT_FL_TRACEPOINT;
911 	call->class->reg = trace_event_reg;
912 	call->class->probe = trace_event_raw_event_synth;
913 	call->data = event;
914 	call->tp = event->tp;
915 
916 	ret = trace_add_event_call(call);
917 	if (ret) {
918 		pr_warn("Failed to register synthetic event: %s\n",
919 			trace_event_name(call));
920 		goto err;
921 	}
922 
923 	ret = set_synth_event_print_fmt(call);
924 	/* unregister_trace_event() will be called inside */
925 	if (ret < 0)
926 		trace_remove_event_call(call);
927  out:
928 	return ret;
929  err:
930 	unregister_trace_event(&call->event);
931 	goto out;
932 }
933 
934 static int unregister_synth_event(struct synth_event *event)
935 {
936 	struct trace_event_call *call = &event->call;
937 	int ret;
938 
939 	ret = trace_remove_event_call(call);
940 
941 	return ret;
942 }
943 
944 static void free_synth_event(struct synth_event *event)
945 {
946 	unsigned int i;
947 
948 	if (!event)
949 		return;
950 
951 	for (i = 0; i < event->n_fields; i++)
952 		free_synth_field(event->fields[i]);
953 
954 	kfree(event->fields);
955 	kfree(event->dynamic_fields);
956 	kfree(event->name);
957 	kfree(event->class.system);
958 	free_synth_tracepoint(event->tp);
959 	free_synth_event_print_fmt(&event->call);
960 	kfree(event);
961 }
962 
963 static struct synth_event *alloc_synth_event(const char *name, int n_fields,
964 					     struct synth_field **fields)
965 {
966 	unsigned int i, j, n_dynamic_fields = 0;
967 	struct synth_event *event;
968 
969 	event = kzalloc(sizeof(*event), GFP_KERNEL);
970 	if (!event) {
971 		event = ERR_PTR(-ENOMEM);
972 		goto out;
973 	}
974 
975 	event->name = kstrdup(name, GFP_KERNEL);
976 	if (!event->name) {
977 		kfree(event);
978 		event = ERR_PTR(-ENOMEM);
979 		goto out;
980 	}
981 
982 	event->fields = kcalloc(n_fields, sizeof(*event->fields), GFP_KERNEL);
983 	if (!event->fields) {
984 		free_synth_event(event);
985 		event = ERR_PTR(-ENOMEM);
986 		goto out;
987 	}
988 
989 	for (i = 0; i < n_fields; i++)
990 		if (fields[i]->is_dynamic)
991 			n_dynamic_fields++;
992 
993 	if (n_dynamic_fields) {
994 		event->dynamic_fields = kcalloc(n_dynamic_fields,
995 						sizeof(*event->dynamic_fields),
996 						GFP_KERNEL);
997 		if (!event->dynamic_fields) {
998 			free_synth_event(event);
999 			event = ERR_PTR(-ENOMEM);
1000 			goto out;
1001 		}
1002 	}
1003 
1004 	dyn_event_init(&event->devent, &synth_event_ops);
1005 
1006 	for (i = 0, j = 0; i < n_fields; i++) {
1007 		fields[i]->field_pos = i;
1008 		event->fields[i] = fields[i];
1009 
1010 		if (fields[i]->is_dynamic)
1011 			event->dynamic_fields[j++] = fields[i];
1012 	}
1013 	event->n_dynamic_fields = j;
1014 	event->n_fields = n_fields;
1015  out:
1016 	return event;
1017 }
1018 
1019 static int synth_event_check_arg_fn(void *data)
1020 {
1021 	struct dynevent_arg_pair *arg_pair = data;
1022 	int size;
1023 
1024 	size = synth_field_size((char *)arg_pair->lhs);
1025 	if (size == 0) {
1026 		if (strstr((char *)arg_pair->lhs, "["))
1027 			return 0;
1028 	}
1029 
1030 	return size ? 0 : -EINVAL;
1031 }
1032 
1033 /**
1034  * synth_event_add_field - Add a new field to a synthetic event cmd
1035  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1036  * @type: The type of the new field to add
1037  * @name: The name of the new field to add
1038  *
1039  * Add a new field to a synthetic event cmd object.  Field ordering is in
1040  * the same order the fields are added.
1041  *
1042  * See synth_field_size() for available types. If field_name contains
1043  * [n] the field is considered to be an array.
1044  *
1045  * Return: 0 if successful, error otherwise.
1046  */
1047 int synth_event_add_field(struct dynevent_cmd *cmd, const char *type,
1048 			  const char *name)
1049 {
1050 	struct dynevent_arg_pair arg_pair;
1051 	int ret;
1052 
1053 	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1054 		return -EINVAL;
1055 
1056 	if (!type || !name)
1057 		return -EINVAL;
1058 
1059 	dynevent_arg_pair_init(&arg_pair, 0, ';');
1060 
1061 	arg_pair.lhs = type;
1062 	arg_pair.rhs = name;
1063 
1064 	ret = dynevent_arg_pair_add(cmd, &arg_pair, synth_event_check_arg_fn);
1065 	if (ret)
1066 		return ret;
1067 
1068 	if (++cmd->n_fields > SYNTH_FIELDS_MAX)
1069 		ret = -EINVAL;
1070 
1071 	return ret;
1072 }
1073 EXPORT_SYMBOL_GPL(synth_event_add_field);
1074 
1075 /**
1076  * synth_event_add_field_str - Add a new field to a synthetic event cmd
1077  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1078  * @type_name: The type and name of the new field to add, as a single string
1079  *
1080  * Add a new field to a synthetic event cmd object, as a single
1081  * string.  The @type_name string is expected to be of the form 'type
1082  * name', which will be appended by ';'.  No sanity checking is done -
1083  * what's passed in is assumed to already be well-formed.  Field
1084  * ordering is in the same order the fields are added.
1085  *
1086  * See synth_field_size() for available types. If field_name contains
1087  * [n] the field is considered to be an array.
1088  *
1089  * Return: 0 if successful, error otherwise.
1090  */
1091 int synth_event_add_field_str(struct dynevent_cmd *cmd, const char *type_name)
1092 {
1093 	struct dynevent_arg arg;
1094 	int ret;
1095 
1096 	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1097 		return -EINVAL;
1098 
1099 	if (!type_name)
1100 		return -EINVAL;
1101 
1102 	dynevent_arg_init(&arg, ';');
1103 
1104 	arg.str = type_name;
1105 
1106 	ret = dynevent_arg_add(cmd, &arg, NULL);
1107 	if (ret)
1108 		return ret;
1109 
1110 	if (++cmd->n_fields > SYNTH_FIELDS_MAX)
1111 		ret = -EINVAL;
1112 
1113 	return ret;
1114 }
1115 EXPORT_SYMBOL_GPL(synth_event_add_field_str);
1116 
1117 /**
1118  * synth_event_add_fields - Add multiple fields to a synthetic event cmd
1119  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1120  * @fields: An array of type/name field descriptions
1121  * @n_fields: The number of field descriptions contained in the fields array
1122  *
1123  * Add a new set of fields to a synthetic event cmd object.  The event
1124  * fields that will be defined for the event should be passed in as an
1125  * array of struct synth_field_desc, and the number of elements in the
1126  * array passed in as n_fields.  Field ordering will retain the
1127  * ordering given in the fields array.
1128  *
1129  * See synth_field_size() for available types. If field_name contains
1130  * [n] the field is considered to be an array.
1131  *
1132  * Return: 0 if successful, error otherwise.
1133  */
1134 int synth_event_add_fields(struct dynevent_cmd *cmd,
1135 			   struct synth_field_desc *fields,
1136 			   unsigned int n_fields)
1137 {
1138 	unsigned int i;
1139 	int ret = 0;
1140 
1141 	for (i = 0; i < n_fields; i++) {
1142 		if (fields[i].type == NULL || fields[i].name == NULL) {
1143 			ret = -EINVAL;
1144 			break;
1145 		}
1146 
1147 		ret = synth_event_add_field(cmd, fields[i].type, fields[i].name);
1148 		if (ret)
1149 			break;
1150 	}
1151 
1152 	return ret;
1153 }
1154 EXPORT_SYMBOL_GPL(synth_event_add_fields);
1155 
1156 /**
1157  * __synth_event_gen_cmd_start - Start a synthetic event command from arg list
1158  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1159  * @name: The name of the synthetic event
1160  * @mod: The module creating the event, NULL if not created from a module
1161  * @args: Variable number of arg (pairs), one pair for each field
1162  *
1163  * NOTE: Users normally won't want to call this function directly, but
1164  * rather use the synth_event_gen_cmd_start() wrapper, which
1165  * automatically adds a NULL to the end of the arg list.  If this
1166  * function is used directly, make sure the last arg in the variable
1167  * arg list is NULL.
1168  *
1169  * Generate a synthetic event command to be executed by
1170  * synth_event_gen_cmd_end().  This function can be used to generate
1171  * the complete command or only the first part of it; in the latter
1172  * case, synth_event_add_field(), synth_event_add_field_str(), or
1173  * synth_event_add_fields() can be used to add more fields following
1174  * this.
1175  *
1176  * There should be an even number variable args, each pair consisting
1177  * of a type followed by a field name.
1178  *
1179  * See synth_field_size() for available types. If field_name contains
1180  * [n] the field is considered to be an array.
1181  *
1182  * Return: 0 if successful, error otherwise.
1183  */
1184 int __synth_event_gen_cmd_start(struct dynevent_cmd *cmd, const char *name,
1185 				struct module *mod, ...)
1186 {
1187 	struct dynevent_arg arg;
1188 	va_list args;
1189 	int ret;
1190 
1191 	cmd->event_name = name;
1192 	cmd->private_data = mod;
1193 
1194 	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1195 		return -EINVAL;
1196 
1197 	dynevent_arg_init(&arg, 0);
1198 	arg.str = name;
1199 	ret = dynevent_arg_add(cmd, &arg, NULL);
1200 	if (ret)
1201 		return ret;
1202 
1203 	va_start(args, mod);
1204 	for (;;) {
1205 		const char *type, *name;
1206 
1207 		type = va_arg(args, const char *);
1208 		if (!type)
1209 			break;
1210 		name = va_arg(args, const char *);
1211 		if (!name)
1212 			break;
1213 
1214 		if (++cmd->n_fields > SYNTH_FIELDS_MAX) {
1215 			ret = -EINVAL;
1216 			break;
1217 		}
1218 
1219 		ret = synth_event_add_field(cmd, type, name);
1220 		if (ret)
1221 			break;
1222 	}
1223 	va_end(args);
1224 
1225 	return ret;
1226 }
1227 EXPORT_SYMBOL_GPL(__synth_event_gen_cmd_start);
1228 
1229 /**
1230  * synth_event_gen_cmd_array_start - Start synthetic event command from an array
1231  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1232  * @name: The name of the synthetic event
1233  * @mod: The module creating the event, NULL if not created from a module
1234  * @fields: An array of type/name field descriptions
1235  * @n_fields: The number of field descriptions contained in the fields array
1236  *
1237  * Generate a synthetic event command to be executed by
1238  * synth_event_gen_cmd_end().  This function can be used to generate
1239  * the complete command or only the first part of it; in the latter
1240  * case, synth_event_add_field(), synth_event_add_field_str(), or
1241  * synth_event_add_fields() can be used to add more fields following
1242  * this.
1243  *
1244  * The event fields that will be defined for the event should be
1245  * passed in as an array of struct synth_field_desc, and the number of
1246  * elements in the array passed in as n_fields.  Field ordering will
1247  * retain the ordering given in the fields array.
1248  *
1249  * See synth_field_size() for available types. If field_name contains
1250  * [n] the field is considered to be an array.
1251  *
1252  * Return: 0 if successful, error otherwise.
1253  */
1254 int synth_event_gen_cmd_array_start(struct dynevent_cmd *cmd, const char *name,
1255 				    struct module *mod,
1256 				    struct synth_field_desc *fields,
1257 				    unsigned int n_fields)
1258 {
1259 	struct dynevent_arg arg;
1260 	unsigned int i;
1261 	int ret = 0;
1262 
1263 	cmd->event_name = name;
1264 	cmd->private_data = mod;
1265 
1266 	if (cmd->type != DYNEVENT_TYPE_SYNTH)
1267 		return -EINVAL;
1268 
1269 	if (n_fields > SYNTH_FIELDS_MAX)
1270 		return -EINVAL;
1271 
1272 	dynevent_arg_init(&arg, 0);
1273 	arg.str = name;
1274 	ret = dynevent_arg_add(cmd, &arg, NULL);
1275 	if (ret)
1276 		return ret;
1277 
1278 	for (i = 0; i < n_fields; i++) {
1279 		if (fields[i].type == NULL || fields[i].name == NULL)
1280 			return -EINVAL;
1281 
1282 		ret = synth_event_add_field(cmd, fields[i].type, fields[i].name);
1283 		if (ret)
1284 			break;
1285 	}
1286 
1287 	return ret;
1288 }
1289 EXPORT_SYMBOL_GPL(synth_event_gen_cmd_array_start);
1290 
1291 static int __create_synth_event(const char *name, const char *raw_fields)
1292 {
1293 	char **argv, *field_str, *tmp_fields, *saved_fields = NULL;
1294 	struct synth_field *field, *fields[SYNTH_FIELDS_MAX];
1295 	int consumed, cmd_version = 1, n_fields_this_loop;
1296 	int i, argc, n_fields = 0, ret = 0;
1297 	struct synth_event *event = NULL;
1298 
1299 	/*
1300 	 * Argument syntax:
1301 	 *  - Add synthetic event: <event_name> field[;field] ...
1302 	 *  - Remove synthetic event: !<event_name> field[;field] ...
1303 	 *      where 'field' = type field_name
1304 	 */
1305 
1306 	if (name[0] == '\0') {
1307 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1308 		return -EINVAL;
1309 	}
1310 
1311 	if (!is_good_name(name)) {
1312 		synth_err(SYNTH_ERR_BAD_NAME, errpos(name));
1313 		return -EINVAL;
1314 	}
1315 
1316 	mutex_lock(&event_mutex);
1317 
1318 	event = find_synth_event(name);
1319 	if (event) {
1320 		synth_err(SYNTH_ERR_EVENT_EXISTS, errpos(name));
1321 		ret = -EEXIST;
1322 		goto err;
1323 	}
1324 
1325 	tmp_fields = saved_fields = kstrdup(raw_fields, GFP_KERNEL);
1326 	if (!tmp_fields) {
1327 		ret = -ENOMEM;
1328 		goto err;
1329 	}
1330 
1331 	while ((field_str = strsep(&tmp_fields, ";")) != NULL) {
1332 		argv = argv_split(GFP_KERNEL, field_str, &argc);
1333 		if (!argv) {
1334 			ret = -ENOMEM;
1335 			goto err;
1336 		}
1337 
1338 		if (!argc) {
1339 			argv_free(argv);
1340 			continue;
1341 		}
1342 
1343 		n_fields_this_loop = 0;
1344 		consumed = 0;
1345 		while (argc > consumed) {
1346 			int field_version;
1347 
1348 			field = parse_synth_field(argc - consumed,
1349 						  argv + consumed, &consumed,
1350 						  &field_version);
1351 			if (IS_ERR(field)) {
1352 				ret = PTR_ERR(field);
1353 				goto err_free_arg;
1354 			}
1355 
1356 			/*
1357 			 * Track the highest version of any field we
1358 			 * found in the command.
1359 			 */
1360 			if (field_version > cmd_version)
1361 				cmd_version = field_version;
1362 
1363 			/*
1364 			 * Now sort out what is and isn't valid for
1365 			 * each supported version.
1366 			 *
1367 			 * If we see more than 1 field per loop, it
1368 			 * means we have multiple fields between
1369 			 * semicolons, and that's something we no
1370 			 * longer support in a version 2 or greater
1371 			 * command.
1372 			 */
1373 			if (cmd_version > 1 && n_fields_this_loop >= 1) {
1374 				synth_err(SYNTH_ERR_INVALID_CMD, errpos(field_str));
1375 				ret = -EINVAL;
1376 				goto err_free_arg;
1377 			}
1378 
1379 			if (n_fields == SYNTH_FIELDS_MAX) {
1380 				synth_err(SYNTH_ERR_TOO_MANY_FIELDS, 0);
1381 				ret = -EINVAL;
1382 				goto err_free_arg;
1383 			}
1384 			fields[n_fields++] = field;
1385 
1386 			n_fields_this_loop++;
1387 		}
1388 		argv_free(argv);
1389 
1390 		if (consumed < argc) {
1391 			synth_err(SYNTH_ERR_INVALID_CMD, 0);
1392 			ret = -EINVAL;
1393 			goto err;
1394 		}
1395 
1396 	}
1397 
1398 	if (n_fields == 0) {
1399 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1400 		ret = -EINVAL;
1401 		goto err;
1402 	}
1403 
1404 	event = alloc_synth_event(name, n_fields, fields);
1405 	if (IS_ERR(event)) {
1406 		ret = PTR_ERR(event);
1407 		event = NULL;
1408 		goto err;
1409 	}
1410 	ret = register_synth_event(event);
1411 	if (!ret)
1412 		dyn_event_add(&event->devent, &event->call);
1413 	else
1414 		free_synth_event(event);
1415  out:
1416 	mutex_unlock(&event_mutex);
1417 
1418 	kfree(saved_fields);
1419 
1420 	return ret;
1421  err_free_arg:
1422 	argv_free(argv);
1423  err:
1424 	for (i = 0; i < n_fields; i++)
1425 		free_synth_field(fields[i]);
1426 
1427 	goto out;
1428 }
1429 
1430 /**
1431  * synth_event_create - Create a new synthetic event
1432  * @name: The name of the new synthetic event
1433  * @fields: An array of type/name field descriptions
1434  * @n_fields: The number of field descriptions contained in the fields array
1435  * @mod: The module creating the event, NULL if not created from a module
1436  *
1437  * Create a new synthetic event with the given name under the
1438  * trace/events/synthetic/ directory.  The event fields that will be
1439  * defined for the event should be passed in as an array of struct
1440  * synth_field_desc, and the number elements in the array passed in as
1441  * n_fields. Field ordering will retain the ordering given in the
1442  * fields array.
1443  *
1444  * If the new synthetic event is being created from a module, the mod
1445  * param must be non-NULL.  This will ensure that the trace buffer
1446  * won't contain unreadable events.
1447  *
1448  * The new synth event should be deleted using synth_event_delete()
1449  * function.  The new synthetic event can be generated from modules or
1450  * other kernel code using trace_synth_event() and related functions.
1451  *
1452  * Return: 0 if successful, error otherwise.
1453  */
1454 int synth_event_create(const char *name, struct synth_field_desc *fields,
1455 		       unsigned int n_fields, struct module *mod)
1456 {
1457 	struct dynevent_cmd cmd;
1458 	char *buf;
1459 	int ret;
1460 
1461 	buf = kzalloc(MAX_DYNEVENT_CMD_LEN, GFP_KERNEL);
1462 	if (!buf)
1463 		return -ENOMEM;
1464 
1465 	synth_event_cmd_init(&cmd, buf, MAX_DYNEVENT_CMD_LEN);
1466 
1467 	ret = synth_event_gen_cmd_array_start(&cmd, name, mod,
1468 					      fields, n_fields);
1469 	if (ret)
1470 		goto out;
1471 
1472 	ret = synth_event_gen_cmd_end(&cmd);
1473  out:
1474 	kfree(buf);
1475 
1476 	return ret;
1477 }
1478 EXPORT_SYMBOL_GPL(synth_event_create);
1479 
1480 static int destroy_synth_event(struct synth_event *se)
1481 {
1482 	int ret;
1483 
1484 	if (se->ref)
1485 		return -EBUSY;
1486 
1487 	if (trace_event_dyn_busy(&se->call))
1488 		return -EBUSY;
1489 
1490 	ret = unregister_synth_event(se);
1491 	if (!ret) {
1492 		dyn_event_remove(&se->devent);
1493 		free_synth_event(se);
1494 	}
1495 
1496 	return ret;
1497 }
1498 
1499 /**
1500  * synth_event_delete - Delete a synthetic event
1501  * @event_name: The name of the new synthetic event
1502  *
1503  * Delete a synthetic event that was created with synth_event_create().
1504  *
1505  * Return: 0 if successful, error otherwise.
1506  */
1507 int synth_event_delete(const char *event_name)
1508 {
1509 	struct synth_event *se = NULL;
1510 	struct module *mod = NULL;
1511 	int ret = -ENOENT;
1512 
1513 	mutex_lock(&event_mutex);
1514 	se = find_synth_event(event_name);
1515 	if (se) {
1516 		mod = se->mod;
1517 		ret = destroy_synth_event(se);
1518 	}
1519 	mutex_unlock(&event_mutex);
1520 
1521 	if (mod) {
1522 		/*
1523 		 * It is safest to reset the ring buffer if the module
1524 		 * being unloaded registered any events that were
1525 		 * used. The only worry is if a new module gets
1526 		 * loaded, and takes on the same id as the events of
1527 		 * this module. When printing out the buffer, traced
1528 		 * events left over from this module may be passed to
1529 		 * the new module events and unexpected results may
1530 		 * occur.
1531 		 */
1532 		tracing_reset_all_online_cpus();
1533 	}
1534 
1535 	return ret;
1536 }
1537 EXPORT_SYMBOL_GPL(synth_event_delete);
1538 
1539 static int check_command(const char *raw_command)
1540 {
1541 	char **argv = NULL, *cmd, *saved_cmd, *name_and_field;
1542 	int argc, ret = 0;
1543 
1544 	cmd = saved_cmd = kstrdup(raw_command, GFP_KERNEL);
1545 	if (!cmd)
1546 		return -ENOMEM;
1547 
1548 	name_and_field = strsep(&cmd, ";");
1549 	if (!name_and_field) {
1550 		ret = -EINVAL;
1551 		goto free;
1552 	}
1553 
1554 	if (name_and_field[0] == '!')
1555 		goto free;
1556 
1557 	argv = argv_split(GFP_KERNEL, name_and_field, &argc);
1558 	if (!argv) {
1559 		ret = -ENOMEM;
1560 		goto free;
1561 	}
1562 	argv_free(argv);
1563 
1564 	if (argc < 3)
1565 		ret = -EINVAL;
1566 free:
1567 	kfree(saved_cmd);
1568 
1569 	return ret;
1570 }
1571 
1572 static int create_or_delete_synth_event(const char *raw_command)
1573 {
1574 	char *name = NULL, *fields, *p;
1575 	int ret = 0;
1576 
1577 	raw_command = skip_spaces(raw_command);
1578 	if (raw_command[0] == '\0')
1579 		return ret;
1580 
1581 	last_cmd_set(raw_command);
1582 
1583 	ret = check_command(raw_command);
1584 	if (ret) {
1585 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1586 		return ret;
1587 	}
1588 
1589 	p = strpbrk(raw_command, " \t");
1590 	if (!p && raw_command[0] != '!') {
1591 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
1592 		ret = -EINVAL;
1593 		goto free;
1594 	}
1595 
1596 	name = kmemdup_nul(raw_command, p ? p - raw_command : strlen(raw_command), GFP_KERNEL);
1597 	if (!name)
1598 		return -ENOMEM;
1599 
1600 	if (name[0] == '!') {
1601 		ret = synth_event_delete(name + 1);
1602 		goto free;
1603 	}
1604 
1605 	fields = skip_spaces(p);
1606 
1607 	ret = __create_synth_event(name, fields);
1608 free:
1609 	kfree(name);
1610 
1611 	return ret;
1612 }
1613 
1614 static int synth_event_run_command(struct dynevent_cmd *cmd)
1615 {
1616 	struct synth_event *se;
1617 	int ret;
1618 
1619 	ret = create_or_delete_synth_event(cmd->seq.buffer);
1620 	if (ret)
1621 		return ret;
1622 
1623 	se = find_synth_event(cmd->event_name);
1624 	if (WARN_ON(!se))
1625 		return -ENOENT;
1626 
1627 	se->mod = cmd->private_data;
1628 
1629 	return ret;
1630 }
1631 
1632 /**
1633  * synth_event_cmd_init - Initialize a synthetic event command object
1634  * @cmd: A pointer to the dynevent_cmd struct representing the new event
1635  * @buf: A pointer to the buffer used to build the command
1636  * @maxlen: The length of the buffer passed in @buf
1637  *
1638  * Initialize a synthetic event command object.  Use this before
1639  * calling any of the other dyenvent_cmd functions.
1640  */
1641 void synth_event_cmd_init(struct dynevent_cmd *cmd, char *buf, int maxlen)
1642 {
1643 	dynevent_cmd_init(cmd, buf, maxlen, DYNEVENT_TYPE_SYNTH,
1644 			  synth_event_run_command);
1645 }
1646 EXPORT_SYMBOL_GPL(synth_event_cmd_init);
1647 
1648 static inline int
1649 __synth_event_trace_init(struct trace_event_file *file,
1650 			 struct synth_event_trace_state *trace_state)
1651 {
1652 	int ret = 0;
1653 
1654 	memset(trace_state, '\0', sizeof(*trace_state));
1655 
1656 	/*
1657 	 * Normal event tracing doesn't get called at all unless the
1658 	 * ENABLED bit is set (which attaches the probe thus allowing
1659 	 * this code to be called, etc).  Because this is called
1660 	 * directly by the user, we don't have that but we still need
1661 	 * to honor not logging when disabled.  For the iterated
1662 	 * trace case, we save the enabled state upon start and just
1663 	 * ignore the following data calls.
1664 	 */
1665 	if (!(file->flags & EVENT_FILE_FL_ENABLED) ||
1666 	    trace_trigger_soft_disabled(file)) {
1667 		trace_state->disabled = true;
1668 		ret = -ENOENT;
1669 		goto out;
1670 	}
1671 
1672 	trace_state->event = file->event_call->data;
1673 out:
1674 	return ret;
1675 }
1676 
1677 static inline int
1678 __synth_event_trace_start(struct trace_event_file *file,
1679 			  struct synth_event_trace_state *trace_state,
1680 			  int dynamic_fields_size)
1681 {
1682 	int entry_size, fields_size = 0;
1683 	int ret = 0;
1684 
1685 	fields_size = trace_state->event->n_u64 * sizeof(u64);
1686 	fields_size += dynamic_fields_size;
1687 
1688 	/*
1689 	 * Avoid ring buffer recursion detection, as this event
1690 	 * is being performed within another event.
1691 	 */
1692 	trace_state->buffer = file->tr->array_buffer.buffer;
1693 	ring_buffer_nest_start(trace_state->buffer);
1694 
1695 	entry_size = sizeof(*trace_state->entry) + fields_size;
1696 	trace_state->entry = trace_event_buffer_reserve(&trace_state->fbuffer,
1697 							file,
1698 							entry_size);
1699 	if (!trace_state->entry) {
1700 		ring_buffer_nest_end(trace_state->buffer);
1701 		ret = -EINVAL;
1702 	}
1703 
1704 	return ret;
1705 }
1706 
1707 static inline void
1708 __synth_event_trace_end(struct synth_event_trace_state *trace_state)
1709 {
1710 	trace_event_buffer_commit(&trace_state->fbuffer);
1711 
1712 	ring_buffer_nest_end(trace_state->buffer);
1713 }
1714 
1715 /**
1716  * synth_event_trace - Trace a synthetic event
1717  * @file: The trace_event_file representing the synthetic event
1718  * @n_vals: The number of values in vals
1719  * @args: Variable number of args containing the event values
1720  *
1721  * Trace a synthetic event using the values passed in the variable
1722  * argument list.
1723  *
1724  * The argument list should be a list 'n_vals' u64 values.  The number
1725  * of vals must match the number of field in the synthetic event, and
1726  * must be in the same order as the synthetic event fields.
1727  *
1728  * All vals should be cast to u64, and string vals are just pointers
1729  * to strings, cast to u64.  Strings will be copied into space
1730  * reserved in the event for the string, using these pointers.
1731  *
1732  * Return: 0 on success, err otherwise.
1733  */
1734 int synth_event_trace(struct trace_event_file *file, unsigned int n_vals, ...)
1735 {
1736 	unsigned int i, n_u64, len, data_size = 0;
1737 	struct synth_event_trace_state state;
1738 	va_list args;
1739 	int ret;
1740 
1741 	ret = __synth_event_trace_init(file, &state);
1742 	if (ret) {
1743 		if (ret == -ENOENT)
1744 			ret = 0; /* just disabled, not really an error */
1745 		return ret;
1746 	}
1747 
1748 	if (state.event->n_dynamic_fields) {
1749 		va_start(args, n_vals);
1750 
1751 		for (i = 0; i < state.event->n_fields; i++) {
1752 			u64 val = va_arg(args, u64);
1753 
1754 			if (state.event->fields[i]->is_string &&
1755 			    state.event->fields[i]->is_dynamic) {
1756 				char *str_val = (char *)(long)val;
1757 
1758 				data_size += strlen(str_val) + 1;
1759 			}
1760 		}
1761 
1762 		va_end(args);
1763 	}
1764 
1765 	ret = __synth_event_trace_start(file, &state, data_size);
1766 	if (ret)
1767 		return ret;
1768 
1769 	if (n_vals != state.event->n_fields) {
1770 		ret = -EINVAL;
1771 		goto out;
1772 	}
1773 
1774 	data_size = 0;
1775 
1776 	va_start(args, n_vals);
1777 	for (i = 0, n_u64 = 0; i < state.event->n_fields; i++) {
1778 		u64 val;
1779 
1780 		val = va_arg(args, u64);
1781 
1782 		if (state.event->fields[i]->is_string) {
1783 			char *str_val = (char *)(long)val;
1784 
1785 			len = trace_string(state.entry, state.event, str_val,
1786 					   state.event->fields[i]->is_dynamic,
1787 					   data_size, &n_u64);
1788 			data_size += len; /* only dynamic string increments */
1789 		} else {
1790 			struct synth_field *field = state.event->fields[i];
1791 
1792 			switch (field->size) {
1793 			case 1:
1794 				*(u8 *)&state.entry->fields[n_u64] = (u8)val;
1795 				break;
1796 
1797 			case 2:
1798 				*(u16 *)&state.entry->fields[n_u64] = (u16)val;
1799 				break;
1800 
1801 			case 4:
1802 				*(u32 *)&state.entry->fields[n_u64] = (u32)val;
1803 				break;
1804 
1805 			default:
1806 				state.entry->fields[n_u64] = val;
1807 				break;
1808 			}
1809 			n_u64++;
1810 		}
1811 	}
1812 	va_end(args);
1813 out:
1814 	__synth_event_trace_end(&state);
1815 
1816 	return ret;
1817 }
1818 EXPORT_SYMBOL_GPL(synth_event_trace);
1819 
1820 /**
1821  * synth_event_trace_array - Trace a synthetic event from an array
1822  * @file: The trace_event_file representing the synthetic event
1823  * @vals: Array of values
1824  * @n_vals: The number of values in vals
1825  *
1826  * Trace a synthetic event using the values passed in as 'vals'.
1827  *
1828  * The 'vals' array is just an array of 'n_vals' u64.  The number of
1829  * vals must match the number of field in the synthetic event, and
1830  * must be in the same order as the synthetic event fields.
1831  *
1832  * All vals should be cast to u64, and string vals are just pointers
1833  * to strings, cast to u64.  Strings will be copied into space
1834  * reserved in the event for the string, using these pointers.
1835  *
1836  * Return: 0 on success, err otherwise.
1837  */
1838 int synth_event_trace_array(struct trace_event_file *file, u64 *vals,
1839 			    unsigned int n_vals)
1840 {
1841 	unsigned int i, n_u64, field_pos, len, data_size = 0;
1842 	struct synth_event_trace_state state;
1843 	char *str_val;
1844 	int ret;
1845 
1846 	ret = __synth_event_trace_init(file, &state);
1847 	if (ret) {
1848 		if (ret == -ENOENT)
1849 			ret = 0; /* just disabled, not really an error */
1850 		return ret;
1851 	}
1852 
1853 	if (state.event->n_dynamic_fields) {
1854 		for (i = 0; i < state.event->n_dynamic_fields; i++) {
1855 			field_pos = state.event->dynamic_fields[i]->field_pos;
1856 			str_val = (char *)(long)vals[field_pos];
1857 			len = strlen(str_val) + 1;
1858 			data_size += len;
1859 		}
1860 	}
1861 
1862 	ret = __synth_event_trace_start(file, &state, data_size);
1863 	if (ret)
1864 		return ret;
1865 
1866 	if (n_vals != state.event->n_fields) {
1867 		ret = -EINVAL;
1868 		goto out;
1869 	}
1870 
1871 	data_size = 0;
1872 
1873 	for (i = 0, n_u64 = 0; i < state.event->n_fields; i++) {
1874 		if (state.event->fields[i]->is_string) {
1875 			char *str_val = (char *)(long)vals[i];
1876 
1877 			len = trace_string(state.entry, state.event, str_val,
1878 					   state.event->fields[i]->is_dynamic,
1879 					   data_size, &n_u64);
1880 			data_size += len; /* only dynamic string increments */
1881 		} else {
1882 			struct synth_field *field = state.event->fields[i];
1883 			u64 val = vals[i];
1884 
1885 			switch (field->size) {
1886 			case 1:
1887 				*(u8 *)&state.entry->fields[n_u64] = (u8)val;
1888 				break;
1889 
1890 			case 2:
1891 				*(u16 *)&state.entry->fields[n_u64] = (u16)val;
1892 				break;
1893 
1894 			case 4:
1895 				*(u32 *)&state.entry->fields[n_u64] = (u32)val;
1896 				break;
1897 
1898 			default:
1899 				state.entry->fields[n_u64] = val;
1900 				break;
1901 			}
1902 			n_u64++;
1903 		}
1904 	}
1905 out:
1906 	__synth_event_trace_end(&state);
1907 
1908 	return ret;
1909 }
1910 EXPORT_SYMBOL_GPL(synth_event_trace_array);
1911 
1912 /**
1913  * synth_event_trace_start - Start piecewise synthetic event trace
1914  * @file: The trace_event_file representing the synthetic event
1915  * @trace_state: A pointer to object tracking the piecewise trace state
1916  *
1917  * Start the trace of a synthetic event field-by-field rather than all
1918  * at once.
1919  *
1920  * This function 'opens' an event trace, which means space is reserved
1921  * for the event in the trace buffer, after which the event's
1922  * individual field values can be set through either
1923  * synth_event_add_next_val() or synth_event_add_val().
1924  *
1925  * A pointer to a trace_state object is passed in, which will keep
1926  * track of the current event trace state until the event trace is
1927  * closed (and the event finally traced) using
1928  * synth_event_trace_end().
1929  *
1930  * Note that synth_event_trace_end() must be called after all values
1931  * have been added for each event trace, regardless of whether adding
1932  * all field values succeeded or not.
1933  *
1934  * Note also that for a given event trace, all fields must be added
1935  * using either synth_event_add_next_val() or synth_event_add_val()
1936  * but not both together or interleaved.
1937  *
1938  * Return: 0 on success, err otherwise.
1939  */
1940 int synth_event_trace_start(struct trace_event_file *file,
1941 			    struct synth_event_trace_state *trace_state)
1942 {
1943 	int ret;
1944 
1945 	if (!trace_state)
1946 		return -EINVAL;
1947 
1948 	ret = __synth_event_trace_init(file, trace_state);
1949 	if (ret) {
1950 		if (ret == -ENOENT)
1951 			ret = 0; /* just disabled, not really an error */
1952 		return ret;
1953 	}
1954 
1955 	if (trace_state->event->n_dynamic_fields)
1956 		return -ENOTSUPP;
1957 
1958 	ret = __synth_event_trace_start(file, trace_state, 0);
1959 
1960 	return ret;
1961 }
1962 EXPORT_SYMBOL_GPL(synth_event_trace_start);
1963 
1964 static int __synth_event_add_val(const char *field_name, u64 val,
1965 				 struct synth_event_trace_state *trace_state)
1966 {
1967 	struct synth_field *field = NULL;
1968 	struct synth_trace_event *entry;
1969 	struct synth_event *event;
1970 	int i, ret = 0;
1971 
1972 	if (!trace_state) {
1973 		ret = -EINVAL;
1974 		goto out;
1975 	}
1976 
1977 	/* can't mix add_next_synth_val() with add_synth_val() */
1978 	if (field_name) {
1979 		if (trace_state->add_next) {
1980 			ret = -EINVAL;
1981 			goto out;
1982 		}
1983 		trace_state->add_name = true;
1984 	} else {
1985 		if (trace_state->add_name) {
1986 			ret = -EINVAL;
1987 			goto out;
1988 		}
1989 		trace_state->add_next = true;
1990 	}
1991 
1992 	if (trace_state->disabled)
1993 		goto out;
1994 
1995 	event = trace_state->event;
1996 	if (trace_state->add_name) {
1997 		for (i = 0; i < event->n_fields; i++) {
1998 			field = event->fields[i];
1999 			if (strcmp(field->name, field_name) == 0)
2000 				break;
2001 		}
2002 		if (!field) {
2003 			ret = -EINVAL;
2004 			goto out;
2005 		}
2006 	} else {
2007 		if (trace_state->cur_field >= event->n_fields) {
2008 			ret = -EINVAL;
2009 			goto out;
2010 		}
2011 		field = event->fields[trace_state->cur_field++];
2012 	}
2013 
2014 	entry = trace_state->entry;
2015 	if (field->is_string) {
2016 		char *str_val = (char *)(long)val;
2017 		char *str_field;
2018 
2019 		if (field->is_dynamic) { /* add_val can't do dynamic strings */
2020 			ret = -EINVAL;
2021 			goto out;
2022 		}
2023 
2024 		if (!str_val) {
2025 			ret = -EINVAL;
2026 			goto out;
2027 		}
2028 
2029 		str_field = (char *)&entry->fields[field->offset];
2030 		strscpy(str_field, str_val, STR_VAR_LEN_MAX);
2031 	} else {
2032 		switch (field->size) {
2033 		case 1:
2034 			*(u8 *)&trace_state->entry->fields[field->offset] = (u8)val;
2035 			break;
2036 
2037 		case 2:
2038 			*(u16 *)&trace_state->entry->fields[field->offset] = (u16)val;
2039 			break;
2040 
2041 		case 4:
2042 			*(u32 *)&trace_state->entry->fields[field->offset] = (u32)val;
2043 			break;
2044 
2045 		default:
2046 			trace_state->entry->fields[field->offset] = val;
2047 			break;
2048 		}
2049 	}
2050  out:
2051 	return ret;
2052 }
2053 
2054 /**
2055  * synth_event_add_next_val - Add the next field's value to an open synth trace
2056  * @val: The value to set the next field to
2057  * @trace_state: A pointer to object tracking the piecewise trace state
2058  *
2059  * Set the value of the next field in an event that's been opened by
2060  * synth_event_trace_start().
2061  *
2062  * The val param should be the value cast to u64.  If the value points
2063  * to a string, the val param should be a char * cast to u64.
2064  *
2065  * This function assumes all the fields in an event are to be set one
2066  * after another - successive calls to this function are made, one for
2067  * each field, in the order of the fields in the event, until all
2068  * fields have been set.  If you'd rather set each field individually
2069  * without regard to ordering, synth_event_add_val() can be used
2070  * instead.
2071  *
2072  * Note however that synth_event_add_next_val() and
2073  * synth_event_add_val() can't be intermixed for a given event trace -
2074  * one or the other but not both can be used at the same time.
2075  *
2076  * Note also that synth_event_trace_end() must be called after all
2077  * values have been added for each event trace, regardless of whether
2078  * adding all field values succeeded or not.
2079  *
2080  * Return: 0 on success, err otherwise.
2081  */
2082 int synth_event_add_next_val(u64 val,
2083 			     struct synth_event_trace_state *trace_state)
2084 {
2085 	return __synth_event_add_val(NULL, val, trace_state);
2086 }
2087 EXPORT_SYMBOL_GPL(synth_event_add_next_val);
2088 
2089 /**
2090  * synth_event_add_val - Add a named field's value to an open synth trace
2091  * @field_name: The name of the synthetic event field value to set
2092  * @val: The value to set the named field to
2093  * @trace_state: A pointer to object tracking the piecewise trace state
2094  *
2095  * Set the value of the named field in an event that's been opened by
2096  * synth_event_trace_start().
2097  *
2098  * The val param should be the value cast to u64.  If the value points
2099  * to a string, the val param should be a char * cast to u64.
2100  *
2101  * This function looks up the field name, and if found, sets the field
2102  * to the specified value.  This lookup makes this function more
2103  * expensive than synth_event_add_next_val(), so use that or the
2104  * none-piecewise synth_event_trace() instead if efficiency is more
2105  * important.
2106  *
2107  * Note however that synth_event_add_next_val() and
2108  * synth_event_add_val() can't be intermixed for a given event trace -
2109  * one or the other but not both can be used at the same time.
2110  *
2111  * Note also that synth_event_trace_end() must be called after all
2112  * values have been added for each event trace, regardless of whether
2113  * adding all field values succeeded or not.
2114  *
2115  * Return: 0 on success, err otherwise.
2116  */
2117 int synth_event_add_val(const char *field_name, u64 val,
2118 			struct synth_event_trace_state *trace_state)
2119 {
2120 	return __synth_event_add_val(field_name, val, trace_state);
2121 }
2122 EXPORT_SYMBOL_GPL(synth_event_add_val);
2123 
2124 /**
2125  * synth_event_trace_end - End piecewise synthetic event trace
2126  * @trace_state: A pointer to object tracking the piecewise trace state
2127  *
2128  * End the trace of a synthetic event opened by
2129  * synth_event_trace__start().
2130  *
2131  * This function 'closes' an event trace, which basically means that
2132  * it commits the reserved event and cleans up other loose ends.
2133  *
2134  * A pointer to a trace_state object is passed in, which will keep
2135  * track of the current event trace state opened with
2136  * synth_event_trace_start().
2137  *
2138  * Note that this function must be called after all values have been
2139  * added for each event trace, regardless of whether adding all field
2140  * values succeeded or not.
2141  *
2142  * Return: 0 on success, err otherwise.
2143  */
2144 int synth_event_trace_end(struct synth_event_trace_state *trace_state)
2145 {
2146 	if (!trace_state)
2147 		return -EINVAL;
2148 
2149 	__synth_event_trace_end(trace_state);
2150 
2151 	return 0;
2152 }
2153 EXPORT_SYMBOL_GPL(synth_event_trace_end);
2154 
2155 static int create_synth_event(const char *raw_command)
2156 {
2157 	char *fields, *p;
2158 	const char *name;
2159 	int len, ret = 0;
2160 
2161 	raw_command = skip_spaces(raw_command);
2162 	if (raw_command[0] == '\0')
2163 		return ret;
2164 
2165 	last_cmd_set(raw_command);
2166 
2167 	name = raw_command;
2168 
2169 	/* Don't try to process if not our system */
2170 	if (name[0] != 's' || name[1] != ':')
2171 		return -ECANCELED;
2172 	name += 2;
2173 
2174 	p = strpbrk(raw_command, " \t");
2175 	if (!p) {
2176 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
2177 		return -EINVAL;
2178 	}
2179 
2180 	fields = skip_spaces(p);
2181 
2182 	/* This interface accepts group name prefix */
2183 	if (strchr(name, '/')) {
2184 		len = str_has_prefix(name, SYNTH_SYSTEM "/");
2185 		if (len == 0) {
2186 			synth_err(SYNTH_ERR_INVALID_DYN_CMD, 0);
2187 			return -EINVAL;
2188 		}
2189 		name += len;
2190 	}
2191 
2192 	len = name - raw_command;
2193 
2194 	ret = check_command(raw_command + len);
2195 	if (ret) {
2196 		synth_err(SYNTH_ERR_INVALID_CMD, 0);
2197 		return ret;
2198 	}
2199 
2200 	name = kmemdup_nul(raw_command + len, p - raw_command - len, GFP_KERNEL);
2201 	if (!name)
2202 		return -ENOMEM;
2203 
2204 	ret = __create_synth_event(name, fields);
2205 
2206 	kfree(name);
2207 
2208 	return ret;
2209 }
2210 
2211 static int synth_event_release(struct dyn_event *ev)
2212 {
2213 	struct synth_event *event = to_synth_event(ev);
2214 	int ret;
2215 
2216 	if (event->ref)
2217 		return -EBUSY;
2218 
2219 	if (trace_event_dyn_busy(&event->call))
2220 		return -EBUSY;
2221 
2222 	ret = unregister_synth_event(event);
2223 	if (ret)
2224 		return ret;
2225 
2226 	dyn_event_remove(ev);
2227 	free_synth_event(event);
2228 	return 0;
2229 }
2230 
2231 static int __synth_event_show(struct seq_file *m, struct synth_event *event)
2232 {
2233 	struct synth_field *field;
2234 	unsigned int i;
2235 	char *type, *t;
2236 
2237 	seq_printf(m, "%s\t", event->name);
2238 
2239 	for (i = 0; i < event->n_fields; i++) {
2240 		field = event->fields[i];
2241 
2242 		type = field->type;
2243 		t = strstr(type, "__data_loc");
2244 		if (t) { /* __data_loc belongs in format but not event desc */
2245 			t += sizeof("__data_loc");
2246 			type = t;
2247 		}
2248 
2249 		/* parameter values */
2250 		seq_printf(m, "%s %s%s", type, field->name,
2251 			   i == event->n_fields - 1 ? "" : "; ");
2252 	}
2253 
2254 	seq_putc(m, '\n');
2255 
2256 	return 0;
2257 }
2258 
2259 static int synth_event_show(struct seq_file *m, struct dyn_event *ev)
2260 {
2261 	struct synth_event *event = to_synth_event(ev);
2262 
2263 	seq_printf(m, "s:%s/", event->class.system);
2264 
2265 	return __synth_event_show(m, event);
2266 }
2267 
2268 static int synth_events_seq_show(struct seq_file *m, void *v)
2269 {
2270 	struct dyn_event *ev = v;
2271 
2272 	if (!is_synth_event(ev))
2273 		return 0;
2274 
2275 	return __synth_event_show(m, to_synth_event(ev));
2276 }
2277 
2278 static const struct seq_operations synth_events_seq_op = {
2279 	.start	= dyn_event_seq_start,
2280 	.next	= dyn_event_seq_next,
2281 	.stop	= dyn_event_seq_stop,
2282 	.show	= synth_events_seq_show,
2283 };
2284 
2285 static int synth_events_open(struct inode *inode, struct file *file)
2286 {
2287 	int ret;
2288 
2289 	ret = security_locked_down(LOCKDOWN_TRACEFS);
2290 	if (ret)
2291 		return ret;
2292 
2293 	if ((file->f_mode & FMODE_WRITE) && (file->f_flags & O_TRUNC)) {
2294 		ret = dyn_events_release_all(&synth_event_ops);
2295 		if (ret < 0)
2296 			return ret;
2297 	}
2298 
2299 	return seq_open(file, &synth_events_seq_op);
2300 }
2301 
2302 static ssize_t synth_events_write(struct file *file,
2303 				  const char __user *buffer,
2304 				  size_t count, loff_t *ppos)
2305 {
2306 	return trace_parse_run_command(file, buffer, count, ppos,
2307 				       create_or_delete_synth_event);
2308 }
2309 
2310 static const struct file_operations synth_events_fops = {
2311 	.open           = synth_events_open,
2312 	.write		= synth_events_write,
2313 	.read           = seq_read,
2314 	.llseek         = seq_lseek,
2315 	.release        = seq_release,
2316 };
2317 
2318 /*
2319  * Register dynevent at core_initcall. This allows kernel to setup kprobe
2320  * events in postcore_initcall without tracefs.
2321  */
2322 static __init int trace_events_synth_init_early(void)
2323 {
2324 	int err = 0;
2325 
2326 	err = dyn_event_register(&synth_event_ops);
2327 	if (err)
2328 		pr_warn("Could not register synth_event_ops\n");
2329 
2330 	return err;
2331 }
2332 core_initcall(trace_events_synth_init_early);
2333 
2334 static __init int trace_events_synth_init(void)
2335 {
2336 	struct dentry *entry = NULL;
2337 	int err = 0;
2338 	err = tracing_init_dentry();
2339 	if (err)
2340 		goto err;
2341 
2342 	entry = tracefs_create_file("synthetic_events", TRACE_MODE_WRITE,
2343 				    NULL, NULL, &synth_events_fops);
2344 	if (!entry) {
2345 		err = -ENODEV;
2346 		goto err;
2347 	}
2348 
2349 	return err;
2350  err:
2351 	pr_warn("Could not create tracefs 'synthetic_events' entry\n");
2352 
2353 	return err;
2354 }
2355 
2356 fs_initcall(trace_events_synth_init);
2357