1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright(C) 2015-2018 Linaro Limited.
4  *
5  * Author: Tor Jeremiassen <tor@ti.com>
6  * Author: Mathieu Poirier <mathieu.poirier@linaro.org>
7  */
8 
9 #include <asm/bug.h>
10 #include <linux/coresight-pmu.h>
11 #include <linux/err.h>
12 #include <linux/list.h>
13 #include <linux/zalloc.h>
14 #include <stdlib.h>
15 #include <opencsd/c_api/opencsd_c_api.h>
16 
17 #include "cs-etm.h"
18 #include "cs-etm-decoder.h"
19 #include "debug.h"
20 #include "intlist.h"
21 
22 /* use raw logging */
23 #ifdef CS_DEBUG_RAW
24 #define CS_LOG_RAW_FRAMES
25 #ifdef CS_RAW_PACKED
26 #define CS_RAW_DEBUG_FLAGS (OCSD_DFRMTR_UNPACKED_RAW_OUT | \
27 			    OCSD_DFRMTR_PACKED_RAW_OUT)
28 #else
29 #define CS_RAW_DEBUG_FLAGS (OCSD_DFRMTR_UNPACKED_RAW_OUT)
30 #endif
31 #endif
32 
33 struct cs_etm_decoder {
34 	void *data;
35 	void (*packet_printer)(const char *msg);
36 	bool suppress_printing;
37 	dcd_tree_handle_t dcd_tree;
38 	cs_etm_mem_cb_type mem_access;
39 	ocsd_datapath_resp_t prev_return;
40 	const char *decoder_name;
41 };
42 
43 static u32
44 cs_etm_decoder__mem_access(const void *context,
45 			   const ocsd_vaddr_t address,
46 			   const ocsd_mem_space_acc_t mem_space __maybe_unused,
47 			   const u8 trace_chan_id,
48 			   const u32 req_size,
49 			   u8 *buffer)
50 {
51 	struct cs_etm_decoder *decoder = (struct cs_etm_decoder *) context;
52 
53 	return decoder->mem_access(decoder->data, trace_chan_id,
54 				   address, req_size, buffer);
55 }
56 
57 int cs_etm_decoder__add_mem_access_cb(struct cs_etm_decoder *decoder,
58 				      u64 start, u64 end,
59 				      cs_etm_mem_cb_type cb_func)
60 {
61 	decoder->mem_access = cb_func;
62 
63 	if (ocsd_dt_add_callback_trcid_mem_acc(decoder->dcd_tree, start, end,
64 					       OCSD_MEM_SPACE_ANY,
65 					       cs_etm_decoder__mem_access,
66 					       decoder))
67 		return -1;
68 
69 	return 0;
70 }
71 
72 int cs_etm_decoder__reset(struct cs_etm_decoder *decoder)
73 {
74 	ocsd_datapath_resp_t dp_ret;
75 
76 	decoder->prev_return = OCSD_RESP_CONT;
77 	decoder->suppress_printing = true;
78 	dp_ret = ocsd_dt_process_data(decoder->dcd_tree, OCSD_OP_RESET,
79 				      0, 0, NULL, NULL);
80 	decoder->suppress_printing = false;
81 	if (OCSD_DATA_RESP_IS_FATAL(dp_ret))
82 		return -1;
83 
84 	return 0;
85 }
86 
87 int cs_etm_decoder__get_packet(struct cs_etm_packet_queue *packet_queue,
88 			       struct cs_etm_packet *packet)
89 {
90 	if (!packet_queue || !packet)
91 		return -EINVAL;
92 
93 	/* Nothing to do, might as well just return */
94 	if (packet_queue->packet_count == 0)
95 		return 0;
96 	/*
97 	 * The queueing process in function cs_etm_decoder__buffer_packet()
98 	 * increments the tail *before* using it.  This is somewhat counter
99 	 * intuitive but it has the advantage of centralizing tail management
100 	 * at a single location.  Because of that we need to follow the same
101 	 * heuristic with the head, i.e we increment it before using its
102 	 * value.  Otherwise the first element of the packet queue is not
103 	 * used.
104 	 */
105 	packet_queue->head = (packet_queue->head + 1) &
106 			     (CS_ETM_PACKET_MAX_BUFFER - 1);
107 
108 	*packet = packet_queue->packet_buffer[packet_queue->head];
109 
110 	packet_queue->packet_count--;
111 
112 	return 1;
113 }
114 
115 static int cs_etm_decoder__gen_etmv3_config(struct cs_etm_trace_params *params,
116 					    ocsd_etmv3_cfg *config)
117 {
118 	config->reg_idr = params->etmv3.reg_idr;
119 	config->reg_ctrl = params->etmv3.reg_ctrl;
120 	config->reg_ccer = params->etmv3.reg_ccer;
121 	config->reg_trc_id = params->etmv3.reg_trc_id;
122 	config->arch_ver = ARCH_V7;
123 	config->core_prof = profile_CortexA;
124 
125 	return 0;
126 }
127 
128 #define TRCIDR1_TRCARCHMIN_SHIFT 4
129 #define TRCIDR1_TRCARCHMIN_MASK  GENMASK(7, 4)
130 #define TRCIDR1_TRCARCHMIN(x)    (((x) & TRCIDR1_TRCARCHMIN_MASK) >> TRCIDR1_TRCARCHMIN_SHIFT)
131 
132 static enum _ocsd_arch_version cs_etm_decoder__get_etmv4_arch_ver(u32 reg_idr1)
133 {
134 	/*
135 	 * For ETMv4 if the trace minor version is 4 or more then we can assume
136 	 * the architecture is ARCH_AA64 rather than just V8.
137 	 * ARCH_V8 = V8 architecture
138 	 * ARCH_AA64 = Min v8r3 plus additional AA64 PE features
139 	 */
140 	return TRCIDR1_TRCARCHMIN(reg_idr1) >= 4 ? ARCH_AA64 : ARCH_V8;
141 }
142 
143 static void cs_etm_decoder__gen_etmv4_config(struct cs_etm_trace_params *params,
144 					     ocsd_etmv4_cfg *config)
145 {
146 	config->reg_configr = params->etmv4.reg_configr;
147 	config->reg_traceidr = params->etmv4.reg_traceidr;
148 	config->reg_idr0 = params->etmv4.reg_idr0;
149 	config->reg_idr1 = params->etmv4.reg_idr1;
150 	config->reg_idr2 = params->etmv4.reg_idr2;
151 	config->reg_idr8 = params->etmv4.reg_idr8;
152 	config->reg_idr9 = 0;
153 	config->reg_idr10 = 0;
154 	config->reg_idr11 = 0;
155 	config->reg_idr12 = 0;
156 	config->reg_idr13 = 0;
157 	config->arch_ver = cs_etm_decoder__get_etmv4_arch_ver(params->etmv4.reg_idr1);
158 	config->core_prof = profile_CortexA;
159 }
160 
161 static void cs_etm_decoder__gen_ete_config(struct cs_etm_trace_params *params,
162 					   ocsd_ete_cfg *config)
163 {
164 	config->reg_configr = params->ete.reg_configr;
165 	config->reg_traceidr = params->ete.reg_traceidr;
166 	config->reg_idr0 = params->ete.reg_idr0;
167 	config->reg_idr1 = params->ete.reg_idr1;
168 	config->reg_idr2 = params->ete.reg_idr2;
169 	config->reg_idr8 = params->ete.reg_idr8;
170 	config->reg_devarch = params->ete.reg_devarch;
171 	config->arch_ver = ARCH_AA64;
172 	config->core_prof = profile_CortexA;
173 }
174 
175 static void cs_etm_decoder__print_str_cb(const void *p_context,
176 					 const char *msg,
177 					 const int str_len)
178 {
179 	const struct cs_etm_decoder *decoder = p_context;
180 
181 	if (p_context && str_len && !decoder->suppress_printing)
182 		decoder->packet_printer(msg);
183 }
184 
185 static int
186 cs_etm_decoder__init_def_logger_printing(struct cs_etm_decoder_params *d_params,
187 					 struct cs_etm_decoder *decoder)
188 {
189 	int ret = 0;
190 
191 	if (d_params->packet_printer == NULL)
192 		return -1;
193 
194 	decoder->packet_printer = d_params->packet_printer;
195 
196 	/*
197 	 * Set up a library default logger to process any printers
198 	 * (packet/raw frame) we add later.
199 	 */
200 	ret = ocsd_def_errlog_init(OCSD_ERR_SEV_ERROR, 1);
201 	if (ret != 0)
202 		return -1;
203 
204 	/* no stdout / err / file output */
205 	ret = ocsd_def_errlog_config_output(C_API_MSGLOGOUT_FLG_NONE, NULL);
206 	if (ret != 0)
207 		return -1;
208 
209 	/*
210 	 * Set the string CB for the default logger, passes strings to
211 	 * perf print logger.
212 	 */
213 	ret = ocsd_def_errlog_set_strprint_cb(decoder->dcd_tree,
214 					      (void *)decoder,
215 					      cs_etm_decoder__print_str_cb);
216 	if (ret != 0)
217 		ret = -1;
218 
219 	return 0;
220 }
221 
222 #ifdef CS_LOG_RAW_FRAMES
223 static void
224 cs_etm_decoder__init_raw_frame_logging(struct cs_etm_decoder_params *d_params,
225 				       struct cs_etm_decoder *decoder)
226 {
227 	/* Only log these during a --dump operation */
228 	if (d_params->operation == CS_ETM_OPERATION_PRINT) {
229 		/* set up a library default logger to process the
230 		 *  raw frame printer we add later
231 		 */
232 		ocsd_def_errlog_init(OCSD_ERR_SEV_ERROR, 1);
233 
234 		/* no stdout / err / file output */
235 		ocsd_def_errlog_config_output(C_API_MSGLOGOUT_FLG_NONE, NULL);
236 
237 		/* set the string CB for the default logger,
238 		 * passes strings to perf print logger.
239 		 */
240 		ocsd_def_errlog_set_strprint_cb(decoder->dcd_tree,
241 						(void *)decoder,
242 						cs_etm_decoder__print_str_cb);
243 
244 		/* use the built in library printer for the raw frames */
245 		ocsd_dt_set_raw_frame_printer(decoder->dcd_tree,
246 					      CS_RAW_DEBUG_FLAGS);
247 	}
248 }
249 #else
250 static void
251 cs_etm_decoder__init_raw_frame_logging(
252 		struct cs_etm_decoder_params *d_params __maybe_unused,
253 		struct cs_etm_decoder *decoder __maybe_unused)
254 {
255 }
256 #endif
257 
258 static ocsd_datapath_resp_t
259 cs_etm_decoder__do_soft_timestamp(struct cs_etm_queue *etmq,
260 				  struct cs_etm_packet_queue *packet_queue,
261 				  const uint8_t trace_chan_id)
262 {
263 	/* No timestamp packet has been received, nothing to do */
264 	if (!packet_queue->cs_timestamp)
265 		return OCSD_RESP_CONT;
266 
267 	packet_queue->cs_timestamp = packet_queue->next_cs_timestamp;
268 
269 	/* Estimate the timestamp for the next range packet */
270 	packet_queue->next_cs_timestamp += packet_queue->instr_count;
271 	packet_queue->instr_count = 0;
272 
273 	/* Tell the front end which traceid_queue needs attention */
274 	cs_etm__etmq_set_traceid_queue_timestamp(etmq, trace_chan_id);
275 
276 	return OCSD_RESP_WAIT;
277 }
278 
279 static ocsd_datapath_resp_t
280 cs_etm_decoder__do_hard_timestamp(struct cs_etm_queue *etmq,
281 				  const ocsd_generic_trace_elem *elem,
282 				  const uint8_t trace_chan_id,
283 				  const ocsd_trc_index_t indx)
284 {
285 	struct cs_etm_packet_queue *packet_queue;
286 
287 	/* First get the packet queue for this traceID */
288 	packet_queue = cs_etm__etmq_get_packet_queue(etmq, trace_chan_id);
289 	if (!packet_queue)
290 		return OCSD_RESP_FATAL_SYS_ERR;
291 
292 	/*
293 	 * We've seen a timestamp packet before - simply record the new value.
294 	 * Function do_soft_timestamp() will report the value to the front end,
295 	 * hence asking the decoder to keep decoding rather than stopping.
296 	 */
297 	if (packet_queue->cs_timestamp) {
298 		packet_queue->next_cs_timestamp = elem->timestamp;
299 		return OCSD_RESP_CONT;
300 	}
301 
302 
303 	if (!elem->timestamp) {
304 		/*
305 		 * Zero timestamps can be seen due to misconfiguration or hardware bugs.
306 		 * Warn once, and don't try to subtract instr_count as it would result in an
307 		 * underflow.
308 		 */
309 		packet_queue->cs_timestamp = 0;
310 		if (!cs_etm__etmq_is_timeless(etmq))
311 			pr_warning_once("Zero Coresight timestamp found at Idx:%" OCSD_TRC_IDX_STR
312 					". Decoding may be improved by prepending 'Z' to your current --itrace arguments.\n",
313 					indx);
314 
315 	} else if (packet_queue->instr_count > elem->timestamp) {
316 		/*
317 		 * Sanity check that the elem->timestamp - packet_queue->instr_count would not
318 		 * result in an underflow. Warn and clamp at 0 if it would.
319 		 */
320 		packet_queue->cs_timestamp = 0;
321 		pr_err("Timestamp calculation underflow at Idx:%" OCSD_TRC_IDX_STR "\n", indx);
322 	} else {
323 		/*
324 		 * This is the first timestamp we've seen since the beginning of traces
325 		 * or a discontinuity.  Since timestamps packets are generated *after*
326 		 * range packets have been generated, we need to estimate the time at
327 		 * which instructions started by subtracting the number of instructions
328 		 * executed to the timestamp.
329 		 */
330 		packet_queue->cs_timestamp = elem->timestamp - packet_queue->instr_count;
331 	}
332 	packet_queue->next_cs_timestamp = elem->timestamp;
333 	packet_queue->instr_count = 0;
334 
335 	/* Tell the front end which traceid_queue needs attention */
336 	cs_etm__etmq_set_traceid_queue_timestamp(etmq, trace_chan_id);
337 
338 	/* Halt processing until we are being told to proceed */
339 	return OCSD_RESP_WAIT;
340 }
341 
342 static void
343 cs_etm_decoder__reset_timestamp(struct cs_etm_packet_queue *packet_queue)
344 {
345 	packet_queue->cs_timestamp = 0;
346 	packet_queue->next_cs_timestamp = 0;
347 	packet_queue->instr_count = 0;
348 }
349 
350 static ocsd_datapath_resp_t
351 cs_etm_decoder__buffer_packet(struct cs_etm_packet_queue *packet_queue,
352 			      const u8 trace_chan_id,
353 			      enum cs_etm_sample_type sample_type)
354 {
355 	u32 et = 0;
356 	int cpu;
357 
358 	if (packet_queue->packet_count >= CS_ETM_PACKET_MAX_BUFFER - 1)
359 		return OCSD_RESP_FATAL_SYS_ERR;
360 
361 	if (cs_etm__get_cpu(trace_chan_id, &cpu) < 0)
362 		return OCSD_RESP_FATAL_SYS_ERR;
363 
364 	et = packet_queue->tail;
365 	et = (et + 1) & (CS_ETM_PACKET_MAX_BUFFER - 1);
366 	packet_queue->tail = et;
367 	packet_queue->packet_count++;
368 
369 	packet_queue->packet_buffer[et].sample_type = sample_type;
370 	packet_queue->packet_buffer[et].isa = CS_ETM_ISA_UNKNOWN;
371 	packet_queue->packet_buffer[et].cpu = cpu;
372 	packet_queue->packet_buffer[et].start_addr = CS_ETM_INVAL_ADDR;
373 	packet_queue->packet_buffer[et].end_addr = CS_ETM_INVAL_ADDR;
374 	packet_queue->packet_buffer[et].instr_count = 0;
375 	packet_queue->packet_buffer[et].last_instr_taken_branch = false;
376 	packet_queue->packet_buffer[et].last_instr_size = 0;
377 	packet_queue->packet_buffer[et].last_instr_type = 0;
378 	packet_queue->packet_buffer[et].last_instr_subtype = 0;
379 	packet_queue->packet_buffer[et].last_instr_cond = 0;
380 	packet_queue->packet_buffer[et].flags = 0;
381 	packet_queue->packet_buffer[et].exception_number = UINT32_MAX;
382 	packet_queue->packet_buffer[et].trace_chan_id = trace_chan_id;
383 
384 	if (packet_queue->packet_count == CS_ETM_PACKET_MAX_BUFFER - 1)
385 		return OCSD_RESP_WAIT;
386 
387 	return OCSD_RESP_CONT;
388 }
389 
390 static ocsd_datapath_resp_t
391 cs_etm_decoder__buffer_range(struct cs_etm_queue *etmq,
392 			     struct cs_etm_packet_queue *packet_queue,
393 			     const ocsd_generic_trace_elem *elem,
394 			     const uint8_t trace_chan_id)
395 {
396 	int ret = 0;
397 	struct cs_etm_packet *packet;
398 
399 	ret = cs_etm_decoder__buffer_packet(packet_queue, trace_chan_id,
400 					    CS_ETM_RANGE);
401 	if (ret != OCSD_RESP_CONT && ret != OCSD_RESP_WAIT)
402 		return ret;
403 
404 	packet = &packet_queue->packet_buffer[packet_queue->tail];
405 
406 	switch (elem->isa) {
407 	case ocsd_isa_aarch64:
408 		packet->isa = CS_ETM_ISA_A64;
409 		break;
410 	case ocsd_isa_arm:
411 		packet->isa = CS_ETM_ISA_A32;
412 		break;
413 	case ocsd_isa_thumb2:
414 		packet->isa = CS_ETM_ISA_T32;
415 		break;
416 	case ocsd_isa_tee:
417 	case ocsd_isa_jazelle:
418 	case ocsd_isa_custom:
419 	case ocsd_isa_unknown:
420 	default:
421 		packet->isa = CS_ETM_ISA_UNKNOWN;
422 	}
423 
424 	packet->start_addr = elem->st_addr;
425 	packet->end_addr = elem->en_addr;
426 	packet->instr_count = elem->num_instr_range;
427 	packet->last_instr_type = elem->last_i_type;
428 	packet->last_instr_subtype = elem->last_i_subtype;
429 	packet->last_instr_cond = elem->last_instr_cond;
430 
431 	if (elem->last_i_type == OCSD_INSTR_BR || elem->last_i_type == OCSD_INSTR_BR_INDIRECT)
432 		packet->last_instr_taken_branch = elem->last_instr_exec;
433 	else
434 		packet->last_instr_taken_branch = false;
435 
436 	packet->last_instr_size = elem->last_instr_sz;
437 
438 	/* per-thread scenario, no need to generate a timestamp */
439 	if (cs_etm__etmq_is_timeless(etmq))
440 		goto out;
441 
442 	/*
443 	 * The packet queue is full and we haven't seen a timestamp (had we
444 	 * seen one the packet queue wouldn't be full).  Let the front end
445 	 * deal with it.
446 	 */
447 	if (ret == OCSD_RESP_WAIT)
448 		goto out;
449 
450 	packet_queue->instr_count += elem->num_instr_range;
451 	/* Tell the front end we have a new timestamp to process */
452 	ret = cs_etm_decoder__do_soft_timestamp(etmq, packet_queue,
453 						trace_chan_id);
454 out:
455 	return ret;
456 }
457 
458 static ocsd_datapath_resp_t
459 cs_etm_decoder__buffer_discontinuity(struct cs_etm_packet_queue *queue,
460 				     const uint8_t trace_chan_id)
461 {
462 	/*
463 	 * Something happened and who knows when we'll get new traces so
464 	 * reset time statistics.
465 	 */
466 	cs_etm_decoder__reset_timestamp(queue);
467 	return cs_etm_decoder__buffer_packet(queue, trace_chan_id,
468 					     CS_ETM_DISCONTINUITY);
469 }
470 
471 static ocsd_datapath_resp_t
472 cs_etm_decoder__buffer_exception(struct cs_etm_packet_queue *queue,
473 				 const ocsd_generic_trace_elem *elem,
474 				 const uint8_t trace_chan_id)
475 {	int ret = 0;
476 	struct cs_etm_packet *packet;
477 
478 	ret = cs_etm_decoder__buffer_packet(queue, trace_chan_id,
479 					    CS_ETM_EXCEPTION);
480 	if (ret != OCSD_RESP_CONT && ret != OCSD_RESP_WAIT)
481 		return ret;
482 
483 	packet = &queue->packet_buffer[queue->tail];
484 	packet->exception_number = elem->exception_number;
485 
486 	return ret;
487 }
488 
489 static ocsd_datapath_resp_t
490 cs_etm_decoder__buffer_exception_ret(struct cs_etm_packet_queue *queue,
491 				     const uint8_t trace_chan_id)
492 {
493 	return cs_etm_decoder__buffer_packet(queue, trace_chan_id,
494 					     CS_ETM_EXCEPTION_RET);
495 }
496 
497 static ocsd_datapath_resp_t
498 cs_etm_decoder__set_tid(struct cs_etm_queue *etmq,
499 			struct cs_etm_packet_queue *packet_queue,
500 			const ocsd_generic_trace_elem *elem,
501 			const uint8_t trace_chan_id)
502 {
503 	pid_t tid = -1;
504 	static u64 pid_fmt;
505 	int ret;
506 
507 	/*
508 	 * As all the ETMs run at the same exception level, the system should
509 	 * have the same PID format crossing CPUs.  So cache the PID format
510 	 * and reuse it for sequential decoding.
511 	 */
512 	if (!pid_fmt) {
513 		ret = cs_etm__get_pid_fmt(trace_chan_id, &pid_fmt);
514 		if (ret)
515 			return OCSD_RESP_FATAL_SYS_ERR;
516 	}
517 
518 	/*
519 	 * Process the PE_CONTEXT packets if we have a valid contextID or VMID.
520 	 * If the kernel is running at EL2, the PID is traced in CONTEXTIDR_EL2
521 	 * as VMID, Bit ETM_OPT_CTXTID2 is set in this case.
522 	 */
523 	switch (pid_fmt) {
524 	case BIT(ETM_OPT_CTXTID):
525 		if (elem->context.ctxt_id_valid)
526 			tid = elem->context.context_id;
527 		break;
528 	case BIT(ETM_OPT_CTXTID2):
529 		if (elem->context.vmid_valid)
530 			tid = elem->context.vmid;
531 		break;
532 	default:
533 		break;
534 	}
535 
536 	if (tid == -1)
537 		return OCSD_RESP_CONT;
538 
539 	if (cs_etm__etmq_set_tid(etmq, tid, trace_chan_id))
540 		return OCSD_RESP_FATAL_SYS_ERR;
541 
542 	/*
543 	 * A timestamp is generated after a PE_CONTEXT element so make sure
544 	 * to rely on that coming one.
545 	 */
546 	cs_etm_decoder__reset_timestamp(packet_queue);
547 
548 	return OCSD_RESP_CONT;
549 }
550 
551 static ocsd_datapath_resp_t cs_etm_decoder__gen_trace_elem_printer(
552 				const void *context,
553 				const ocsd_trc_index_t indx,
554 				const u8 trace_chan_id __maybe_unused,
555 				const ocsd_generic_trace_elem *elem)
556 {
557 	ocsd_datapath_resp_t resp = OCSD_RESP_CONT;
558 	struct cs_etm_decoder *decoder = (struct cs_etm_decoder *) context;
559 	struct cs_etm_queue *etmq = decoder->data;
560 	struct cs_etm_packet_queue *packet_queue;
561 
562 	/* First get the packet queue for this traceID */
563 	packet_queue = cs_etm__etmq_get_packet_queue(etmq, trace_chan_id);
564 	if (!packet_queue)
565 		return OCSD_RESP_FATAL_SYS_ERR;
566 
567 	switch (elem->elem_type) {
568 	case OCSD_GEN_TRC_ELEM_UNKNOWN:
569 		break;
570 	case OCSD_GEN_TRC_ELEM_EO_TRACE:
571 	case OCSD_GEN_TRC_ELEM_NO_SYNC:
572 	case OCSD_GEN_TRC_ELEM_TRACE_ON:
573 		resp = cs_etm_decoder__buffer_discontinuity(packet_queue,
574 							    trace_chan_id);
575 		break;
576 	case OCSD_GEN_TRC_ELEM_INSTR_RANGE:
577 		resp = cs_etm_decoder__buffer_range(etmq, packet_queue, elem,
578 						    trace_chan_id);
579 		break;
580 	case OCSD_GEN_TRC_ELEM_EXCEPTION:
581 		resp = cs_etm_decoder__buffer_exception(packet_queue, elem,
582 							trace_chan_id);
583 		break;
584 	case OCSD_GEN_TRC_ELEM_EXCEPTION_RET:
585 		resp = cs_etm_decoder__buffer_exception_ret(packet_queue,
586 							    trace_chan_id);
587 		break;
588 	case OCSD_GEN_TRC_ELEM_TIMESTAMP:
589 		resp = cs_etm_decoder__do_hard_timestamp(etmq, elem,
590 							 trace_chan_id,
591 							 indx);
592 		break;
593 	case OCSD_GEN_TRC_ELEM_PE_CONTEXT:
594 		resp = cs_etm_decoder__set_tid(etmq, packet_queue,
595 					       elem, trace_chan_id);
596 		break;
597 	/* Unused packet types */
598 	case OCSD_GEN_TRC_ELEM_I_RANGE_NOPATH:
599 	case OCSD_GEN_TRC_ELEM_ADDR_NACC:
600 	case OCSD_GEN_TRC_ELEM_CYCLE_COUNT:
601 	case OCSD_GEN_TRC_ELEM_ADDR_UNKNOWN:
602 	case OCSD_GEN_TRC_ELEM_EVENT:
603 	case OCSD_GEN_TRC_ELEM_SWTRACE:
604 	case OCSD_GEN_TRC_ELEM_CUSTOM:
605 	case OCSD_GEN_TRC_ELEM_SYNC_MARKER:
606 	case OCSD_GEN_TRC_ELEM_MEMTRANS:
607 	default:
608 		break;
609 	}
610 
611 	return resp;
612 }
613 
614 static int
615 cs_etm_decoder__create_etm_decoder(struct cs_etm_decoder_params *d_params,
616 				   struct cs_etm_trace_params *t_params,
617 				   struct cs_etm_decoder *decoder)
618 {
619 	ocsd_etmv3_cfg config_etmv3;
620 	ocsd_etmv4_cfg trace_config_etmv4;
621 	ocsd_ete_cfg trace_config_ete;
622 	void *trace_config;
623 	u8 csid;
624 
625 	switch (t_params->protocol) {
626 	case CS_ETM_PROTO_ETMV3:
627 	case CS_ETM_PROTO_PTM:
628 		cs_etm_decoder__gen_etmv3_config(t_params, &config_etmv3);
629 		decoder->decoder_name = (t_params->protocol == CS_ETM_PROTO_ETMV3) ?
630 							OCSD_BUILTIN_DCD_ETMV3 :
631 							OCSD_BUILTIN_DCD_PTM;
632 		trace_config = &config_etmv3;
633 		break;
634 	case CS_ETM_PROTO_ETMV4i:
635 		cs_etm_decoder__gen_etmv4_config(t_params, &trace_config_etmv4);
636 		decoder->decoder_name = OCSD_BUILTIN_DCD_ETMV4I;
637 		trace_config = &trace_config_etmv4;
638 		break;
639 	case CS_ETM_PROTO_ETE:
640 		cs_etm_decoder__gen_ete_config(t_params, &trace_config_ete);
641 		decoder->decoder_name = OCSD_BUILTIN_DCD_ETE;
642 		trace_config = &trace_config_ete;
643 		break;
644 	default:
645 		return -1;
646 	}
647 
648 	if (d_params->operation == CS_ETM_OPERATION_DECODE) {
649 		if (ocsd_dt_create_decoder(decoder->dcd_tree,
650 					   decoder->decoder_name,
651 					   OCSD_CREATE_FLG_FULL_DECODER,
652 					   trace_config, &csid))
653 			return -1;
654 
655 		if (ocsd_dt_set_gen_elem_outfn(decoder->dcd_tree,
656 					       cs_etm_decoder__gen_trace_elem_printer,
657 					       decoder))
658 			return -1;
659 
660 		return 0;
661 	} else if (d_params->operation == CS_ETM_OPERATION_PRINT) {
662 		if (ocsd_dt_create_decoder(decoder->dcd_tree, decoder->decoder_name,
663 					   OCSD_CREATE_FLG_PACKET_PROC,
664 					   trace_config, &csid))
665 			return -1;
666 
667 		if (ocsd_dt_set_pkt_protocol_printer(decoder->dcd_tree, csid, 0))
668 			return -1;
669 
670 		return 0;
671 	}
672 
673 	return -1;
674 }
675 
676 struct cs_etm_decoder *
677 cs_etm_decoder__new(int decoders, struct cs_etm_decoder_params *d_params,
678 		    struct cs_etm_trace_params t_params[])
679 {
680 	struct cs_etm_decoder *decoder;
681 	ocsd_dcd_tree_src_t format;
682 	u32 flags;
683 	int i, ret;
684 
685 	if ((!t_params) || (!d_params))
686 		return NULL;
687 
688 	decoder = zalloc(sizeof(*decoder));
689 
690 	if (!decoder)
691 		return NULL;
692 
693 	decoder->data = d_params->data;
694 	decoder->prev_return = OCSD_RESP_CONT;
695 	format = (d_params->formatted ? OCSD_TRC_SRC_FRAME_FORMATTED :
696 					 OCSD_TRC_SRC_SINGLE);
697 	flags = 0;
698 	flags |= (d_params->fsyncs ? OCSD_DFRMTR_HAS_FSYNCS : 0);
699 	flags |= (d_params->hsyncs ? OCSD_DFRMTR_HAS_HSYNCS : 0);
700 	flags |= (d_params->frame_aligned ? OCSD_DFRMTR_FRAME_MEM_ALIGN : 0);
701 
702 	/*
703 	 * Drivers may add barrier frames when used with perf, set up to
704 	 * handle this. Barriers const of FSYNC packet repeated 4 times.
705 	 */
706 	flags |= OCSD_DFRMTR_RESET_ON_4X_FSYNC;
707 
708 	/* Create decode tree for the data source */
709 	decoder->dcd_tree = ocsd_create_dcd_tree(format, flags);
710 
711 	if (decoder->dcd_tree == 0)
712 		goto err_free_decoder;
713 
714 	/* init library print logging support */
715 	ret = cs_etm_decoder__init_def_logger_printing(d_params, decoder);
716 	if (ret != 0)
717 		goto err_free_decoder;
718 
719 	/* init raw frame logging if required */
720 	cs_etm_decoder__init_raw_frame_logging(d_params, decoder);
721 
722 	for (i = 0; i < decoders; i++) {
723 		ret = cs_etm_decoder__create_etm_decoder(d_params,
724 							 &t_params[i],
725 							 decoder);
726 		if (ret != 0)
727 			goto err_free_decoder;
728 	}
729 
730 	return decoder;
731 
732 err_free_decoder:
733 	cs_etm_decoder__free(decoder);
734 	return NULL;
735 }
736 
737 int cs_etm_decoder__process_data_block(struct cs_etm_decoder *decoder,
738 				       u64 indx, const u8 *buf,
739 				       size_t len, size_t *consumed)
740 {
741 	int ret = 0;
742 	ocsd_datapath_resp_t cur = OCSD_RESP_CONT;
743 	ocsd_datapath_resp_t prev_return = decoder->prev_return;
744 	size_t processed = 0;
745 	u32 count;
746 
747 	while (processed < len) {
748 		if (OCSD_DATA_RESP_IS_WAIT(prev_return)) {
749 			cur = ocsd_dt_process_data(decoder->dcd_tree,
750 						   OCSD_OP_FLUSH,
751 						   0,
752 						   0,
753 						   NULL,
754 						   NULL);
755 		} else if (OCSD_DATA_RESP_IS_CONT(prev_return)) {
756 			cur = ocsd_dt_process_data(decoder->dcd_tree,
757 						   OCSD_OP_DATA,
758 						   indx + processed,
759 						   len - processed,
760 						   &buf[processed],
761 						   &count);
762 			processed += count;
763 		} else {
764 			ret = -EINVAL;
765 			break;
766 		}
767 
768 		/*
769 		 * Return to the input code if the packet buffer is full.
770 		 * Flushing will get done once the packet buffer has been
771 		 * processed.
772 		 */
773 		if (OCSD_DATA_RESP_IS_WAIT(cur))
774 			break;
775 
776 		prev_return = cur;
777 	}
778 
779 	decoder->prev_return = cur;
780 	*consumed = processed;
781 
782 	return ret;
783 }
784 
785 void cs_etm_decoder__free(struct cs_etm_decoder *decoder)
786 {
787 	if (!decoder)
788 		return;
789 
790 	ocsd_destroy_dcd_tree(decoder->dcd_tree);
791 	decoder->dcd_tree = NULL;
792 	free(decoder);
793 }
794 
795 const char *cs_etm_decoder__get_name(struct cs_etm_decoder *decoder)
796 {
797 	return decoder->decoder_name;
798 }
799