1 // SPDX-License-Identifier: GPL-2.0 2 /* 3 * xHCI host controller driver 4 * 5 * Copyright (C) 2013 Xenia Ragiadakou 6 * 7 * Author: Xenia Ragiadakou 8 * Email : burzalodowa@gmail.com 9 */ 10 11 #undef TRACE_SYSTEM 12 #define TRACE_SYSTEM xhci-hcd 13 14 /* 15 * The TRACE_SYSTEM_VAR defaults to TRACE_SYSTEM, but must be a 16 * legitimate C variable. It is not exported to user space. 17 */ 18 #undef TRACE_SYSTEM_VAR 19 #define TRACE_SYSTEM_VAR xhci_hcd 20 21 #if !defined(__XHCI_TRACE_H) || defined(TRACE_HEADER_MULTI_READ) 22 #define __XHCI_TRACE_H 23 24 #include <linux/tracepoint.h> 25 #include "xhci.h" 26 #include "xhci-dbgcap.h" 27 28 #define XHCI_MSG_MAX 500 29 30 DECLARE_EVENT_CLASS(xhci_log_msg, 31 TP_PROTO(struct va_format *vaf), 32 TP_ARGS(vaf), 33 TP_STRUCT__entry(__dynamic_array(char, msg, XHCI_MSG_MAX)), 34 TP_fast_assign( 35 vsnprintf(__get_str(msg), XHCI_MSG_MAX, vaf->fmt, *vaf->va); 36 ), 37 TP_printk("%s", __get_str(msg)) 38 ); 39 40 DEFINE_EVENT(xhci_log_msg, xhci_dbg_address, 41 TP_PROTO(struct va_format *vaf), 42 TP_ARGS(vaf) 43 ); 44 45 DEFINE_EVENT(xhci_log_msg, xhci_dbg_context_change, 46 TP_PROTO(struct va_format *vaf), 47 TP_ARGS(vaf) 48 ); 49 50 DEFINE_EVENT(xhci_log_msg, xhci_dbg_quirks, 51 TP_PROTO(struct va_format *vaf), 52 TP_ARGS(vaf) 53 ); 54 55 DEFINE_EVENT(xhci_log_msg, xhci_dbg_reset_ep, 56 TP_PROTO(struct va_format *vaf), 57 TP_ARGS(vaf) 58 ); 59 60 DEFINE_EVENT(xhci_log_msg, xhci_dbg_cancel_urb, 61 TP_PROTO(struct va_format *vaf), 62 TP_ARGS(vaf) 63 ); 64 65 DEFINE_EVENT(xhci_log_msg, xhci_dbg_init, 66 TP_PROTO(struct va_format *vaf), 67 TP_ARGS(vaf) 68 ); 69 70 DEFINE_EVENT(xhci_log_msg, xhci_dbg_ring_expansion, 71 TP_PROTO(struct va_format *vaf), 72 TP_ARGS(vaf) 73 ); 74 75 DECLARE_EVENT_CLASS(xhci_log_ctx, 76 TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx, 77 unsigned int ep_num), 78 TP_ARGS(xhci, ctx, ep_num), 79 TP_STRUCT__entry( 80 __field(int, ctx_64) 81 __field(unsigned, ctx_type) 82 __field(dma_addr_t, ctx_dma) 83 __field(u8 *, ctx_va) 84 __field(unsigned, ctx_ep_num) 85 __field(int, slot_id) 86 __dynamic_array(u32, ctx_data, 87 ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 8) * 88 ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1)) 89 ), 90 TP_fast_assign( 91 struct usb_device *udev; 92 93 udev = to_usb_device(xhci_to_hcd(xhci)->self.controller); 94 __entry->ctx_64 = HCC_64BYTE_CONTEXT(xhci->hcc_params); 95 __entry->ctx_type = ctx->type; 96 __entry->ctx_dma = ctx->dma; 97 __entry->ctx_va = ctx->bytes; 98 __entry->slot_id = udev->slot_id; 99 __entry->ctx_ep_num = ep_num; 100 memcpy(__get_dynamic_array(ctx_data), ctx->bytes, 101 ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 32) * 102 ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1)); 103 ), 104 TP_printk("ctx_64=%d, ctx_type=%u, ctx_dma=@%llx, ctx_va=@%p", 105 __entry->ctx_64, __entry->ctx_type, 106 (unsigned long long) __entry->ctx_dma, __entry->ctx_va 107 ) 108 ); 109 110 DEFINE_EVENT(xhci_log_ctx, xhci_address_ctx, 111 TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx, 112 unsigned int ep_num), 113 TP_ARGS(xhci, ctx, ep_num) 114 ); 115 116 DECLARE_EVENT_CLASS(xhci_log_trb, 117 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 118 TP_ARGS(ring, trb), 119 TP_STRUCT__entry( 120 __field(u32, type) 121 __field(u32, field0) 122 __field(u32, field1) 123 __field(u32, field2) 124 __field(u32, field3) 125 ), 126 TP_fast_assign( 127 __entry->type = ring->type; 128 __entry->field0 = le32_to_cpu(trb->field[0]); 129 __entry->field1 = le32_to_cpu(trb->field[1]); 130 __entry->field2 = le32_to_cpu(trb->field[2]); 131 __entry->field3 = le32_to_cpu(trb->field[3]); 132 ), 133 TP_printk("%s: %s", xhci_ring_type_string(__entry->type), 134 xhci_decode_trb(__entry->field0, __entry->field1, 135 __entry->field2, __entry->field3) 136 ) 137 ); 138 139 DEFINE_EVENT(xhci_log_trb, xhci_handle_event, 140 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 141 TP_ARGS(ring, trb) 142 ); 143 144 DEFINE_EVENT(xhci_log_trb, xhci_handle_command, 145 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 146 TP_ARGS(ring, trb) 147 ); 148 149 DEFINE_EVENT(xhci_log_trb, xhci_handle_transfer, 150 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 151 TP_ARGS(ring, trb) 152 ); 153 154 DEFINE_EVENT(xhci_log_trb, xhci_queue_trb, 155 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 156 TP_ARGS(ring, trb) 157 ); 158 159 DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_event, 160 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 161 TP_ARGS(ring, trb) 162 ); 163 164 DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_transfer, 165 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 166 TP_ARGS(ring, trb) 167 ); 168 169 DEFINE_EVENT(xhci_log_trb, xhci_dbc_gadget_ep_queue, 170 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 171 TP_ARGS(ring, trb) 172 ); 173 174 DECLARE_EVENT_CLASS(xhci_log_free_virt_dev, 175 TP_PROTO(struct xhci_virt_device *vdev), 176 TP_ARGS(vdev), 177 TP_STRUCT__entry( 178 __field(void *, vdev) 179 __field(unsigned long long, out_ctx) 180 __field(unsigned long long, in_ctx) 181 __field(u8, fake_port) 182 __field(u8, real_port) 183 __field(u16, current_mel) 184 185 ), 186 TP_fast_assign( 187 __entry->vdev = vdev; 188 __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma; 189 __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma; 190 __entry->fake_port = (u8) vdev->fake_port; 191 __entry->real_port = (u8) vdev->real_port; 192 __entry->current_mel = (u16) vdev->current_mel; 193 ), 194 TP_printk("vdev %p ctx %llx | %llx fake_port %d real_port %d current_mel %d", 195 __entry->vdev, __entry->in_ctx, __entry->out_ctx, 196 __entry->fake_port, __entry->real_port, __entry->current_mel 197 ) 198 ); 199 200 DEFINE_EVENT(xhci_log_free_virt_dev, xhci_free_virt_device, 201 TP_PROTO(struct xhci_virt_device *vdev), 202 TP_ARGS(vdev) 203 ); 204 205 DECLARE_EVENT_CLASS(xhci_log_virt_dev, 206 TP_PROTO(struct xhci_virt_device *vdev), 207 TP_ARGS(vdev), 208 TP_STRUCT__entry( 209 __field(void *, vdev) 210 __field(unsigned long long, out_ctx) 211 __field(unsigned long long, in_ctx) 212 __field(int, devnum) 213 __field(int, state) 214 __field(int, speed) 215 __field(u8, portnum) 216 __field(u8, level) 217 __field(int, slot_id) 218 ), 219 TP_fast_assign( 220 __entry->vdev = vdev; 221 __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma; 222 __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma; 223 __entry->devnum = vdev->udev->devnum; 224 __entry->state = vdev->udev->state; 225 __entry->speed = vdev->udev->speed; 226 __entry->portnum = vdev->udev->portnum; 227 __entry->level = vdev->udev->level; 228 __entry->slot_id = vdev->udev->slot_id; 229 ), 230 TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d", 231 __entry->vdev, __entry->in_ctx, __entry->out_ctx, 232 __entry->devnum, __entry->state, __entry->speed, 233 __entry->portnum, __entry->level, __entry->slot_id 234 ) 235 ); 236 237 DEFINE_EVENT(xhci_log_virt_dev, xhci_alloc_virt_device, 238 TP_PROTO(struct xhci_virt_device *vdev), 239 TP_ARGS(vdev) 240 ); 241 242 DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_device, 243 TP_PROTO(struct xhci_virt_device *vdev), 244 TP_ARGS(vdev) 245 ); 246 247 DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_addressable_virt_device, 248 TP_PROTO(struct xhci_virt_device *vdev), 249 TP_ARGS(vdev) 250 ); 251 252 DEFINE_EVENT(xhci_log_virt_dev, xhci_stop_device, 253 TP_PROTO(struct xhci_virt_device *vdev), 254 TP_ARGS(vdev) 255 ); 256 257 DECLARE_EVENT_CLASS(xhci_log_urb, 258 TP_PROTO(struct urb *urb), 259 TP_ARGS(urb), 260 TP_STRUCT__entry( 261 __field(void *, urb) 262 __field(unsigned int, pipe) 263 __field(unsigned int, stream) 264 __field(int, status) 265 __field(unsigned int, flags) 266 __field(int, num_mapped_sgs) 267 __field(int, num_sgs) 268 __field(int, length) 269 __field(int, actual) 270 __field(int, epnum) 271 __field(int, dir_in) 272 __field(int, type) 273 __field(int, slot_id) 274 ), 275 TP_fast_assign( 276 __entry->urb = urb; 277 __entry->pipe = urb->pipe; 278 __entry->stream = urb->stream_id; 279 __entry->status = urb->status; 280 __entry->flags = urb->transfer_flags; 281 __entry->num_mapped_sgs = urb->num_mapped_sgs; 282 __entry->num_sgs = urb->num_sgs; 283 __entry->length = urb->transfer_buffer_length; 284 __entry->actual = urb->actual_length; 285 __entry->epnum = usb_endpoint_num(&urb->ep->desc); 286 __entry->dir_in = usb_endpoint_dir_in(&urb->ep->desc); 287 __entry->type = usb_endpoint_type(&urb->ep->desc); 288 __entry->slot_id = urb->dev->slot_id; 289 ), 290 TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x", 291 __entry->epnum, __entry->dir_in ? "in" : "out", 292 ({ char *s; 293 switch (__entry->type) { 294 case USB_ENDPOINT_XFER_INT: 295 s = "intr"; 296 break; 297 case USB_ENDPOINT_XFER_CONTROL: 298 s = "control"; 299 break; 300 case USB_ENDPOINT_XFER_BULK: 301 s = "bulk"; 302 break; 303 case USB_ENDPOINT_XFER_ISOC: 304 s = "isoc"; 305 break; 306 default: 307 s = "UNKNOWN"; 308 } s; }), __entry->urb, __entry->pipe, __entry->slot_id, 309 __entry->actual, __entry->length, __entry->num_mapped_sgs, 310 __entry->num_sgs, __entry->stream, __entry->flags 311 ) 312 ); 313 314 DEFINE_EVENT(xhci_log_urb, xhci_urb_enqueue, 315 TP_PROTO(struct urb *urb), 316 TP_ARGS(urb) 317 ); 318 319 DEFINE_EVENT(xhci_log_urb, xhci_urb_giveback, 320 TP_PROTO(struct urb *urb), 321 TP_ARGS(urb) 322 ); 323 324 DEFINE_EVENT(xhci_log_urb, xhci_urb_dequeue, 325 TP_PROTO(struct urb *urb), 326 TP_ARGS(urb) 327 ); 328 329 DECLARE_EVENT_CLASS(xhci_log_ep_ctx, 330 TP_PROTO(struct xhci_ep_ctx *ctx), 331 TP_ARGS(ctx), 332 TP_STRUCT__entry( 333 __field(u32, info) 334 __field(u32, info2) 335 __field(u64, deq) 336 __field(u32, tx_info) 337 ), 338 TP_fast_assign( 339 __entry->info = le32_to_cpu(ctx->ep_info); 340 __entry->info2 = le32_to_cpu(ctx->ep_info2); 341 __entry->deq = le64_to_cpu(ctx->deq); 342 __entry->tx_info = le32_to_cpu(ctx->tx_info); 343 ), 344 TP_printk("%s", xhci_decode_ep_context(__entry->info, 345 __entry->info2, __entry->deq, __entry->tx_info) 346 ) 347 ); 348 349 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_stop_ep, 350 TP_PROTO(struct xhci_ep_ctx *ctx), 351 TP_ARGS(ctx) 352 ); 353 354 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_set_deq_ep, 355 TP_PROTO(struct xhci_ep_ctx *ctx), 356 TP_ARGS(ctx) 357 ); 358 359 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_reset_ep, 360 TP_PROTO(struct xhci_ep_ctx *ctx), 361 TP_ARGS(ctx) 362 ); 363 364 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_config_ep, 365 TP_PROTO(struct xhci_ep_ctx *ctx), 366 TP_ARGS(ctx) 367 ); 368 369 DEFINE_EVENT(xhci_log_ep_ctx, xhci_add_endpoint, 370 TP_PROTO(struct xhci_ep_ctx *ctx), 371 TP_ARGS(ctx) 372 ); 373 374 DECLARE_EVENT_CLASS(xhci_log_slot_ctx, 375 TP_PROTO(struct xhci_slot_ctx *ctx), 376 TP_ARGS(ctx), 377 TP_STRUCT__entry( 378 __field(u32, info) 379 __field(u32, info2) 380 __field(u32, tt_info) 381 __field(u32, state) 382 ), 383 TP_fast_assign( 384 __entry->info = le32_to_cpu(ctx->dev_info); 385 __entry->info2 = le32_to_cpu(ctx->dev_info2); 386 __entry->tt_info = le64_to_cpu(ctx->tt_info); 387 __entry->state = le32_to_cpu(ctx->dev_state); 388 ), 389 TP_printk("%s", xhci_decode_slot_context(__entry->info, 390 __entry->info2, __entry->tt_info, 391 __entry->state) 392 ) 393 ); 394 395 DEFINE_EVENT(xhci_log_slot_ctx, xhci_alloc_dev, 396 TP_PROTO(struct xhci_slot_ctx *ctx), 397 TP_ARGS(ctx) 398 ); 399 400 DEFINE_EVENT(xhci_log_slot_ctx, xhci_free_dev, 401 TP_PROTO(struct xhci_slot_ctx *ctx), 402 TP_ARGS(ctx) 403 ); 404 405 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_disable_slot, 406 TP_PROTO(struct xhci_slot_ctx *ctx), 407 TP_ARGS(ctx) 408 ); 409 410 DEFINE_EVENT(xhci_log_slot_ctx, xhci_discover_or_reset_device, 411 TP_PROTO(struct xhci_slot_ctx *ctx), 412 TP_ARGS(ctx) 413 ); 414 415 DEFINE_EVENT(xhci_log_slot_ctx, xhci_setup_device_slot, 416 TP_PROTO(struct xhci_slot_ctx *ctx), 417 TP_ARGS(ctx) 418 ); 419 420 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_addr_dev, 421 TP_PROTO(struct xhci_slot_ctx *ctx), 422 TP_ARGS(ctx) 423 ); 424 425 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_reset_dev, 426 TP_PROTO(struct xhci_slot_ctx *ctx), 427 TP_ARGS(ctx) 428 ); 429 430 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_set_deq, 431 TP_PROTO(struct xhci_slot_ctx *ctx), 432 TP_ARGS(ctx) 433 ); 434 435 DEFINE_EVENT(xhci_log_slot_ctx, xhci_configure_endpoint, 436 TP_PROTO(struct xhci_slot_ctx *ctx), 437 TP_ARGS(ctx) 438 ); 439 440 DECLARE_EVENT_CLASS(xhci_log_ctrl_ctx, 441 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 442 TP_ARGS(ctrl_ctx), 443 TP_STRUCT__entry( 444 __field(u32, drop) 445 __field(u32, add) 446 ), 447 TP_fast_assign( 448 __entry->drop = le32_to_cpu(ctrl_ctx->drop_flags); 449 __entry->add = le32_to_cpu(ctrl_ctx->add_flags); 450 ), 451 TP_printk("%s", xhci_decode_ctrl_ctx(__entry->drop, __entry->add) 452 ) 453 ); 454 455 DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_address_ctrl_ctx, 456 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 457 TP_ARGS(ctrl_ctx) 458 ); 459 460 DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_configure_endpoint_ctrl_ctx, 461 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 462 TP_ARGS(ctrl_ctx) 463 ); 464 465 DECLARE_EVENT_CLASS(xhci_log_ring, 466 TP_PROTO(struct xhci_ring *ring), 467 TP_ARGS(ring), 468 TP_STRUCT__entry( 469 __field(u32, type) 470 __field(void *, ring) 471 __field(dma_addr_t, enq) 472 __field(dma_addr_t, deq) 473 __field(dma_addr_t, enq_seg) 474 __field(dma_addr_t, deq_seg) 475 __field(unsigned int, num_segs) 476 __field(unsigned int, stream_id) 477 __field(unsigned int, cycle_state) 478 __field(unsigned int, num_trbs_free) 479 __field(unsigned int, bounce_buf_len) 480 ), 481 TP_fast_assign( 482 __entry->ring = ring; 483 __entry->type = ring->type; 484 __entry->num_segs = ring->num_segs; 485 __entry->stream_id = ring->stream_id; 486 __entry->enq_seg = ring->enq_seg->dma; 487 __entry->deq_seg = ring->deq_seg->dma; 488 __entry->cycle_state = ring->cycle_state; 489 __entry->num_trbs_free = ring->num_trbs_free; 490 __entry->bounce_buf_len = ring->bounce_buf_len; 491 __entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue); 492 __entry->deq = xhci_trb_virt_to_dma(ring->deq_seg, ring->dequeue); 493 ), 494 TP_printk("%s %p: enq %pad(%pad) deq %pad(%pad) segs %d stream %d free_trbs %d bounce %d cycle %d", 495 xhci_ring_type_string(__entry->type), __entry->ring, 496 &__entry->enq, &__entry->enq_seg, 497 &__entry->deq, &__entry->deq_seg, 498 __entry->num_segs, 499 __entry->stream_id, 500 __entry->num_trbs_free, 501 __entry->bounce_buf_len, 502 __entry->cycle_state 503 ) 504 ); 505 506 DEFINE_EVENT(xhci_log_ring, xhci_ring_alloc, 507 TP_PROTO(struct xhci_ring *ring), 508 TP_ARGS(ring) 509 ); 510 511 DEFINE_EVENT(xhci_log_ring, xhci_ring_free, 512 TP_PROTO(struct xhci_ring *ring), 513 TP_ARGS(ring) 514 ); 515 516 DEFINE_EVENT(xhci_log_ring, xhci_ring_expansion, 517 TP_PROTO(struct xhci_ring *ring), 518 TP_ARGS(ring) 519 ); 520 521 DEFINE_EVENT(xhci_log_ring, xhci_inc_enq, 522 TP_PROTO(struct xhci_ring *ring), 523 TP_ARGS(ring) 524 ); 525 526 DEFINE_EVENT(xhci_log_ring, xhci_inc_deq, 527 TP_PROTO(struct xhci_ring *ring), 528 TP_ARGS(ring) 529 ); 530 531 DECLARE_EVENT_CLASS(xhci_log_portsc, 532 TP_PROTO(u32 portnum, u32 portsc), 533 TP_ARGS(portnum, portsc), 534 TP_STRUCT__entry( 535 __field(u32, portnum) 536 __field(u32, portsc) 537 ), 538 TP_fast_assign( 539 __entry->portnum = portnum; 540 __entry->portsc = portsc; 541 ), 542 TP_printk("port-%d: %s", 543 __entry->portnum, 544 xhci_decode_portsc(__entry->portsc) 545 ) 546 ); 547 548 DEFINE_EVENT(xhci_log_portsc, xhci_handle_port_status, 549 TP_PROTO(u32 portnum, u32 portsc), 550 TP_ARGS(portnum, portsc) 551 ); 552 553 DEFINE_EVENT(xhci_log_portsc, xhci_get_port_status, 554 TP_PROTO(u32 portnum, u32 portsc), 555 TP_ARGS(portnum, portsc) 556 ); 557 558 DEFINE_EVENT(xhci_log_portsc, xhci_hub_status_data, 559 TP_PROTO(u32 portnum, u32 portsc), 560 TP_ARGS(portnum, portsc) 561 ); 562 563 DECLARE_EVENT_CLASS(xhci_dbc_log_request, 564 TP_PROTO(struct dbc_request *req), 565 TP_ARGS(req), 566 TP_STRUCT__entry( 567 __field(struct dbc_request *, req) 568 __field(bool, dir) 569 __field(unsigned int, actual) 570 __field(unsigned int, length) 571 __field(int, status) 572 ), 573 TP_fast_assign( 574 __entry->req = req; 575 __entry->dir = req->direction; 576 __entry->actual = req->actual; 577 __entry->length = req->length; 578 __entry->status = req->status; 579 ), 580 TP_printk("%s: req %p length %u/%u ==> %d", 581 __entry->dir ? "bulk-in" : "bulk-out", 582 __entry->req, __entry->actual, 583 __entry->length, __entry->status 584 ) 585 ); 586 587 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_alloc_request, 588 TP_PROTO(struct dbc_request *req), 589 TP_ARGS(req) 590 ); 591 592 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_free_request, 593 TP_PROTO(struct dbc_request *req), 594 TP_ARGS(req) 595 ); 596 597 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_queue_request, 598 TP_PROTO(struct dbc_request *req), 599 TP_ARGS(req) 600 ); 601 602 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_giveback_request, 603 TP_PROTO(struct dbc_request *req), 604 TP_ARGS(req) 605 ); 606 #endif /* __XHCI_TRACE_H */ 607 608 /* this part must be outside header guard */ 609 610 #undef TRACE_INCLUDE_PATH 611 #define TRACE_INCLUDE_PATH . 612 613 #undef TRACE_INCLUDE_FILE 614 #define TRACE_INCLUDE_FILE xhci-trace 615 616 #include <trace/define_trace.h> 617