1 /* SPDX-License-Identifier: GPL-2.0 */ 2 /* 3 * xHCI host controller driver 4 * 5 * Copyright (C) 2013 Xenia Ragiadakou 6 * 7 * Author: Xenia Ragiadakou 8 * Email : burzalodowa@gmail.com 9 */ 10 11 #undef TRACE_SYSTEM 12 #define TRACE_SYSTEM xhci-hcd 13 14 /* 15 * The TRACE_SYSTEM_VAR defaults to TRACE_SYSTEM, but must be a 16 * legitimate C variable. It is not exported to user space. 17 */ 18 #undef TRACE_SYSTEM_VAR 19 #define TRACE_SYSTEM_VAR xhci_hcd 20 21 #if !defined(__XHCI_TRACE_H) || defined(TRACE_HEADER_MULTI_READ) 22 #define __XHCI_TRACE_H 23 24 #include <linux/tracepoint.h> 25 #include "xhci.h" 26 #include "xhci-dbgcap.h" 27 28 DECLARE_EVENT_CLASS(xhci_log_msg, 29 TP_PROTO(struct va_format *vaf), 30 TP_ARGS(vaf), 31 TP_STRUCT__entry(__vstring(msg, vaf->fmt, vaf->va)), 32 TP_fast_assign( 33 __assign_vstr(msg, vaf->fmt, vaf->va); 34 ), 35 TP_printk("%s", __get_str(msg)) 36 ); 37 38 DEFINE_EVENT(xhci_log_msg, xhci_dbg_address, 39 TP_PROTO(struct va_format *vaf), 40 TP_ARGS(vaf) 41 ); 42 43 DEFINE_EVENT(xhci_log_msg, xhci_dbg_context_change, 44 TP_PROTO(struct va_format *vaf), 45 TP_ARGS(vaf) 46 ); 47 48 DEFINE_EVENT(xhci_log_msg, xhci_dbg_quirks, 49 TP_PROTO(struct va_format *vaf), 50 TP_ARGS(vaf) 51 ); 52 53 DEFINE_EVENT(xhci_log_msg, xhci_dbg_reset_ep, 54 TP_PROTO(struct va_format *vaf), 55 TP_ARGS(vaf) 56 ); 57 58 DEFINE_EVENT(xhci_log_msg, xhci_dbg_cancel_urb, 59 TP_PROTO(struct va_format *vaf), 60 TP_ARGS(vaf) 61 ); 62 63 DEFINE_EVENT(xhci_log_msg, xhci_dbg_init, 64 TP_PROTO(struct va_format *vaf), 65 TP_ARGS(vaf) 66 ); 67 68 DEFINE_EVENT(xhci_log_msg, xhci_dbg_ring_expansion, 69 TP_PROTO(struct va_format *vaf), 70 TP_ARGS(vaf) 71 ); 72 73 DECLARE_EVENT_CLASS(xhci_log_ctx, 74 TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx, 75 unsigned int ep_num), 76 TP_ARGS(xhci, ctx, ep_num), 77 TP_STRUCT__entry( 78 __field(int, ctx_64) 79 __field(unsigned, ctx_type) 80 __field(dma_addr_t, ctx_dma) 81 __field(u8 *, ctx_va) 82 __field(unsigned, ctx_ep_num) 83 __dynamic_array(u32, ctx_data, 84 ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 8) * 85 ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1)) 86 ), 87 TP_fast_assign( 88 89 __entry->ctx_64 = HCC_64BYTE_CONTEXT(xhci->hcc_params); 90 __entry->ctx_type = ctx->type; 91 __entry->ctx_dma = ctx->dma; 92 __entry->ctx_va = ctx->bytes; 93 __entry->ctx_ep_num = ep_num; 94 memcpy(__get_dynamic_array(ctx_data), ctx->bytes, 95 ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 32) * 96 ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1)); 97 ), 98 TP_printk("ctx_64=%d, ctx_type=%u, ctx_dma=@%llx, ctx_va=@%p", 99 __entry->ctx_64, __entry->ctx_type, 100 (unsigned long long) __entry->ctx_dma, __entry->ctx_va 101 ) 102 ); 103 104 DEFINE_EVENT(xhci_log_ctx, xhci_address_ctx, 105 TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx, 106 unsigned int ep_num), 107 TP_ARGS(xhci, ctx, ep_num) 108 ); 109 110 DECLARE_EVENT_CLASS(xhci_log_trb, 111 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 112 TP_ARGS(ring, trb), 113 TP_STRUCT__entry( 114 __field(u32, type) 115 __field(u32, field0) 116 __field(u32, field1) 117 __field(u32, field2) 118 __field(u32, field3) 119 ), 120 TP_fast_assign( 121 __entry->type = ring->type; 122 __entry->field0 = le32_to_cpu(trb->field[0]); 123 __entry->field1 = le32_to_cpu(trb->field[1]); 124 __entry->field2 = le32_to_cpu(trb->field[2]); 125 __entry->field3 = le32_to_cpu(trb->field[3]); 126 ), 127 TP_printk("%s: %s", xhci_ring_type_string(__entry->type), 128 xhci_decode_trb(__get_buf(XHCI_MSG_MAX), XHCI_MSG_MAX, __entry->field0, 129 __entry->field1, __entry->field2, __entry->field3) 130 ) 131 ); 132 133 DEFINE_EVENT(xhci_log_trb, xhci_handle_event, 134 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 135 TP_ARGS(ring, trb) 136 ); 137 138 DEFINE_EVENT(xhci_log_trb, xhci_handle_command, 139 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 140 TP_ARGS(ring, trb) 141 ); 142 143 DEFINE_EVENT(xhci_log_trb, xhci_handle_transfer, 144 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 145 TP_ARGS(ring, trb) 146 ); 147 148 DEFINE_EVENT(xhci_log_trb, xhci_queue_trb, 149 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 150 TP_ARGS(ring, trb) 151 ); 152 153 DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_event, 154 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 155 TP_ARGS(ring, trb) 156 ); 157 158 DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_transfer, 159 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 160 TP_ARGS(ring, trb) 161 ); 162 163 DEFINE_EVENT(xhci_log_trb, xhci_dbc_gadget_ep_queue, 164 TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb), 165 TP_ARGS(ring, trb) 166 ); 167 168 DECLARE_EVENT_CLASS(xhci_log_free_virt_dev, 169 TP_PROTO(struct xhci_virt_device *vdev), 170 TP_ARGS(vdev), 171 TP_STRUCT__entry( 172 __field(void *, vdev) 173 __field(unsigned long long, out_ctx) 174 __field(unsigned long long, in_ctx) 175 __field(u8, fake_port) 176 __field(u8, real_port) 177 __field(u16, current_mel) 178 179 ), 180 TP_fast_assign( 181 __entry->vdev = vdev; 182 __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma; 183 __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma; 184 __entry->fake_port = (u8) vdev->fake_port; 185 __entry->real_port = (u8) vdev->real_port; 186 __entry->current_mel = (u16) vdev->current_mel; 187 ), 188 TP_printk("vdev %p ctx %llx | %llx fake_port %d real_port %d current_mel %d", 189 __entry->vdev, __entry->in_ctx, __entry->out_ctx, 190 __entry->fake_port, __entry->real_port, __entry->current_mel 191 ) 192 ); 193 194 DEFINE_EVENT(xhci_log_free_virt_dev, xhci_free_virt_device, 195 TP_PROTO(struct xhci_virt_device *vdev), 196 TP_ARGS(vdev) 197 ); 198 199 DECLARE_EVENT_CLASS(xhci_log_virt_dev, 200 TP_PROTO(struct xhci_virt_device *vdev), 201 TP_ARGS(vdev), 202 TP_STRUCT__entry( 203 __field(void *, vdev) 204 __field(unsigned long long, out_ctx) 205 __field(unsigned long long, in_ctx) 206 __field(int, devnum) 207 __field(int, state) 208 __field(int, speed) 209 __field(u8, portnum) 210 __field(u8, level) 211 __field(int, slot_id) 212 ), 213 TP_fast_assign( 214 __entry->vdev = vdev; 215 __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma; 216 __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma; 217 __entry->devnum = vdev->udev->devnum; 218 __entry->state = vdev->udev->state; 219 __entry->speed = vdev->udev->speed; 220 __entry->portnum = vdev->udev->portnum; 221 __entry->level = vdev->udev->level; 222 __entry->slot_id = vdev->udev->slot_id; 223 ), 224 TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d", 225 __entry->vdev, __entry->in_ctx, __entry->out_ctx, 226 __entry->devnum, __entry->state, __entry->speed, 227 __entry->portnum, __entry->level, __entry->slot_id 228 ) 229 ); 230 231 DEFINE_EVENT(xhci_log_virt_dev, xhci_alloc_virt_device, 232 TP_PROTO(struct xhci_virt_device *vdev), 233 TP_ARGS(vdev) 234 ); 235 236 DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_device, 237 TP_PROTO(struct xhci_virt_device *vdev), 238 TP_ARGS(vdev) 239 ); 240 241 DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_addressable_virt_device, 242 TP_PROTO(struct xhci_virt_device *vdev), 243 TP_ARGS(vdev) 244 ); 245 246 DEFINE_EVENT(xhci_log_virt_dev, xhci_stop_device, 247 TP_PROTO(struct xhci_virt_device *vdev), 248 TP_ARGS(vdev) 249 ); 250 251 DECLARE_EVENT_CLASS(xhci_log_urb, 252 TP_PROTO(struct urb *urb), 253 TP_ARGS(urb), 254 TP_STRUCT__entry( 255 __field(void *, urb) 256 __field(unsigned int, pipe) 257 __field(unsigned int, stream) 258 __field(int, status) 259 __field(unsigned int, flags) 260 __field(int, num_mapped_sgs) 261 __field(int, num_sgs) 262 __field(int, length) 263 __field(int, actual) 264 __field(int, epnum) 265 __field(int, dir_in) 266 __field(int, type) 267 __field(int, slot_id) 268 ), 269 TP_fast_assign( 270 __entry->urb = urb; 271 __entry->pipe = urb->pipe; 272 __entry->stream = urb->stream_id; 273 __entry->status = urb->status; 274 __entry->flags = urb->transfer_flags; 275 __entry->num_mapped_sgs = urb->num_mapped_sgs; 276 __entry->num_sgs = urb->num_sgs; 277 __entry->length = urb->transfer_buffer_length; 278 __entry->actual = urb->actual_length; 279 __entry->epnum = usb_endpoint_num(&urb->ep->desc); 280 __entry->dir_in = usb_endpoint_dir_in(&urb->ep->desc); 281 __entry->type = usb_endpoint_type(&urb->ep->desc); 282 __entry->slot_id = urb->dev->slot_id; 283 ), 284 TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x", 285 __entry->epnum, __entry->dir_in ? "in" : "out", 286 __print_symbolic(__entry->type, 287 { USB_ENDPOINT_XFER_INT, "intr" }, 288 { USB_ENDPOINT_XFER_CONTROL, "control" }, 289 { USB_ENDPOINT_XFER_BULK, "bulk" }, 290 { USB_ENDPOINT_XFER_ISOC, "isoc" }), 291 __entry->urb, __entry->pipe, __entry->slot_id, 292 __entry->actual, __entry->length, __entry->num_mapped_sgs, 293 __entry->num_sgs, __entry->stream, __entry->flags 294 ) 295 ); 296 297 DEFINE_EVENT(xhci_log_urb, xhci_urb_enqueue, 298 TP_PROTO(struct urb *urb), 299 TP_ARGS(urb) 300 ); 301 302 DEFINE_EVENT(xhci_log_urb, xhci_urb_giveback, 303 TP_PROTO(struct urb *urb), 304 TP_ARGS(urb) 305 ); 306 307 DEFINE_EVENT(xhci_log_urb, xhci_urb_dequeue, 308 TP_PROTO(struct urb *urb), 309 TP_ARGS(urb) 310 ); 311 312 DECLARE_EVENT_CLASS(xhci_log_ep_ctx, 313 TP_PROTO(struct xhci_ep_ctx *ctx), 314 TP_ARGS(ctx), 315 TP_STRUCT__entry( 316 __field(u32, info) 317 __field(u32, info2) 318 __field(u64, deq) 319 __field(u32, tx_info) 320 ), 321 TP_fast_assign( 322 __entry->info = le32_to_cpu(ctx->ep_info); 323 __entry->info2 = le32_to_cpu(ctx->ep_info2); 324 __entry->deq = le64_to_cpu(ctx->deq); 325 __entry->tx_info = le32_to_cpu(ctx->tx_info); 326 ), 327 TP_printk("%s", xhci_decode_ep_context(__get_buf(XHCI_MSG_MAX), 328 __entry->info, __entry->info2, __entry->deq, __entry->tx_info) 329 ) 330 ); 331 332 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_stop_ep, 333 TP_PROTO(struct xhci_ep_ctx *ctx), 334 TP_ARGS(ctx) 335 ); 336 337 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_set_deq_ep, 338 TP_PROTO(struct xhci_ep_ctx *ctx), 339 TP_ARGS(ctx) 340 ); 341 342 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_reset_ep, 343 TP_PROTO(struct xhci_ep_ctx *ctx), 344 TP_ARGS(ctx) 345 ); 346 347 DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_config_ep, 348 TP_PROTO(struct xhci_ep_ctx *ctx), 349 TP_ARGS(ctx) 350 ); 351 352 DEFINE_EVENT(xhci_log_ep_ctx, xhci_add_endpoint, 353 TP_PROTO(struct xhci_ep_ctx *ctx), 354 TP_ARGS(ctx) 355 ); 356 357 DECLARE_EVENT_CLASS(xhci_log_slot_ctx, 358 TP_PROTO(struct xhci_slot_ctx *ctx), 359 TP_ARGS(ctx), 360 TP_STRUCT__entry( 361 __field(u32, info) 362 __field(u32, info2) 363 __field(u32, tt_info) 364 __field(u32, state) 365 ), 366 TP_fast_assign( 367 __entry->info = le32_to_cpu(ctx->dev_info); 368 __entry->info2 = le32_to_cpu(ctx->dev_info2); 369 __entry->tt_info = le64_to_cpu(ctx->tt_info); 370 __entry->state = le32_to_cpu(ctx->dev_state); 371 ), 372 TP_printk("%s", xhci_decode_slot_context(__get_buf(XHCI_MSG_MAX), 373 __entry->info, __entry->info2, 374 __entry->tt_info, __entry->state) 375 ) 376 ); 377 378 DEFINE_EVENT(xhci_log_slot_ctx, xhci_alloc_dev, 379 TP_PROTO(struct xhci_slot_ctx *ctx), 380 TP_ARGS(ctx) 381 ); 382 383 DEFINE_EVENT(xhci_log_slot_ctx, xhci_free_dev, 384 TP_PROTO(struct xhci_slot_ctx *ctx), 385 TP_ARGS(ctx) 386 ); 387 388 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_disable_slot, 389 TP_PROTO(struct xhci_slot_ctx *ctx), 390 TP_ARGS(ctx) 391 ); 392 393 DEFINE_EVENT(xhci_log_slot_ctx, xhci_discover_or_reset_device, 394 TP_PROTO(struct xhci_slot_ctx *ctx), 395 TP_ARGS(ctx) 396 ); 397 398 DEFINE_EVENT(xhci_log_slot_ctx, xhci_setup_device_slot, 399 TP_PROTO(struct xhci_slot_ctx *ctx), 400 TP_ARGS(ctx) 401 ); 402 403 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_addr_dev, 404 TP_PROTO(struct xhci_slot_ctx *ctx), 405 TP_ARGS(ctx) 406 ); 407 408 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_reset_dev, 409 TP_PROTO(struct xhci_slot_ctx *ctx), 410 TP_ARGS(ctx) 411 ); 412 413 DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_set_deq, 414 TP_PROTO(struct xhci_slot_ctx *ctx), 415 TP_ARGS(ctx) 416 ); 417 418 DEFINE_EVENT(xhci_log_slot_ctx, xhci_configure_endpoint, 419 TP_PROTO(struct xhci_slot_ctx *ctx), 420 TP_ARGS(ctx) 421 ); 422 423 DECLARE_EVENT_CLASS(xhci_log_ctrl_ctx, 424 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 425 TP_ARGS(ctrl_ctx), 426 TP_STRUCT__entry( 427 __field(u32, drop) 428 __field(u32, add) 429 ), 430 TP_fast_assign( 431 __entry->drop = le32_to_cpu(ctrl_ctx->drop_flags); 432 __entry->add = le32_to_cpu(ctrl_ctx->add_flags); 433 ), 434 TP_printk("%s", xhci_decode_ctrl_ctx(__get_buf(XHCI_MSG_MAX), __entry->drop, __entry->add) 435 ) 436 ); 437 438 DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_address_ctrl_ctx, 439 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 440 TP_ARGS(ctrl_ctx) 441 ); 442 443 DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_configure_endpoint_ctrl_ctx, 444 TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx), 445 TP_ARGS(ctrl_ctx) 446 ); 447 448 DECLARE_EVENT_CLASS(xhci_log_ring, 449 TP_PROTO(struct xhci_ring *ring), 450 TP_ARGS(ring), 451 TP_STRUCT__entry( 452 __field(u32, type) 453 __field(void *, ring) 454 __field(dma_addr_t, enq) 455 __field(dma_addr_t, deq) 456 __field(dma_addr_t, enq_seg) 457 __field(dma_addr_t, deq_seg) 458 __field(unsigned int, num_segs) 459 __field(unsigned int, stream_id) 460 __field(unsigned int, cycle_state) 461 __field(unsigned int, bounce_buf_len) 462 ), 463 TP_fast_assign( 464 __entry->ring = ring; 465 __entry->type = ring->type; 466 __entry->num_segs = ring->num_segs; 467 __entry->stream_id = ring->stream_id; 468 __entry->enq_seg = ring->enq_seg->dma; 469 __entry->deq_seg = ring->deq_seg->dma; 470 __entry->cycle_state = ring->cycle_state; 471 __entry->bounce_buf_len = ring->bounce_buf_len; 472 __entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue); 473 __entry->deq = xhci_trb_virt_to_dma(ring->deq_seg, ring->dequeue); 474 ), 475 TP_printk("%s %p: enq %pad(%pad) deq %pad(%pad) segs %d stream %d bounce %d cycle %d", 476 xhci_ring_type_string(__entry->type), __entry->ring, 477 &__entry->enq, &__entry->enq_seg, 478 &__entry->deq, &__entry->deq_seg, 479 __entry->num_segs, 480 __entry->stream_id, 481 __entry->bounce_buf_len, 482 __entry->cycle_state 483 ) 484 ); 485 486 DEFINE_EVENT(xhci_log_ring, xhci_ring_alloc, 487 TP_PROTO(struct xhci_ring *ring), 488 TP_ARGS(ring) 489 ); 490 491 DEFINE_EVENT(xhci_log_ring, xhci_ring_free, 492 TP_PROTO(struct xhci_ring *ring), 493 TP_ARGS(ring) 494 ); 495 496 DEFINE_EVENT(xhci_log_ring, xhci_ring_expansion, 497 TP_PROTO(struct xhci_ring *ring), 498 TP_ARGS(ring) 499 ); 500 501 DEFINE_EVENT(xhci_log_ring, xhci_inc_enq, 502 TP_PROTO(struct xhci_ring *ring), 503 TP_ARGS(ring) 504 ); 505 506 DEFINE_EVENT(xhci_log_ring, xhci_inc_deq, 507 TP_PROTO(struct xhci_ring *ring), 508 TP_ARGS(ring) 509 ); 510 511 DECLARE_EVENT_CLASS(xhci_log_portsc, 512 TP_PROTO(u32 portnum, u32 portsc), 513 TP_ARGS(portnum, portsc), 514 TP_STRUCT__entry( 515 __field(u32, portnum) 516 __field(u32, portsc) 517 ), 518 TP_fast_assign( 519 __entry->portnum = portnum; 520 __entry->portsc = portsc; 521 ), 522 TP_printk("port-%d: %s", 523 __entry->portnum, 524 xhci_decode_portsc(__get_buf(XHCI_MSG_MAX), __entry->portsc) 525 ) 526 ); 527 528 DEFINE_EVENT(xhci_log_portsc, xhci_handle_port_status, 529 TP_PROTO(u32 portnum, u32 portsc), 530 TP_ARGS(portnum, portsc) 531 ); 532 533 DEFINE_EVENT(xhci_log_portsc, xhci_get_port_status, 534 TP_PROTO(u32 portnum, u32 portsc), 535 TP_ARGS(portnum, portsc) 536 ); 537 538 DEFINE_EVENT(xhci_log_portsc, xhci_hub_status_data, 539 TP_PROTO(u32 portnum, u32 portsc), 540 TP_ARGS(portnum, portsc) 541 ); 542 543 DECLARE_EVENT_CLASS(xhci_log_doorbell, 544 TP_PROTO(u32 slot, u32 doorbell), 545 TP_ARGS(slot, doorbell), 546 TP_STRUCT__entry( 547 __field(u32, slot) 548 __field(u32, doorbell) 549 ), 550 TP_fast_assign( 551 __entry->slot = slot; 552 __entry->doorbell = doorbell; 553 ), 554 TP_printk("Ring doorbell for %s", 555 xhci_decode_doorbell(__get_buf(XHCI_MSG_MAX), __entry->slot, __entry->doorbell) 556 ) 557 ); 558 559 DEFINE_EVENT(xhci_log_doorbell, xhci_ring_ep_doorbell, 560 TP_PROTO(u32 slot, u32 doorbell), 561 TP_ARGS(slot, doorbell) 562 ); 563 564 DEFINE_EVENT(xhci_log_doorbell, xhci_ring_host_doorbell, 565 TP_PROTO(u32 slot, u32 doorbell), 566 TP_ARGS(slot, doorbell) 567 ); 568 569 DECLARE_EVENT_CLASS(xhci_dbc_log_request, 570 TP_PROTO(struct dbc_request *req), 571 TP_ARGS(req), 572 TP_STRUCT__entry( 573 __field(struct dbc_request *, req) 574 __field(bool, dir) 575 __field(unsigned int, actual) 576 __field(unsigned int, length) 577 __field(int, status) 578 ), 579 TP_fast_assign( 580 __entry->req = req; 581 __entry->dir = req->direction; 582 __entry->actual = req->actual; 583 __entry->length = req->length; 584 __entry->status = req->status; 585 ), 586 TP_printk("%s: req %p length %u/%u ==> %d", 587 __entry->dir ? "bulk-in" : "bulk-out", 588 __entry->req, __entry->actual, 589 __entry->length, __entry->status 590 ) 591 ); 592 593 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_alloc_request, 594 TP_PROTO(struct dbc_request *req), 595 TP_ARGS(req) 596 ); 597 598 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_free_request, 599 TP_PROTO(struct dbc_request *req), 600 TP_ARGS(req) 601 ); 602 603 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_queue_request, 604 TP_PROTO(struct dbc_request *req), 605 TP_ARGS(req) 606 ); 607 608 DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_giveback_request, 609 TP_PROTO(struct dbc_request *req), 610 TP_ARGS(req) 611 ); 612 #endif /* __XHCI_TRACE_H */ 613 614 /* this part must be outside header guard */ 615 616 #undef TRACE_INCLUDE_PATH 617 #define TRACE_INCLUDE_PATH . 618 619 #undef TRACE_INCLUDE_FILE 620 #define TRACE_INCLUDE_FILE xhci-trace 621 622 #include <trace/define_trace.h> 623