Lines Matching refs:hwreq

357 static int add_td_to_list(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq,  in add_td_to_list()  argument
378 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in add_td_to_list()
380 if (hwreq->req.length == 0 in add_td_to_list()
381 || hwreq->req.length % hwep->ep.maxpacket) in add_td_to_list()
387 temp = (u32) (sg_dma_address(s) + hwreq->req.actual); in add_td_to_list()
390 temp = (u32) (hwreq->req.dma + hwreq->req.actual); in add_td_to_list()
402 hwreq->req.actual += length; in add_td_to_list()
404 if (!list_empty(&hwreq->tds)) { in add_td_to_list()
406 lastnode = list_entry(hwreq->tds.prev, in add_td_to_list()
412 list_add_tail(&node->td, &hwreq->tds); in add_td_to_list()
427 struct ci_hw_req *hwreq) in prepare_td_for_non_sg() argument
429 unsigned int rest = hwreq->req.length; in prepare_td_for_non_sg()
434 ret = add_td_to_list(hwep, hwreq, 0, NULL); in prepare_td_for_non_sg()
443 if (hwreq->req.dma % PAGE_SIZE) in prepare_td_for_non_sg()
447 unsigned int count = min(hwreq->req.length - hwreq->req.actual, in prepare_td_for_non_sg()
450 ret = add_td_to_list(hwep, hwreq, count, NULL); in prepare_td_for_non_sg()
457 if (hwreq->req.zero && hwreq->req.length && hwep->dir == TX in prepare_td_for_non_sg()
458 && (hwreq->req.length % hwep->ep.maxpacket == 0)) { in prepare_td_for_non_sg()
459 ret = add_td_to_list(hwep, hwreq, 0, NULL); in prepare_td_for_non_sg()
467 static int prepare_td_per_sg(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq, in prepare_td_per_sg() argument
473 hwreq->req.actual = 0; in prepare_td_per_sg()
478 ret = add_td_to_list(hwep, hwreq, count, s); in prepare_td_per_sg()
507 static int prepare_td_for_sg(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in prepare_td_for_sg() argument
509 struct usb_request *req = &hwreq->req; in prepare_td_for_sg()
529 ret = prepare_td_per_sg(hwep, hwreq, s); in prepare_td_for_sg()
533 node = list_entry(hwreq->tds.prev, in prepare_td_for_sg()
550 static int _hardware_enqueue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_enqueue() argument
557 if (hwreq->req.status == -EALREADY) in _hardware_enqueue()
560 hwreq->req.status = -EALREADY; in _hardware_enqueue()
563 &hwreq->req, hwep->dir); in _hardware_enqueue()
567 if (hwreq->req.num_mapped_sgs) in _hardware_enqueue()
568 ret = prepare_td_for_sg(hwep, hwreq); in _hardware_enqueue()
570 ret = prepare_td_for_non_sg(hwep, hwreq); in _hardware_enqueue()
575 lastnode = list_entry(hwreq->tds.prev, in _hardware_enqueue()
579 if (!hwreq->req.no_interrupt) in _hardware_enqueue()
582 list_for_each_entry_safe(firstnode, lastnode, &hwreq->tds, td) in _hardware_enqueue()
583 trace_ci_prepare_td(hwep, hwreq, firstnode); in _hardware_enqueue()
585 firstnode = list_first_entry(&hwreq->tds, struct td_node, td); in _hardware_enqueue()
589 hwreq->req.actual = 0; in _hardware_enqueue()
627 u32 mul = hwreq->req.length / hwep->ep.maxpacket; in _hardware_enqueue()
629 if (hwreq->req.length == 0 in _hardware_enqueue()
630 || hwreq->req.length % hwep->ep.maxpacket) in _hardware_enqueue()
672 static int _hardware_dequeue(struct ci_hw_ep *hwep, struct ci_hw_req *hwreq) in _hardware_dequeue() argument
677 unsigned actual = hwreq->req.length; in _hardware_dequeue()
680 if (hwreq->req.status != -EALREADY) in _hardware_dequeue()
683 hwreq->req.status = 0; in _hardware_dequeue()
685 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _hardware_dequeue()
687 trace_ci_complete_td(hwep, hwreq, node); in _hardware_dequeue()
694 hwreq->req.status = -EALREADY; in _hardware_dequeue()
702 hwreq->req.status = tmptoken & TD_STATUS; in _hardware_dequeue()
703 if ((TD_STATUS_HALTED & hwreq->req.status)) { in _hardware_dequeue()
704 hwreq->req.status = -EPIPE; in _hardware_dequeue()
706 } else if ((TD_STATUS_DT_ERR & hwreq->req.status)) { in _hardware_dequeue()
707 hwreq->req.status = -EPROTO; in _hardware_dequeue()
709 } else if ((TD_STATUS_TR_ERR & hwreq->req.status)) { in _hardware_dequeue()
710 hwreq->req.status = -EILSEQ; in _hardware_dequeue()
716 hwreq->req.status = -EPROTO; in _hardware_dequeue()
733 &hwreq->req, hwep->dir); in _hardware_dequeue()
735 hwreq->req.actual += actual; in _hardware_dequeue()
737 if (hwreq->req.status) in _hardware_dequeue()
738 return hwreq->req.status; in _hardware_dequeue()
740 return hwreq->req.actual; in _hardware_dequeue()
763 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, in _ep_nuke() local
766 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in _ep_nuke()
773 list_del_init(&hwreq->queue); in _ep_nuke()
774 hwreq->req.status = -ESHUTDOWN; in _ep_nuke()
776 if (hwreq->req.complete != NULL) { in _ep_nuke()
778 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in _ep_nuke()
932 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in _ep_queue() local
951 hwreq->req.length > hwep->ep.mult * hwep->ep.maxpacket) { in _ep_queue()
957 if (!list_empty(&hwreq->queue)) { in _ep_queue()
963 hwreq->req.status = -EINPROGRESS; in _ep_queue()
964 hwreq->req.actual = 0; in _ep_queue()
966 retval = _hardware_enqueue(hwep, hwreq); in _ep_queue()
971 list_add_tail(&hwreq->queue, &hwep->qh.queue); in _ep_queue()
1104 struct ci_hw_req *hwreq, *hwreqtemp; in isr_tr_complete_low() local
1108 list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue, in isr_tr_complete_low()
1110 retval = _hardware_dequeue(hwep, hwreq); in isr_tr_complete_low()
1113 list_del_init(&hwreq->queue); in isr_tr_complete_low()
1114 if (hwreq->req.complete != NULL) { in isr_tr_complete_low()
1117 hwreq->req.length) in isr_tr_complete_low()
1119 usb_gadget_giveback_request(&hweptemp->ep, &hwreq->req); in isr_tr_complete_low()
1466 struct ci_hw_req *hwreq; in ep_alloc_request() local
1471 hwreq = kzalloc(sizeof(struct ci_hw_req), gfp_flags); in ep_alloc_request()
1472 if (hwreq != NULL) { in ep_alloc_request()
1473 INIT_LIST_HEAD(&hwreq->queue); in ep_alloc_request()
1474 INIT_LIST_HEAD(&hwreq->tds); in ep_alloc_request()
1477 return (hwreq == NULL) ? NULL : &hwreq->req; in ep_alloc_request()
1488 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_free_request() local
1494 } else if (!list_empty(&hwreq->queue)) { in ep_free_request()
1501 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_free_request()
1508 kfree(hwreq); in ep_free_request()
1546 struct ci_hw_req *hwreq = container_of(req, struct ci_hw_req, req); in ep_dequeue() local
1550 if (ep == NULL || req == NULL || hwreq->req.status != -EALREADY || in ep_dequeue()
1551 hwep->ep.desc == NULL || list_empty(&hwreq->queue) || in ep_dequeue()
1559 list_for_each_entry_safe(node, tmpnode, &hwreq->tds, td) { in ep_dequeue()
1566 list_del_init(&hwreq->queue); in ep_dequeue()
1572 if (hwreq->req.complete != NULL) { in ep_dequeue()
1574 usb_gadget_giveback_request(&hwep->ep, &hwreq->req); in ep_dequeue()