Home
last modified time | relevance | path

Searched refs:txreq (Results 1 – 17 of 17) sorted by relevance

/openbmc/qemu/hw/net/
H A Dxen_nic.c131 netif_tx_request_t txreq; in net_tx_packets() local
147 memcpy(&txreq, RING_GET_REQUEST(&netdev->tx_ring, rc), sizeof(txreq)); in net_tx_packets()
157 net_tx_error(netdev, &txreq, rc); in net_tx_packets()
160 if (txreq.flags & NETTXF_more_data) { in net_tx_packets()
168 if (txreq.size < 14) { in net_tx_packets()
175 if ((txreq.offset + txreq.size) > XEN_PAGE_SIZE) { in net_tx_packets()
182 trace_xen_netdev_tx(netdev->dev, txreq.gref, txreq.offset, in net_tx_packets()
183 txreq.size, txreq.flags, in net_tx_packets()
203 memcpy(tmpbuf, page + txreq.offset, txreq.size); in net_tx_packets()
206 txreq.size); in net_tx_packets()
[all …]
/openbmc/linux/drivers/infiniband/hw/hfi1/
H A Dvnic_sdma.c29 struct sdma_txreq txreq; member
41 struct vnic_txreq *tx = container_of(txreq, struct vnic_txreq, txreq); in vnic_sdma_complete()
44 sdma_txclean(vnic_sdma->dd, txreq); in vnic_sdma_complete()
56 &tx->txreq, in build_vnic_ulp_payload()
67 &tx->txreq, in build_vnic_ulp_payload()
93 &tx->txreq, in build_vnic_tx_desc()
108 &tx->txreq, in build_vnic_tx_desc()
156 &tx->txreq, vnic_sdma->pkts_sent); in hfi1_vnic_send_dma()
168 sdma_txclean(dd, &tx->txreq); in hfi1_vnic_send_dma()
188 struct sdma_txreq *txreq, in hfi1_vnic_sdma_sleep() argument
[all …]
H A Dipoib_tx.c190 struct ipoib_txreq *tx = container_of(txreq, struct ipoib_txreq, txreq); in hfi1_ipoib_sdma_complete()
203 struct sdma_txreq *txreq = &tx->txreq; in hfi1_ipoib_build_ulp_payload() local
218 txreq, in hfi1_ipoib_build_ulp_payload()
234 struct sdma_txreq *txreq = &tx->txreq; in hfi1_ipoib_build_tx_desc() local
246 txreq, in hfi1_ipoib_build_tx_desc()
377 tx->txreq.num_desc = 0; in hfi1_ipoib_send_dma_common()
446 &tx->txreq, in hfi1_ipoib_submit_tx()
615 struct sdma_txreq *txreq, in hfi1_ipoib_sdma_sleep() argument
772 struct sdma_txreq *txreq; in hfi1_ipoib_drain_tx_list() local
777 container_of(txreq, struct ipoib_txreq, txreq); in hfi1_ipoib_drain_tx_list()
[all …]
H A Dverbs_txreq.h18 struct sdma_txreq txreq; member
53 tx->txreq.num_desc = 0; in get_txreq()
56 tx->txreq.flags = 0; in get_txreq()
66 return container_of(stx, struct verbs_txreq, txreq); in get_waiting_verbs_txreq()
H A Duser_sdma.c59 struct sdma_txreq *txreq,
67 struct sdma_txreq *txreq, in defer_packet_queue() argument
76 if (sdma_progress(sde, seq, txreq)) in defer_packet_queue()
664 ret = sdma_txinit_ahg(&tx->txreq, SDMA_TXREQ_F_AHG_COPY, in user_sdma_txadd_ahg()
671 sdma_txclean(pq->dd, &tx->txreq); in user_sdma_txadd_ahg()
784 ret = sdma_txinit(&tx->txreq, 0, sizeof(req->hdr) + in user_sdma_send_pkts()
810 list_add_tail(&tx->txreq.list, &req->txps); in user_sdma_send_pkts()
837 sdma_txclean(pq->dd, &tx->txreq); in user_sdma_send_pkts()
1135 sdma_txinit_ahg(&tx->txreq, in set_txreq_header_ahg()
1157 container_of(txreq, struct user_sdma_txreq, txreq); in user_sdma_txreq_cb()
[all …]
H A Dverbs.c590 container_of(cookie, struct verbs_txreq, txreq); in verbs_sdma_complete()
637 list_add_tail(&ps->s_txreq->txreq.list, in wait_kmem()
671 &tx->txreq, in build_verbs_ulp_payload()
740 &tx->txreq, in build_verbs_tx_desc()
754 &tx->txreq, in build_verbs_tx_desc()
761 &tx->txreq, in build_verbs_tx_desc()
781 ret = sdma_txadd_daddr(sde->dd, &tx->txreq, sde->dd->sdma_pad_phys, in build_verbs_tx_desc()
823 if (!sdma_txreq_built(&tx->txreq)) { in hfi1_verbs_send_dma()
852 ret = sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent); in hfi1_verbs_send_dma()
900 list_add_tail(&ps->s_txreq->txreq.list, in pio_wait()
[all …]
H A Dipoib.h57 struct sdma_txreq txreq; member
H A Dverbs_txreq.c27 sdma_txclean(dd_from_dev(dev), &tx->txreq); in hfi1_put_txreq()
H A Duser_sdma.h187 struct sdma_txreq txreq; member
H A Dpin_system.c325 ret = sdma_txadd_page(pq->dd, &tx->txreq, in add_mapping_to_sdma_packet()
H A Dqp.c118 container_of(tx, struct verbs_txreq, txreq)); in flush_list_head()
443 struct verbs_txreq *tx = container_of(stx, struct verbs_txreq, txreq); in iowait_sleep()
H A Drc.c349 ps->s_txreq->txreq.flags |= SDMA_TXREQ_F_VIP; in make_rc_ack()
H A Dtid_rdma.c5318 ps->s_txreq->txreq.flags |= SDMA_TXREQ_F_VIP; in make_tid_rdma_ack()
/openbmc/linux/drivers/infiniband/hw/qib/
H A Dqib_sdma.c496 tx->txreq.start_idx = 0; in complete_sdma_err_req()
497 tx->txreq.next_descq_idx = 0; in complete_sdma_err_req()
498 list_add_tail(&tx->txreq.list, &ppd->sdma_activelist); in complete_sdma_err_req()
547 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF) in qib_sdma_verbs_send()
563 tx->txreq.start_idx = tail; in qib_sdma_verbs_send()
580 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_USELARGEBUF) in qib_sdma_verbs_send()
601 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_HEADTOHOST) in qib_sdma_verbs_send()
603 if (tx->txreq.flags & QIB_SDMA_TXREQ_F_INTREQ) in qib_sdma_verbs_send()
607 tx->txreq.next_descq_idx = tail; in qib_sdma_verbs_send()
609 ppd->sdma_descq_added += tx->txreq.sg_count; in qib_sdma_verbs_send()
[all …]
H A Dqib_verbs.c628 tx->txreq.flags &= ~QIB_SDMA_TXREQ_F_FREEBUF; in qib_put_txreq()
630 tx->txreq.addr, tx->hdr_dwords << 2, in qib_put_txreq()
638 list_add(&tx->txreq.list, &dev->txreq_free); in qib_put_txreq()
686 if (qpp->s_tx->txreq.sg_count > avail) in qib_verbs_sdma_desc_avail()
688 avail -= qpp->s_tx->txreq.sg_count; in qib_verbs_sdma_desc_avail()
804 tx->txreq.callback = sdma_complete; in qib_verbs_send_dma()
808 tx->txreq.flags = QIB_SDMA_TXREQ_F_INTREQ; in qib_verbs_send_dma()
828 tx->txreq.sg_count = ndesc; in qib_verbs_send_dma()
829 tx->txreq.addr = dev->pio_hdrs_phys + in qib_verbs_send_dma()
851 tx->txreq.flags |= QIB_SDMA_TXREQ_F_FREEBUF; in qib_verbs_send_dma()
[all …]
H A Dqib.h248 struct qib_sdma_txreq txreq; member
/openbmc/linux/drivers/net/xen-netback/
H A Dnetback.c924 struct xen_netif_tx_request txreq; in xenvif_tx_build_gops() local
949 RING_COPY_REQUEST(&queue->tx, idx, &txreq); in xenvif_tx_build_gops()
953 tx_credit_exceeded(queue, txreq.size)) in xenvif_tx_build_gops()
956 queue->remaining_credit -= txreq.size; in xenvif_tx_build_gops()
963 if (txreq.flags & XEN_NETTXF_extra_info) { in xenvif_tx_build_gops()
997 XEN_NETBACK_TX_COPY_LEN : txreq.size; in xenvif_tx_build_gops()
1007 if (unlikely(txreq.size < ETH_HLEN)) { in xenvif_tx_build_gops()
1009 "Bad packet size: %d\n", txreq.size); in xenvif_tx_build_gops()
1015 if (unlikely((txreq.offset + txreq.size) > XEN_PAGE_SIZE)) { in xenvif_tx_build_gops()
1017 txreq.offset, txreq.size); in xenvif_tx_build_gops()
[all …]