Lines Matching refs:tx_info
461 struct sk_buff *skb, u32 tx_info) in mt76_dma_tx_queue_skb_raw() argument
481 mt76_dma_add_buf(dev, q, &buf, 1, tx_info, skb, NULL); in mt76_dma_tx_queue_skb_raw()
500 struct mt76_tx_info tx_info = { in mt76_dma_tx_queue_skb() local
528 tx_info.buf[n].addr = t->dma_addr; in mt76_dma_tx_queue_skb()
529 tx_info.buf[n++].len = dev->drv->txwi_size; in mt76_dma_tx_queue_skb()
530 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
531 tx_info.buf[n++].len = len; in mt76_dma_tx_queue_skb()
534 if (n == ARRAY_SIZE(tx_info.buf)) in mt76_dma_tx_queue_skb()
542 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
543 tx_info.buf[n++].len = iter->len; in mt76_dma_tx_queue_skb()
545 tx_info.nbuf = n; in mt76_dma_tx_queue_skb()
547 if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) { in mt76_dma_tx_queue_skb()
554 ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info); in mt76_dma_tx_queue_skb()
560 return mt76_dma_add_buf(dev, q, tx_info.buf, tx_info.nbuf, in mt76_dma_tx_queue_skb()
561 tx_info.info, tx_info.skb, t); in mt76_dma_tx_queue_skb()
565 dma_unmap_single(dev->dma_dev, tx_info.buf[n].addr, in mt76_dma_tx_queue_skb()
566 tx_info.buf[n].len, DMA_TO_DEVICE); in mt76_dma_tx_queue_skb()
574 if (tx_info.skb == phy->test.tx_skb) in mt76_dma_tx_queue_skb()
582 status.skb = tx_info.skb; in mt76_dma_tx_queue_skb()
583 hw = mt76_tx_status_get_hw(dev, tx_info.skb); in mt76_dma_tx_queue_skb()