Lines Matching refs:pdev_dev
323 unmap_skb(adap->pdev_dev, d->skb, d->addr); in free_tx_desc()
444 dma_unmap_page(adap->pdev_dev, get_buf_addr(d), in free_rx_bufs()
471 dma_unmap_page(adap->pdev_dev, get_buf_addr(d), in unmap_rx_buf()
552 node = dev_to_node(adap->pdev_dev); in refill_fl()
567 mapping = dma_map_page(adap->pdev_dev, pg, 0, in refill_fl()
570 if (unlikely(dma_mapping_error(adap->pdev_dev, mapping))) { in refill_fl()
598 mapping = dma_map_page(adap->pdev_dev, pg, 0, PAGE_SIZE, in refill_fl()
600 if (unlikely(dma_mapping_error(adap->pdev_dev, mapping))) { in refill_fl()
1568 dev_err(adap->pdev_dev, in cxgb4_eth_xmit()
1586 unlikely(cxgb4_map_skb(adap->pdev_dev, skb, sgl_sdesc->addr) < 0)) { in cxgb4_eth_xmit()
1880 dev_err(adapter->pdev_dev, in cxgb4_vf_eth_xmit()
1892 unlikely(cxgb4_map_skb(adapter->pdev_dev, skb, in cxgb4_vf_eth_xmit()
2139 unmap_skb(adap->pdev_dev, d->skb, d->addr); in cxgb4_eosw_txq_free_desc()
2348 ret = cxgb4_map_skb(adap->pdev_dev, skb, d->addr); in ethofld_hard_xmit()
2991 else if (cxgb4_map_skb(q->adap->pdev_dev, skb, in service_ofldq()
3942 dma_sync_single_for_cpu(q->adap->pdev_dev, in process_responses()
4387 iq->desc = alloc_ring(adap->pdev_dev, iq->size, iq->iqe_len, 0, in t4_sge_alloc_rxq()
4389 dev_to_node(adap->pdev_dev)); in t4_sge_alloc_rxq()
4430 fl->desc = alloc_ring(adap->pdev_dev, fl->size, sizeof(__be64), in t4_sge_alloc_rxq()
4433 dev_to_node(adap->pdev_dev)); in t4_sge_alloc_rxq()
4541 dev_warn(adap->pdev_dev, "Failed to set Congestion" in t4_sge_alloc_rxq()
4552 dma_free_coherent(adap->pdev_dev, iq->size * iq->iqe_len, in t4_sge_alloc_rxq()
4559 dma_free_coherent(adap->pdev_dev, flsz * sizeof(struct tx_desc), in t4_sge_alloc_rxq()
4603 txq->q.desc = alloc_ring(adap->pdev_dev, txq->q.size, in t4_sge_alloc_eth_txq()
4663 dma_free_coherent(adap->pdev_dev, in t4_sge_alloc_eth_txq()
4696 txq->q.desc = alloc_ring(adap->pdev_dev, nentries, in t4_sge_alloc_ctrl_txq()
4698 NULL, 0, dev_to_node(adap->pdev_dev)); in t4_sge_alloc_ctrl_txq()
4725 dma_free_coherent(adap->pdev_dev, in t4_sge_alloc_ctrl_txq()
4765 q->desc = alloc_ring(adap->pdev_dev, q->size, sizeof(struct tx_desc), in t4_sge_alloc_ofld_txq()
4798 dma_free_coherent(adap->pdev_dev, in t4_sge_alloc_ofld_txq()
4856 dma_free_coherent(adap->pdev_dev, in free_txq()
4873 dma_free_coherent(adap->pdev_dev, (rq->size + 1) * rq->iqe_len, in free_rspq_fl()
4882 dma_free_coherent(adap->pdev_dev, fl->size * 8 + s->stat_len, in free_rspq_fl()
5091 dev_err(adap->pdev_dev, "bad SGE CPL MODE\n"); in t4_sge_init_soft()
5124 dev_err(adap->pdev_dev, "bad SGE FL page buffer sizes [%d, %d]\n", in t4_sge_init_soft()
5133 dev_err(adap->pdev_dev, "bad SGE FL MTU sizes [%d, %d]\n", in t4_sge_init_soft()
5217 dev_err(adap->pdev_dev, "Unsupported Chip version %d\n", in t4_sge_init()