/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos_sq.c | 23 if (!pfvf->qset.pool) in otx2_qos_aura_pool_free() 26 pool = &pfvf->qset.pool[pool_id]; in otx2_qos_aura_pool_free() 35 struct otx2_qset *qset = &pfvf->qset; in otx2_qos_sq_aura_pool_init() local 57 pool = &pfvf->qset.pool[pool_id]; in otx2_qos_sq_aura_pool_init() 76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init() 118 struct otx2_qset *qset = &pfvf->qset; in otx2_qos_sq_free_sqbs() local 124 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 140 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 261 struct otx2_qset *qset = &pfvf->qset; in otx2_qos_disable_sq() local 273 sq = &pfvf->qset.sq[sq_idx]; in otx2_qos_disable_sq() [all …]
|
H A D | otx2_common.c | 80 if (!pfvf->qset.rq) in otx2_update_rq_stats() 91 if (!pfvf->qset.sq) in otx2_update_sq_stats() 855 struct otx2_qset *qset = &pfvf->qset; in otx2_rq_init() local 925 struct otx2_qset *qset = &pfvf->qset; in otx2_sq_init() local 931 sq = &qset->sq[qidx]; in otx2_sq_init() 987 struct otx2_qset *qset = &pfvf->qset; in otx2_cq_init() local 1187 struct otx2_qset *qset = &pfvf->qset; in otx2_sq_free_sqbs() local 1429 struct otx2_qset *qset = &pfvf->qset; in otx2_sq_aura_pool_init() local 1550 struct otx2_qset *qset = &pfvf->qset; in otx2_config_npa() local 1564 if (!qset->pool) in otx2_config_npa() [all …]
|
H A D | otx2_pf.c | 1391 struct otx2_qset *qset = &pf->qset; in otx2_disable_napi() local 1405 struct otx2_qset *qset = &pf->qset; in otx2_free_cq_res() local 1419 struct otx2_qset *qset = &pf->qset; in otx2_free_sq_res() local 1594 struct otx2_qset *qset = &pf->qset; in otx2_free_hw_resources() local 1764 struct otx2_qset *qset = &pf->qset; in otx2_open() local 1784 qset->rqe_cnt = qset->rqe_cnt ? qset->rqe_cnt : Q_COUNT(Q_SIZE_256); in otx2_open() 1786 qset->sqe_cnt = qset->sqe_cnt ? qset->sqe_cnt : Q_COUNT(Q_SIZE_4K); in otx2_open() 1789 qset->cq = kcalloc(pf->qset.cq_cnt, in otx2_open() 1791 if (!qset->cq) in otx2_open() 1796 if (!qset->sq) in otx2_open() [all …]
|
H A D | otx2_txrx.c | 463 sq = &pfvf->qset.sq[qidx]; in otx2_tx_napi_handler() 478 otx2_snd_pkt_handler(pfvf, cq, &pfvf->qset.sq[qidx], in otx2_tx_napi_handler() 536 struct otx2_qset *qset; in otx2_napi_handler() local 542 qset = &pfvf->qset; in otx2_napi_handler() 548 cq = &qset->cq[cq_idx]; in otx2_napi_handler() 1228 pool = &pfvf->qset.pool[pool_id]; in otx2_cleanup_rx_cqes() 1263 sq = &pfvf->qset.sq[qidx]; in otx2_cleanup_tx_cqes() 1328 sq = &pfvf->qset.sq[sq_idx]; in otx2_free_pending_sqe() 1376 sq = &pfvf->qset.sq[qidx]; in otx2_xdp_sq_append_pkt()
|
H A D | otx2_ethtool.c | 83 static void otx2_get_qset_strings(struct otx2_nic *pfvf, u8 **data, int qset) in otx2_get_qset_strings() argument 85 int start_qidx = qset * pfvf->hw.rx_queues; in otx2_get_qset_strings() 164 *((*data)++) = ((u64 *)&pfvf->qset.rq[qidx].stats) in otx2_get_qset_stats() 175 *((*data)++) = ((u64 *)&pfvf->qset.sq[qidx].stats) in otx2_get_qset_stats() 382 struct otx2_qset *qs = &pfvf->qset; in otx2_get_ringparam() 402 struct otx2_qset *qs = &pfvf->qset; in otx2_set_ringparam()
|
H A D | cn10k.c | 99 aq->sq.cq_limit = ((SEND_CQ_SKID * 256) / (pfvf->qset.sqe_cnt)); in cn10k_sq_aq_init()
|
H A D | otx2_common.h | 472 struct otx2_qset qset; member 599 pfvf->qset.rqe_cnt = Q_COUNT(Q_SIZE_1K); in otx2_setup_dev_hw_settings()
|
H A D | otx2_vf.c | 395 sq = &vf->qset.sq[qidx]; in otx2vf_xmit()
|
/openbmc/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_idc.c | 70 if (WARN_ON(!pf || !qset)) in ice_add_rdma_qset() 87 max_rdmaqs[qset->tc]++; in ice_add_rdma_qset() 88 qs_handle = qset->qs_handle; in ice_add_rdma_qset() 97 status = ice_ena_vsi_rdma_qset(vsi->port_info, vsi->idx, qset->tc, in ice_add_rdma_qset() 103 vsi->qset_handle[qset->tc] = qset->qs_handle; in ice_add_rdma_qset() 104 qset->teid = qset_teid; in ice_add_rdma_qset() 121 if (WARN_ON(!pf || !qset)) in ice_del_rdma_qset() 124 vsi = ice_find_vsi(pf, qset->vport_id); in ice_del_rdma_qset() 130 q_id = qset->qs_handle; in ice_del_rdma_qset() 131 teid = qset->teid; in ice_del_rdma_qset() [all …]
|
/openbmc/linux/drivers/net/ethernet/fungible/funeth/ |
H A D | funeth_main.c | 468 free_rxqs(rxqs, qset->nrxqs, qset->rxq_start, qset->state); in fun_free_rings() 469 free_txqs(qset->txqs, qset->ntxqs, qset->txq_start, qset->state); in fun_free_rings() 470 free_xdpqs(xdpqs, qset->nxdpqs, qset->xdpq_start, qset->state); in fun_free_rings() 475 qset->rxqs = rxqs; in fun_free_rings() 485 err = fun_alloc_queue_irqs(netdev, qset->ntxqs, qset->nrxqs); in fun_alloc_rings() 494 xdpqs = alloc_xdpqs(netdev, qset->nxdpqs, qset->sq_depth, in fun_alloc_rings() 495 qset->xdpq_start, qset->state); in fun_alloc_rings() 503 err = alloc_txqs(netdev, txqs, qset->ntxqs, qset->sq_depth, in fun_alloc_rings() 504 qset->txq_start, qset->state); in fun_alloc_rings() 508 err = alloc_rxqs(netdev, rxqs, qset->nrxqs, qset->cq_depth, in fun_alloc_rings() [all …]
|
/openbmc/linux/drivers/infiniband/hw/irdma/ |
H A D | main.c | 170 struct iidc_rdma_qset_params qset = {}; in irdma_lan_register_qset() local 173 qset.qs_handle = tc_node->qs_handle; in irdma_lan_register_qset() 174 qset.tc = tc_node->traffic_class; in irdma_lan_register_qset() 175 qset.vport_id = vsi->vsi_idx; in irdma_lan_register_qset() 176 ret = ice_add_rdma_qset(pf, &qset); in irdma_lan_register_qset() 182 tc_node->l2_sched_node_id = qset.teid; in irdma_lan_register_qset() 200 qset.qs_handle = tc_node->qs_handle; in irdma_lan_unregister_qset() 201 qset.tc = tc_node->traffic_class; in irdma_lan_unregister_qset() 202 qset.vport_id = vsi->vsi_idx; in irdma_lan_unregister_qset() 203 qset.teid = tc_node->l2_sched_node_id; in irdma_lan_unregister_qset() [all …]
|
H A D | i40iw_if.c | 113 u16 qset; in i40iw_open() local 136 qset = cdev_info->params.qos.prio_qos[i].qs_handle; in i40iw_open() 138 l2params.qs_handle_list[i] = qset; in i40iw_open() 140 last_qset = qset; in i40iw_open() 141 else if ((qset != last_qset) && (qset != IRDMA_NO_QSET)) in i40iw_open()
|
/openbmc/linux/include/linux/net/intel/ |
H A D | iidc.h | 76 int ice_add_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset); 77 int ice_del_rdma_qset(struct ice_pf *pf, struct iidc_rdma_qset_params *qset);
|
/openbmc/linux/drivers/net/ethernet/cavium/thunder/ |
H A D | nic_main.c | 478 u8 qset, rq_idx = 0; in nic_config_cpi() local 504 qset = cfg->vf_id; in nic_config_cpi() 508 (qset << 3) | rq_idx); in nic_config_cpi() 567 u8 qset, idx = 0; in nic_config_rss() local 579 qset = nic->vf_sqs[cfg->vf_id][svf - 1]; in nic_config_rss() 581 qset = cfg->vf_id; in nic_config_rss() 583 (qset << 3) | (cfg->ind_tbl[idx] & 0x7)); in nic_config_rss()
|
H A D | nicvf_ethtool.c | 212 static void nicvf_get_qset_strings(struct nicvf *nic, u8 **data, int qset) in nicvf_get_qset_strings() argument 215 int start_qidx = qset * MAX_RCV_QUEUES_PER_QS; in nicvf_get_qset_strings()
|
/openbmc/linux/drivers/net/ethernet/chelsio/cxgb3/ |
H A D | cxgb3_main.c | 670 &adap->params.sge.qset[qset_idx], ntxq, dev, in setup_sge_qsets() 1957 const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset]; in get_sge_param() 1991 q = &adapter->params.sge.qset[pi->first_qset]; in set_sge_param() 2017 qsp = &adapter->params.sge.qset[i]; in set_coalesce() 2032 struct qset_params *q = adapter->params.sge.qset; in get_coalesce() 2202 q = &adapter->params.sge.qset[t.qset_idx]; in cxgb_siocdevprivate() 2236 qset[i]; in cxgb_siocdevprivate() 2278 q = &adapter->params.sge.qset[q1 + t.qset_idx]; in cxgb_siocdevprivate()
|
H A D | common.h | 329 struct qset_params qset[SGE_QSETS]; member
|
H A D | sge.c | 3369 struct qset_params *q = p->qset + i; in t3_sge_prep()
|
/openbmc/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | cxgb4_debugfs.c | 2601 static inline struct port_info *ethqset2pinfo(struct adapter *adap, int qset) in ethqset2pinfo() argument 2608 if (qset >= pi->first_qset && in ethqset2pinfo() 2609 qset < pi->first_qset + pi->nqsets) in ethqset2pinfo()
|