/openbmc/linux/drivers/infiniband/hw/usnic/ |
H A D | usnic_abi.h | 80 u32 cq_cnt; member
|
H A D | usnic_ib_verbs.c | 152 resp.cq_cnt = chunk->cnt; in usnic_ib_fill_create_qp_resp() 467 int cq_cnt; in usnic_ib_create_qp() local 501 cq_cnt = (init_attr->send_cq == init_attr->recv_cq) ? 1 : 2; in usnic_ib_create_qp() 503 usnic_vnic_res_spec_update(&res_spec, USNIC_VNIC_RES_TYPE_CQ, cq_cnt); in usnic_ib_create_qp()
|
/openbmc/linux/drivers/net/ethernet/apm/xgene/ |
H A D | xgene_enet_main.c | 887 if (!pdata->cq_cnt) { in xgene_enet_set_irq_name() 896 for (i = 0; i < pdata->cq_cnt; i++) { in xgene_enet_set_irq_name() 922 for (i = 0; i < pdata->cq_cnt; i++) { in xgene_enet_register_irq() 952 for (i = 0; i < pdata->cq_cnt; i++) { in xgene_enet_free_irq() 969 for (i = 0; i < pdata->cq_cnt; i++) { in xgene_enet_napi_enable() 985 for (i = 0; i < pdata->cq_cnt; i++) { in xgene_enet_napi_disable() 1071 if (pdata->cq_cnt) in xgene_enet_delete_desc_rings() 1161 if (ring->cp_ring && pdata->cq_cnt) in xgene_enet_free_desc_rings() 1422 if (!pdata->cq_cnt) { in xgene_enet_create_desc_rings() 1640 pdata->cq_cnt = max_irqs / 2; in xgene_enet_get_irqs() [all …]
|
H A D | xgene_enet_main.h | 200 u8 cq_cnt; member
|
/openbmc/linux/drivers/accel/habanalabs/common/ |
H A D | hw_queue.c | 633 int rc = 0, i, cq_cnt; in hl_hw_queue_schedule_cs() local 653 for (i = 0, cq_cnt = 0 ; i < max_queues ; i++, q++) { in hl_hw_queue_schedule_cs() 685 cq_cnt++; in hl_hw_queue_schedule_cs() 784 for (i = 0 ; (i < max_queues) && (cq_cnt > 0) ; i++, q++) { in hl_hw_queue_schedule_cs() 790 cq_cnt--; in hl_hw_queue_schedule_cs()
|
H A D | device.c | 2062 int i, rc, cq_cnt, user_interrupt_cnt, cq_ready_cnt; in hl_device_init() local 2109 cq_cnt = hdev->asic_prop.completion_queues_count; in hl_device_init() 2116 if (cq_cnt) { in hl_device_init() 2117 hdev->completion_queue = kcalloc(cq_cnt, in hl_device_init() 2129 for (i = 0, cq_ready_cnt = 0 ; i < cq_cnt ; i++, cq_ready_cnt++) { in hl_device_init()
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.h | 148 u16 cq_cnt; member
|
H A D | otx2_common.c | 1117 for (qidx = 0; qidx < pfvf->qset.cq_cnt; qidx++) { in otx2_config_nix_queues() 1127 pfvf->refill_wrk = devm_kcalloc(pfvf->dev, pfvf->qset.cq_cnt, in otx2_config_nix_queues() 1132 for (qidx = 0; qidx < pfvf->qset.cq_cnt; qidx++) { in otx2_config_nix_queues() 1156 nixlf->cq_cnt = pfvf->qset.cq_cnt; in otx2_config_nix()
|
H A D | otx2_pf.c | 1278 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1428 for (qidx = 0; qidx < qset->cq_cnt; qidx++) { in otx2_free_cq_res() 1643 for (qidx = 0; qidx < qset->cq_cnt; qidx++) { in otx2_free_hw_resources() 1794 pf->qset.cq_cnt = pf->hw.rx_queues + otx2_get_total_tx_queues(pf); in otx2_open() 1806 qset->cq = kcalloc(pf->qset.cq_cnt, in otx2_open() 2035 for (wrk = 0; wrk < pf->qset.cq_cnt; wrk++) in otx2_stop()
|
/openbmc/linux/drivers/net/ethernet/cavium/thunder/ |
H A D | nicvf_main.c | 456 nic->snicvf[sqs]->qs->cq_cnt = in nicvf_request_sqs() 995 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err() 1138 for (irq = 0; irq < nic->qs->cq_cnt; irq++) { in nicvf_register_interrupts() 1315 for (qidx = 0; qidx < nic->qs->cq_cnt; qidx++) { in nicvf_free_cq_poll() 1372 for (qidx = 0; qidx < nic->qs->cq_cnt; qidx++) { in nicvf_stop() 1467 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_open() 1543 for (qidx = 0; qidx < qs->cq_cnt; qidx++) in nicvf_open() 1567 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_open() 1826 nic->qs->cq_cnt = max_t(u8, nic->qs->rq_cnt, nic->qs->sq_cnt); in nicvf_set_xdp_queues()
|
H A D | nicvf_queues.c | 999 for (qidx = 0; qidx < qs->cq_cnt; qidx++) in nicvf_free_resources() 1026 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_alloc_resources() 1050 qs->cq_cnt = max_t(u8, qs->rq_cnt, qs->sq_cnt); in nicvf_set_qset_resources() 1089 for (qidx = 0; qidx < qs->cq_cnt; qidx++) in nicvf_config_data_transfer() 1102 for (qidx = 0; qidx < qs->cq_cnt; qidx++) in nicvf_config_data_transfer()
|
H A D | nicvf_queues.h | 293 u8 cq_cnt; member
|
H A D | nicvf_ethtool.c | 773 nic->qs->cq_cnt = max(nic->qs->rq_cnt, nic->qs->sq_cnt); in nicvf_set_channels()
|
/openbmc/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 614 dev->eq_tbl[0].cq_cnt++; in ocrdma_create_mq() 1002 int budget = eq->cq_cnt; in ocrdma_irq_handler() 1746 int i, selected_eq = 0, cq_cnt = 0; in ocrdma_bind_eq() local 1750 cq_cnt = dev->eq_tbl[0].cq_cnt; in ocrdma_bind_eq() 1756 if (dev->eq_tbl[i].cq_cnt < cq_cnt) { in ocrdma_bind_eq() 1757 cq_cnt = dev->eq_tbl[i].cq_cnt; in ocrdma_bind_eq() 1762 dev->eq_tbl[selected_eq].cq_cnt += 1; in ocrdma_bind_eq() 1775 dev->eq_tbl[i].cq_cnt -= 1; in ocrdma_unbind_eq()
|
H A D | ocrdma.h | 150 int cq_cnt; member
|
/openbmc/linux/drivers/infiniband/hw/irdma/ |
H A D | uk.c | 906 void irdma_uk_cq_set_resized_cnt(struct irdma_cq_uk *cq, u16 cq_cnt) in irdma_uk_cq_set_resized_cnt() argument 917 sw_cq_sel += cq_cnt; in irdma_uk_cq_set_resized_cnt()
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_nix.c | 1337 if (!req->rq_cnt || !req->sq_cnt || !req->cq_cnt) in rvu_mbox_handler_nix_lf_alloc() 1429 err = qmem_alloc(rvu->dev, &pfvf->cq_ctx, req->cq_cnt, hwctx_size); in rvu_mbox_handler_nix_lf_alloc() 1433 pfvf->cq_bmap = kcalloc(req->cq_cnt, sizeof(long), GFP_KERNEL); in rvu_mbox_handler_nix_lf_alloc() 1440 cfg = BIT_ULL(36) | (req->cq_cnt - 1) | req->way_mask << 20; in rvu_mbox_handler_nix_lf_alloc()
|
H A D | mbox.h | 860 u32 cq_cnt; /* No of completion queues */ member
|
/openbmc/linux/drivers/accel/habanalabs/goya/ |
H A D | goya.c | 2357 int cq_cnt = hdev->asic_prop.completion_queues_count; in goya_enable_msix() local 2372 for (i = 0, irq_cnt_init = 0 ; i < cq_cnt ; i++, irq_cnt_init++) { in goya_enable_msix()
|