/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 687 buf->cqe_size = cqe_size; in alloc_cq_frag_buf() 742 if ((ucmd.cqe_size != 64 && ucmd.cqe_size != 128) || in create_cq_user() 746 *cqe_size = ucmd.cqe_size; in create_cq_user() 809 *cqe_size); in create_cq_user() 833 *cqe_size); in create_cq_user() 957 int cqe_size; in mlx5_ib_create_cq() local 1000 cq->cqe_size = cqe_size; in mlx5_ib_create_cq() 1168 if (ucmd.cqe_size && SIZE_MAX / ucmd.cqe_size <= entries - 1) in resize_user() 1180 *cqe_size = ucmd.cqe_size; in resize_user() 1274 int cqe_size; in mlx5_ib_resize_cq() local [all …]
|
H A D | mlx5_ib.h | 564 int cqe_size; member 588 int cqe_size; member
|
/openbmc/linux/drivers/infiniband/hw/hns/ |
H A D | hns_roce_cq.c | 207 buf_attr.region[0].size = hr_cq->cq_depth * hr_cq->cqe_size; in alloc_cq_buf() 336 hr_cq->cqe_size = hr_dev->caps.cqe_sz; in set_cqe_size() 340 if (udata->inlen >= offsetofend(typeof(*ucmd), cqe_size)) { in set_cqe_size() 341 if (ucmd->cqe_size != HNS_ROCE_V2_CQE_SIZE && in set_cqe_size() 342 ucmd->cqe_size != HNS_ROCE_V3_CQE_SIZE) { in set_cqe_size() 344 "invalid cqe size %u.\n", ucmd->cqe_size); in set_cqe_size() 348 hr_cq->cqe_size = ucmd->cqe_size; in set_cqe_size() 350 hr_cq->cqe_size = HNS_ROCE_V2_CQE_SIZE; in set_cqe_size()
|
H A D | hns_roce_restrack.c | 27 if (rdma_nl_put_driver_u32(msg, "cqe_size", hr_cq->cqe_size)) in hns_roce_fill_res_cq_entry()
|
/openbmc/linux/include/uapi/rdma/ |
H A D | hns-abi.h | 42 __u32 cqe_size; member 102 __u32 cqe_size; member
|
H A D | mlx4-abi.h | 71 __u32 cqe_size; member
|
H A D | mlx5-abi.h | 289 __u32 cqe_size; member 305 __u16 cqe_size; member
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | cq.c | 290 static int mlx4_init_user_cqes(void *buf, int entries, int cqe_size) in mlx4_init_user_cqes() argument 292 int entries_per_copy = PAGE_SIZE / cqe_size; in mlx4_init_user_cqes() 318 array_size(entries, cqe_size)) ? in mlx4_init_user_cqes() 330 int cqe_size) in mlx4_init_kernel_cqes() argument 335 memset(buf->direct.buf, 0xcc, entries * cqe_size); in mlx4_init_kernel_cqes() 395 dev->caps.cqe_size); in mlx4_cq_alloc() 400 dev->caps.cqe_size); in mlx4_cq_alloc()
|
H A D | fw.h | 203 u16 cqe_size; /* For use only when CQE stride feature enabled */ member
|
H A D | en_cq.c | 62 cq->buf_size = cq->size * mdev->dev->caps.cqe_size; in mlx4_en_create_cq()
|
H A D | en_tx.c | 457 cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor; in mlx4_en_process_tx_cq() 510 cqe = mlx4_en_get_cqe(buf, index, priv->cqe_size) + factor; in mlx4_en_process_tx_cq()
|
H A D | en_rx.c | 741 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq() 960 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq()
|
/openbmc/linux/net/ethtool/ |
H A D | rings.c | 103 (kr->cqe_size && in rings_fill_reply() 104 (nla_put_u32(skb, ETHTOOL_A_RINGS_CQE_SIZE, kr->cqe_size))) || in rings_fill_reply() 205 ethnl_update_u32(&kernel_ringparam.cqe_size, in ethnl_set_rings()
|
/openbmc/linux/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 105 err = mlx4_buf_alloc(dev->dev, nent * dev->dev->caps.cqe_size, in mlx4_ib_alloc_cq_buf() 111 buf->entry_size = dev->dev->caps.cqe_size; in mlx4_ib_alloc_cq_buf() 143 int cqe_size = dev->dev->caps.cqe_size; in mlx4_ib_get_cq_umem() local 147 *umem = ib_umem_get(&dev->ib_dev, buf_addr, cqe * cqe_size, in mlx4_ib_get_cq_umem() 358 int cqe_size = cq->buf.entry_size; in mlx4_ib_cq_resize_copy_cqes() local 359 int cqe_inc = cqe_size == 64 ? 1 : 0; in mlx4_ib_cq_resize_copy_cqes() 368 memcpy(new_cqe, get_cqe(cq, i & cq->ibcq.cqe), cqe_size); in mlx4_ib_cq_resize_copy_cqes()
|
/openbmc/linux/drivers/vfio/pci/mlx5/ |
H A D | cmd.c | 856 int cqe_size) in alloc_cq_frag_buf() argument 859 u8 log_wq_stride = 6 + (cqe_size == 128 ? 1 : 0); in alloc_cq_frag_buf() 860 u8 log_wq_sz = ilog2(cqe_size); in alloc_cq_frag_buf() 863 err = mlx5_frag_buf_alloc_node(mdev, nent * cqe_size, frag_buf, in alloc_cq_frag_buf() 869 buf->cqe_size = cqe_size; in alloc_cq_frag_buf() 882 cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64; in init_cq_frag_buf() 950 int cqe_size = cache_line_size() == 128 ? 128 : 64; in mlx5vf_create_cq() local 967 cq->mcq.cqe_sz = cqe_size; in mlx5vf_create_cq() 968 err = alloc_cq_frag_buf(mdev, &cq->buf, ncqe, cqe_size); in mlx5vf_create_cq()
|
H A D | cmd.h | 112 int cqe_size; member
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.h | 132 u16 cqe_size; member
|
H A D | otx2_ethtool.c | 389 kernel_ring->cqe_size = pfvf->hw.xqe_size; in otx2_get_ringparam() 400 u32 xqe_size = kernel_ring->cqe_size; in otx2_set_ringparam()
|
/openbmc/linux/drivers/ufs/core/ |
H A D | ufs-mcq.c | 201 size_t utrdl_size, cqe_size; in ufshcd_mcq_memory_alloc() local 217 cqe_size = sizeof(struct cq_entry) * hwq->max_entries; in ufshcd_mcq_memory_alloc() 218 hwq->cqe_base_addr = dmam_alloc_coherent(hba->dev, cqe_size, in ufshcd_mcq_memory_alloc()
|
/openbmc/linux/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_qp.c | 320 size_t cqe_dma_size, cqe_size; in alloc_rq_cqe() local 324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe() 325 rq->cqe = vzalloc(cqe_size); in alloc_rq_cqe()
|
/openbmc/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_hw.c | 1786 u32 hw_pages, cqe_size, page_size, cqe_count; in ocrdma_mbx_create_cq() local 1799 cqe_size = OCRDMA_DPP_CQE_SIZE; in ocrdma_mbx_create_cq() 1804 cqe_size = sizeof(struct ocrdma_cqe); in ocrdma_mbx_create_cq() 1808 cq->len = roundup(max_hw_cqe * cqe_size, OCRDMA_MIN_Q_PAGE_SIZE); in ocrdma_mbx_create_cq() 1828 cqe_count = cq->len / cqe_size; in ocrdma_mbx_create_cq() 1857 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size); in ocrdma_mbx_create_cq() 1859 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size) - 1; in ocrdma_mbx_create_cq()
|
/openbmc/linux/drivers/scsi/bnx2i/ |
H A D | bnx2i.h | 661 u32 cqe_size; member
|
H A D | bnx2i_hwi.c | 171 if (cq_index > ep->qp.cqe_size * 2) in bnx2i_arm_cq_event_coalescing() 172 cq_index -= ep->qp.cqe_size * 2; in bnx2i_arm_cq_event_coalescing() 1123 ep->qp.cqe_size = hba->max_cqes; in bnx2i_alloc_qp_resc() 2063 if (qp->cqe_exp_seq_sn == (qp->cqe_size * 2 + 1)) in bnx2i_process_new_cqes()
|
/openbmc/linux/include/linux/ |
H A D | ethtool.h | 86 u32 cqe_size; member
|
/openbmc/linux/tools/net/ynl/generated/ |
H A D | ethtool-user.h | 1997 __u32 cqe_size:1; member 2015 __u32 cqe_size; member 2110 __u32 cqe_size:1; member 2128 __u32 cqe_size; member 2237 __u32 cqe_size) in ethtool_rings_set_req_set_cqe_size() argument 2239 req->_present.cqe_size = 1; in ethtool_rings_set_req_set_cqe_size() 2240 req->cqe_size = cqe_size; in ethtool_rings_set_req_set_cqe_size()
|