Lines Matching refs:nvme_rdma_qe
50 struct nvme_rdma_qe { struct
65 struct nvme_rdma_qe sqe; argument
86 struct nvme_rdma_qe *rsp_ring;
111 struct nvme_rdma_qe async_event_sqe;
176 static void nvme_rdma_free_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, in nvme_rdma_free_qe()
183 static int nvme_rdma_alloc_qe(struct ib_device *ibdev, struct nvme_rdma_qe *qe, in nvme_rdma_alloc_qe()
201 struct nvme_rdma_qe *ring, size_t ib_queue_size, in nvme_rdma_free_ring()
211 static struct nvme_rdma_qe *nvme_rdma_alloc_ring(struct ib_device *ibdev, in nvme_rdma_alloc_ring()
215 struct nvme_rdma_qe *ring; in nvme_rdma_alloc_ring()
218 ring = kcalloc(ib_queue_size, sizeof(struct nvme_rdma_qe), GFP_KERNEL); in nvme_rdma_alloc_ring()
1572 struct nvme_rdma_qe *qe = in nvme_rdma_send_done()
1573 container_of(wc->wr_cqe, struct nvme_rdma_qe, cqe); in nvme_rdma_send_done()
1584 struct nvme_rdma_qe *qe, struct ib_sge *sge, u32 num_sge, in nvme_rdma_post_send()
1615 struct nvme_rdma_qe *qe) in nvme_rdma_post_recv()
1660 struct nvme_rdma_qe *sqe = &ctrl->async_event_sqe; in nvme_rdma_submit_async_event()
1728 struct nvme_rdma_qe *qe = in nvme_rdma_recv_done()
1729 container_of(wc->wr_cqe, struct nvme_rdma_qe, cqe); in nvme_rdma_recv_done()
1985 struct nvme_rdma_qe *sqe = &req->sqe; in nvme_rdma_queue_rq()