Lines Matching refs:sqe

65 	struct nvme_rdma_qe	sqe;  member
290 kfree(req->sqe.data); in nvme_rdma_exit_request()
303 req->sqe.data = kzalloc(sizeof(struct nvme_command), GFP_KERNEL); in nvme_rdma_init_request()
304 if (!req->sqe.data) in nvme_rdma_init_request()
314 nvme_req(rq)->cmd = req->sqe.data; in nvme_rdma_init_request()
1575 container_of(qe, struct nvme_rdma_request, sqe); in nvme_rdma_send_done()
1660 struct nvme_rdma_qe *sqe = &ctrl->async_event_sqe; in nvme_rdma_submit_async_event() local
1661 struct nvme_command *cmd = sqe->data; in nvme_rdma_submit_async_event()
1665 ib_dma_sync_single_for_cpu(dev, sqe->dma, sizeof(*cmd), DMA_TO_DEVICE); in nvme_rdma_submit_async_event()
1673 sqe->cqe.done = nvme_rdma_async_done; in nvme_rdma_submit_async_event()
1675 ib_dma_sync_single_for_device(dev, sqe->dma, sizeof(*cmd), in nvme_rdma_submit_async_event()
1678 ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL); in nvme_rdma_submit_async_event()
1985 struct nvme_rdma_qe *sqe = &req->sqe; in nvme_rdma_queue_rq() local
1999 req->sqe.dma = ib_dma_map_single(dev, req->sqe.data, in nvme_rdma_queue_rq()
2002 err = ib_dma_mapping_error(dev, req->sqe.dma); in nvme_rdma_queue_rq()
2006 ib_dma_sync_single_for_cpu(dev, sqe->dma, in nvme_rdma_queue_rq()
2031 sqe->cqe.done = nvme_rdma_send_done; in nvme_rdma_queue_rq()
2033 ib_dma_sync_single_for_device(dev, sqe->dma, in nvme_rdma_queue_rq()
2036 err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge, in nvme_rdma_queue_rq()
2054 ib_dma_unmap_single(dev, req->sqe.dma, sizeof(struct nvme_command), in nvme_rdma_queue_rq()
2107 ib_dma_unmap_single(ibdev, req->sqe.dma, sizeof(struct nvme_command), in nvme_rdma_complete_rq()