/openbmc/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_qp.c | 541 const struct ib_recv_wr *recv_wr) in erdma_post_recv_one() argument 550 if (recv_wr->num_sge == 0) { in erdma_post_recv_one() 552 } else if (recv_wr->num_sge == 1) { in erdma_post_recv_one() 553 rqe->stag = cpu_to_le32(recv_wr->sg_list[0].lkey); in erdma_post_recv_one() 554 rqe->to = cpu_to_le64(recv_wr->sg_list[0].addr); in erdma_post_recv_one() 555 rqe->length = cpu_to_le32(recv_wr->sg_list[0].length); in erdma_post_recv_one() 564 recv_wr->wr_id; in erdma_post_recv_one() 570 int erdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *recv_wr, in erdma_post_recv() argument 573 const struct ib_recv_wr *wr = recv_wr; in erdma_post_recv()
|
H A D | erdma_verbs.h | 355 int erdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *recv_wr,
|
/openbmc/linux/drivers/infiniband/core/ |
H A D | cma_trace.h | 187 __field(u32, recv_wr) 199 __entry->recv_wr = qp_init_attr->cap.max_recv_wr; 219 __entry->recv_wr, __entry->qp_num, __entry->rc
|
H A D | mad.c | 2677 struct ib_recv_wr recv_wr; in ib_mad_post_receive_mads() local 2684 recv_wr.next = NULL; in ib_mad_post_receive_mads() 2685 recv_wr.sg_list = &sg_list; in ib_mad_post_receive_mads() 2686 recv_wr.num_sge = 1; in ib_mad_post_receive_mads() 2715 recv_wr.wr_cqe = &mad_priv->header.mad_list.cqe; in ib_mad_post_receive_mads() 2722 ret = ib_post_recv(qp_info->qp, &recv_wr, NULL); in ib_mad_post_receive_mads()
|
/openbmc/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | ib_verbs.h | 207 int bnxt_re_post_srq_recv(struct ib_srq *srq, const struct ib_recv_wr *recv_wr, 218 int bnxt_re_post_recv(struct ib_qp *qp, const struct ib_recv_wr *recv_wr,
|
/openbmc/linux/fs/smb/client/ |
H A D | smbdirect.c | 1010 struct ib_recv_wr recv_wr; in smbd_post_recv() local 1024 recv_wr.wr_cqe = &response->cqe; in smbd_post_recv() 1025 recv_wr.next = NULL; in smbd_post_recv() 1026 recv_wr.sg_list = &response->sge; in smbd_post_recv() 1027 recv_wr.num_sge = 1; in smbd_post_recv() 1029 rc = ib_post_recv(info->id->qp, &recv_wr, NULL); in smbd_post_recv()
|
/openbmc/linux/drivers/infiniband/hw/mlx4/ |
H A D | mad.c | 1319 struct ib_recv_wr recv_wr; in mlx4_ib_post_pv_qp_buf() local 1330 recv_wr.next = NULL; in mlx4_ib_post_pv_qp_buf() 1331 recv_wr.sg_list = &sg_list; in mlx4_ib_post_pv_qp_buf() 1332 recv_wr.num_sge = 1; in mlx4_ib_post_pv_qp_buf() 1333 recv_wr.wr_id = (u64) index | MLX4_TUN_WRID_RECV | in mlx4_ib_post_pv_qp_buf() 1337 return ib_post_recv(tun_qp->qp, &recv_wr, &bad_recv_wr); in mlx4_ib_post_pv_qp_buf()
|
/openbmc/linux/include/rdma/ |
H A D | ib_verbs.h | 2352 int (*post_recv)(struct ib_qp *qp, const struct ib_recv_wr *recv_wr, 2360 const struct ib_recv_wr *recv_wr, 3718 * @recv_wr: A list of work requests to post on the receive queue. 3723 const struct ib_recv_wr *recv_wr, in ib_post_srq_recv() 3728 return srq->device->ops.post_srq_recv(srq, recv_wr, in ib_post_srq_recv() 3857 * @recv_wr: A list of work requests to post on the receive queue. 3862 const struct ib_recv_wr *recv_wr, in ib_post_recv() 3867 return qp->device->ops.post_recv(qp, recv_wr, bad_recv_wr ? : &dummy); 3721 ib_post_srq_recv(struct ib_srq * srq,const struct ib_recv_wr * recv_wr,const struct ib_recv_wr ** bad_recv_wr) ib_post_srq_recv() argument 3860 ib_post_recv(struct ib_qp * qp,const struct ib_recv_wr * recv_wr,const struct ib_recv_wr ** bad_recv_wr) ib_post_recv() argument
|
/openbmc/linux/include/uapi/rdma/ |
H A D | ib_user_verbs.h | 862 struct ib_uverbs_recv_wr recv_wr[]; member
|
/openbmc/qemu/migration/ |
H A D | rdma.c | 1767 struct ibv_recv_wr recv_wr = { in qemu_rdma_post_recv_control() local 1774 if (ibv_post_recv(rdma->qp, &recv_wr, &bad_wr)) { in qemu_rdma_post_recv_control()
|