/openbmc/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_resp.c | 1067 struct rxe_cqe cqe; in do_complete() local 1068 struct ib_wc *wc = &cqe.ibwc; in do_complete() 1069 struct ib_uverbs_wc *uwc = &cqe.uibwc; in do_complete() 1077 memset(&cqe, 0, sizeof(cqe)); in do_complete() 1159 if (rxe_cq_post(qp->rcq, &cqe, pkt ? bth_se(pkt) : 1)) in do_complete() 1441 struct rxe_cqe cqe = {}; in flush_recv_wqe() local 1442 struct ib_wc *wc = &cqe.ibwc; in flush_recv_wqe() 1443 struct ib_uverbs_wc *uwc = &cqe.uibwc; in flush_recv_wqe() 1456 err = rxe_cq_post(qp->rcq, &cqe, 0); in flush_recv_wqe()
|
H A D | rxe_verbs.c | 1084 err = rxe_cq_chk_attr(rxe, NULL, attr->cqe, attr->comp_vector); in rxe_create_cq() 1096 err = rxe_cq_from_init(rxe, cq, attr->cqe, attr->comp_vector, udata, in rxe_create_cq() 1114 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument 1130 err = rxe_cq_chk_attr(rxe, cq, cqe, 0); in rxe_resize_cq() 1136 err = rxe_cq_resize_queue(cq, cqe, uresp, udata); in rxe_resize_cq() 1153 struct rxe_cqe *cqe; in rxe_poll_cq() local 1158 cqe = queue_head(cq->queue, QUEUE_TYPE_TO_ULP); in rxe_poll_cq() 1159 if (!cqe) in rxe_poll_cq() 1162 memcpy(wc++, &cqe->ibwc, sizeof(*wc)); in rxe_poll_cq()
|
/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | umr.h | 80 struct ib_cqe cqe; member
|
H A D | umr.c | 227 static int mlx5r_umr_post_send(struct ib_qp *ibqp, u32 mkey, struct ib_cqe *cqe, in mlx5r_umr_post_send() argument 261 id.ib_cqe = cqe; in mlx5r_umr_post_send() 276 container_of(wc->wr_cqe, struct mlx5_ib_umr_context, cqe); in mlx5r_umr_done() 284 context->cqe.done = mlx5r_umr_done; in mlx5r_umr_init_context() 316 err = mlx5r_umr_post_send(umrc->qp, mkey, &umr_context.cqe, wqe, in mlx5r_umr_post_send_wait()
|
/openbmc/linux/drivers/ufs/core/ |
H A D | ufshcd-priv.h | 65 struct cq_entry *cqe); 405 struct cq_entry *cqe = q->cqe_base_addr; in ufshcd_mcq_cur_cqe() local 407 return cqe + q->cq_head_slot; in ufshcd_mcq_cur_cqe()
|
/openbmc/qemu/hw/ufs/ |
H A D | ufs.c | 445 req->cqe.utp_addr = in ufs_mcq_process_cq() 448 req->cqe.utp_addr |= req->sq->sqid; in ufs_mcq_process_cq() 449 req->cqe.resp_len = req->utrd.response_upiu_length; in ufs_mcq_process_cq() 450 req->cqe.resp_off = req->utrd.response_upiu_offset; in ufs_mcq_process_cq() 451 req->cqe.prdt_len = req->utrd.prd_table_length; in ufs_mcq_process_cq() 452 req->cqe.prdt_off = req->utrd.prd_table_offset; in ufs_mcq_process_cq() 453 req->cqe.status = req->utrd.header.dword_2 & 0xf; in ufs_mcq_process_cq() 454 req->cqe.error = 0; in ufs_mcq_process_cq() 456 ret = ufs_addr_write(u, cq->addr + tail, &req->cqe, sizeof(req->cqe)); in ufs_mcq_process_cq() 462 tail = (tail + sizeof(req->cqe)) % (cq->size * sizeof(req->cqe)); in ufs_mcq_process_cq()
|
/openbmc/linux/drivers/scsi/qedi/ |
H A D | qedi.h | 178 struct iscsi_cqe_solicited cqe; member 373 union iscsi_cqe cqe; member
|
H A D | qedi_main.c | 1205 static int qedi_queue_cqe(struct qedi_ctx *qedi, union iscsi_cqe *cqe, in qedi_queue_cqe() argument 1214 iscsi_cid = cqe->cqe_common.conn_id; in qedi_queue_cqe() 1223 switch (cqe->cqe_common.cqe_type) { in qedi_queue_cqe() 1226 qedi_cmd = qedi_get_cmd_from_tid(qedi, cqe->cqe_solicited.itid); in qedi_queue_cqe() 1233 memcpy(&qedi_cmd->cqe_work.cqe, cqe, sizeof(union iscsi_cqe)); in qedi_queue_cqe() 1248 memcpy(&qedi_work->cqe, cqe, sizeof(union iscsi_cqe)); in qedi_queue_cqe() 1269 union iscsi_cqe *cqe; in qedi_process_completions() local 1293 cqe = &que->cq[que->cq_cons_idx]; in qedi_process_completions() 1297 cqe, prod_idx, que->cq_cons_idx); in qedi_process_completions() 1299 ret = qedi_queue_cqe(qedi, cqe, fp->sb_id, p); in qedi_process_completions() [all …]
|
/openbmc/linux/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_spq.c | 453 struct eth_slow_path_rx_cqe *cqe, in qed_cqe_completion() argument 463 return qed_spq_completion(p_hwfn, cqe->echo, 0, NULL); in qed_cqe_completion() 467 struct eth_slow_path_rx_cqe *cqe) in qed_eth_cqe_completion() argument 471 rc = qed_cqe_completion(p_hwfn, cqe, PROTOCOLID_ETH); in qed_eth_cqe_completion() 475 cqe->ramrod_cmd_id); in qed_eth_cqe_completion()
|
/openbmc/linux/include/linux/ |
H A D | nvme-tcp.h | 149 struct nvme_completion cqe; member
|
H A D | io_uring_types.h | 551 struct io_cqe cqe; member 602 struct io_uring_cqe cqe; member
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/rep/ |
H A D | tc.c | 686 void mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq, in mlx5e_rep_tc_receive() argument 698 reg_c0 = (be32_to_cpu(cqe->sop_drop_qpn) & MLX5E_TC_FLOW_ID_MASK); in mlx5e_rep_tc_receive() 710 reg_c1 = be32_to_cpu(cqe->ft_metadata); in mlx5e_rep_tc_receive() 730 if (!mlx5e_tc_update_skb(cqe, skb, mapping_ctx, reg_c0, ct_priv, in mlx5e_rep_tc_receive()
|
/openbmc/linux/drivers/scsi/qedf/ |
H A D | qedf_main.c | 2236 struct fcoe_cqe *cqe; in qedf_process_completions() local 2260 cqe = &que->cq[que->cq_cons_idx]; in qedf_process_completions() 2280 xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK; in qedf_process_completions() 2308 memcpy(&io_work->cqe, cqe, sizeof(struct fcoe_cqe)); in qedf_process_completions() 2707 comp_type = (io_work->cqe.cqe_data >> in qedf_fp_io_handler() 2824 xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK; in qedf_process_cqe() 2862 qedf_scsi_completion(qedf, cqe, io_req); in qedf_process_cqe() 2865 qedf_process_els_compl(qedf, cqe, io_req); in qedf_process_cqe() 2868 qedf_process_tmf_compl(qedf, cqe, io_req); in qedf_process_cqe() 2879 qedf_process_error_detect(qedf, cqe, io_req); in qedf_process_cqe() [all …]
|
/openbmc/linux/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_verbs.c | 176 cq_attr.cqe = size; in ipoib_transport_dev_init() 185 cq_attr.cqe = ipoib_sendq_size; in ipoib_transport_dev_init()
|
/openbmc/linux/include/uapi/rdma/ |
H A D | ib_user_verbs.h | 419 __u32 cqe; member 433 __u32 cqe; member 443 __u32 cqe; member 456 __u32 cqe; member 461 __u32 cqe; member
|
/openbmc/linux/io_uring/ |
H A D | notif.c | 18 notif->cqe.res |= IORING_NOTIF_USAGE_ZC_COPIED; in io_notif_complete_tw_ext()
|
/openbmc/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.h | 74 int ocrdma_resize_cq(struct ib_cq *, int cqe, struct ib_udata *);
|
/openbmc/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | bnxt_re.h | 97 struct bnxt_qplib_cqe cqe; member
|
/openbmc/linux/drivers/scsi/elx/efct/ |
H A D | efct_hw_queues.c | 522 u8 *cqe) in efct_hw_rqpair_process_rq() argument 533 rq_status = sli_fc_rqe_rqid_and_index(&hw->sli, cqe, in efct_hw_rqpair_process_rq() 590 sli_fc_rqe_length(&hw->sli, cqe, &h_len, &p_len); in efct_hw_rqpair_process_rq() 593 seq->fcfi = sli_fc_rqe_fcfi(&hw->sli, cqe); in efct_hw_rqpair_process_rq()
|
/openbmc/linux/drivers/infiniband/core/ |
H A D | rw.c | 498 u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_wrs() argument 533 last_wr->wr_cqe = cqe; in rdma_rw_ctx_wrs() 556 struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_post() argument 560 first_wr = rdma_rw_ctx_wrs(ctx, qp, port_num, cqe, chain_wr); in rdma_rw_ctx_post()
|
H A D | mad_priv.h | 67 struct ib_cqe cqe; member
|
/openbmc/linux/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_dev.h | 37 typedef void (*fun_admin_event_cb)(struct fun_dev *fdev, void *cqe);
|
/openbmc/linux/drivers/scsi/elx/libefc_sli/ |
H A D | sli4.h | 3833 struct sli4_fc_async_rcqe *rcqe = cqe; in sli_fc_rqe_length() 3847 sli_fc_rqe_fcfi(struct sli4 *sli4, void *cqe) in sli_fc_rqe_fcfi() argument 3849 u8 code = ((u8 *)cqe)[SLI4_CQE_CODE_OFFSET]; in sli_fc_rqe_fcfi() 3854 struct sli4_fc_async_rcqe *rcqe = cqe; in sli_fc_rqe_fcfi() 3860 struct sli4_fc_async_rcqe_v1 *rcqev1 = cqe; in sli_fc_rqe_fcfi() 3866 struct sli4_fc_optimized_write_cmd_cqe *opt_wr = cqe; in sli_fc_rqe_fcfi() 4038 u8 *cqe, enum sli4_qentry *etype, 4040 u32 sli_fc_response_length(struct sli4 *sli4, u8 *cqe); 4041 u32 sli_fc_io_length(struct sli4 *sli4, u8 *cqe); 4042 int sli_fc_els_did(struct sli4 *sli4, u8 *cqe, u32 *d_id); [all …]
|
/openbmc/linux/drivers/infiniband/hw/qedr/ |
H A D | verbs.c | 920 int entries = attr->cqe; in qedr_create_cq() 975 cq->ibcq.cqe = chain_entries; in qedr_create_cq() 4413 union rdma_cqe *cqe; in qedr_poll_cq() local 4430 cqe = cq->latest_cqe; in qedr_poll_cq() 4439 qp = cqe_get_qp(cqe); in qedr_poll_cq() 4447 switch (cqe_get_type(cqe)) { in qedr_poll_cq() 4450 &cqe->req); in qedr_poll_cq() 4455 &cqe->resp, &update); in qedr_poll_cq() 4459 wc, &cqe->resp); in qedr_poll_cq() 4465 cqe_get_type(cqe)); in qedr_poll_cq() [all …]
|
/openbmc/linux/tools/testing/selftests/x86/ |
H A D | lam.c | 471 struct io_uring_cqe *cqe; in handle_uring_cq() local 482 cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask]; in handle_uring_cq() 483 fi = (struct file_io *)cqe->user_data; in handle_uring_cq() 484 if (cqe->res < 0) in handle_uring_cq()
|