Home
last modified time | relevance | path

Searched refs:cqe (Results 101 – 125 of 252) sorted by relevance

1234567891011

/openbmc/linux/drivers/infiniband/sw/rxe/
H A Drxe_resp.c1067 struct rxe_cqe cqe; in do_complete() local
1068 struct ib_wc *wc = &cqe.ibwc; in do_complete()
1069 struct ib_uverbs_wc *uwc = &cqe.uibwc; in do_complete()
1077 memset(&cqe, 0, sizeof(cqe)); in do_complete()
1159 if (rxe_cq_post(qp->rcq, &cqe, pkt ? bth_se(pkt) : 1)) in do_complete()
1441 struct rxe_cqe cqe = {}; in flush_recv_wqe() local
1442 struct ib_wc *wc = &cqe.ibwc; in flush_recv_wqe()
1443 struct ib_uverbs_wc *uwc = &cqe.uibwc; in flush_recv_wqe()
1456 err = rxe_cq_post(qp->rcq, &cqe, 0); in flush_recv_wqe()
H A Drxe_verbs.c1084 err = rxe_cq_chk_attr(rxe, NULL, attr->cqe, attr->comp_vector); in rxe_create_cq()
1096 err = rxe_cq_from_init(rxe, cq, attr->cqe, attr->comp_vector, udata, in rxe_create_cq()
1114 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument
1130 err = rxe_cq_chk_attr(rxe, cq, cqe, 0); in rxe_resize_cq()
1136 err = rxe_cq_resize_queue(cq, cqe, uresp, udata); in rxe_resize_cq()
1153 struct rxe_cqe *cqe; in rxe_poll_cq() local
1158 cqe = queue_head(cq->queue, QUEUE_TYPE_TO_ULP); in rxe_poll_cq()
1159 if (!cqe) in rxe_poll_cq()
1162 memcpy(wc++, &cqe->ibwc, sizeof(*wc)); in rxe_poll_cq()
/openbmc/linux/drivers/infiniband/hw/mlx5/
H A Dumr.h80 struct ib_cqe cqe; member
H A Dumr.c227 static int mlx5r_umr_post_send(struct ib_qp *ibqp, u32 mkey, struct ib_cqe *cqe, in mlx5r_umr_post_send() argument
261 id.ib_cqe = cqe; in mlx5r_umr_post_send()
276 container_of(wc->wr_cqe, struct mlx5_ib_umr_context, cqe); in mlx5r_umr_done()
284 context->cqe.done = mlx5r_umr_done; in mlx5r_umr_init_context()
316 err = mlx5r_umr_post_send(umrc->qp, mkey, &umr_context.cqe, wqe, in mlx5r_umr_post_send_wait()
/openbmc/linux/drivers/ufs/core/
H A Dufshcd-priv.h65 struct cq_entry *cqe);
405 struct cq_entry *cqe = q->cqe_base_addr; in ufshcd_mcq_cur_cqe() local
407 return cqe + q->cq_head_slot; in ufshcd_mcq_cur_cqe()
/openbmc/qemu/hw/ufs/
H A Dufs.c445 req->cqe.utp_addr = in ufs_mcq_process_cq()
448 req->cqe.utp_addr |= req->sq->sqid; in ufs_mcq_process_cq()
449 req->cqe.resp_len = req->utrd.response_upiu_length; in ufs_mcq_process_cq()
450 req->cqe.resp_off = req->utrd.response_upiu_offset; in ufs_mcq_process_cq()
451 req->cqe.prdt_len = req->utrd.prd_table_length; in ufs_mcq_process_cq()
452 req->cqe.prdt_off = req->utrd.prd_table_offset; in ufs_mcq_process_cq()
453 req->cqe.status = req->utrd.header.dword_2 & 0xf; in ufs_mcq_process_cq()
454 req->cqe.error = 0; in ufs_mcq_process_cq()
456 ret = ufs_addr_write(u, cq->addr + tail, &req->cqe, sizeof(req->cqe)); in ufs_mcq_process_cq()
462 tail = (tail + sizeof(req->cqe)) % (cq->size * sizeof(req->cqe)); in ufs_mcq_process_cq()
/openbmc/linux/drivers/scsi/qedi/
H A Dqedi.h178 struct iscsi_cqe_solicited cqe; member
373 union iscsi_cqe cqe; member
H A Dqedi_main.c1205 static int qedi_queue_cqe(struct qedi_ctx *qedi, union iscsi_cqe *cqe, in qedi_queue_cqe() argument
1214 iscsi_cid = cqe->cqe_common.conn_id; in qedi_queue_cqe()
1223 switch (cqe->cqe_common.cqe_type) { in qedi_queue_cqe()
1226 qedi_cmd = qedi_get_cmd_from_tid(qedi, cqe->cqe_solicited.itid); in qedi_queue_cqe()
1233 memcpy(&qedi_cmd->cqe_work.cqe, cqe, sizeof(union iscsi_cqe)); in qedi_queue_cqe()
1248 memcpy(&qedi_work->cqe, cqe, sizeof(union iscsi_cqe)); in qedi_queue_cqe()
1269 union iscsi_cqe *cqe; in qedi_process_completions() local
1293 cqe = &que->cq[que->cq_cons_idx]; in qedi_process_completions()
1297 cqe, prod_idx, que->cq_cons_idx); in qedi_process_completions()
1299 ret = qedi_queue_cqe(qedi, cqe, fp->sb_id, p); in qedi_process_completions()
[all …]
/openbmc/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_spq.c453 struct eth_slow_path_rx_cqe *cqe, in qed_cqe_completion() argument
463 return qed_spq_completion(p_hwfn, cqe->echo, 0, NULL); in qed_cqe_completion()
467 struct eth_slow_path_rx_cqe *cqe) in qed_eth_cqe_completion() argument
471 rc = qed_cqe_completion(p_hwfn, cqe, PROTOCOLID_ETH); in qed_eth_cqe_completion()
475 cqe->ramrod_cmd_id); in qed_eth_cqe_completion()
/openbmc/linux/include/linux/
H A Dnvme-tcp.h149 struct nvme_completion cqe; member
H A Dio_uring_types.h551 struct io_cqe cqe; member
602 struct io_uring_cqe cqe; member
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/rep/
H A Dtc.c686 void mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq, in mlx5e_rep_tc_receive() argument
698 reg_c0 = (be32_to_cpu(cqe->sop_drop_qpn) & MLX5E_TC_FLOW_ID_MASK); in mlx5e_rep_tc_receive()
710 reg_c1 = be32_to_cpu(cqe->ft_metadata); in mlx5e_rep_tc_receive()
730 if (!mlx5e_tc_update_skb(cqe, skb, mapping_ctx, reg_c0, ct_priv, in mlx5e_rep_tc_receive()
/openbmc/linux/drivers/scsi/qedf/
H A Dqedf_main.c2236 struct fcoe_cqe *cqe; in qedf_process_completions() local
2260 cqe = &que->cq[que->cq_cons_idx]; in qedf_process_completions()
2280 xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK; in qedf_process_completions()
2308 memcpy(&io_work->cqe, cqe, sizeof(struct fcoe_cqe)); in qedf_process_completions()
2707 comp_type = (io_work->cqe.cqe_data >> in qedf_fp_io_handler()
2824 xid = cqe->cqe_data & FCOE_CQE_TASK_ID_MASK; in qedf_process_cqe()
2862 qedf_scsi_completion(qedf, cqe, io_req); in qedf_process_cqe()
2865 qedf_process_els_compl(qedf, cqe, io_req); in qedf_process_cqe()
2868 qedf_process_tmf_compl(qedf, cqe, io_req); in qedf_process_cqe()
2879 qedf_process_error_detect(qedf, cqe, io_req); in qedf_process_cqe()
[all …]
/openbmc/linux/drivers/infiniband/ulp/ipoib/
H A Dipoib_verbs.c176 cq_attr.cqe = size; in ipoib_transport_dev_init()
185 cq_attr.cqe = ipoib_sendq_size; in ipoib_transport_dev_init()
/openbmc/linux/include/uapi/rdma/
H A Dib_user_verbs.h419 __u32 cqe; member
433 __u32 cqe; member
443 __u32 cqe; member
456 __u32 cqe; member
461 __u32 cqe; member
/openbmc/linux/io_uring/
H A Dnotif.c18 notif->cqe.res |= IORING_NOTIF_USAGE_ZC_COPIED; in io_notif_complete_tw_ext()
/openbmc/linux/drivers/infiniband/hw/ocrdma/
H A Docrdma_verbs.h74 int ocrdma_resize_cq(struct ib_cq *, int cqe, struct ib_udata *);
/openbmc/linux/drivers/infiniband/hw/bnxt_re/
H A Dbnxt_re.h97 struct bnxt_qplib_cqe cqe; member
/openbmc/linux/drivers/scsi/elx/efct/
H A Defct_hw_queues.c522 u8 *cqe) in efct_hw_rqpair_process_rq() argument
533 rq_status = sli_fc_rqe_rqid_and_index(&hw->sli, cqe, in efct_hw_rqpair_process_rq()
590 sli_fc_rqe_length(&hw->sli, cqe, &h_len, &p_len); in efct_hw_rqpair_process_rq()
593 seq->fcfi = sli_fc_rqe_fcfi(&hw->sli, cqe); in efct_hw_rqpair_process_rq()
/openbmc/linux/drivers/infiniband/core/
H A Drw.c498 u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_wrs() argument
533 last_wr->wr_cqe = cqe; in rdma_rw_ctx_wrs()
556 struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_post() argument
560 first_wr = rdma_rw_ctx_wrs(ctx, qp, port_num, cqe, chain_wr); in rdma_rw_ctx_post()
H A Dmad_priv.h67 struct ib_cqe cqe; member
/openbmc/linux/drivers/net/ethernet/fungible/funcore/
H A Dfun_dev.h37 typedef void (*fun_admin_event_cb)(struct fun_dev *fdev, void *cqe);
/openbmc/linux/drivers/scsi/elx/libefc_sli/
H A Dsli4.h3833 struct sli4_fc_async_rcqe *rcqe = cqe; in sli_fc_rqe_length()
3847 sli_fc_rqe_fcfi(struct sli4 *sli4, void *cqe) in sli_fc_rqe_fcfi() argument
3849 u8 code = ((u8 *)cqe)[SLI4_CQE_CODE_OFFSET]; in sli_fc_rqe_fcfi()
3854 struct sli4_fc_async_rcqe *rcqe = cqe; in sli_fc_rqe_fcfi()
3860 struct sli4_fc_async_rcqe_v1 *rcqev1 = cqe; in sli_fc_rqe_fcfi()
3866 struct sli4_fc_optimized_write_cmd_cqe *opt_wr = cqe; in sli_fc_rqe_fcfi()
4038 u8 *cqe, enum sli4_qentry *etype,
4040 u32 sli_fc_response_length(struct sli4 *sli4, u8 *cqe);
4041 u32 sli_fc_io_length(struct sli4 *sli4, u8 *cqe);
4042 int sli_fc_els_did(struct sli4 *sli4, u8 *cqe, u32 *d_id);
[all …]
/openbmc/linux/drivers/infiniband/hw/qedr/
H A Dverbs.c920 int entries = attr->cqe; in qedr_create_cq()
975 cq->ibcq.cqe = chain_entries; in qedr_create_cq()
4413 union rdma_cqe *cqe; in qedr_poll_cq() local
4430 cqe = cq->latest_cqe; in qedr_poll_cq()
4439 qp = cqe_get_qp(cqe); in qedr_poll_cq()
4447 switch (cqe_get_type(cqe)) { in qedr_poll_cq()
4450 &cqe->req); in qedr_poll_cq()
4455 &cqe->resp, &update); in qedr_poll_cq()
4459 wc, &cqe->resp); in qedr_poll_cq()
4465 cqe_get_type(cqe)); in qedr_poll_cq()
[all …]
/openbmc/linux/tools/testing/selftests/x86/
H A Dlam.c471 struct io_uring_cqe *cqe; in handle_uring_cq() local
482 cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask]; in handle_uring_cq()
483 fi = (struct file_io *)cqe->user_data; in handle_uring_cq()
484 if (cqe->res < 0) in handle_uring_cq()

1234567891011