Home
last modified time | relevance | path

Searched refs:cqe (Results 1 – 25 of 259) sorted by relevance

1234567891011

/openbmc/linux/drivers/infiniband/hw/mthca/
H A Dmthca_cq.c176 return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe; in cqe_sw()
195 be32_to_cpu(cqe[0]), be32_to_cpu(cqe[1]), be32_to_cpu(cqe[2]), in dump_cqe()
196 be32_to_cpu(cqe[3]), be32_to_cpu(cqe[4]), be32_to_cpu(cqe[5]), in dump_cqe()
197 be32_to_cpu(cqe[6]), be32_to_cpu(cqe[7])); in dump_cqe()
304 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean()
336 cq->ibcq.cqe < cq->resize_buf->cqe) { in mthca_cq_resize_copy_cqes()
383 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe), in handle_error_cqe()
494 if (!cqe) in mthca_poll_one()
512 is_send = is_error ? cqe->opcode & 0x01 : cqe->is_send & 0x80; in mthca_poll_one()
646 set_cqe_hw(cqe); in mthca_poll_one()
[all …]
/openbmc/linux/drivers/infiniband/hw/mlx4/
H A Dcq.c81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local
85 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe()
362 cqe = get_cqe(cq, i & cq->ibcq.cqe); in mlx4_ib_cq_resize_copy_cqes()
373 cqe = get_cqe(cq, ++i & cq->ibcq.cqe); in mlx4_ib_cq_resize_copy_cqes()
430 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_resize_cq()
447 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_resize_cq()
674 if (!cqe) in mlx4_ib_poll_one()
678 cqe++; in mlx4_ib_poll_one()
699 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_poll_one()
939 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in __mlx4_ib_cq_clean()
[all …]
/openbmc/linux/drivers/infiniband/sw/siw/
H A Dsiw_cq.c50 struct siw_cqe *cqe; in siw_reap_cqe() local
55 cqe = &cq->queue[cq->cq_get % cq->num_cqe]; in siw_reap_cqe()
58 wc->wr_id = cqe->id; in siw_reap_cqe()
59 wc->byte_len = cqe->bytes; in siw_reap_cqe()
67 if (cqe->flags & SIW_WQE_REM_INVAL) { in siw_reap_cqe()
71 wc->qp = cqe->base_qp; in siw_reap_cqe()
72 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe()
77 cqe->flags, (void *)(uintptr_t)cqe->id); in siw_reap_cqe()
86 u8 opcode = cqe->opcode; in siw_reap_cqe()
87 u16 status = cqe->status; in siw_reap_cqe()
[all …]
/openbmc/linux/drivers/infiniband/sw/rxe/
H A Drxe_cq.c12 int cqe, int comp_vector) in rxe_cq_chk_attr() argument
16 if (cqe <= 0) { in rxe_cq_chk_attr()
17 rxe_dbg_dev(rxe, "cqe(%d) <= 0\n", cqe); in rxe_cq_chk_attr()
21 if (cqe > rxe->attr.max_cqe) { in rxe_cq_chk_attr()
23 cqe, rxe->attr.max_cqe); in rxe_cq_chk_attr()
29 if (cqe < count) { in rxe_cq_chk_attr()
31 cqe, count); in rxe_cq_chk_attr()
50 cq->queue = rxe_queue_init(rxe, &cqe, in rxe_cq_from_init()
68 cq->ibcq.cqe = cqe; in rxe_cq_from_init()
82 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue()
[all …]
/openbmc/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_cq.c105 int entries = attr->cqe; in pvrdma_create_cq()
132 cq->ibcq.cqe = entries; in pvrdma_create_cq()
186 cmd->cqe = entries; in pvrdma_create_cq()
195 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq()
293 cq->ibcq.cqe); in _pvrdma_flush_cqe()
309 *cqe = *curr_cqe; in _pvrdma_flush_cqe()
315 cq->ibcq.cqe); in _pvrdma_flush_cqe()
329 struct pvrdma_cqe *cqe; in pvrdma_poll_one() local
358 wc->wr_id = cqe->wr_id; in pvrdma_poll_one()
365 wc->slid = cqe->slid; in pvrdma_poll_one()
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci_hw.h116 return mlxsw_pci_cqe##v0##_##name##_get(cqe); \
118 return mlxsw_pci_cqe##v1##_##name##_get(cqe); \
120 return mlxsw_pci_cqe##v2##_##name##_get(cqe); \
124 char *cqe, u32 val) \
129 mlxsw_pci_cqe##v0##_##name##_set(cqe, val); \
132 mlxsw_pci_cqe##v1##_##name##_set(cqe, val); \
135 mlxsw_pci_cqe##v2##_##name##_set(cqe, val); \
153 MLXSW_ITEM32(pci, cqe, system_port, 0x00, 0, 16);
164 MLXSW_ITEM32(pci, cqe, wqe_counter, 0x04, 16, 16);
171 MLXSW_ITEM32(pci, cqe, byte_count, 0x04, 0, 14);
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_rx.c140 cqe->op_own = op_own; in mlx5e_cqes_update_owner()
148 cqe->op_own = op_own; in mlx5e_cqes_update_owner()
999 struct mlx5_cqe64 *cqe; in mlx5e_poll_ico_cq() local
1007 if (likely(!cqe)) in mlx5e_poll_ico_cq()
1667 mxbuf->cqe = cqe; in mlx5e_fill_mxbuf()
2299 u16 data_bcnt = mpwrq_get_cqe_byte_cnt(cqe) - cqe->shampo.header_size; in mlx5e_handle_rx_cqe_mpwrq_shampo()
2450 if (!cqe) in mlx5e_rx_cq_process_enhanced_cqe_comp()
2471 title_cqe = cqe; in mlx5e_rx_cq_process_enhanced_cqe_comp()
2476 rq, cqe); in mlx5e_rx_cq_process_enhanced_cqe_comp()
2494 struct mlx5_cqe64 *cqe; in mlx5e_rx_cq_process_basic_cqe_comp() local
[all …]
H A Dwq.h202 struct mlx5_cqe64 *cqe = mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_cqwq_get_wqe() local
205 cqe += wq->fbc.log_stride == 7; in mlx5_cqwq_get_wqe()
207 return cqe; in mlx5_cqwq_get_wqe()
233 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5_cqwq_get_cqe() local
234 u8 cqe_ownership_bit = cqe->op_own & MLX5_CQE_OWNER_MASK; in mlx5_cqwq_get_cqe()
243 return cqe; in mlx5_cqwq_get_cqe()
251 struct mlx5_cqe64 *cqe; in mlx5_cqwq_get_cqe_enahnced_comp() local
253 cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5_cqwq_get_cqe_enahnced_comp()
254 if (cqe->validity_iteration_count != sw_validity_iteration_count) in mlx5_cqwq_get_cqe_enahnced_comp()
260 return cqe; in mlx5_cqwq_get_cqe_enahnced_comp()
/openbmc/linux/drivers/infiniband/hw/cxgb4/
H A Dcq.c190 memset(&cqe, 0, sizeof(cqe)); in insert_recv_cqe()
224 memset(&cqe, 0, sizeof(cqe)); in insert_sq_cqe()
432 if ((CQE_OPCODE(cqe) == FW_RI_RDMA_WRITE) && RQ_TYPE(cqe)) in cqe_completes_wr()
435 if ((CQE_OPCODE(cqe) == FW_RI_READ_RESP) && SQ_TYPE(cqe)) in cqe_completes_wr()
438 if (CQE_SEND_OPCODE(cqe) && RQ_TYPE(cqe) && t4_rq_empty(wq)) in cqe_completes_wr()
783 CQE_TYPE(&cqe), CQE_OPCODE(&cqe), in __c4iw_poll_cq_one()
784 CQE_STATUS(&cqe), CQE_LEN(&cqe), in __c4iw_poll_cq_one()
785 CQE_WRID_HI(&cqe), CQE_WRID_LOW(&cqe), in __c4iw_poll_cq_one()
812 CQE_OPCODE(&cqe), CQE_QPID(&cqe)); in __c4iw_poll_cq_one()
849 CQE_OPCODE(&cqe), CQE_QPID(&cqe)); in __c4iw_poll_cq_one()
[all …]
/openbmc/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c81 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local
84 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe()
88 return cqe; in get_sw_cqe()
276 cqe, sizeof(*cqe), false); in dump_cqe()
459 void *cqe; in mlx5_poll_one() local
464 if (!cqe) in mlx5_poll_one()
467 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in mlx5_poll_one()
869 void *cqe; in init_cq_frag_buf() local
874 cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64; in init_cq_frag_buf()
1096 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in __mlx5_ib_cq_clean()
[all …]
/openbmc/qemu/hw/rdma/vmw/
H A Dpvrdma_qp_ops.c31 struct pvrdma_cqe cqe; member
72 cqe1->wr_id = cqe->wr_id; in pvrdma_post_cqe()
73 cqe1->qp = cqe->qp ? cqe->qp : wc->qp_num; in pvrdma_post_cqe()
74 cqe1->opcode = cqe->opcode; in pvrdma_post_cqe()
160 comp_ctx->cqe.wr_id = wqe->hdr.wr_id; in pvrdma_qp_send()
161 comp_ctx->cqe.qp = qp_handle; in pvrdma_qp_send()
162 comp_ctx->cqe.opcode = IBV_WC_SEND; in pvrdma_qp_send()
224 comp_ctx->cqe.qp = qp_handle; in pvrdma_qp_recv()
225 comp_ctx->cqe.opcode = IBV_WC_RECV; in pvrdma_qp_recv()
266 comp_ctx->cqe.qp = 0; in pvrdma_srq_recv()
[all …]
/openbmc/linux/drivers/infiniband/hw/erdma/
H A Derdma_cq.c11 __be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci, in get_next_valid_cqe() local
14 be32_to_cpu(READ_ONCE(*cqe))); in get_next_valid_cqe()
16 return owner ^ !!(cq->kern_cq.ci & cq->depth) ? cqe : NULL; in get_next_valid_cqe()
115 struct erdma_cqe *cqe; in erdma_poll_one_cqe() local
122 cqe = get_next_valid_cqe(cq); in erdma_poll_one_cqe()
123 if (!cqe) in erdma_poll_one_cqe()
131 qpn = be32_to_cpu(cqe->qpn); in erdma_poll_one_cqe()
132 wqe_idx = be32_to_cpu(cqe->qe_idx); in erdma_poll_one_cqe()
133 cqe_hdr = be32_to_cpu(cqe->hdr); in erdma_poll_one_cqe()
158 wc->byte_len = be32_to_cpu(cqe->size); in erdma_poll_one_cqe()
[all …]
/openbmc/linux/drivers/infiniband/sw/rdmavt/
H A Dcq.c55 head = cq->ibcq.cqe; in rvt_cq_enter()
249 cq->ibcq.cqe = entries; in rvt_create_cq()
350 if (cqe < 1 || cqe > rdi->dparms.props.max_cqe) in rvt_resize_cq()
393 if (head > (u32)cq->ibcq.cqe) in rvt_resize_cq()
394 head = (u32)cq->ibcq.cqe; in rvt_resize_cq()
395 if (tail > (u32)cq->ibcq.cqe) in rvt_resize_cq()
396 tail = (u32)cq->ibcq.cqe; in rvt_resize_cq()
401 if (unlikely((u32)cqe < n)) { in rvt_resize_cq()
415 cq->ibcq.cqe = cqe; in rvt_resize_cq()
493 tail = (u32)cq->ibcq.cqe; in rvt_poll_cq()
[all …]
/openbmc/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c2287 memset(cqe, 0, sizeof(*cqe)); in __flush_sq()
2294 cqe++; in __flush_sq()
2337 memset(cqe, 0, sizeof(*cqe)); in __flush_rq()
2343 cqe++; in __flush_rq()
2507 memset(cqe, 0, sizeof(*cqe)); in bnxt_qplib_cq_process_req()
2525 sq->swq_last, cqe->wr_id, cqe->status); in bnxt_qplib_cq_process_req()
2624 cqe++; in bnxt_qplib_cq_process_res_rc()
2641 cqe++; in bnxt_qplib_cq_process_res_rc()
2710 cqe++; in bnxt_qplib_cq_process_res_ud()
2728 cqe++; in bnxt_qplib_cq_process_res_ud()
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
H A Dmacsec.h32 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) in mlx5e_macsec_is_rx_flow() argument
34 return MLX5_MACSEC_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata)); in mlx5e_macsec_is_rx_flow()
38 struct mlx5_cqe64 *cqe);
46 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; } in mlx5e_macsec_is_rx_flow() argument
49 struct mlx5_cqe64 *cqe) in mlx5e_macsec_offload_handle_rx_skb() argument
/openbmc/linux/drivers/infiniband/ulp/iser/
H A Discsi_iser.h244 struct ib_cqe cqe; member
271 struct ib_cqe cqe; member
291 struct ib_cqe cqe; member
557 iser_rx(struct ib_cqe *cqe) in iser_rx() argument
559 return container_of(cqe, struct iser_rx_desc, cqe); in iser_rx()
563 iser_tx(struct ib_cqe *cqe) in iser_tx() argument
565 return container_of(cqe, struct iser_tx_desc, cqe); in iser_tx()
569 iser_login(struct ib_cqe *cqe) in iser_login() argument
571 return container_of(cqe, struct iser_login_desc, cqe); in iser_login()
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dotx2_txrx.c117 struct nix_cqe_tx_s *cqe) in otx2_xdp_snd_pkt_handler() argument
249 hash = cqe->hdr.flow_tag; in otx2_set_rxhash()
333 if (cqe->sg.segs) in otx2_check_rcv_errors()
393 struct nix_cqe_rx_s *cqe; in otx2_rx_napi_handler() local
406 !cqe->sg.seg_addr) { in otx2_rx_napi_handler()
417 cqe->sg.seg_addr = 0x00; in otx2_rx_napi_handler()
452 struct nix_cqe_tx_s *cqe; in otx2_tx_napi_handler() local
467 if (unlikely(!cqe)) { in otx2_tx_napi_handler()
1235 if (!cqe) in otx2_cleanup_rx_cqes()
1237 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes()
[all …]
/openbmc/linux/drivers/net/ethernet/qlogic/qede/
H A Dqede_fp.c661 cqe->header_len; in qede_set_gro_params()
871 qede_get_rxhash(tpa_info->skb, cqe->bitfields, cqe->rss_hash); in qede_tpa_start()
877 if (likely(cqe->bw_ext_bd_len_list[0])) in qede_tpa_start()
963 for (i = 0; cqe->len_list[i]; i++) in qede_tpa_cont()
965 le16_to_cpu(cqe->len_list[i])); in qede_tpa_cont()
988 for (i = 0; cqe->len_list[i]; i++) in qede_tpa_end()
990 le16_to_cpu(cqe->len_list[i])); in qede_tpa_end()
1226 union eth_rx_cqe *cqe, in qede_rx_process_tpa_cqe() argument
1251 union eth_rx_cqe *cqe; in qede_rx_process_cqe() local
1280 fp_cqe = &cqe->fast_path_regular; in qede_rx_process_cqe()
[all …]
/openbmc/linux/io_uring/
H A Dfdinfo.c126 struct io_uring_cqe *cqe = &r->cqes[(entry & cq_mask) << cq_shift]; in io_uring_show_fdinfo() local
129 entry & cq_mask, cqe->user_data, cqe->res, in io_uring_show_fdinfo()
130 cqe->flags); in io_uring_show_fdinfo()
133 cqe->big_cqe[0], cqe->big_cqe[1]); in io_uring_show_fdinfo()
204 struct io_uring_cqe *cqe = &ocqe->cqe; in io_uring_show_fdinfo() local
207 cqe->user_data, cqe->res, cqe->flags); in io_uring_show_fdinfo()
H A Dio_uring.h160 struct io_uring_cqe *cqe; in io_fill_cqe_req() local
167 if (unlikely(!io_get_cqe(ctx, &cqe))) in io_fill_cqe_req()
171 trace_io_uring_complete(req->ctx, req, req->cqe.user_data, in io_fill_cqe_req()
172 req->cqe.res, req->cqe.flags, in io_fill_cqe_req()
175 memcpy(cqe, &req->cqe, sizeof(*cqe)); in io_fill_cqe_req()
177 memcpy(cqe->big_cqe, &req->big_cqe, sizeof(*cqe)); in io_fill_cqe_req()
194 req->cqe.res = res; in io_req_set_res()
195 req->cqe.flags = cflags; in io_req_set_res()
/openbmc/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_hw_qp.c326 if (!rq->cqe) in alloc_rq_cqe()
336 sizeof(*rq->cqe[i]), in alloc_rq_cqe()
338 if (!rq->cqe[i]) in alloc_rq_cqe()
346 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[j]), rq->cqe[j], in alloc_rq_cqe()
352 vfree(rq->cqe); in alloc_rq_cqe()
368 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[i]), rq->cqe[i], in free_rq_cqe()
372 vfree(rq->cqe); in free_rq_cqe()
850 cqe = rq->cqe[*cons_idx]; in hinic_rq_read_wqe()
902 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_put_wqe() local
925 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_get_sge() local
[all …]
/openbmc/linux/drivers/infiniband/hw/mana/
H A Dcq.c29 if (attr->cqe > MAX_SEND_BUFFERS_PER_QUEUE) { in mana_ib_create_cq()
30 ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe); in mana_ib_create_cq()
34 cq->cqe = attr->cqe; in mana_ib_create_cq()
35 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE, in mana_ib_create_cq()
/openbmc/linux/drivers/scsi/qedi/
H A Dqedi_fw.c31 union iscsi_cqe *cqe, in qedi_process_logout_resp() argument
82 union iscsi_cqe *cqe, in qedi_process_text_resp() argument
178 union iscsi_cqe *cqe, in qedi_process_tmf_resp() argument
331 idx = cqe->rqe_opaque; in qedi_get_rq_bdq_buf()
366 idx = cqe->rqe_opaque; in qedi_put_rq_bdq_buf()
409 union iscsi_cqe *cqe, in qedi_process_nopin_mesg() argument
576 union iscsi_cqe *cqe, in qedi_scsi_completion() argument
674 union iscsi_cqe *cqe, in qedi_mtask_completion() argument
734 u32 proto_itt = cqe->itid; in qedi_process_cmd_cleanup_resp()
743 iscsi_cid = cqe->conn_id; in qedi_process_cmd_cleanup_resp()
[all …]
/openbmc/linux/tools/testing/selftests/net/
H A Dio_uring_zerocopy_tx.c361 struct io_uring_cqe *cqe; in do_tx() local
425 if (cqe->user_data != NONZC_TAG && in do_tx()
426 cqe->user_data != ZC_TAG) in do_tx()
430 if (cqe->flags & IORING_CQE_F_MORE) in do_tx()
440 if (cqe->user_data != ZC_TAG) in do_tx()
444 if (cqe->res >= 0) { in do_tx()
446 bytes += cqe->res; in do_tx()
447 } else if (cqe->res != -EAGAIN) { in do_tx()
448 error(1, cqe->res, "send failed"); in do_tx()
455 ret = io_uring_wait_cqe(&ring, &cqe); in do_tx()
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/
H A Den_rx.c666 struct mlx4_cqe *cqe; member
680 mlx4_en_get_cqe_ts(_ctx->cqe)); in mlx4_en_xdp_rx_timestamp()
688 struct mlx4_cqe *cqe = _ctx->cqe; in mlx4_en_xdp_rx_hash() local
696 status = cqe->status; in mlx4_en_xdp_rx_hash()
705 if (cqe->ipv6_ext_mask) in mlx4_en_xdp_rx_hash()
723 struct mlx4_cqe *cqe; in mlx4_en_process_rx_cq() local
812 length = be32_to_cpu(cqe->byte_cnt); in mlx4_en_process_rx_cq()
831 mxbuf.cqe = cqe; in mlx4_en_process_rx_cq()
936 if ((cqe->vlan_my_qpn & in mlx4_en_process_rx_cq()
940 be16_to_cpu(cqe->sl_vid)); in mlx4_en_process_rx_cq()
[all …]

1234567891011