/openbmc/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_cq.c | 176 return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe; in cqe_sw() 195 be32_to_cpu(cqe[0]), be32_to_cpu(cqe[1]), be32_to_cpu(cqe[2]), in dump_cqe() 196 be32_to_cpu(cqe[3]), be32_to_cpu(cqe[4]), be32_to_cpu(cqe[5]), in dump_cqe() 197 be32_to_cpu(cqe[6]), be32_to_cpu(cqe[7])); in dump_cqe() 304 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 336 cq->ibcq.cqe < cq->resize_buf->cqe) { in mthca_cq_resize_copy_cqes() 383 be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe), in handle_error_cqe() 494 if (!cqe) in mthca_poll_one() 512 is_send = is_error ? cqe->opcode & 0x01 : cqe->is_send & 0x80; in mthca_poll_one() 646 set_cqe_hw(cqe); in mthca_poll_one() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mlx4/ |
H A D | cq.c | 81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 85 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe() 362 cqe = get_cqe(cq, i & cq->ibcq.cqe); in mlx4_ib_cq_resize_copy_cqes() 373 cqe = get_cqe(cq, ++i & cq->ibcq.cqe); in mlx4_ib_cq_resize_copy_cqes() 430 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_resize_cq() 447 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_resize_cq() 674 if (!cqe) in mlx4_ib_poll_one() 678 cqe++; in mlx4_ib_poll_one() 699 cq->ibcq.cqe = cq->resize_buf->cqe; in mlx4_ib_poll_one() 939 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in __mlx4_ib_cq_clean() [all …]
|
/openbmc/linux/drivers/infiniband/sw/siw/ |
H A D | siw_cq.c | 50 struct siw_cqe *cqe; in siw_reap_cqe() local 55 cqe = &cq->queue[cq->cq_get % cq->num_cqe]; in siw_reap_cqe() 58 wc->wr_id = cqe->id; in siw_reap_cqe() 59 wc->byte_len = cqe->bytes; in siw_reap_cqe() 67 if (cqe->flags & SIW_WQE_REM_INVAL) { in siw_reap_cqe() 71 wc->qp = cqe->base_qp; in siw_reap_cqe() 72 wc->opcode = map_wc_opcode[cqe->opcode]; in siw_reap_cqe() 77 cqe->flags, (void *)(uintptr_t)cqe->id); in siw_reap_cqe() 86 u8 opcode = cqe->opcode; in siw_reap_cqe() 87 u16 status = cqe->status; in siw_reap_cqe() [all …]
|
/openbmc/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_cq.c | 12 int cqe, int comp_vector) in rxe_cq_chk_attr() argument 16 if (cqe <= 0) { in rxe_cq_chk_attr() 17 rxe_dbg_dev(rxe, "cqe(%d) <= 0\n", cqe); in rxe_cq_chk_attr() 21 if (cqe > rxe->attr.max_cqe) { in rxe_cq_chk_attr() 23 cqe, rxe->attr.max_cqe); in rxe_cq_chk_attr() 29 if (cqe < count) { in rxe_cq_chk_attr() 31 cqe, count); in rxe_cq_chk_attr() 50 cq->queue = rxe_queue_init(rxe, &cqe, in rxe_cq_from_init() 68 cq->ibcq.cqe = cqe; in rxe_cq_from_init() 82 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue() [all …]
|
/openbmc/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_cq.c | 105 int entries = attr->cqe; in pvrdma_create_cq() 132 cq->ibcq.cqe = entries; in pvrdma_create_cq() 186 cmd->cqe = entries; in pvrdma_create_cq() 195 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq() 293 cq->ibcq.cqe); in _pvrdma_flush_cqe() 309 *cqe = *curr_cqe; in _pvrdma_flush_cqe() 315 cq->ibcq.cqe); in _pvrdma_flush_cqe() 329 struct pvrdma_cqe *cqe; in pvrdma_poll_one() local 358 wc->wr_id = cqe->wr_id; in pvrdma_poll_one() 365 wc->slid = cqe->slid; in pvrdma_poll_one() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci_hw.h | 116 return mlxsw_pci_cqe##v0##_##name##_get(cqe); \ 118 return mlxsw_pci_cqe##v1##_##name##_get(cqe); \ 120 return mlxsw_pci_cqe##v2##_##name##_get(cqe); \ 124 char *cqe, u32 val) \ 129 mlxsw_pci_cqe##v0##_##name##_set(cqe, val); \ 132 mlxsw_pci_cqe##v1##_##name##_set(cqe, val); \ 135 mlxsw_pci_cqe##v2##_##name##_set(cqe, val); \ 153 MLXSW_ITEM32(pci, cqe, system_port, 0x00, 0, 16); 164 MLXSW_ITEM32(pci, cqe, wqe_counter, 0x04, 16, 16); 171 MLXSW_ITEM32(pci, cqe, byte_count, 0x04, 0, 14); [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_rx.c | 140 cqe->op_own = op_own; in mlx5e_cqes_update_owner() 148 cqe->op_own = op_own; in mlx5e_cqes_update_owner() 999 struct mlx5_cqe64 *cqe; in mlx5e_poll_ico_cq() local 1007 if (likely(!cqe)) in mlx5e_poll_ico_cq() 1667 mxbuf->cqe = cqe; in mlx5e_fill_mxbuf() 2299 u16 data_bcnt = mpwrq_get_cqe_byte_cnt(cqe) - cqe->shampo.header_size; in mlx5e_handle_rx_cqe_mpwrq_shampo() 2454 if (!cqe) in mlx5e_rx_cq_process_enhanced_cqe_comp() 2475 title_cqe = cqe; in mlx5e_rx_cq_process_enhanced_cqe_comp() 2480 rq, cqe); in mlx5e_rx_cq_process_enhanced_cqe_comp() 2498 struct mlx5_cqe64 *cqe; in mlx5e_rx_cq_process_basic_cqe_comp() local [all …]
|
H A D | wq.h | 202 struct mlx5_cqe64 *cqe = mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_cqwq_get_wqe() local 205 cqe += wq->fbc.log_stride == 7; in mlx5_cqwq_get_wqe() 207 return cqe; in mlx5_cqwq_get_wqe() 233 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5_cqwq_get_cqe() local 234 u8 cqe_ownership_bit = cqe->op_own & MLX5_CQE_OWNER_MASK; in mlx5_cqwq_get_cqe() 243 return cqe; in mlx5_cqwq_get_cqe() 251 struct mlx5_cqe64 *cqe; in mlx5_cqwq_get_cqe_enahnced_comp() local 253 cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5_cqwq_get_cqe_enahnced_comp() 254 if (cqe->validity_iteration_count != sw_validity_iteration_count) in mlx5_cqwq_get_cqe_enahnced_comp() 260 return cqe; in mlx5_cqwq_get_cqe_enahnced_comp()
|
/openbmc/linux/drivers/infiniband/hw/cxgb4/ |
H A D | cq.c | 190 memset(&cqe, 0, sizeof(cqe)); in insert_recv_cqe() 224 memset(&cqe, 0, sizeof(cqe)); in insert_sq_cqe() 432 if ((CQE_OPCODE(cqe) == FW_RI_RDMA_WRITE) && RQ_TYPE(cqe)) in cqe_completes_wr() 435 if ((CQE_OPCODE(cqe) == FW_RI_READ_RESP) && SQ_TYPE(cqe)) in cqe_completes_wr() 438 if (CQE_SEND_OPCODE(cqe) && RQ_TYPE(cqe) && t4_rq_empty(wq)) in cqe_completes_wr() 783 CQE_TYPE(&cqe), CQE_OPCODE(&cqe), in __c4iw_poll_cq_one() 784 CQE_STATUS(&cqe), CQE_LEN(&cqe), in __c4iw_poll_cq_one() 785 CQE_WRID_HI(&cqe), CQE_WRID_LOW(&cqe), in __c4iw_poll_cq_one() 812 CQE_OPCODE(&cqe), CQE_QPID(&cqe)); in __c4iw_poll_cq_one() 849 CQE_OPCODE(&cqe), CQE_QPID(&cqe)); in __c4iw_poll_cq_one() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 81 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() local 84 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe() 88 return cqe; in get_sw_cqe() 276 cqe, sizeof(*cqe), false); in dump_cqe() 459 void *cqe; in mlx5_poll_one() local 464 if (!cqe) in mlx5_poll_one() 467 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in mlx5_poll_one() 869 void *cqe; in init_cq_frag_buf() local 874 cqe64 = buf->cqe_size == 64 ? cqe : cqe + 64; in init_cq_frag_buf() 1096 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in __mlx5_ib_cq_clean() [all …]
|
/openbmc/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_cq.c | 11 __be32 *cqe = get_queue_entry(cq->kern_cq.qbuf, cq->kern_cq.ci, in get_next_valid_cqe() local 14 be32_to_cpu(READ_ONCE(*cqe))); in get_next_valid_cqe() 16 return owner ^ !!(cq->kern_cq.ci & cq->depth) ? cqe : NULL; in get_next_valid_cqe() 115 struct erdma_cqe *cqe; in erdma_poll_one_cqe() local 122 cqe = get_next_valid_cqe(cq); in erdma_poll_one_cqe() 123 if (!cqe) in erdma_poll_one_cqe() 131 qpn = be32_to_cpu(cqe->qpn); in erdma_poll_one_cqe() 132 wqe_idx = be32_to_cpu(cqe->qe_idx); in erdma_poll_one_cqe() 133 cqe_hdr = be32_to_cpu(cqe->hdr); in erdma_poll_one_cqe() 158 wc->byte_len = be32_to_cpu(cqe->size); in erdma_poll_one_cqe() [all …]
|
/openbmc/linux/drivers/infiniband/sw/rdmavt/ |
H A D | cq.c | 55 head = cq->ibcq.cqe; in rvt_cq_enter() 249 cq->ibcq.cqe = entries; in rvt_create_cq() 350 if (cqe < 1 || cqe > rdi->dparms.props.max_cqe) in rvt_resize_cq() 393 if (head > (u32)cq->ibcq.cqe) in rvt_resize_cq() 394 head = (u32)cq->ibcq.cqe; in rvt_resize_cq() 395 if (tail > (u32)cq->ibcq.cqe) in rvt_resize_cq() 396 tail = (u32)cq->ibcq.cqe; in rvt_resize_cq() 401 if (unlikely((u32)cqe < n)) { in rvt_resize_cq() 415 cq->ibcq.cqe = cqe; in rvt_resize_cq() 493 tail = (u32)cq->ibcq.cqe; in rvt_poll_cq() [all …]
|
/openbmc/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.c | 2291 memset(cqe, 0, sizeof(*cqe)); in __flush_sq() 2298 cqe++; in __flush_sq() 2341 memset(cqe, 0, sizeof(*cqe)); in __flush_rq() 2347 cqe++; in __flush_rq() 2511 memset(cqe, 0, sizeof(*cqe)); in bnxt_qplib_cq_process_req() 2529 sq->swq_last, cqe->wr_id, cqe->status); in bnxt_qplib_cq_process_req() 2628 cqe++; in bnxt_qplib_cq_process_res_rc() 2645 cqe++; in bnxt_qplib_cq_process_res_rc() 2714 cqe++; in bnxt_qplib_cq_process_res_ud() 2732 cqe++; in bnxt_qplib_cq_process_res_ud() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | macsec.h | 32 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) in mlx5e_macsec_is_rx_flow() argument 34 return MLX5_MACSEC_METADATA_MARKER(be32_to_cpu(cqe->ft_metadata)); in mlx5e_macsec_is_rx_flow() 38 struct mlx5_cqe64 *cqe); 46 static inline bool mlx5e_macsec_is_rx_flow(struct mlx5_cqe64 *cqe) { return false; } in mlx5e_macsec_is_rx_flow() argument 49 struct mlx5_cqe64 *cqe) in mlx5e_macsec_offload_handle_rx_skb() argument
|
/openbmc/linux/drivers/infiniband/ulp/iser/ |
H A D | iscsi_iser.h | 244 struct ib_cqe cqe; member 271 struct ib_cqe cqe; member 291 struct ib_cqe cqe; member 557 iser_rx(struct ib_cqe *cqe) in iser_rx() argument 559 return container_of(cqe, struct iser_rx_desc, cqe); in iser_rx() 563 iser_tx(struct ib_cqe *cqe) in iser_tx() argument 565 return container_of(cqe, struct iser_tx_desc, cqe); in iser_tx() 569 iser_login(struct ib_cqe *cqe) in iser_login() argument 571 return container_of(cqe, struct iser_login_desc, cqe); in iser_login()
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | otx2_txrx.c | 117 struct nix_cqe_tx_s *cqe) in otx2_xdp_snd_pkt_handler() argument 249 hash = cqe->hdr.flow_tag; in otx2_set_rxhash() 333 if (cqe->sg.segs) in otx2_check_rcv_errors() 393 struct nix_cqe_rx_s *cqe; in otx2_rx_napi_handler() local 406 !cqe->sg.seg_addr) { in otx2_rx_napi_handler() 417 cqe->sg.seg_addr = 0x00; in otx2_rx_napi_handler() 452 struct nix_cqe_tx_s *cqe; in otx2_tx_napi_handler() local 467 if (unlikely(!cqe)) { in otx2_tx_napi_handler() 1235 if (!cqe) in otx2_cleanup_rx_cqes() 1237 if (cqe->sg.segs > 1) { in otx2_cleanup_rx_cqes() [all …]
|
/openbmc/linux/io_uring/ |
H A D | fdinfo.c | 126 struct io_uring_cqe *cqe = &r->cqes[(entry & cq_mask) << cq_shift]; in io_uring_show_fdinfo() local 129 entry & cq_mask, cqe->user_data, cqe->res, in io_uring_show_fdinfo() 130 cqe->flags); in io_uring_show_fdinfo() 133 cqe->big_cqe[0], cqe->big_cqe[1]); in io_uring_show_fdinfo() 204 struct io_uring_cqe *cqe = &ocqe->cqe; in io_uring_show_fdinfo() local 207 cqe->user_data, cqe->res, cqe->flags); in io_uring_show_fdinfo()
|
H A D | io_uring.h | 160 struct io_uring_cqe *cqe; in io_fill_cqe_req() local 167 if (unlikely(!io_get_cqe(ctx, &cqe))) in io_fill_cqe_req() 171 trace_io_uring_complete(req->ctx, req, req->cqe.user_data, in io_fill_cqe_req() 172 req->cqe.res, req->cqe.flags, in io_fill_cqe_req() 175 memcpy(cqe, &req->cqe, sizeof(*cqe)); in io_fill_cqe_req() 177 memcpy(cqe->big_cqe, &req->big_cqe, sizeof(*cqe)); in io_fill_cqe_req() 194 req->cqe.res = res; in io_req_set_res() 195 req->cqe.flags = cflags; in io_req_set_res()
|
/openbmc/linux/drivers/net/ethernet/qlogic/qede/ |
H A D | qede_fp.c | 661 cqe->header_len; in qede_set_gro_params() 871 qede_get_rxhash(tpa_info->skb, cqe->bitfields, cqe->rss_hash); in qede_tpa_start() 877 if (likely(cqe->bw_ext_bd_len_list[0])) in qede_tpa_start() 963 for (i = 0; cqe->len_list[i]; i++) in qede_tpa_cont() 965 le16_to_cpu(cqe->len_list[i])); in qede_tpa_cont() 988 for (i = 0; cqe->len_list[i]; i++) in qede_tpa_end() 990 le16_to_cpu(cqe->len_list[i])); in qede_tpa_end() 1226 union eth_rx_cqe *cqe, in qede_rx_process_tpa_cqe() argument 1251 union eth_rx_cqe *cqe; in qede_rx_process_cqe() local 1280 fp_cqe = &cqe->fast_path_regular; in qede_rx_process_cqe() [all …]
|
/openbmc/linux/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_hw_qp.c | 326 if (!rq->cqe) in alloc_rq_cqe() 336 sizeof(*rq->cqe[i]), in alloc_rq_cqe() 338 if (!rq->cqe[i]) in alloc_rq_cqe() 346 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[j]), rq->cqe[j], in alloc_rq_cqe() 352 vfree(rq->cqe); in alloc_rq_cqe() 368 dma_free_coherent(&pdev->dev, sizeof(*rq->cqe[i]), rq->cqe[i], in free_rq_cqe() 372 vfree(rq->cqe); in free_rq_cqe() 850 cqe = rq->cqe[*cons_idx]; in hinic_rq_read_wqe() 902 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_put_wqe() local 925 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_get_sge() local [all …]
|
/openbmc/linux/drivers/infiniband/hw/mana/ |
H A D | cq.c | 29 if (attr->cqe > MAX_SEND_BUFFERS_PER_QUEUE) { in mana_ib_create_cq() 30 ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe); in mana_ib_create_cq() 34 cq->cqe = attr->cqe; in mana_ib_create_cq() 35 cq->umem = ib_umem_get(ibdev, ucmd.buf_addr, cq->cqe * COMP_ENTRY_SIZE, in mana_ib_create_cq()
|
/openbmc/linux/drivers/scsi/qedi/ |
H A D | qedi_fw.c | 31 union iscsi_cqe *cqe, in qedi_process_logout_resp() argument 82 union iscsi_cqe *cqe, in qedi_process_text_resp() argument 178 union iscsi_cqe *cqe, in qedi_process_tmf_resp() argument 331 idx = cqe->rqe_opaque; in qedi_get_rq_bdq_buf() 366 idx = cqe->rqe_opaque; in qedi_put_rq_bdq_buf() 409 union iscsi_cqe *cqe, in qedi_process_nopin_mesg() argument 576 union iscsi_cqe *cqe, in qedi_scsi_completion() argument 674 union iscsi_cqe *cqe, in qedi_mtask_completion() argument 734 u32 proto_itt = cqe->itid; in qedi_process_cmd_cleanup_resp() 743 iscsi_cid = cqe->conn_id; in qedi_process_cmd_cleanup_resp() [all …]
|
/openbmc/linux/tools/testing/selftests/net/ |
H A D | io_uring_zerocopy_tx.c | 361 struct io_uring_cqe *cqe; in do_tx() local 425 if (cqe->user_data != NONZC_TAG && in do_tx() 426 cqe->user_data != ZC_TAG) in do_tx() 430 if (cqe->flags & IORING_CQE_F_MORE) in do_tx() 440 if (cqe->user_data != ZC_TAG) in do_tx() 444 if (cqe->res >= 0) { in do_tx() 446 bytes += cqe->res; in do_tx() 447 } else if (cqe->res != -EAGAIN) { in do_tx() 448 error(1, cqe->res, "send failed"); in do_tx() 455 ret = io_uring_wait_cqe(&ring, &cqe); in do_tx() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_rx.c | 666 struct mlx4_cqe *cqe; member 680 mlx4_en_get_cqe_ts(_ctx->cqe)); in mlx4_en_xdp_rx_timestamp() 688 struct mlx4_cqe *cqe = _ctx->cqe; in mlx4_en_xdp_rx_hash() local 696 status = cqe->status; in mlx4_en_xdp_rx_hash() 705 if (cqe->ipv6_ext_mask) in mlx4_en_xdp_rx_hash() 723 struct mlx4_cqe *cqe; in mlx4_en_process_rx_cq() local 812 length = be32_to_cpu(cqe->byte_cnt); in mlx4_en_process_rx_cq() 831 mxbuf.cqe = cqe; in mlx4_en_process_rx_cq() 936 if ((cqe->vlan_my_qpn & in mlx4_en_process_rx_cq() 940 be16_to_cpu(cqe->sl_vid)); in mlx4_en_process_rx_cq() [all …]
|
/openbmc/linux/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_verbs.c | 969 int entries = attr->cqe; in ocrdma_create_cq() 1023 ibcq->cqe = new_cnt; in ocrdma_resize_cq() 1036 cqe = cq->va; in ocrdma_flush_cq() 1046 cqe++; in ocrdma_flush_cq() 1593 struct ocrdma_cqe *cqe; in ocrdma_discard_cqes() local 1612 cqe = cq->va + cur_getp; in ocrdma_discard_cqes() 1643 cqe->cmn.qpn = 0; in ocrdma_discard_cqes() 2440 if (is_cqe_for_sq(cqe)) { in ocrdma_set_cqe_status_flushed() 2694 if (is_cqe_imm(cqe)) { in ocrdma_poll_success_rcqe() 2760 struct ocrdma_cqe *cqe; in ocrdma_poll_hwcq() local [all …]
|