/openbmc/linux/io_uring/ |
H A D | openclose.h | 6 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 10 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 13 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | msg_ring.c | 249 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_msg_ring_prep() argument 253 if (unlikely(sqe->buf_index || sqe->personality)) in io_msg_ring_prep() 257 msg->user_data = READ_ONCE(sqe->off); in io_msg_ring_prep() 258 msg->len = READ_ONCE(sqe->len); in io_msg_ring_prep() 259 msg->cmd = READ_ONCE(sqe->addr); in io_msg_ring_prep() 260 msg->src_fd = READ_ONCE(sqe->addr3); in io_msg_ring_prep() 261 msg->dst_fd = READ_ONCE(sqe->file_index); in io_msg_ring_prep() 262 msg->flags = READ_ONCE(sqe->msg_ring_flags); in io_msg_ring_prep()
|
H A D | kbuf.c | 304 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_remove_buffers_prep() argument 309 if (sqe->rw_flags || sqe->addr || sqe->len || sqe->off || in io_remove_buffers_prep() 310 sqe->splice_fd_in) in io_remove_buffers_prep() 313 tmp = READ_ONCE(sqe->fd); in io_remove_buffers_prep() 319 p->bgid = READ_ONCE(sqe->buf_group); in io_remove_buffers_prep() 353 if (sqe->rw_flags || sqe->splice_fd_in) in io_provide_buffers_prep() 356 tmp = READ_ONCE(sqe->fd); in io_provide_buffers_prep() 360 p->addr = READ_ONCE(sqe->addr); in io_provide_buffers_prep() 361 p->len = READ_ONCE(sqe->len); in io_provide_buffers_prep() 373 p->bgid = READ_ONCE(sqe->buf_group); in io_provide_buffers_prep() [all …]
|
H A D | timeout.h | 32 int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 33 int io_link_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 35 int io_timeout_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | cancel.c | 129 int io_async_cancel_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_async_cancel_prep() argument 135 if (sqe->off || sqe->splice_fd_in) in io_async_cancel_prep() 138 cancel->addr = READ_ONCE(sqe->addr); in io_async_cancel_prep() 139 cancel->flags = READ_ONCE(sqe->cancel_flags); in io_async_cancel_prep() 145 cancel->fd = READ_ONCE(sqe->fd); in io_async_cancel_prep() 150 cancel->opcode = READ_ONCE(sqe->len); in io_async_cancel_prep()
|
H A D | nop.c | 13 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_nop_prep() argument 15 if (READ_ONCE(sqe->rw_flags)) in io_nop_prep()
|
H A D | splice.h | 3 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 6 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | advise.h | 3 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 6 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | poll.c | 889 static __poll_t io_poll_parse_events(const struct io_uring_sqe *sqe, in io_poll_parse_events() argument 894 events = READ_ONCE(sqe->poll32_events); in io_poll_parse_events() 911 if (sqe->buf_index || sqe->splice_fd_in) in io_poll_remove_prep() 913 flags = READ_ONCE(sqe->len); in io_poll_remove_prep() 921 upd->old_user_data = READ_ONCE(sqe->addr); in io_poll_remove_prep() 925 upd->new_user_data = READ_ONCE(sqe->off); in io_poll_remove_prep() 929 upd->events = io_poll_parse_events(sqe, flags); in io_poll_remove_prep() 930 else if (sqe->poll32_events) in io_poll_remove_prep() 941 if (sqe->buf_index || sqe->off || sqe->addr) in io_poll_add_prep() 943 flags = READ_ONCE(sqe->len); in io_poll_add_prep() [all …]
|
H A D | poll.h | 27 int io_poll_add_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 30 int io_poll_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | kbuf.h | 48 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe); 51 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | nop.h | 3 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | epoll.h | 4 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
H A D | msg_ring.h | 3 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
|
/openbmc/linux/drivers/scsi/qedi/ |
H A D | qedi_fw_api.c | 98 if (!task_params->sqe) in init_sqe() 101 memset(task_params->sqe, 0, sizeof(*task_params->sqe)); in init_sqe() 104 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe() 115 init_dif_context_flags(&task_params->sqe->prot_flags, in init_sqe() 118 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe() 134 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_NUM_SGES, in init_sqe() 141 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 147 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe() 152 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe() 178 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_RESPONSE, in init_sqe() [all …]
|
/openbmc/linux/include/linux/ |
H A D | io_uring.h | 27 const struct io_uring_sqe *sqe; member 39 static inline const void *io_uring_sqe_cmd(const struct io_uring_sqe *sqe) in io_uring_sqe_cmd() argument 41 return sqe->cmd; in io_uring_sqe_cmd()
|
/openbmc/linux/drivers/infiniband/sw/siw/ |
H A D | siw.h | 193 struct siw_sqe sqe; member 478 #define tx_type(wqe) ((wqe)->sqe.opcode) 480 #define tx_flags(wqe) ((wqe)->sqe.flags) 525 void siw_read_to_orq(struct siw_sqe *rreq, struct siw_sqe *sqe); 526 int siw_sqe_complete(struct siw_qp *qp, struct siw_sqe *sqe, u32 bytes, 630 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() local 632 return READ_ONCE(sqe->flags) == 0; in siw_sq_empty() 637 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next() local 639 if (READ_ONCE(sqe->flags) & SIW_WQE_VALID) in sq_get_next() 640 return sqe; in sq_get_next()
|
H A D | siw_qp_rx.c | 176 srx->ddp_stag = wqe->sqe.sge[0].lkey; in siw_rresp_check_ntoh() 177 srx->ddp_to = wqe->sqe.sge[0].laddr; in siw_rresp_check_ntoh() 695 resp = &tx_work->sqe; in siw_init_rresp() 758 wqe->sqe.id = orqe->id; in siw_orqe_start_rx() 759 wqe->sqe.opcode = orqe->opcode; in siw_orqe_start_rx() 760 wqe->sqe.sge[0].laddr = orqe->sge[0].laddr; in siw_orqe_start_rx() 761 wqe->sqe.sge[0].lkey = orqe->sge[0].lkey; in siw_orqe_start_rx() 763 wqe->sqe.flags = orqe->flags; in siw_orqe_start_rx() 764 wqe->sqe.num_sge = 1; in siw_orqe_start_rx() 827 sge = wqe->sqe.sge; /* there is only one */ in siw_proc_rresp() [all …]
|
/openbmc/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | qplib_fp.c | 1862 sqe->flags = wqe->flags; in bnxt_qplib_post_send() 1863 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send() 1880 sqe->wqe_type = wqe->type; in bnxt_qplib_post_send() 1881 sqe->flags = wqe->flags; in bnxt_qplib_post_send() 1882 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send() 1912 sqe->flags = wqe->flags; in bnxt_qplib_post_send() 1913 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send() 1932 sqe->flags = wqe->flags; in bnxt_qplib_post_send() 1949 sqe->flags = wqe->flags; in bnxt_qplib_post_send() 1960 sqe->flags = wqe->flags; in bnxt_qplib_post_send() [all …]
|
/openbmc/linux/drivers/infiniband/hw/cxgb4/ |
H A D | restrack.c | 96 struct t4_swsqe *sqe) in fill_swsqe() argument 100 if (rdma_nl_put_driver_u32(msg, "opcode", sqe->opcode)) in fill_swsqe() 102 if (rdma_nl_put_driver_u32(msg, "complete", sqe->complete)) in fill_swsqe() 104 if (sqe->complete && in fill_swsqe() 105 rdma_nl_put_driver_u32(msg, "cqe_status", CQE_STATUS(&sqe->cqe))) in fill_swsqe() 107 if (rdma_nl_put_driver_u32(msg, "signaled", sqe->signaled)) in fill_swsqe() 109 if (rdma_nl_put_driver_u32(msg, "flushed", sqe->flushed)) in fill_swsqe()
|
/openbmc/linux/drivers/scsi/bnx2i/ |
H A D | bnx2i.h | 498 struct sqe { struct 634 struct sqe *sq_virt; 638 struct sqe *sq_prod_qe; 639 struct sqe *sq_cons_qe; 640 struct sqe *sq_first_qe; 641 struct sqe *sq_last_qe;
|
/openbmc/linux/drivers/scsi/qedf/ |
H A D | qedf_io.c | 588 struct fcoe_wqe *sqe) in qedf_init_task() argument 624 io_req->task_params->sqe = sqe; in qedf_init_task() 703 io_req->task_params->sqe = sqe; in qedf_init_mp_task() 854 struct fcoe_wqe *sqe; in qedf_post_io_req() local 901 sqe = &fcport->sq[sqe_idx]; in qedf_post_io_req() 1860 struct fcoe_wqe *sqe; in qedf_initiate_abts() local 1942 sqe = &fcport->sq[sqe_idx]; in qedf_initiate_abts() 1944 io_req->task_params->sqe = sqe; in qedf_initiate_abts() 2157 struct fcoe_wqe *sqe; in qedf_initiate_cleanup() local 2229 io_req->task_params->sqe = sqe; in qedf_initiate_cleanup() [all …]
|
/openbmc/linux/tools/testing/selftests/x86/ |
H A D | lam.c | 510 struct io_uring_sqe *sqe; in handle_uring_sq() local 540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq() 541 sqe->fd = file_fd; in handle_uring_sq() 542 sqe->flags = 0; in handle_uring_sq() 543 sqe->opcode = IORING_OP_READV; in handle_uring_sq() 544 sqe->addr = (unsigned long)fi->iovecs; in handle_uring_sq() 545 sqe->len = blocks; in handle_uring_sq() 546 sqe->off = 0; in handle_uring_sq() 547 sqe->user_data = (uint64_t)fi; in handle_uring_sq()
|
/openbmc/linux/drivers/nvme/host/ |
H A D | rdma.c | 65 struct nvme_rdma_qe sqe; member 290 kfree(req->sqe.data); in nvme_rdma_exit_request() 304 if (!req->sqe.data) in nvme_rdma_init_request() 314 nvme_req(rq)->cmd = req->sqe.data; in nvme_rdma_init_request() 1661 struct nvme_command *cmd = sqe->data; in nvme_rdma_submit_async_event() 1673 sqe->cqe.done = nvme_rdma_async_done; in nvme_rdma_submit_async_event() 1985 struct nvme_rdma_qe *sqe = &req->sqe; in nvme_rdma_queue_rq() local 1999 req->sqe.dma = ib_dma_map_single(dev, req->sqe.data, in nvme_rdma_queue_rq() 2006 ib_dma_sync_single_for_cpu(dev, sqe->dma, in nvme_rdma_queue_rq() 2031 sqe->cqe.done = nvme_rdma_send_done; in nvme_rdma_queue_rq() [all …]
|
/openbmc/linux/drivers/dma/ |
H A D | hisi_dma.c | 141 struct hisi_dma_sqe sqe; member 492 desc->sqe.length = cpu_to_le32(len); in hisi_dma_prep_dma_memcpy() 493 desc->sqe.src_addr = cpu_to_le64(src); in hisi_dma_prep_dma_memcpy() 494 desc->sqe.dst_addr = cpu_to_le64(dst); in hisi_dma_prep_dma_memcpy() 508 struct hisi_dma_sqe *sqe = chan->sq + chan->sq_tail; in hisi_dma_start_transfer() local 522 memcpy(sqe, &desc->sqe, sizeof(struct hisi_dma_sqe)); in hisi_dma_start_transfer() 525 sqe->dw0 = cpu_to_le32(FIELD_PREP(OPCODE_MASK, OPCODE_M2M)); in hisi_dma_start_transfer() 526 sqe->dw0 |= cpu_to_le32(LOCAL_IRQ_EN); in hisi_dma_start_transfer()
|