Home
last modified time | relevance | path

Searched refs:sqe (Results 26 – 50 of 89) sorted by relevance

1234

/openbmc/linux/io_uring/
H A Dopenclose.h6 int io_openat_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
10 int io_openat2_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
13 int io_close_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dmsg_ring.c249 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_msg_ring_prep() argument
253 if (unlikely(sqe->buf_index || sqe->personality)) in io_msg_ring_prep()
257 msg->user_data = READ_ONCE(sqe->off); in io_msg_ring_prep()
258 msg->len = READ_ONCE(sqe->len); in io_msg_ring_prep()
259 msg->cmd = READ_ONCE(sqe->addr); in io_msg_ring_prep()
260 msg->src_fd = READ_ONCE(sqe->addr3); in io_msg_ring_prep()
261 msg->dst_fd = READ_ONCE(sqe->file_index); in io_msg_ring_prep()
262 msg->flags = READ_ONCE(sqe->msg_ring_flags); in io_msg_ring_prep()
H A Dkbuf.c304 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_remove_buffers_prep() argument
309 if (sqe->rw_flags || sqe->addr || sqe->len || sqe->off || in io_remove_buffers_prep()
310 sqe->splice_fd_in) in io_remove_buffers_prep()
313 tmp = READ_ONCE(sqe->fd); in io_remove_buffers_prep()
319 p->bgid = READ_ONCE(sqe->buf_group); in io_remove_buffers_prep()
353 if (sqe->rw_flags || sqe->splice_fd_in) in io_provide_buffers_prep()
356 tmp = READ_ONCE(sqe->fd); in io_provide_buffers_prep()
360 p->addr = READ_ONCE(sqe->addr); in io_provide_buffers_prep()
361 p->len = READ_ONCE(sqe->len); in io_provide_buffers_prep()
373 p->bgid = READ_ONCE(sqe->buf_group); in io_provide_buffers_prep()
[all …]
H A Dtimeout.h32 int io_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
33 int io_link_timeout_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
35 int io_timeout_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dcancel.c129 int io_async_cancel_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_async_cancel_prep() argument
135 if (sqe->off || sqe->splice_fd_in) in io_async_cancel_prep()
138 cancel->addr = READ_ONCE(sqe->addr); in io_async_cancel_prep()
139 cancel->flags = READ_ONCE(sqe->cancel_flags); in io_async_cancel_prep()
145 cancel->fd = READ_ONCE(sqe->fd); in io_async_cancel_prep()
150 cancel->opcode = READ_ONCE(sqe->len); in io_async_cancel_prep()
H A Dnop.c13 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe) in io_nop_prep() argument
15 if (READ_ONCE(sqe->rw_flags)) in io_nop_prep()
H A Dsplice.h3 int io_tee_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_splice_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dadvise.h3 int io_madvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
6 int io_fadvise_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dpoll.c889 static __poll_t io_poll_parse_events(const struct io_uring_sqe *sqe, in io_poll_parse_events() argument
894 events = READ_ONCE(sqe->poll32_events); in io_poll_parse_events()
911 if (sqe->buf_index || sqe->splice_fd_in) in io_poll_remove_prep()
913 flags = READ_ONCE(sqe->len); in io_poll_remove_prep()
921 upd->old_user_data = READ_ONCE(sqe->addr); in io_poll_remove_prep()
925 upd->new_user_data = READ_ONCE(sqe->off); in io_poll_remove_prep()
929 upd->events = io_poll_parse_events(sqe, flags); in io_poll_remove_prep()
930 else if (sqe->poll32_events) in io_poll_remove_prep()
941 if (sqe->buf_index || sqe->off || sqe->addr) in io_poll_add_prep()
943 flags = READ_ONCE(sqe->len); in io_poll_add_prep()
[all …]
H A Dpoll.h27 int io_poll_add_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
30 int io_poll_remove_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dkbuf.h48 int io_remove_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
51 int io_provide_buffers_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dnop.h3 int io_nop_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Depoll.h4 int io_epoll_ctl_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
H A Dmsg_ring.h3 int io_msg_ring_prep(struct io_kiocb *req, const struct io_uring_sqe *sqe);
/openbmc/linux/drivers/scsi/qedi/
H A Dqedi_fw_api.c98 if (!task_params->sqe) in init_sqe()
101 memset(task_params->sqe, 0, sizeof(*task_params->sqe)); in init_sqe()
104 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
115 init_dif_context_flags(&task_params->sqe->prot_flags, in init_sqe()
118 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
134 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_NUM_SGES, in init_sqe()
141 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe()
147 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_WQE_TYPE, in init_sqe()
152 SET_FIELD(task_params->sqe->contlen_cdbsize, in init_sqe()
178 SET_FIELD(task_params->sqe->flags, ISCSI_WQE_RESPONSE, in init_sqe()
[all …]
/openbmc/linux/include/linux/
H A Dio_uring.h27 const struct io_uring_sqe *sqe; member
39 static inline const void *io_uring_sqe_cmd(const struct io_uring_sqe *sqe) in io_uring_sqe_cmd() argument
41 return sqe->cmd; in io_uring_sqe_cmd()
/openbmc/linux/drivers/infiniband/sw/siw/
H A Dsiw.h193 struct siw_sqe sqe; member
478 #define tx_type(wqe) ((wqe)->sqe.opcode)
480 #define tx_flags(wqe) ((wqe)->sqe.flags)
525 void siw_read_to_orq(struct siw_sqe *rreq, struct siw_sqe *sqe);
526 int siw_sqe_complete(struct siw_qp *qp, struct siw_sqe *sqe, u32 bytes,
630 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_empty() local
632 return READ_ONCE(sqe->flags) == 0; in siw_sq_empty()
637 struct siw_sqe *sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in sq_get_next() local
639 if (READ_ONCE(sqe->flags) & SIW_WQE_VALID) in sq_get_next()
640 return sqe; in sq_get_next()
H A Dsiw_qp_rx.c176 srx->ddp_stag = wqe->sqe.sge[0].lkey; in siw_rresp_check_ntoh()
177 srx->ddp_to = wqe->sqe.sge[0].laddr; in siw_rresp_check_ntoh()
695 resp = &tx_work->sqe; in siw_init_rresp()
758 wqe->sqe.id = orqe->id; in siw_orqe_start_rx()
759 wqe->sqe.opcode = orqe->opcode; in siw_orqe_start_rx()
760 wqe->sqe.sge[0].laddr = orqe->sge[0].laddr; in siw_orqe_start_rx()
761 wqe->sqe.sge[0].lkey = orqe->sge[0].lkey; in siw_orqe_start_rx()
763 wqe->sqe.flags = orqe->flags; in siw_orqe_start_rx()
764 wqe->sqe.num_sge = 1; in siw_orqe_start_rx()
827 sge = wqe->sqe.sge; /* there is only one */ in siw_proc_rresp()
[all …]
/openbmc/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c1862 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1863 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send()
1880 sqe->wqe_type = wqe->type; in bnxt_qplib_post_send()
1881 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1882 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send()
1912 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1913 sqe->wqe_size = wqe_sz; in bnxt_qplib_post_send()
1932 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1949 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
1960 sqe->flags = wqe->flags; in bnxt_qplib_post_send()
[all …]
/openbmc/linux/drivers/infiniband/hw/cxgb4/
H A Drestrack.c96 struct t4_swsqe *sqe) in fill_swsqe() argument
100 if (rdma_nl_put_driver_u32(msg, "opcode", sqe->opcode)) in fill_swsqe()
102 if (rdma_nl_put_driver_u32(msg, "complete", sqe->complete)) in fill_swsqe()
104 if (sqe->complete && in fill_swsqe()
105 rdma_nl_put_driver_u32(msg, "cqe_status", CQE_STATUS(&sqe->cqe))) in fill_swsqe()
107 if (rdma_nl_put_driver_u32(msg, "signaled", sqe->signaled)) in fill_swsqe()
109 if (rdma_nl_put_driver_u32(msg, "flushed", sqe->flushed)) in fill_swsqe()
/openbmc/linux/drivers/scsi/bnx2i/
H A Dbnx2i.h498 struct sqe { struct
634 struct sqe *sq_virt;
638 struct sqe *sq_prod_qe;
639 struct sqe *sq_cons_qe;
640 struct sqe *sq_first_qe;
641 struct sqe *sq_last_qe;
/openbmc/linux/drivers/scsi/qedf/
H A Dqedf_io.c588 struct fcoe_wqe *sqe) in qedf_init_task() argument
624 io_req->task_params->sqe = sqe; in qedf_init_task()
703 io_req->task_params->sqe = sqe; in qedf_init_mp_task()
854 struct fcoe_wqe *sqe; in qedf_post_io_req() local
901 sqe = &fcport->sq[sqe_idx]; in qedf_post_io_req()
1860 struct fcoe_wqe *sqe; in qedf_initiate_abts() local
1942 sqe = &fcport->sq[sqe_idx]; in qedf_initiate_abts()
1944 io_req->task_params->sqe = sqe; in qedf_initiate_abts()
2157 struct fcoe_wqe *sqe; in qedf_initiate_cleanup() local
2229 io_req->task_params->sqe = sqe; in qedf_initiate_cleanup()
[all …]
/openbmc/linux/tools/testing/selftests/x86/
H A Dlam.c510 struct io_uring_sqe *sqe; in handle_uring_sq() local
540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()
541 sqe->fd = file_fd; in handle_uring_sq()
542 sqe->flags = 0; in handle_uring_sq()
543 sqe->opcode = IORING_OP_READV; in handle_uring_sq()
544 sqe->addr = (unsigned long)fi->iovecs; in handle_uring_sq()
545 sqe->len = blocks; in handle_uring_sq()
546 sqe->off = 0; in handle_uring_sq()
547 sqe->user_data = (uint64_t)fi; in handle_uring_sq()
/openbmc/linux/drivers/nvme/host/
H A Drdma.c65 struct nvme_rdma_qe sqe; member
290 kfree(req->sqe.data); in nvme_rdma_exit_request()
304 if (!req->sqe.data) in nvme_rdma_init_request()
314 nvme_req(rq)->cmd = req->sqe.data; in nvme_rdma_init_request()
1661 struct nvme_command *cmd = sqe->data; in nvme_rdma_submit_async_event()
1673 sqe->cqe.done = nvme_rdma_async_done; in nvme_rdma_submit_async_event()
1985 struct nvme_rdma_qe *sqe = &req->sqe; in nvme_rdma_queue_rq() local
1999 req->sqe.dma = ib_dma_map_single(dev, req->sqe.data, in nvme_rdma_queue_rq()
2006 ib_dma_sync_single_for_cpu(dev, sqe->dma, in nvme_rdma_queue_rq()
2031 sqe->cqe.done = nvme_rdma_send_done; in nvme_rdma_queue_rq()
[all …]
/openbmc/linux/drivers/dma/
H A Dhisi_dma.c141 struct hisi_dma_sqe sqe; member
492 desc->sqe.length = cpu_to_le32(len); in hisi_dma_prep_dma_memcpy()
493 desc->sqe.src_addr = cpu_to_le64(src); in hisi_dma_prep_dma_memcpy()
494 desc->sqe.dst_addr = cpu_to_le64(dst); in hisi_dma_prep_dma_memcpy()
508 struct hisi_dma_sqe *sqe = chan->sq + chan->sq_tail; in hisi_dma_start_transfer() local
522 memcpy(sqe, &desc->sqe, sizeof(struct hisi_dma_sqe)); in hisi_dma_start_transfer()
525 sqe->dw0 = cpu_to_le32(FIELD_PREP(OPCODE_MASK, OPCODE_M2M)); in hisi_dma_start_transfer()
526 sqe->dw0 |= cpu_to_le32(LOCAL_IRQ_EN); in hisi_dma_start_transfer()

1234