Home
last modified time | relevance | path

Searched refs:sq_ring (Results 1 – 8 of 8) sorted by relevance

/openbmc/linux/drivers/infiniband/hw/irdma/
H A Duk.c65 if (!qp->sq_ring.head) in irdma_nop_1()
68 wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_nop_1()
143 qp->initial_ring.head = qp->sq_ring.head; in irdma_uk_qp_post_wr()
173 IRDMA_SQ_RING_FREE_QUANTA(qp->sq_ring)) in irdma_qp_get_next_send_wqe()
178 IRDMA_RING_MOVE_HEAD_NOCHECK(qp->sq_ring); in irdma_qp_get_next_send_wqe()
1128 if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) { in irdma_uk_cq_poll_cmpl()
1143 IRDMA_RING_SET_TAIL(qp->sq_ring, in irdma_uk_cq_poll_cmpl()
1156 tail = qp->sq_ring.tail; in irdma_uk_cq_poll_cmpl()
1162 IRDMA_RING_SET_TAIL(qp->sq_ring, in irdma_uk_cq_poll_cmpl()
1174 if (!IRDMA_RING_MORE_WORK(qp->sq_ring)) in irdma_uk_cq_poll_cmpl()
[all …]
H A Dutils.c539 pending_work = IRDMA_RING_USED_QUANTA(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op()
540 wqe_idx = IRDMA_RING_CURRENT_TAIL(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op()
546 wqe_idx = (wqe_idx + 1) % IRDMA_RING_SIZE(cqp->sc_cqp.sq_ring); in irdma_cleanup_pending_cqp_op()
2454 struct irdma_ring *sq_ring = &qp->sq_ring; in irdma_generate_flush_completions() local
2468 while (IRDMA_RING_MORE_WORK(*sq_ring)) { in irdma_generate_flush_completions()
2476 wqe_idx = sq_ring->tail; in irdma_generate_flush_completions()
2485 IRDMA_RING_SET_TAIL(*sq_ring, in irdma_generate_flush_completions()
2486 sq_ring->tail + qp->sq_wrtrk_array[sq_ring->tail].quanta); in irdma_generate_flush_completions()
H A Dctrl.c415 qp->hw_sq_size, qp->qp_uk.sq_ring.size); in irdma_sc_qp_init()
2082 cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size); in irdma_sc_gather_stats()
2763 IRDMA_RING_MOVE_TAIL(cqp->sq_ring); in irdma_cqp_poll_registers()
3123 IRDMA_RING_INIT(cqp->sq_ring, cqp->sq_size); in irdma_sc_cqp_init()
3250 cqp->sq_ring.head, cqp->sq_ring.tail, cqp->sq_ring.size); in irdma_sc_cqp_post_sq()
3266 if (IRDMA_RING_FULL_ERR(cqp->sq_ring)) { in irdma_sc_cqp_get_next_send_wqe_idx()
3269 cqp->sq_ring.head, cqp->sq_ring.tail, in irdma_sc_cqp_get_next_send_wqe_idx()
3270 cqp->sq_ring.size); in irdma_sc_cqp_get_next_send_wqe_idx()
3273 IRDMA_ATOMIC_RING_MOVE_HEAD(cqp->sq_ring, *wqe_idx, ret_code); in irdma_sc_cqp_get_next_send_wqe_idx()
3402 IRDMA_RING_MOVE_TAIL(cqp->sq_ring); in irdma_sc_ccq_get_cqe_info()
[all …]
H A Duser.h322 struct irdma_ring sq_ring; member
H A Dpuda.c196 *wqe_idx = IRDMA_RING_CURRENT_HEAD(qp->sq_ring); in irdma_puda_get_next_send_wqe()
199 IRDMA_RING_MOVE_HEAD(qp->sq_ring, ret_code); in irdma_puda_get_next_send_wqe()
399 IRDMA_RING_SET_TAIL(qp->sq_ring, info.wqe_idx); in irdma_puda_poll_cmpl()
687 IRDMA_RING_INIT(ukqp->sq_ring, ukqp->sq_size); in irdma_puda_qp_create()
H A Dhw.c2569 if (IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) { in irdma_hw_flush_wqes_callback()
2632 if (IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) { in irdma_hw_flush_wqes()
2670 if (!IRDMA_RING_MORE_WORK(qp->qp_uk.sq_ring)) in irdma_hw_flush_wqes()
H A Dtype.h363 struct irdma_ring sq_ring; member
/openbmc/linux/tools/testing/selftests/x86/
H A Dlam.c104 struct io_uring_queue sq_ring; member
386 struct io_uring_queue *sring = &s->sq_ring; in mmap_io_uring()
429 s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe), in mmap_io_uring()
432 if (s->sq_ring.queue.sqes == MAP_FAILED) { in mmap_io_uring()
508 struct io_uring_queue *sring = &ring->sq_ring; in handle_uring_sq()
538 index = tail & *ring->sq_ring.ring_mask; in handle_uring_sq()
540 sqe = &ring->sq_ring.queue.sqes[index]; in handle_uring_sq()