Home
last modified time | relevance | path

Searched refs:io_cq (Results 1 – 17 of 17) sorted by relevance

/openbmc/linux/drivers/net/ethernet/amazon/ena/
H A Dena_eth_com.h60 int ena_com_rx_pkt(struct ena_com_io_cq *io_cq,
68 bool ena_com_cq_empty(struct ena_com_io_cq *io_cq);
70 static inline void ena_com_unmask_intr(struct ena_com_io_cq *io_cq, in ena_com_unmask_intr() argument
73 writel(intr_reg->intr_control, io_cq->unmask_reg); in ena_com_unmask_intr()
174 static inline int ena_com_update_dev_comp_head(struct ena_com_io_cq *io_cq) in ena_com_update_dev_comp_head() argument
179 if (unlikely(io_cq->cq_head_db_reg)) { in ena_com_update_dev_comp_head()
180 head = io_cq->head; in ena_com_update_dev_comp_head()
181 unreported_comp = head - io_cq->last_head_update; in ena_com_update_dev_comp_head()
182 need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH); in ena_com_update_dev_comp_head()
185 netdev_dbg(ena_com_io_cq_to_ena_dev(io_cq)->net_device, in ena_com_update_dev_comp_head()
[all …]
H A Dena_eth_com.c9 struct ena_com_io_cq *io_cq) in ena_com_get_next_rx_cdesc() argument
15 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc()
16 expected_phase = io_cq->phase; in ena_com_get_next_rx_cdesc()
18 cdesc = (struct ena_eth_io_rx_cdesc_base *)(io_cq->cdesc_addr.virt_addr in ena_com_get_next_rx_cdesc()
19 + (head_masked * io_cq->cdesc_entry_size_in_bytes)); in ena_com_get_next_rx_cdesc()
224 ena_com_rx_cdesc_idx_to_ptr(struct ena_com_io_cq *io_cq, u16 idx) in ena_com_rx_cdesc_idx_to_ptr() argument
226 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr()
228 ((uintptr_t)io_cq->cdesc_addr.virt_addr + in ena_com_rx_cdesc_idx_to_ptr()
229 idx * io_cq->cdesc_entry_size_in_bytes); in ena_com_rx_cdesc_idx_to_ptr()
232 static u16 ena_com_cdesc_rx_pkt_get(struct ena_com_io_cq *io_cq, in ena_com_cdesc_rx_pkt_get() argument
[all …]
H A Dena_com.c390 struct ena_com_io_cq *io_cq) in ena_com_init_io_cq() argument
394 memset(&io_cq->cdesc_addr, 0x0, sizeof(io_cq->cdesc_addr)); in ena_com_init_io_cq()
397 io_cq->cdesc_entry_size_in_bytes = in ena_com_init_io_cq()
398 (io_cq->direction == ENA_COM_IO_QUEUE_DIRECTION_TX) ? in ena_com_init_io_cq()
402 size = io_cq->cdesc_entry_size_in_bytes * io_cq->q_depth; in ena_com_init_io_cq()
404 io_cq->cdesc_addr.virt_addr = in ena_com_init_io_cq()
405 dma_alloc_coherent(ena_dev->dmadev, size, &io_cq->cdesc_addr.phys_addr, GFP_KERNEL); in ena_com_init_io_cq()
406 if (!io_cq->cdesc_addr.virt_addr) { in ena_com_init_io_cq()
407 io_cq->cdesc_addr.virt_addr = in ena_com_init_io_cq()
408 dma_alloc_coherent(ena_dev->dmadev, size, &io_cq->cdesc_addr.phys_addr, in ena_com_init_io_cq()
[all …]
H A Dena_com.h454 struct ena_com_io_cq **io_cq);
844 struct ena_com_io_cq *io_cq);
855 struct ena_com_io_cq *io_cq);
951 static inline struct ena_com_dev *ena_com_io_cq_to_ena_dev(struct ena_com_io_cq *io_cq) in ena_com_io_cq_to_ena_dev() argument
953 return container_of(io_cq, struct ena_com_dev, io_cq_queues[io_cq->qid]); in ena_com_io_cq_to_ena_dev()
/openbmc/linux/block/
H A Dblk-ioc.c37 struct io_cq *icq = container_of(head, struct io_cq, __rcu_head); in icq_free_icq_rcu()
46 static void ioc_exit_icq(struct io_cq *icq) in ioc_exit_icq()
61 struct io_cq *icq; in ioc_exit_icqs()
73 static void ioc_destroy_icq(struct io_cq *icq) in ioc_destroy_icq()
119 struct io_cq *icq = hlist_entry(ioc->icq_list.first, in ioc_release_fn()
120 struct io_cq, ioc_node); in ioc_release_fn()
175 struct io_cq *icq = in ioc_clear_queue()
176 list_first_entry(&q->icq_list, struct io_cq, q_node); in ioc_clear_queue()
324 struct io_cq *ioc_lookup_icq(struct request_queue *q) in ioc_lookup_icq()
327 struct io_cq *icq; in ioc_lookup_icq()
[all …]
H A Delevator.h9 struct io_cq;
49 void (*init_icq)(struct io_cq *);
50 void (*exit_icq)(struct io_cq *);
H A Dblk.h359 struct io_cq *ioc_find_get_icq(struct request_queue *q);
360 struct io_cq *ioc_lookup_icq(struct request_queue *q);
H A Delevator.c511 if (WARN_ON(e->icq_size < sizeof(struct io_cq)) || in elv_register()
512 WARN_ON(e->icq_align < __alignof__(struct io_cq))) in elv_register()
H A Dbfq-iosched.h479 struct io_cq icq; /* must be the first member */
H A Dbfq-iosched.c445 static struct bfq_io_cq *icq_to_bic(struct io_cq *icq) in icq_to_bic()
5479 static void bfq_exit_icq(struct io_cq *icq) in bfq_exit_icq()
/openbmc/linux/include/linux/
H A Diocontext.h73 struct io_cq { struct
110 struct io_cq __rcu *icq_hint; argument
H A Dblk-mq.h175 struct io_cq *icq;
/openbmc/linux/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h471 cq = phba->sli4_hba.hdwq[wqidx].io_cq; in lpfc_debug_dump_cq()
648 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) in lpfc_debug_dump_cq_by_id()
653 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq); in lpfc_debug_dump_cq_by_id()
H A Dlpfc_sli4.h700 struct lpfc_queue *io_cq; /* Fast-path FCP & NVME compl queue */ member
H A Dlpfc_init.c10397 phba->sli4_hba.hdwq[idx].io_cq = qdesc; in lpfc_alloc_io_wq_cq()
10844 lpfc_sli4_queue_free(hdwq[idx].io_cq); in lpfc_sli4_release_hdwq()
10847 hdwq[idx].io_cq = NULL; in lpfc_sli4_release_hdwq()
11184 qp[qidx].io_cq, in lpfc_sli4_queue_setup()
11493 lpfc_cq_destroy(phba, qp->io_cq); in lpfc_sli4_queue_unset()
H A Dlpfc_debugfs.c3964 qp = phba->sli4_hba.hdwq[eqidx].io_cq; in lpfc_idiag_cqs_for_eq()
4428 qp = phba->sli4_hba.hdwq[qidx].io_cq; in lpfc_idiag_queacc_write()
H A Dlpfc_sli.c6170 sli4_hba->sli4_write_cq_db(phba, qp->io_cq, 0, in lpfc_sli4_arm_cqeq_intr()