Home
last modified time | relevance | path

Searched refs:q (Results 76 – 100 of 1913) sorted by relevance

12345678910>>...77

/openbmc/linux/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_mqd_manager_vi.c80 struct queue_properties *q) in allocate_mqd() argument
93 struct queue_properties *q) in init_mqd() argument
123 set_priority(m, q); in init_mqd()
129 if (q->tba_addr) { in init_mqd()
193 q->doorbell_off << in __update_mqd()
222 m->cp_hqd_vmid = q->vmid; in __update_mqd()
235 set_priority(m, q); in __update_mqd()
237 q->is_active = QUEUE_IS_ACTIVE(*q); in __update_mqd()
321 struct queue_properties *q) in init_mqd_hiq() argument
342 struct queue_properties *q) in init_mqd_sdma() argument
[all …]
H A Dkfd_mqd_manager_cik.c90 struct queue_properties *q) in init_mqd() argument
131 set_priority(m, q); in init_mqd()
144 struct queue_properties *q) in init_mqd_sdma() argument
198 m->cp_hqd_vmid = q->vmid; in __update_mqd()
204 set_priority(m, q); in __update_mqd()
206 q->is_active = QUEUE_IS_ACTIVE(*q); in __update_mqd()
248 q->is_active = QUEUE_IS_ACTIVE(*q); in update_mqd_sdma()
330 struct queue_properties *q) in init_mqd_hiq() argument
358 m->cp_hqd_vmid = q->vmid; in update_mqd_hiq()
360 q->is_active = QUEUE_IS_ACTIVE(*q); in update_mqd_hiq()
[all …]
/openbmc/qemu/block/
H A Dnvme.c176 q->head = q->tail = 0; in nvme_init_queue()
197 trace_nvme_free_queue_pair(q->index, q, &q->cq, &q->sq); in nvme_free_queue_pair()
295 *q->sq.doorbell = cpu_to_le32(q->sq.tail); in nvme_kick()
296 q->inflight += q->need_kick; in nvme_kick()
304 req = &q->reqs[q->free_req_head]; in nvme_get_free_req_nofail_locked()
340 q->free_req_head = req - q->reqs; in nvme_put_free_req_locked()
419 q->cq_phase = !q->cq_phase; in nvme_process_completion()
444 *q->cq.doorbell = cpu_to_le32(q->cq.head); in nvme_process_completion()
463 *q->cq.doorbell = cpu_to_le32(q->cq.head); in nvme_process_completion_bh()
506 q->sq.tail = (q->sq.tail + 1) % NVME_QUEUE_SIZE; in nvme_submit_command()
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlxsw/
H A Dpci.c153 int index = q->producer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_producer_get()
155 if ((u16) (q->producer_counter - q->consumer_counter) == q->count) in mlxsw_pci_queue_elem_info_producer_get()
163 int index = q->consumer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_consumer_get()
175 return owner_bit != !!(q->consumer_counter & q->count); in mlxsw_pci_elem_hw_owned()
269 q->consumer_counter + q->count); in mlxsw_pci_queue_doorbell_consumer_ring()
713 struct mlxsw_pci_queue *q = from_tasklet(q, t, tasklet); in mlxsw_pci_cq_tasklet() local
826 struct mlxsw_pci_queue *q = from_tasklet(q, t, tasklet); in mlxsw_pci_eq_tasklet() local
938 q->count = q_ops->elem_count_f ? q_ops->elem_count_f(q) : in mlxsw_pci_queue_init()
940 q->elem_size = q_ops->elem_size_f ? q_ops->elem_size_f(q) : in mlxsw_pci_queue_init()
955 q->elem_info = kcalloc(q->count, sizeof(*q->elem_info), GFP_KERNEL); in mlxsw_pci_queue_init()
[all …]
/openbmc/linux/net/ipv4/
H A Dinet_fragment.c64 q->fragments_tail = skb; in fragrun_append_to_last()
73 if (q->last_run_head) in fragrun_create()
81 q->last_run_head = skb; in fragrun_create()
263 f->destructor(q); in inet_frag_destroy_rcu()
304 fqdir = q->fqdir; in inet_frag_destroy()
322 if (!q) in inet_frag_alloc()
325 q->fqdir = fqdir; in inet_frag_alloc()
333 return q; in inet_frag_alloc()
344 if (!q) { in inet_frag_create()
354 inet_frag_kill(q); in inet_frag_create()
[all …]
/openbmc/linux/net/sched/
H A Dsch_generic.c94 q->q.qlen--; in __skb_dequeue_bad_txq()
135 q->q.qlen++; in qdisc_enqueue_skb_bad_txq()
164 q->q.qlen++; in dev_requeue_skb()
186 struct sk_buff *nskb = q->dequeue(q); in try_bulk_dequeue_skb()
211 nskb = q->dequeue(q); in try_bulk_dequeue_skb_slow()
267 q->q.qlen--; in dequeue_skb()
292 skb = q->dequeue(q); in dequeue_skb()
360 dev->name, ret, q->q.qlen); in sch_direct_xmit()
660 .q.lock = __SPIN_LOCK_UNLOCKED(noop_qdisc.q.lock),
825 q->qlen = 0; in pfifo_fast_reset()
[all …]
H A Dsch_teql.c81 if (q->q.qlen < dev->tx_queue_len) { in teql_enqueue()
82 __skb_queue_tail(&q->q, skb); in teql_enqueue()
110 sch->q.qlen = dat->q.qlen + q->q.qlen; in teql_dequeue()
181 q->m = m; in teql_qdisc_init()
183 skb_queue_head_init(&q->q); in teql_qdisc_init()
292 q = start; in teql_master_xmit()
293 if (!q) in teql_master_xmit()
336 } while ((q = NEXT_SLAVE(q)) != start); in teql_master_xmit()
389 } while ((q = NEXT_SLAVE(q)) != m->slaves); in teql_master_open()
420 if (q) { in teql_master_mtu()
[all …]
H A Dsch_taprio.c583 sch->q.qlen++; in taprio_enqueue_one()
763 sch->q.qlen--; in taprio_dequeue_from_txq()
1651 (q->clockid != -1 && q->clockid != clockid)) { in taprio_parse_clockid()
1869 q->flags = err; in taprio_change()
1958 hrtimer_init(&q->advance_timer, q->clockid, HRTIMER_MODE_ABS); in taprio_change()
2024 if (q->qdiscs) { in taprio_reset()
2048 if (q->qdiscs) { in taprio_destroy()
2082 q->root = sch; in taprio_init()
2087 q->clockid = -1; in taprio_init()
2102 q->qdiscs = kcalloc(dev->num_tx_queues, sizeof(q->qdiscs[0]), in taprio_init()
[all …]
/openbmc/linux/lib/raid6/
H A Dsse2.c39 u8 *p, *q; in raid6_sse21_gen_syndrome() local
44 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse21_gen_syndrome()
91 u8 *p, *q; in raid6_sse21_xor_syndrome() local
96 q = dptr[disks-1]; /* RS syndrome */ in raid6_sse21_xor_syndrome()
149 u8 *p, *q; in raid6_sse22_gen_syndrome() local
154 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse22_gen_syndrome()
202 u8 *p, *q; in raid6_sse22_xor_syndrome() local
207 q = dptr[disks-1]; /* RS syndrome */ in raid6_sse22_xor_syndrome()
281 u8 *p, *q; in raid6_sse24_gen_syndrome() local
286 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse24_gen_syndrome()
[all …]
H A Davx2.c37 u8 *p, *q; in raid6_avx21_gen_syndrome() local
42 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx21_gen_syndrome()
86 u8 *p, *q; in raid6_avx21_xor_syndrome() local
91 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx21_xor_syndrome()
144 u8 *p, *q; in raid6_avx22_gen_syndrome() local
149 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx22_gen_syndrome()
196 u8 *p, *q; in raid6_avx22_xor_syndrome() local
201 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx22_xor_syndrome()
276 u8 *p, *q; in raid6_avx24_gen_syndrome() local
281 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx24_gen_syndrome()
[all …]
H A Drecov_loongarch_simd.c32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local
57 ptrs[disks - 1] = q; in raid6_2data_recov_lsx()
178 q += 64; in raid6_2data_recov_lsx()
189 u8 *p, *q, *dq; in raid6_datap_recov_lsx() local
207 ptrs[disks - 1] = q; in raid6_datap_recov_lsx()
281 q += 64; in raid6_datap_recov_lsx()
306 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lasx() local
331 ptrs[disks - 1] = q; in raid6_2data_recov_lasx()
417 q += 64; in raid6_2data_recov_lasx()
428 u8 *p, *q, *dq; in raid6_datap_recov_lasx() local
[all …]
/openbmc/linux/drivers/net/wireless/broadcom/brcm80211/brcmutil/
H A Dutils.c54 q = &pq->q[prec].skblist; in brcmu_pktq_penq()
55 skb_queue_tail(q, p); in brcmu_pktq_penq()
73 q = &pq->q[prec].skblist; in brcmu_pktq_penq_head()
89 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq()
90 p = skb_dequeue(q); in brcmu_pktq_pdeq()
112 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq_match()
115 skb_unlink(p, q); in brcmu_pktq_pdeq_match()
129 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq_tail()
146 q = &pq->q[prec].skblist; in brcmu_pktq_pflush()
149 skb_unlink(p, q); in brcmu_pktq_pflush()
[all …]
/openbmc/linux/include/linux/
H A Dblkdev.h581 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) argument
586 #define blk_queue_nonrot(q) test_bit(QUEUE_FLAG_NONROT, &(q)->queue_flags) argument
593 #define blk_queue_dax(q) test_bit(QUEUE_FLAG_DAX, &(q)->queue_flags) argument
607 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument
624 return q->mq_ops; in queue_is_mq()
630 return q->rpm_status; in queue_rpm_status()
728 if (q->queue_depth) in blk_queue_depth()
1161 if (q && q->limits.logical_block_size) in queue_logical_block_size()
1237 if (q) in bdev_write_zeroes_sectors()
1325 return q ? q->limits.dma_alignment : 511; in queue_dma_alignment()
[all …]
/openbmc/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/
H A Dchcr_ktls.c810 pos = q->q.desc; in chcr_write_cpl_set_tcb_ulp()
827 pos = q->q.desc; in chcr_write_cpl_set_tcb_ulp()
869 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_tcb_cpls()
1033 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_write_tcp_options()
1152 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_wr_complete()
1210 pos = q->q.desc; in chcr_ktls_xmit_wr_complete()
1220 pos = q->q.desc; in chcr_ktls_xmit_wr_complete()
1241 pos = q->q.desc; in chcr_ktls_xmit_wr_complete()
1334 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_wr_short()
1513 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_tx_plaintxt()
[all …]
H A Dchcr_common.h61 int left = (void *)q->stat - pos; in chcr_copy_to_txd()
69 memcpy(q->desc, src + left, length - left); in chcr_copy_to_txd()
70 pos = (void *)q->desc + (length - left); in chcr_copy_to_txd()
83 return q->size - 1 - q->in_use; in chcr_txq_avail()
88 q->in_use += n; in chcr_txq_advance()
89 q->pidx += n; in chcr_txq_advance()
90 if (q->pidx >= q->size) in chcr_txq_advance()
91 q->pidx -= q->size; in chcr_txq_advance()
94 static inline void chcr_eth_txq_stop(struct sge_eth_txq *q) in chcr_eth_txq_stop() argument
96 netif_tx_stop_queue(q->txq); in chcr_eth_txq_stop()
[all …]
/openbmc/linux/block/
H A Dblk-mq.h45 void blk_mq_exit_queue(struct request_queue *q);
82 return xa_load(&q->hctx_table, q->tag_set->map[type].mq_map[cpu]); in blk_mq_map_queue_type()
132 return per_cpu_ptr(q->queue_ctx, cpu); in __blk_mq_get_ctx()
148 struct request_queue *q; member
247 if (q->mq_ops->put_budget) in blk_mq_put_dispatch_budget()
248 q->mq_ops->put_budget(q, budget_token); in blk_mq_put_dispatch_budget()
253 if (q->mq_ops->get_budget) in blk_mq_get_dispatch_budget()
254 return q->mq_ops->get_budget(q); in blk_mq_get_dispatch_budget()
263 if (rq->q->mq_ops->set_rq_budget_token) in blk_mq_set_rq_budget_token()
269 if (rq->q->mq_ops->get_rq_budget_token) in blk_mq_get_rq_budget_token()
[all …]
H A Dblk.h50 if (blk_queue_pm_only(q) && in blk_try_enter_queue()
58 blk_queue_exit(q); in blk_try_enter_queue()
68 if (blk_try_enter_queue(q, false)) in bio_queue_enter()
70 return __bio_queue_enter(q, bio); in bio_queue_enter()
166 return queue_max_segments(rq->q); in blk_rq_get_max_segments()
179 return q->limits.max_sectors; in blk_queue_get_max_sectors()
204 return bvec_gap_to_prev(&req->q->limits, in integrity_req_gap_back_merge()
352 if (req == q->last_merge) in req_set_nomerge()
353 q->last_merge = NULL; in req_set_nomerge()
390 struct request_queue *q) in blk_queue_bounce() argument
[all …]
H A Dblk-ioc.c76 struct request_queue *q = icq->q; in ioc_destroy_icq() local
121 struct request_queue *q = icq->q; in ioc_release_fn() local
125 spin_unlock(&q->queue_lock); in ioc_release_fn()
132 spin_lock(&q->queue_lock); in ioc_release_fn()
339 if (icq && icq->q == q) in ioc_lookup_icq()
343 if (icq && icq->q == q) in ioc_lookup_icq()
371 q->node); in ioc_create_icq()
381 icq->q = q; in ioc_create_icq()
396 icq = ioc_lookup_icq(q); in ioc_create_icq()
431 icq = ioc_lookup_icq(q); in ioc_find_get_icq()
[all …]
/openbmc/linux/drivers/crypto/cavium/zip/
H A Dzip_main.c159 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw()
171 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw()
178 q--; in zip_init_hw()
185 zip->iq[q].sw_tail = zip->iq[q].sw_head; in zip_init_hw()
186 zip->iq[q].hw_tail = zip->iq[q].sw_head; in zip_init_hw()
203 zip->iq[q].sw_head, zip->iq[q].sw_tail, in zip_init_hw()
215 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_init_hw()
222 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw()
234 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_init_hw()
356 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_remove()
[all …]
/openbmc/linux/arch/powerpc/lib/
H A Dtest-code-patching.c206 void *p, *q; in test_translate_branch() local
221 q = p + 4; in test_translate_branch()
231 q = buf + 0x2000000; in test_translate_branch()
243 q = buf + 4; in test_translate_branch()
255 q = buf + 0x1400000; in test_translate_branch()
266 q = buf + 4; in test_translate_branch()
281 q = buf + 4; in test_translate_branch()
291 q = buf + 0x8000; in test_translate_branch()
303 q = buf + 4; in test_translate_branch()
315 q = buf + 0x5000; in test_translate_branch()
[all …]
/openbmc/linux/drivers/media/pci/intel/ipu3/
H A Dipu3-cio2.c235 if (!q->fbpt) in cio2_fbpt_init()
243 dma_free_coherent(dev, CIO2_FBPT_SIZE, q->fbpt, q->fbpt_bus_addr); in cio2_fbpt_exit()
313 if (!q->sensor) in cio2_csi2_calc_timing()
551 entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; in cio2_buffer_done()
562 b = q->bufs[q->bufs_first]; in cio2_buffer_done()
568 q->bufs[q->bufs_first] = NULL; in cio2_buffer_done()
583 q->bufs_first = (q->bufs_first + 1) % CIO2_MAX_BUFFERS; in cio2_buffer_done()
584 entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; in cio2_buffer_done()
994 r = video_device_pipeline_start(&q->vdev, &q->pipe); in cio2_vb2_start_streaming()
1341 q->format.width, q->format.height, in cio2_video_link_validate()
[all …]
/openbmc/linux/drivers/media/common/saa7146/
H A Dsaa7146_fops.c55 if (WARN_ON(!q)) in saa7146_buffer_queue()
59 q->curr = buf; in saa7146_buffer_queue()
87 q->curr = NULL; in saa7146_buffer_finish()
106 if (WARN_ON(!q)) in saa7146_buffer_next()
118 q->curr = buf; in saa7146_buffer_next()
120 buf, q->queue.prev, q->queue.next); in saa7146_buffer_next()
156 struct saa7146_dmaqueue *q = from_timer(q, t, timeout); in saa7146_buffer_timeout() local
163 if (q->curr) { in saa7146_buffer_timeout()
355 q = &dev->vv_data->video_dmaq.q; in saa7146_register_device()
358 q = &dev->vv_data->vbi_dmaq.q; in saa7146_register_device()
[all …]
/openbmc/linux/net/sctp/
H A Doutqueue.c76 return q->sched->dequeue(q); in sctp_outq_dequeue_data()
195 q->asoc = asoc; in sctp_outq_init()
270 sctp_outq_init(q->asoc, q); in sctp_outq_teardown()
312 if (!q->cork) in sctp_outq_tail()
560 q->fast_rtx = 1; in sctp_retransmit()
580 q->asoc->stream.si->generate_ftsn(q, q->asoc->ctsn_ack_point); in sctp_retransmit()
753 q->fast_rtx = 0; in __sctp_outq_flush_rtx()
761 if (q->cork) in sctp_outq_uncork()
762 q->cork = 0; in sctp_outq_uncork()
1195 .q = q, in sctp_outq_flush()
[all …]
/openbmc/linux/drivers/media/v4l2-core/
H A Dvideobuf-dma-contig.c75 struct videobuf_queue *q = map->q; in videobuf_vm_close() local
85 dev_dbg(q->dev, "munmap %p q=%p\n", map, q); in videobuf_vm_close()
86 videobuf_queue_lock(q); in videobuf_vm_close()
89 if (q->streaming) in videobuf_vm_close()
90 videobuf_queue_cancel(q); in videobuf_vm_close()
93 if (NULL == q->bufs[i]) in videobuf_vm_close()
99 mem = q->bufs[i]->priv; in videobuf_vm_close()
119 q->bufs[i]->map = NULL; in videobuf_vm_close()
120 q->bufs[i]->baddr = 0; in videobuf_vm_close()
125 videobuf_queue_unlock(q); in videobuf_vm_close()
[all …]
/openbmc/linux/drivers/scsi/csiostor/
H A Dcsio_wr.c247 q->pidx = q->cidx = q->inc_idx = 0; in csio_wr_alloc_q()
261 q->credits = (qsz - q->wr_sz) / q->wr_sz; in csio_wr_alloc_q()
926 if (unlikely(q->pidx == q->credits)) { in csio_wr_get()
1002 if (q->pidx > q->cidx) in csio_wr_avail_qcredits()
1003 return q->pidx - q->cidx; in csio_wr_avail_qcredits()
1004 else if (q->cidx > q->pidx) in csio_wr_avail_qcredits()
1005 return q->credits - (q->cidx - q->pidx); in csio_wr_avail_qcredits()
1135 void *wr = (void *)((uintptr_t)q->vstart + (q->cidx * q->wr_sz)); in csio_wr_process_iq()
1216 (q->cidx * q->wr_sz)); in csio_wr_process_iq()
1254 q->stats.n_tot_rsps += q->inc_idx; in csio_wr_process_iq()
[all …]

12345678910>>...77