/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_mqd_manager_vi.c | 80 struct queue_properties *q) in allocate_mqd() argument 93 struct queue_properties *q) in init_mqd() argument 123 set_priority(m, q); in init_mqd() 129 if (q->tba_addr) { in init_mqd() 193 q->doorbell_off << in __update_mqd() 222 m->cp_hqd_vmid = q->vmid; in __update_mqd() 235 set_priority(m, q); in __update_mqd() 237 q->is_active = QUEUE_IS_ACTIVE(*q); in __update_mqd() 321 struct queue_properties *q) in init_mqd_hiq() argument 342 struct queue_properties *q) in init_mqd_sdma() argument [all …]
|
H A D | kfd_mqd_manager_cik.c | 90 struct queue_properties *q) in init_mqd() argument 131 set_priority(m, q); in init_mqd() 144 struct queue_properties *q) in init_mqd_sdma() argument 198 m->cp_hqd_vmid = q->vmid; in __update_mqd() 204 set_priority(m, q); in __update_mqd() 206 q->is_active = QUEUE_IS_ACTIVE(*q); in __update_mqd() 248 q->is_active = QUEUE_IS_ACTIVE(*q); in update_mqd_sdma() 330 struct queue_properties *q) in init_mqd_hiq() argument 358 m->cp_hqd_vmid = q->vmid; in update_mqd_hiq() 360 q->is_active = QUEUE_IS_ACTIVE(*q); in update_mqd_hiq() [all …]
|
/openbmc/qemu/block/ |
H A D | nvme.c | 176 q->head = q->tail = 0; in nvme_init_queue() 197 trace_nvme_free_queue_pair(q->index, q, &q->cq, &q->sq); in nvme_free_queue_pair() 295 *q->sq.doorbell = cpu_to_le32(q->sq.tail); in nvme_kick() 296 q->inflight += q->need_kick; in nvme_kick() 304 req = &q->reqs[q->free_req_head]; in nvme_get_free_req_nofail_locked() 340 q->free_req_head = req - q->reqs; in nvme_put_free_req_locked() 419 q->cq_phase = !q->cq_phase; in nvme_process_completion() 444 *q->cq.doorbell = cpu_to_le32(q->cq.head); in nvme_process_completion() 463 *q->cq.doorbell = cpu_to_le32(q->cq.head); in nvme_process_completion_bh() 506 q->sq.tail = (q->sq.tail + 1) % NVME_QUEUE_SIZE; in nvme_submit_command() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci.c | 153 int index = q->producer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_producer_get() 155 if ((u16) (q->producer_counter - q->consumer_counter) == q->count) in mlxsw_pci_queue_elem_info_producer_get() 163 int index = q->consumer_counter & (q->count - 1); in mlxsw_pci_queue_elem_info_consumer_get() 175 return owner_bit != !!(q->consumer_counter & q->count); in mlxsw_pci_elem_hw_owned() 269 q->consumer_counter + q->count); in mlxsw_pci_queue_doorbell_consumer_ring() 713 struct mlxsw_pci_queue *q = from_tasklet(q, t, tasklet); in mlxsw_pci_cq_tasklet() local 826 struct mlxsw_pci_queue *q = from_tasklet(q, t, tasklet); in mlxsw_pci_eq_tasklet() local 938 q->count = q_ops->elem_count_f ? q_ops->elem_count_f(q) : in mlxsw_pci_queue_init() 940 q->elem_size = q_ops->elem_size_f ? q_ops->elem_size_f(q) : in mlxsw_pci_queue_init() 955 q->elem_info = kcalloc(q->count, sizeof(*q->elem_info), GFP_KERNEL); in mlxsw_pci_queue_init() [all …]
|
/openbmc/linux/net/ipv4/ |
H A D | inet_fragment.c | 64 q->fragments_tail = skb; in fragrun_append_to_last() 73 if (q->last_run_head) in fragrun_create() 81 q->last_run_head = skb; in fragrun_create() 263 f->destructor(q); in inet_frag_destroy_rcu() 304 fqdir = q->fqdir; in inet_frag_destroy() 322 if (!q) in inet_frag_alloc() 325 q->fqdir = fqdir; in inet_frag_alloc() 333 return q; in inet_frag_alloc() 344 if (!q) { in inet_frag_create() 354 inet_frag_kill(q); in inet_frag_create() [all …]
|
/openbmc/linux/net/sched/ |
H A D | sch_generic.c | 94 q->q.qlen--; in __skb_dequeue_bad_txq() 135 q->q.qlen++; in qdisc_enqueue_skb_bad_txq() 164 q->q.qlen++; in dev_requeue_skb() 186 struct sk_buff *nskb = q->dequeue(q); in try_bulk_dequeue_skb() 211 nskb = q->dequeue(q); in try_bulk_dequeue_skb_slow() 267 q->q.qlen--; in dequeue_skb() 292 skb = q->dequeue(q); in dequeue_skb() 360 dev->name, ret, q->q.qlen); in sch_direct_xmit() 660 .q.lock = __SPIN_LOCK_UNLOCKED(noop_qdisc.q.lock), 825 q->qlen = 0; in pfifo_fast_reset() [all …]
|
H A D | sch_teql.c | 81 if (q->q.qlen < dev->tx_queue_len) { in teql_enqueue() 82 __skb_queue_tail(&q->q, skb); in teql_enqueue() 110 sch->q.qlen = dat->q.qlen + q->q.qlen; in teql_dequeue() 181 q->m = m; in teql_qdisc_init() 183 skb_queue_head_init(&q->q); in teql_qdisc_init() 292 q = start; in teql_master_xmit() 293 if (!q) in teql_master_xmit() 336 } while ((q = NEXT_SLAVE(q)) != start); in teql_master_xmit() 389 } while ((q = NEXT_SLAVE(q)) != m->slaves); in teql_master_open() 420 if (q) { in teql_master_mtu() [all …]
|
H A D | sch_taprio.c | 583 sch->q.qlen++; in taprio_enqueue_one() 763 sch->q.qlen--; in taprio_dequeue_from_txq() 1651 (q->clockid != -1 && q->clockid != clockid)) { in taprio_parse_clockid() 1869 q->flags = err; in taprio_change() 1958 hrtimer_init(&q->advance_timer, q->clockid, HRTIMER_MODE_ABS); in taprio_change() 2024 if (q->qdiscs) { in taprio_reset() 2048 if (q->qdiscs) { in taprio_destroy() 2082 q->root = sch; in taprio_init() 2087 q->clockid = -1; in taprio_init() 2102 q->qdiscs = kcalloc(dev->num_tx_queues, sizeof(q->qdiscs[0]), in taprio_init() [all …]
|
/openbmc/linux/lib/raid6/ |
H A D | sse2.c | 39 u8 *p, *q; in raid6_sse21_gen_syndrome() local 44 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse21_gen_syndrome() 91 u8 *p, *q; in raid6_sse21_xor_syndrome() local 96 q = dptr[disks-1]; /* RS syndrome */ in raid6_sse21_xor_syndrome() 149 u8 *p, *q; in raid6_sse22_gen_syndrome() local 154 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse22_gen_syndrome() 202 u8 *p, *q; in raid6_sse22_xor_syndrome() local 207 q = dptr[disks-1]; /* RS syndrome */ in raid6_sse22_xor_syndrome() 281 u8 *p, *q; in raid6_sse24_gen_syndrome() local 286 q = dptr[z0+2]; /* RS syndrome */ in raid6_sse24_gen_syndrome() [all …]
|
H A D | avx2.c | 37 u8 *p, *q; in raid6_avx21_gen_syndrome() local 42 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx21_gen_syndrome() 86 u8 *p, *q; in raid6_avx21_xor_syndrome() local 91 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx21_xor_syndrome() 144 u8 *p, *q; in raid6_avx22_gen_syndrome() local 149 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx22_gen_syndrome() 196 u8 *p, *q; in raid6_avx22_xor_syndrome() local 201 q = dptr[disks-1]; /* RS syndrome */ in raid6_avx22_xor_syndrome() 276 u8 *p, *q; in raid6_avx24_gen_syndrome() local 281 q = dptr[z0+2]; /* RS syndrome */ in raid6_avx24_gen_syndrome() [all …]
|
H A D | recov_loongarch_simd.c | 32 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lsx() local 57 ptrs[disks - 1] = q; in raid6_2data_recov_lsx() 178 q += 64; in raid6_2data_recov_lsx() 189 u8 *p, *q, *dq; in raid6_datap_recov_lsx() local 207 ptrs[disks - 1] = q; in raid6_datap_recov_lsx() 281 q += 64; in raid6_datap_recov_lsx() 306 u8 *p, *q, *dp, *dq; in raid6_2data_recov_lasx() local 331 ptrs[disks - 1] = q; in raid6_2data_recov_lasx() 417 q += 64; in raid6_2data_recov_lasx() 428 u8 *p, *q, *dq; in raid6_datap_recov_lasx() local [all …]
|
/openbmc/linux/drivers/net/wireless/broadcom/brcm80211/brcmutil/ |
H A D | utils.c | 54 q = &pq->q[prec].skblist; in brcmu_pktq_penq() 55 skb_queue_tail(q, p); in brcmu_pktq_penq() 73 q = &pq->q[prec].skblist; in brcmu_pktq_penq_head() 89 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq() 90 p = skb_dequeue(q); in brcmu_pktq_pdeq() 112 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq_match() 115 skb_unlink(p, q); in brcmu_pktq_pdeq_match() 129 q = &pq->q[prec].skblist; in brcmu_pktq_pdeq_tail() 146 q = &pq->q[prec].skblist; in brcmu_pktq_pflush() 149 skb_unlink(p, q); in brcmu_pktq_pflush() [all …]
|
/openbmc/linux/include/linux/ |
H A D | blkdev.h | 581 #define blk_queue_dying(q) test_bit(QUEUE_FLAG_DYING, &(q)->queue_flags) argument 586 #define blk_queue_nonrot(q) test_bit(QUEUE_FLAG_NONROT, &(q)->queue_flags) argument 593 #define blk_queue_dax(q) test_bit(QUEUE_FLAG_DAX, &(q)->queue_flags) argument 607 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument 624 return q->mq_ops; in queue_is_mq() 630 return q->rpm_status; in queue_rpm_status() 728 if (q->queue_depth) in blk_queue_depth() 1161 if (q && q->limits.logical_block_size) in queue_logical_block_size() 1237 if (q) in bdev_write_zeroes_sectors() 1325 return q ? q->limits.dma_alignment : 511; in queue_dma_alignment() [all …]
|
/openbmc/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/ |
H A D | chcr_ktls.c | 810 pos = q->q.desc; in chcr_write_cpl_set_tcb_ulp() 827 pos = q->q.desc; in chcr_write_cpl_set_tcb_ulp() 869 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_tcb_cpls() 1033 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_write_tcp_options() 1152 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_wr_complete() 1210 pos = q->q.desc; in chcr_ktls_xmit_wr_complete() 1220 pos = q->q.desc; in chcr_ktls_xmit_wr_complete() 1241 pos = q->q.desc; in chcr_ktls_xmit_wr_complete() 1334 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_xmit_wr_short() 1513 pos = &q->q.desc[q->q.pidx]; in chcr_ktls_tx_plaintxt() [all …]
|
H A D | chcr_common.h | 61 int left = (void *)q->stat - pos; in chcr_copy_to_txd() 69 memcpy(q->desc, src + left, length - left); in chcr_copy_to_txd() 70 pos = (void *)q->desc + (length - left); in chcr_copy_to_txd() 83 return q->size - 1 - q->in_use; in chcr_txq_avail() 88 q->in_use += n; in chcr_txq_advance() 89 q->pidx += n; in chcr_txq_advance() 90 if (q->pidx >= q->size) in chcr_txq_advance() 91 q->pidx -= q->size; in chcr_txq_advance() 94 static inline void chcr_eth_txq_stop(struct sge_eth_txq *q) in chcr_eth_txq_stop() argument 96 netif_tx_stop_queue(q->txq); in chcr_eth_txq_stop() [all …]
|
/openbmc/linux/block/ |
H A D | blk-mq.h | 45 void blk_mq_exit_queue(struct request_queue *q); 82 return xa_load(&q->hctx_table, q->tag_set->map[type].mq_map[cpu]); in blk_mq_map_queue_type() 132 return per_cpu_ptr(q->queue_ctx, cpu); in __blk_mq_get_ctx() 148 struct request_queue *q; member 247 if (q->mq_ops->put_budget) in blk_mq_put_dispatch_budget() 248 q->mq_ops->put_budget(q, budget_token); in blk_mq_put_dispatch_budget() 253 if (q->mq_ops->get_budget) in blk_mq_get_dispatch_budget() 254 return q->mq_ops->get_budget(q); in blk_mq_get_dispatch_budget() 263 if (rq->q->mq_ops->set_rq_budget_token) in blk_mq_set_rq_budget_token() 269 if (rq->q->mq_ops->get_rq_budget_token) in blk_mq_get_rq_budget_token() [all …]
|
H A D | blk.h | 50 if (blk_queue_pm_only(q) && in blk_try_enter_queue() 58 blk_queue_exit(q); in blk_try_enter_queue() 68 if (blk_try_enter_queue(q, false)) in bio_queue_enter() 70 return __bio_queue_enter(q, bio); in bio_queue_enter() 166 return queue_max_segments(rq->q); in blk_rq_get_max_segments() 179 return q->limits.max_sectors; in blk_queue_get_max_sectors() 204 return bvec_gap_to_prev(&req->q->limits, in integrity_req_gap_back_merge() 352 if (req == q->last_merge) in req_set_nomerge() 353 q->last_merge = NULL; in req_set_nomerge() 390 struct request_queue *q) in blk_queue_bounce() argument [all …]
|
H A D | blk-ioc.c | 76 struct request_queue *q = icq->q; in ioc_destroy_icq() local 121 struct request_queue *q = icq->q; in ioc_release_fn() local 125 spin_unlock(&q->queue_lock); in ioc_release_fn() 132 spin_lock(&q->queue_lock); in ioc_release_fn() 339 if (icq && icq->q == q) in ioc_lookup_icq() 343 if (icq && icq->q == q) in ioc_lookup_icq() 371 q->node); in ioc_create_icq() 381 icq->q = q; in ioc_create_icq() 396 icq = ioc_lookup_icq(q); in ioc_create_icq() 431 icq = ioc_lookup_icq(q); in ioc_find_get_icq() [all …]
|
/openbmc/linux/drivers/crypto/cavium/zip/ |
H A D | zip_main.c | 159 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw() 171 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw() 178 q--; in zip_init_hw() 185 zip->iq[q].sw_tail = zip->iq[q].sw_head; in zip_init_hw() 186 zip->iq[q].hw_tail = zip->iq[q].sw_head; in zip_init_hw() 203 zip->iq[q].sw_head, zip->iq[q].sw_tail, in zip_init_hw() 215 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_init_hw() 222 for (q = 0; q < ZIP_NUM_QUEUES; q++) { in zip_init_hw() 234 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_init_hw() 356 for (q = 0; q < ZIP_NUM_QUEUES; q++) in zip_remove() [all …]
|
/openbmc/linux/arch/powerpc/lib/ |
H A D | test-code-patching.c | 206 void *p, *q; in test_translate_branch() local 221 q = p + 4; in test_translate_branch() 231 q = buf + 0x2000000; in test_translate_branch() 243 q = buf + 4; in test_translate_branch() 255 q = buf + 0x1400000; in test_translate_branch() 266 q = buf + 4; in test_translate_branch() 281 q = buf + 4; in test_translate_branch() 291 q = buf + 0x8000; in test_translate_branch() 303 q = buf + 4; in test_translate_branch() 315 q = buf + 0x5000; in test_translate_branch() [all …]
|
/openbmc/linux/drivers/media/pci/intel/ipu3/ |
H A D | ipu3-cio2.c | 235 if (!q->fbpt) in cio2_fbpt_init() 243 dma_free_coherent(dev, CIO2_FBPT_SIZE, q->fbpt, q->fbpt_bus_addr); in cio2_fbpt_exit() 313 if (!q->sensor) in cio2_csi2_calc_timing() 551 entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; in cio2_buffer_done() 562 b = q->bufs[q->bufs_first]; in cio2_buffer_done() 568 q->bufs[q->bufs_first] = NULL; in cio2_buffer_done() 583 q->bufs_first = (q->bufs_first + 1) % CIO2_MAX_BUFFERS; in cio2_buffer_done() 584 entry = &q->fbpt[q->bufs_first * CIO2_MAX_LOPS]; in cio2_buffer_done() 994 r = video_device_pipeline_start(&q->vdev, &q->pipe); in cio2_vb2_start_streaming() 1341 q->format.width, q->format.height, in cio2_video_link_validate() [all …]
|
/openbmc/linux/drivers/media/common/saa7146/ |
H A D | saa7146_fops.c | 55 if (WARN_ON(!q)) in saa7146_buffer_queue() 59 q->curr = buf; in saa7146_buffer_queue() 87 q->curr = NULL; in saa7146_buffer_finish() 106 if (WARN_ON(!q)) in saa7146_buffer_next() 118 q->curr = buf; in saa7146_buffer_next() 120 buf, q->queue.prev, q->queue.next); in saa7146_buffer_next() 156 struct saa7146_dmaqueue *q = from_timer(q, t, timeout); in saa7146_buffer_timeout() local 163 if (q->curr) { in saa7146_buffer_timeout() 355 q = &dev->vv_data->video_dmaq.q; in saa7146_register_device() 358 q = &dev->vv_data->vbi_dmaq.q; in saa7146_register_device() [all …]
|
/openbmc/linux/net/sctp/ |
H A D | outqueue.c | 76 return q->sched->dequeue(q); in sctp_outq_dequeue_data() 195 q->asoc = asoc; in sctp_outq_init() 270 sctp_outq_init(q->asoc, q); in sctp_outq_teardown() 312 if (!q->cork) in sctp_outq_tail() 560 q->fast_rtx = 1; in sctp_retransmit() 580 q->asoc->stream.si->generate_ftsn(q, q->asoc->ctsn_ack_point); in sctp_retransmit() 753 q->fast_rtx = 0; in __sctp_outq_flush_rtx() 761 if (q->cork) in sctp_outq_uncork() 762 q->cork = 0; in sctp_outq_uncork() 1195 .q = q, in sctp_outq_flush() [all …]
|
/openbmc/linux/drivers/media/v4l2-core/ |
H A D | videobuf-dma-contig.c | 75 struct videobuf_queue *q = map->q; in videobuf_vm_close() local 85 dev_dbg(q->dev, "munmap %p q=%p\n", map, q); in videobuf_vm_close() 86 videobuf_queue_lock(q); in videobuf_vm_close() 89 if (q->streaming) in videobuf_vm_close() 90 videobuf_queue_cancel(q); in videobuf_vm_close() 93 if (NULL == q->bufs[i]) in videobuf_vm_close() 99 mem = q->bufs[i]->priv; in videobuf_vm_close() 119 q->bufs[i]->map = NULL; in videobuf_vm_close() 120 q->bufs[i]->baddr = 0; in videobuf_vm_close() 125 videobuf_queue_unlock(q); in videobuf_vm_close() [all …]
|
/openbmc/linux/drivers/scsi/csiostor/ |
H A D | csio_wr.c | 247 q->pidx = q->cidx = q->inc_idx = 0; in csio_wr_alloc_q() 261 q->credits = (qsz - q->wr_sz) / q->wr_sz; in csio_wr_alloc_q() 926 if (unlikely(q->pidx == q->credits)) { in csio_wr_get() 1002 if (q->pidx > q->cidx) in csio_wr_avail_qcredits() 1003 return q->pidx - q->cidx; in csio_wr_avail_qcredits() 1004 else if (q->cidx > q->pidx) in csio_wr_avail_qcredits() 1005 return q->credits - (q->cidx - q->pidx); in csio_wr_avail_qcredits() 1135 void *wr = (void *)((uintptr_t)q->vstart + (q->cidx * q->wr_sz)); in csio_wr_process_iq() 1216 (q->cidx * q->wr_sz)); in csio_wr_process_iq() 1254 q->stats.n_tot_rsps += q->inc_idx; in csio_wr_process_iq() [all …]
|