/openbmc/linux/net/sched/ |
H A D | sch_ingress.c | 53 return q->block; in ingress_tcf_block() 101 q->block_info.chain_head_change_priv = &q->miniqp; in ingress_init() 103 err = tcf_block_get_ext(&q->block, sch, &q->block_info, extack); in ingress_init() 107 mini_qdisc_pair_block_init(&q->miniqp, q->block); in ingress_init() 121 tcf_block_put_ext(q->block, sch, &q->block_info); in ingress_destroy() 266 q->ingress_block_info.chain_head_change_priv = &q->miniqp_ingress; in clsact_init() 268 err = tcf_block_get_ext(&q->ingress_block, sch, &q->ingress_block_info, in clsact_init() 273 mini_qdisc_pair_block_init(&q->miniqp_ingress, q->ingress_block); in clsact_init() 285 q->egress_block_info.chain_head_change_priv = &q->miniqp_egress; in clsact_init() 300 tcf_block_put_ext(q->ingress_block, sch, &q->ingress_block_info); in clsact_destroy() [all …]
|
H A D | sch_codel.c | 94 skb = codel_dequeue(sch, &sch->qstats.backlog, &q->params, &q->vars, in codel_qdisc_dequeue() 101 if (q->stats.drop_count && sch->q.qlen) { in codel_qdisc_dequeue() 102 qdisc_tree_reduce_backlog(sch, q->stats.drop_count, q->stats.drop_len); in codel_qdisc_dequeue() 103 q->stats.drop_count = 0; in codel_qdisc_dequeue() 104 q->stats.drop_len = 0; in codel_qdisc_dequeue() 120 q = qdisc_priv(sch); in codel_qdisc_enqueue() 121 q->drop_overlimit++; in codel_qdisc_enqueue() 172 qlen = sch->q.qlen; in codel_change() 229 q->params.ecn)) in codel_dump() 247 .count = q->vars.count, in codel_dump_stats() [all …]
|
H A D | sch_etf.c | 82 if (q->skip_sock_check) in is_packet_valid() 101 now = q->get_time(); in is_packet_valid() 192 sch->q.qlen++; in etf_enqueue_timesortedlist() 225 sch->q.qlen--; in timesortedlist_drop() 250 sch->q.qlen--; in timesortedlist_remove() 263 now = q->get_time(); in etf_dequeue_timesortedlist() 303 if (!q->offload) in etf_disable_offload() 412 qdisc_watchdog_init_clockid(&q->watchdog, sch, q->clockid); in etf_init() 429 sch->q.qlen--; in timesortedlist_clear() 445 q->last = 0; in etf_reset() [all …]
|
H A D | sch_ets.c | 127 q->prio2band, sizeof(q->prio2band)); in ets_offload_change() 293 if (!ets_class_is_strict(q, cl) && sch->q.qlen) in ets_class_qlen_notify() 400 return &q->classes[q->prio2band[band & TC_PRIO_MAX]]; in ets_classify() 406 return &q->classes[q->prio2band[0]]; in ets_classify() 651 if (q->classes[i].qdisc->q.qlen) { in ets_qdisc_change() 652 list_add_tail(&q->classes[i].alist, &q->active); in ets_qdisc_change() 657 if (i >= q->nstrict && q->classes[i].qdisc->q.qlen) in ets_qdisc_change() 712 for (band = q->nstrict; band < q->nbands; band++) { in ets_qdisc_reset() 713 if (q->classes[band].qdisc->q.qlen) in ets_qdisc_reset() 755 if (q->nbands > q->nstrict) { in ets_qdisc_dump() [all …]
|
H A D | sch_pie.c | 96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue() 99 } else if (q->params.ecn && (q->vars.prob <= MAX_PROB / 10) && in pie_qdisc_enqueue() 104 q->stats.ecn_mark++; in pie_qdisc_enqueue() 122 q->stats.dropped++; in pie_qdisc_enqueue() 164 q->params.tupdate = in pie_change() 191 qlen = sch->q.qlen; in pie_change() 422 struct pie_sched_data *q = from_timer(q, t, adapt_timer); in pie_timer() local 429 pie_calculate_probability(&q->params, &q->vars, sch->qstats.backlog); in pie_timer() 433 mod_timer(&q->adapt_timer, jiffies + q->params.tupdate); in pie_timer() 447 q->sch = sch; in pie_init() [all …]
|
H A D | sch_gred.c | 177 if (!q) { in gred_enqueue() 212 q->vars.qavg = red_calc_qavg(&q->parms, in gred_enqueue() 222 switch (red_action(&q->parms, &q->vars, q->vars.qavg + qavg)) { in gred_enqueue() 238 if (gred_use_harddrop(q) || !gred_use_ecn(q) || in gred_enqueue() 303 if (!q) in gred_reset() 336 if (!q) in gred_offload() 405 kfree(q); in gred_destroy_vq() 498 if (!q) in gred_change_vq() 793 max_p[i] = q ? q->parms.max_P : 0; in gred_dump() 840 qavg = red_calc_qavg(&q->parms, &q->vars, in gred_dump() [all …]
|
H A D | sch_prio.c | 57 return q->queues[q->prio2band[band & TC_PRIO_MAX]]; in prio_classify() 62 if (band >= q->bands) in prio_classify() 63 return q->queues[q->prio2band[0]]; in prio_classify() 65 return q->queues[band]; in prio_classify() 89 sch->q.qlen++; in prio_enqueue() 122 sch->q.qlen--; in prio_dequeue() 170 tcf_block_put(q->block); in prio_destroy() 210 q->bands = qopt->bands; in prio_tune() 238 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); in prio_init() 269 opt.bands = q->bands; in prio_dump() [all …]
|
H A D | sch_htb.c | 594 WARN_ON(cl->level || !cl->leaf.q || !cl->leaf.q->q.qlen); in htb_activate() 628 if (q->direct_queue.qlen < q->direct_qlen) { in htb_enqueue() 892 if (unlikely(cl->leaf.q->q.qlen == 0)) { in htb_dequeue_tree() 908 skb = cl->leaf.q->dequeue(cl->leaf.q); in htb_dequeue_tree() 930 if (!cl->leaf.q->q.qlen) in htb_dequeue_tree() 1008 if (cl->leaf.q && !q->offload) in htb_reset() 1017 memset(q->hlevel, 0, sizeof(q->hlevel)); in htb_reset() 1018 memset(q->row_mask, 0, sizeof(q->row_mask)); in htb_reset() 1095 q->direct_qdiscs = kcalloc(q->num_direct_qdiscs, in htb_init() 1544 struct Qdisc *q = cl->leaf.q; in htb_destroy_class_offload() local [all …]
|
/openbmc/linux/block/ |
H A D | blk-core.c | 271 if (queue_is_mq(q)) in blk_free_queue() 272 blk_mq_release(q); in blk_free_queue() 288 blk_free_queue(q); in blk_put_queue() 387 struct request_queue *q = from_timer(q, t, timeout); in blk_rq_timed_out_timer() local 402 if (!q) in blk_alloc_queue() 408 if (q->id < 0) in blk_alloc_queue() 412 if (!q->stats) in blk_alloc_queue() 415 q->node = node_id; in blk_alloc_queue() 447 return q; in blk_alloc_queue() 1033 if (queue_is_mq(q) && q->mq_ops->busy) in blk_lld_busy() [all …]
|
H A D | blk-rq-qos.h | 74 return rq_qos_id(q, RQ_QOS_WBT); in wbt_rq_qos() 115 if (q->rq_qos) in rq_qos_cleanup() 121 if (q->rq_qos) in rq_qos_done() 122 __rq_qos_done(q->rq_qos, rq); in rq_qos_done() 127 if (q->rq_qos) in rq_qos_issue() 133 if (q->rq_qos) in rq_qos_requeue() 142 if (q->rq_qos) in rq_qos_done_bio() 149 if (q->rq_qos) { in rq_qos_throttle() 158 if (q->rq_qos) in rq_qos_track() 165 if (q->rq_qos) { in rq_qos_merge() [all …]
|
H A D | blk-mq.c | 320 rq->q = q; in blk_rq_init() 356 rq->q = q; in blk_mq_rq_ctx_init() 524 .q = q, in blk_mq_rq_cache_fill() 561 if (!rq || rq->q != q) in blk_mq_alloc_cached_request() 586 .q = q, in blk_mq_alloc_request() 615 .q = q, in blk_mq_alloc_request_hctx() 2822 q = rq->q; in blk_mq_flush_plug_list() 2904 .q = q, in blk_mq_get_new_requests() 2995 if (rq && rq->q != q) in blk_mq_submit_bio() 4707 qe->q = q; in blk_mq_elv_switch_none() [all …]
|
H A D | blk-mq-sysfs.c | 54 struct request_queue *q; in blk_mq_hw_sysfs_show() local 59 q = hctx->queue; in blk_mq_hw_sysfs_show() 64 mutex_lock(&q->sysfs_lock); in blk_mq_hw_sysfs_show() 66 mutex_unlock(&q->sysfs_lock); in blk_mq_hw_sysfs_show() 201 kobject_put(q->mq_kobj); in blk_mq_sysfs_deinit() 214 kobject_get(q->mq_kobj); in blk_mq_sysfs_init() 240 q->mq_sysfs_init_done = true; in blk_mq_sysfs_register() 252 kobject_del(q->mq_kobj); in blk_mq_sysfs_register() 268 kobject_del(q->mq_kobj); in blk_mq_sysfs_unregister() 279 if (!q->mq_sysfs_init_done) in blk_mq_sysfs_unregister_hctxs() [all …]
|
/openbmc/linux/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 1235 q->q.stops++; in eth_txq_stop() 1607 wr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_eth_xmit() 1608 eowr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_eth_xmit() 2658 q->q.stops++; in ctrlq_check_stop() 2696 wr = (void *)&q->q.desc[q->q.pidx]; in cxgb4_selftest_lb_pkt() 2770 wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; in ctrl_xmit() 2808 wr = (struct fw_wr_hdr *)&q->q.desc[q->q.pidx]; in restart_ctrlq() 2910 q->q.stops++; in txq_stop_maperr() 2926 q->q.stops++; in ofldtxq_stop() 2988 pos = (u64 *)&q->q.desc[q->q.pidx]; in service_ofldq() [all …]
|
/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_process_queue_manager.c | 38 if ((pqn->q && pqn->q->properties.queue_id == qid) || in get_queue_by_qid() 110 if (pqn->q) in pqm_set_gws() 208 if (pqn->q) in pqm_uninit() 264 *q = NULL; in init_user_queue() 294 q = NULL; in pqm_create_queue() 346 pqn->q = q; in pqm_create_queue() 367 pqn->q = q; in pqm_create_queue() 420 if (q) { in pqm_create_queue() 463 if (pqn->q) in pqm_destroy_queue() 1023 q = pqn->q; in pqm_debugfs_mqds() [all …]
|
/openbmc/linux/drivers/misc/uacce/ |
H A D | uacce.c | 20 return q->state == UACCE_Q_INIT || q->state == UACCE_Q_STARTED; in uacce_queue_is_valid() 31 ret = q->uacce->ops->start_queue(q); in uacce_start_queue() 47 if ((q->state == UACCE_Q_INIT || q->state == UACCE_Q_STARTED) && in uacce_put_queue() 122 q->pasid = pasid; in uacce_bind_queue() 128 if (!q->handle) in uacce_unbind_queue() 131 q->handle = NULL; in uacce_unbind_queue() 145 if (!q) in uacce_fops_open() 159 q->uacce = uacce; in uacce_fops_open() 162 ret = uacce->ops->get_queue(uacce, q->pasid, q); in uacce_fops_open() 180 kfree(q); in uacce_fops_open() [all …]
|
/openbmc/linux/include/media/ |
H A D | videobuf2-core.h | 432 int (*queue_setup)(struct vb2_queue *q, 437 void (*wait_finish)(struct vb2_queue *q); 671 return q->allow_cache_hints && q->memory == VB2_MEMORY_MMAP; in vb2_queue_allows_cache_hints() 1121 return q->streaming; in vb2_is_streaming() 1139 return q->fileio; in vb2_fileio_is_active() 1150 return (q->num_buffers > 0); in vb2_is_busy() 1159 return q->drv_priv; in vb2_get_drv_priv() 1217 return q->start_streaming_called; in vb2_start_streaming_called() 1226 q->last_buffer_dequeued = false; in vb2_clear_last_buffer_dequeued() 1242 if (index < q->num_buffers) in vb2_get_buffer() [all …]
|
/openbmc/linux/drivers/media/common/videobuf2/ |
H A D | videobuf2-v4l2.c | 147 if (q->is_output) { in __copy_timestamp() 686 if (q->allow_cache_hints && q->io_modes & VB2_MMAP) in fill_buf_caps() 839 if (!q->is_output && in vb2_dqbuf() 906 q->is_multiplanar = V4L2_TYPE_IS_MULTIPLANAR(q->type); in vb2_queue_init_name() 907 q->is_output = V4L2_TYPE_IS_OUTPUT(q->type); in vb2_queue_init_name() 918 strscpy(q->name, name, sizeof(q->name)); in vb2_queue_init_name() 920 q->name[0] = '\0'; in vb2_queue_init_name() 943 if (vb2_is_busy(q)) in vb2_queue_change_type() 946 q->type = type; in vb2_queue_change_type() 1190 struct mutex *lock = q->lock ? q->lock : vdev->lock; in vb2_fop_poll() [all …]
|
/openbmc/linux/drivers/net/wireless/broadcom/b43/ |
H A D | pio.h | 111 return b43_read16(q->dev, q->mmio_base + offset); in b43_piotx_read16() 116 return b43_read32(q->dev, q->mmio_base + offset); in b43_piotx_read32() 119 static inline void b43_piotx_write16(struct b43_pio_txqueue *q, in b43_piotx_write16() argument 122 b43_write16(q->dev, q->mmio_base + offset, value); in b43_piotx_write16() 125 static inline void b43_piotx_write32(struct b43_pio_txqueue *q, in b43_piotx_write32() argument 128 b43_write32(q->dev, q->mmio_base + offset, value); in b43_piotx_write32() 134 return b43_read16(q->dev, q->mmio_base + offset); in b43_piorx_read16() 139 return b43_read32(q->dev, q->mmio_base + offset); in b43_piorx_read32() 145 b43_write16(q->dev, q->mmio_base + offset, value); in b43_piorx_write16() 151 b43_write32(q->dev, q->mmio_base + offset, value); in b43_piorx_write32() [all …]
|
/openbmc/linux/drivers/scsi/arm/ |
H A D | queue.c | 43 #define SET_MAGIC(q,m) ((q)->magic = (m)) argument 44 #define BAD_MAGIC(q,m) ((q)->magic != (m)) argument 62 QE_t *q; in queue_initialise() local 75 if (q) { in queue_initialise() 78 q->SCpnt = NULL; in queue_initialise() 111 QE_t *q; in __queue_add() local 125 q->SCpnt = SCpnt; in __queue_add() 140 QE_t *q; in __queue_remove() local 152 return q->SCpnt; in __queue_remove() 220 if (q->SCpnt->device->id == target && q->SCpnt->device->lun == lun && in queue_remove_tgtluntag() [all …]
|
/openbmc/linux/drivers/net/wireless/mediatek/mt76/ |
H A D | sdio.c | 316 q->head = q->tail = 0; in mt76s_alloc_rx_queue() 373 e = &q->entry[q->tail]; in mt76s_get_next_rx_entry() 374 q->tail = (q->tail + 1) % q->ndesc; in mt76s_get_next_rx_entry() 440 if (!q->entry[q->tail].done) in mt76s_process_tx_queue() 443 entry = q->entry[q->tail]; in mt76s_process_tx_queue() 528 if (q->queued == q->ndesc) in mt76s_tx_queue_skb() 537 q->entry[q->head].buf_sz = len; in mt76s_tx_queue_skb() 542 q->head = (q->head + 1) % q->ndesc; in mt76s_tx_queue_skb() 554 if (q->queued == q->ndesc) in mt76s_tx_queue_skb_raw() 565 q->entry[q->head].skb = skb; in mt76s_tx_queue_skb_raw() [all …]
|
H A D | usb.c | 445 urb = q->entry[q->tail].urb; in mt76u_get_next_rx_entry() 446 q->tail = (q->tail + 1) % q->ndesc; in mt76u_get_next_rx_entry() 581 q->head = (q->head + 1) % q->ndesc; in mt76u_complete_rx() 653 q->head = q->tail = 0; in mt76u_submit_rx_buffers() 776 if (!q->entry[q->tail].done) in mt76u_status_worker() 779 entry = q->entry[q->tail]; in mt76u_status_worker() 863 if (q->queued == q->ndesc) in mt76u_tx_queue_skb() 879 q->head = (q->head + 1) % q->ndesc; in mt76u_tx_queue_skb() 892 while (q->first != q->head) { in mt76u_tx_kick() 905 q->first = (q->first + 1) % q->ndesc; in mt76u_tx_kick() [all …]
|
/openbmc/linux/scripts/basic/ |
H A D | fixdep.c | 201 q = p; in parse_config_file() 202 while (isalnum(*q) || *q == '_') in parse_config_file() 203 q++; in parse_config_file() 207 r = q; in parse_config_file() 210 p = q; in parse_config_file() 329 q = p; in parse_dep_file() 330 while (*q != ' ' && *q != '\t' && *q != '\n' && *q != '#' && *q != ':') { in parse_dep_file() 340 if (*(q + 1) == '#' || *(q + 1) == ':') { in parse_dep_file() 345 q++; in parse_dep_file() 350 q++; in parse_dep_file() [all …]
|
/openbmc/linux/kernel/futex/ |
H A D | requeue.c | 87 q->lock_ptr = &hb2->lock; in requeue_futex() 89 q->key = *key2; in requeue_futex() 121 q->pi_state = pi_state; in futex_requeue_pi_prepare() 226 q->key = *key; in requeue_pi_wake_futex() 228 __futex_unqueue(q); in requeue_pi_wake_futex() 230 WARN_ON(!q->rt_waiter); in requeue_pi_wake_futex() 231 q->rt_waiter = NULL; in requeue_pi_wake_futex() 233 q->lock_ptr = &hb->lock; in requeue_pi_wake_futex() 796 q.bitset = bitset; in futex_wait_requeue_pi() 831 if (q.pi_state && (q.pi_state->owner != current)) { in futex_wait_requeue_pi() [all …]
|
/openbmc/linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
H A D | chcr_ipsec.c | 438 pos = q->q.desc; in copy_esn_pktxt() 485 pos = q->q.desc; in copy_cpltx_pktxt() 531 pos = q->q.desc; in copy_key_cpltx_pktxt() 532 left = 64 * q->q.size; in copy_key_cpltx_pktxt() 582 int qid = q->q.cntxt_id; in ch_ipsec_crypto_wreq() 607 q->q.stops++; in ch_ipsec_crypto_wreq() 696 q->q.stops++; in eth_txq_stop() 703 if (q->pidx >= q->size) in txq_advance() 704 q->pidx -= q->size; in txq_advance() 757 last_desc -= q->q.size; in ch_ipsec_xmit() [all …]
|
/openbmc/linux/net/ipv4/ |
H A D | ip_fragment.c | 84 struct ipq *qp = container_of(q, struct ipq, q); in ip4_frag_init() 89 q->key.v4 = *key; in ip4_frag_init() 100 qp = container_of(q, struct ipq, q); in ip4_frag_free() 219 if (!q) in ip_find() 222 return container_of(q, struct ipq, q); in ip_find() 253 if (!mod_timer(&qp->q.timer, jiffies + qp->q.fqdir->timeout)) { in ip_frag_reinit() 262 qp->q.flags = 0; in ip_frag_reinit() 263 qp->q.len = 0; in ip_frag_reinit() 264 qp->q.meat = 0; in ip_frag_reinit() 316 ((qp->q.flags & INET_FRAG_LAST_IN) && end != qp->q.len)) in ip_frag_queue() [all …]
|