/openbmc/linux/arch/alpha/include/asm/ |
H A D | core_wildfire.h | 227 #define WILDFIRE_QBB(q) ((~((long)(q)) & WILDFIRE_QBB_MASK) << 36) argument 230 #define WILDFIRE_QBB_IO(q) (WILDFIRE_BASE | WILDFIRE_QBB(q)) argument 231 #define WILDFIRE_QBB_HOSE(q,h) (WILDFIRE_QBB_IO(q) | WILDFIRE_HOSE(h)) argument 233 #define WILDFIRE_MEM(q,h) (WILDFIRE_QBB_HOSE(q,h) | 0x000000000UL) argument 234 #define WILDFIRE_CONF(q,h) (WILDFIRE_QBB_HOSE(q,h) | 0x1FE000000UL) argument 235 #define WILDFIRE_IO(q,h) (WILDFIRE_QBB_HOSE(q,h) | 0x1FF000000UL) argument 237 #define WILDFIRE_qsd(q) \ argument 243 #define WILDFIRE_qsa(q) \ argument 246 #define WILDFIRE_iop(q) \ argument 249 #define WILDFIRE_gp(q) \ argument [all …]
|
/openbmc/linux/drivers/platform/chrome/wilco_ec/ |
H A D | event.c | 109 q = kzalloc(struct_size(q, entries, capacity), GFP_KERNEL); in event_queue_new() 110 if (!q) in event_queue_new() 115 return q; in event_queue_new() 121 return q->head == q->tail && !q->entries[q->head]; in event_queue_empty() 127 return q->head == q->tail && q->entries[q->head]; in event_queue_full() 137 ev = q->entries[q->tail]; in event_queue_pop() 138 q->entries[q->tail] = NULL; in event_queue_pop() 139 q->tail = (q->tail + 1) % q->capacity; in event_queue_pop() 155 q->entries[q->head] = ev; in event_queue_push() 156 q->head = (q->head + 1) % q->capacity; in event_queue_push() [all …]
|
/openbmc/linux/include/trace/events/ |
H A D | vb2.h | 12 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 13 TP_ARGS(q, vb), 26 __entry->owner = q->owner; 27 __entry->queued_count = q->queued_count; 29 atomic_read(&q->owned_by_drv_count); 47 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 48 TP_ARGS(q, vb) 52 TP_PROTO(struct vb2_queue *q, struct vb2_buffer *vb), 53 TP_ARGS(q, vb) 58 TP_ARGS(q, vb) [all …]
|
/openbmc/qemu/tests/tcg/s390x/ |
H A D | div.c | 8 int32_t b = 101, q, r; in test_dr() local 14 q = r1; in test_dr() 16 assert(q == -41); in test_dr() 24 uint32_t b = 101, q, r; in test_dlr() local 30 q = r1; in test_dlr() 32 assert(q == 42); in test_dlr() 40 int64_t b = 101, q, r; in test_dsgr() local 46 q = r1; in test_dsgr() 48 assert(q == -41); in test_dsgr() 62 q = r1; in test_dlgr() [all …]
|
/openbmc/linux/drivers/net/ethernet/marvell/octeon_ep/ |
H A D | octep_ctrl_mbox.c | 139 qbuf = (q->hw_q + *pi); in octep_write_mbox_data() 153 qbuf = (q->hw_q + *pi); in octep_write_mbox_data() 174 q = &mbox->h2fq; in octep_ctrl_mbox_send() 175 pi = readl(q->hw_prod); in octep_ctrl_mbox_send() 176 ci = readl(q->hw_cons); in octep_ctrl_mbox_send() 191 writel(pi, q->hw_prod); in octep_ctrl_mbox_send() 204 qbuf = (q->hw_q + *ci); in octep_read_mbox_data() 236 q = &mbox->f2hq; in octep_ctrl_mbox_recv() 237 pi = readl(q->hw_prod); in octep_ctrl_mbox_recv() 238 ci = readl(q->hw_cons); in octep_ctrl_mbox_recv() [all …]
|
/openbmc/linux/block/ |
H A D | blk-flush.c | 167 struct request_queue *q = rq->q; in blk_flush_complete_seq() local 191 spin_lock(&q->requeue_lock); in blk_flush_complete_seq() 219 struct request_queue *q = flush_rq->q; in flush_end_io() local 246 if (!q->elevator) { in flush_end_io() 314 blk_rq_init(q, flush_rq); in blk_kick_flush() 327 if (!q->elevator) { in blk_kick_flush() 352 spin_lock(&q->requeue_lock); in blk_kick_flush() 356 blk_mq_kick_requeue_list(q); in blk_kick_flush() 362 struct request_queue *q = rq->q; in mq_flush_data_end_io() local 368 if (q->elevator) { in mq_flush_data_end_io() [all …]
|
H A D | blk-merge.c | 593 struct request_queue *q = rq->q; in blk_rq_get_max_sectors() local 600 if (!q->limits.chunk_sectors || in blk_rq_get_max_sectors() 632 req_set_nomerge(req->q, req); in ll_new_hw_segment() 647 req_set_nomerge(req->q, req); in ll_back_merge_fn() 666 req_set_nomerge(req->q, req); in ll_front_merge_fn() 687 req_set_nomerge(q, req); in req_attempt_discard_merge() 988 rq_qos_merge(req->q, req, bio); in bio_attempt_back_merge() 1014 rq_qos_merge(req->q, req, bio); in bio_attempt_front_merge() 1044 rq_qos_merge(q, req, bio); in bio_attempt_discard_merge() 1054 req_set_nomerge(q, req); in bio_attempt_discard_merge() [all …]
|
H A D | blk-mq-sched.h | 10 bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, 12 bool blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio, 21 int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e); 22 void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e); 23 void blk_mq_sched_free_rqs(struct request_queue *q); 37 blk_mq_sched_allow_merge(struct request_queue *q, struct request *rq, in blk_mq_sched_allow_merge() argument 41 struct elevator_queue *e = q->elevator; in blk_mq_sched_allow_merge() 44 return e->type->ops.allow_merge(q, rq, bio); in blk_mq_sched_allow_merge() 52 struct elevator_queue *e = rq->q->elevator; in blk_mq_sched_completed_request() 62 struct request_queue *q = rq->q; in blk_mq_sched_requeue_request() local [all …]
|
/openbmc/linux/lib/raid6/ |
H A D | recov_avx2.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx2() local 25 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_avx2() 43 ptrs[disks-1] = q; in raid6_2data_recov_avx2() 57 asm volatile("vmovdqa %0, %%ymm1" : : "m" (q[0])); in raid6_2data_recov_avx2() 130 q += 64; in raid6_2data_recov_avx2() 177 q += 32; in raid6_2data_recov_avx2() 189 u8 *p, *q, *dq; in raid6_datap_recov_avx2() local 194 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_avx2() 206 ptrs[disks-1] = q; in raid6_datap_recov_avx2() 265 q += 64; in raid6_datap_recov_avx2() [all …]
|
H A D | loongarch_simd.c | 36 u8 *p, *q; in raid6_lsx_gen_syndrome() local 41 q = dptr[z0+2]; /* RS syndrome */ in raid6_lsx_gen_syndrome() 118 u8 *p, *q; in raid6_lsx_xor_syndrome() local 123 q = dptr[disks-1]; /* RS syndrome */ in raid6_lsx_xor_syndrome() 237 "+m"(q[d+NSIZE*0]), "+m"(q[d+NSIZE*1]), in raid6_lsx_xor_syndrome() 238 "+m"(q[d+NSIZE*2]), "+m"(q[d+NSIZE*3]) in raid6_lsx_xor_syndrome() 267 u8 *p, *q; in raid6_lasx_gen_syndrome() local 272 q = dptr[z0+2]; /* RS syndrome */ in raid6_lasx_gen_syndrome() 327 u8 *p, *q; in raid6_lasx_xor_syndrome() local 332 q = dptr[disks-1]; /* RS syndrome */ in raid6_lasx_xor_syndrome() [all …]
|
H A D | recov_ssse3.c | 19 u8 *p, *q, *dp, *dq; in raid6_2data_recov_ssse3() local 27 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_ssse3() 45 ptrs[disks-1] = q; in raid6_2data_recov_ssse3() 67 asm volatile("movdqa %0,%%xmm1" : : "m" (q[0])); in raid6_2data_recov_ssse3() 134 q += 32; in raid6_2data_recov_ssse3() 181 q += 16; in raid6_2data_recov_ssse3() 194 u8 *p, *q, *dq; in raid6_datap_recov_ssse3() local 201 q = (u8 *)ptrs[disks-1]; in raid6_datap_recov_ssse3() 213 ptrs[disks-1] = q; in raid6_datap_recov_ssse3() 279 q += 32; in raid6_datap_recov_ssse3() [all …]
|
H A D | recov_avx512.c | 27 u8 *p, *q, *dp, *dq; in raid6_2data_recov_avx512() local 33 q = (u8 *)ptrs[disks-1]; in raid6_2data_recov_avx512() 54 ptrs[disks-1] = q; in raid6_2data_recov_avx512() 77 : "m" (q[0]), "m" (q[64]), "m" (p[0]), in raid6_2data_recov_avx512() 157 q += 128; in raid6_2data_recov_avx512() 218 q += 64; in raid6_2data_recov_avx512() 230 u8 *p, *q, *dq; in raid6_datap_recov_avx512() local 250 ptrs[disks-1] = q; in raid6_datap_recov_avx512() 267 "m" (q[64])); in raid6_datap_recov_avx512() 321 q += 128; in raid6_datap_recov_avx512() [all …]
|
/openbmc/qemu/accel/tcg/ |
H A D | perf.c | 80 if (!q->symbol) { in pretty_symbol() 88 if (!q->offset) { in pretty_symbol() 92 return q->symbol; in pretty_symbol() 95 tmp = snprintf(buf, sizeof(buf), "%s+0x%"PRIx64, q->symbol, q->offset); in pretty_symbol() 261 if (q[insn].file) { in write_jr_code_debug_info() 270 if (q[insn].file) { in write_jr_code_debug_info() 276 fwrite(q[insn].file, strlen(q[insn].file) + 1, 1, jitdump); in write_jr_code_debug_info() 315 struct debuginfo_query *q; in perf_report_code() local 323 q = g_try_malloc0_n(tb->icount, sizeof(*q)); in perf_report_code() 324 if (!q) { in perf_report_code() [all …]
|
/openbmc/linux/drivers/scsi/be2iscsi/ |
H A D | be.h | 51 static inline void *queue_head_node(struct be_queue_info *q) in queue_head_node() argument 53 return q->dma_mem.va + q->head * q->entry_size; in queue_head_node() 58 return q->dma_mem.va + wrb_num * q->entry_size; in queue_get_wrb() 61 static inline void *queue_tail_node(struct be_queue_info *q) in queue_tail_node() argument 63 return q->dma_mem.va + q->tail * q->entry_size; in queue_tail_node() 66 static inline void queue_head_inc(struct be_queue_info *q) in queue_head_inc() argument 68 index_inc(&q->head, q->len); in queue_head_inc() 71 static inline void queue_tail_inc(struct be_queue_info *q) in queue_tail_inc() argument 73 index_inc(&q->tail, q->len); in queue_tail_inc() 89 struct be_queue_info q; member [all …]
|
/openbmc/linux/net/ipv6/netfilter/ |
H A D | nf_conntrack_reasm.c | 159 if (!q) in fq_find() 162 return container_of(q, struct frag_queue, q); in fq_find() 205 if (end < fq->q.len || in nf_ct_frag6_queue() 206 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq->q.len)) { in nf_ct_frag6_queue() 211 fq->q.len = end; in nf_ct_frag6_queue() 224 if (end > fq->q.len) { in nf_ct_frag6_queue() 230 fq->q.len = end; in nf_ct_frag6_queue() 268 fq->q.meat += skb->len; in nf_ct_frag6_queue() 283 fq->q.meat == fq->q.len) { in nf_ct_frag6_queue() 301 inet_frag_kill(&fq->q); in nf_ct_frag6_queue() [all …]
|
/openbmc/linux/lib/ |
H A D | ts_kmp.c | 45 unsigned int i, q = 0, text_len, consumed = state->offset; in kmp_find() local 56 while (q > 0 && kmp->pattern[q] in kmp_find() 58 q = kmp->prefix_tbl[q - 1]; in kmp_find() 59 if (kmp->pattern[q] in kmp_find() 61 q++; in kmp_find() 62 if (unlikely(q == kmp->pattern_len)) { in kmp_find() 77 unsigned int k, q; in compute_prefix_tbl() local 80 for (k = 0, q = 1; q < len; q++) { in compute_prefix_tbl() 82 != (icase ? toupper(pattern[q]) : pattern[q])) in compute_prefix_tbl() 85 == (icase ? toupper(pattern[q]) : pattern[q])) in compute_prefix_tbl() [all …]
|
/openbmc/linux/drivers/net/fddi/skfp/ |
H A D | queue.c | 28 smc->q.ev_put = smc->q.ev_get = smc->q.ev_queue ; in ev_init() 37 smc->q.ev_put->class = class ; in queue_event() 38 smc->q.ev_put->event = event ; in queue_event() 39 if (++smc->q.ev_put == &smc->q.ev_queue[MAX_EVENT]) in queue_event() 40 smc->q.ev_put = smc->q.ev_queue ; in queue_event() 42 if (smc->q.ev_put == smc->q.ev_get) { in queue_event() 70 ev = smc->q.ev_get ; in ev_dispatcher() 72 while (ev != smc->q.ev_put) { in ev_dispatcher() 104 if (++ev == &smc->q.ev_queue[MAX_EVENT]) in ev_dispatcher() 105 ev = smc->q.ev_queue ; in ev_dispatcher() [all …]
|
/openbmc/qemu/hw/arm/ |
H A D | smmuv3-internal.h | 216 #define WRAP_MASK(q) (1 << (q)->log2size) argument 220 #define Q_CONS(q) ((q)->cons & INDEX_MASK(q)) argument 221 #define Q_PROD(q) ((q)->prod & INDEX_MASK(q)) argument 223 #define Q_CONS_ENTRY(q) (Q_BASE(q) + (q)->entry_size * Q_CONS(q)) argument 224 #define Q_PROD_ENTRY(q) (Q_BASE(q) + (q)->entry_size * Q_PROD(q)) argument 226 #define Q_CONS_WRAP(q) (((q)->cons & WRAP_MASK(q)) >> (q)->log2size) argument 227 #define Q_PROD_WRAP(q) (((q)->prod & WRAP_MASK(q)) >> (q)->log2size) argument 231 return ((q->cons ^ q->prod) & WRAP_INDEX_MASK(q)) == WRAP_MASK(q); in smmuv3_q_full() 236 return (q->cons & WRAP_INDEX_MASK(q)) == (q->prod & WRAP_INDEX_MASK(q)); in smmuv3_q_empty() 241 q->prod = (q->prod + 1) & WRAP_INDEX_MASK(q); in queue_prod_incr() [all …]
|
/openbmc/linux/drivers/net/wireless/ath/ath9k/ |
H A D | mac.c | 137 int i, q; in ath9k_hw_abort_tx_dma() local 152 for (q = 0; q < AR_NUM_QCU; q++) { in ath9k_hw_abort_tx_dma() 203 qi = &ah->txq[q]; in ath9k_hw_set_txq_props() 267 qi = &ah->txq[q]; in ath9k_hw_get_txq_props() 298 int q; in ath9k_hw_setuptxqueue() local 308 q = 1; in ath9k_hw_setuptxqueue() 314 q = qinfo->tqi_subtype; in ath9k_hw_setuptxqueue() 323 qi = &ah->txq[q]; in ath9k_hw_setuptxqueue() 333 return q; in ath9k_hw_setuptxqueue() 351 qi = &ah->txq[q]; in ath9k_hw_releasetxqueue() [all …]
|
/openbmc/linux/drivers/firmware/arm_scmi/ |
H A D | raw_mode.c | 253 return raw->q[idx]; in scmi_raw_queue_select() 707 if (wait_event_interruptible(q->wq, !list_empty(&q->msg_q))) in scmi_raw_message_dequeue() 742 struct scmi_raw_queue *q; in scmi_raw_message_receive() local 745 if (!q) in scmi_raw_message_receive() 853 if (!q) in scmi_test_dbg_raw_common_poll() 1040 q = devm_kzalloc(dev, sizeof(*q), GFP_KERNEL); in scmi_raw_queue_init() 1041 if (!q) in scmi_raw_queue_init() 1062 return q; in scmi_raw_queue_init() 1122 if (IS_ERR(q)) { in scmi_raw_mode_setup() 1123 ret = PTR_ERR(q); in scmi_raw_mode_setup() [all …]
|
/openbmc/linux/include/net/ |
H A D | ipv6_frag.h | 24 struct inet_frag_queue q; member 34 struct frag_queue *fq = container_of(q, struct frag_queue, q); in ip6frag_init() 37 q->key.v6 = *key; in ip6frag_init() 72 if (READ_ONCE(fq->q.fqdir->dead)) in ip6frag_expire_frag_queue() 74 spin_lock(&fq->q.lock); in ip6frag_expire_frag_queue() 79 fq->q.flags |= INET_FRAG_DROP; in ip6frag_expire_frag_queue() 80 inet_frag_kill(&fq->q); in ip6frag_expire_frag_queue() 97 head = inet_frag_pull_head(&fq->q); in ip6frag_expire_frag_queue() 102 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue() 109 spin_unlock(&fq->q.lock); in ip6frag_expire_frag_queue() [all …]
|
/openbmc/qemu/hw/input/ |
H A D | ps2.c | 169 q->data[q->wptr] = b; in ps2_queue_noirq() 234 q->data[q->cwptr] = b; in ps2_cqueue_data() 245 q->rptr = (q->rptr - 1) & (PS2_BUFFER_SIZE - 1); in ps2_cqueue_1() 246 q->cwptr = q->rptr; in ps2_cqueue_1() 256 q->cwptr = q->rptr; in ps2_cqueue_2() 267 q->cwptr = q->rptr; in ps2_cqueue_3() 285 q->rptr = q->cwptr; in ps2_cqueue_reset() 559 val = q->data[q->rptr]; in ps2_read_data() 564 if (q->rptr == q->cwptr) { in ps2_read_data() 1046 q->rptr = q->rptr & (PS2_BUFFER_SIZE - 1); in ps2_common_post_load() [all …]
|
/openbmc/linux/net/ipv6/ |
H A D | reassembly.c | 100 if (!q) in fq_find() 103 return container_of(q, struct frag_queue, q); in fq_find() 151 if (end < fq->q.len || in ip6_frag_queue() 152 ((fq->q.flags & INET_FRAG_LAST_IN) && end != fq->q.len)) in ip6_frag_queue() 155 fq->q.len = end; in ip6_frag_queue() 167 if (end > fq->q.len) { in ip6_frag_queue() 171 fq->q.len = end; in ip6_frag_queue() 202 fq->q.meat += skb->len; in ip6_frag_queue() 219 fq->q.meat == fq->q.len) { in ip6_frag_queue() 241 inet_frag_kill(&fq->q); in ip6_frag_queue() [all …]
|
/openbmc/linux/tools/testing/selftests/bpf/progs/ |
H A D | test_spin_lock.c | 55 struct bpf_vqueue *q; in bpf_spin_lock_test() local 80 q = bpf_map_lookup_elem(&vqueue, &key); in bpf_spin_lock_test() 81 if (!q) in bpf_spin_lock_test() 84 bpf_spin_lock(&q->lock); in bpf_spin_lock_test() 85 q->credit += CREDIT_PER_NS(curtime - q->lasttime, q->rate); in bpf_spin_lock_test() 86 q->lasttime = curtime; in bpf_spin_lock_test() 87 if (q->credit > max_credit) in bpf_spin_lock_test() 88 q->credit = max_credit; in bpf_spin_lock_test() 89 q->credit -= pkt_len; in bpf_spin_lock_test() 90 credit = q->credit; in bpf_spin_lock_test() [all …]
|
/openbmc/qemu/tests/tcg/riscv64/ |
H A D | test-div.c | 5 long x, y, q, r; member 17 unsigned long x, y, q, r; member 34 long q, r; in main() local 38 : "=&r" (q), "=r" (r) in main() 41 assert(q == test_s[i].q); in main() 46 unsigned long q, r; in main() local 50 : "=&r" (q), "=r" (r) in main() 53 assert(q == test_u[i].q); in main()
|