/openbmc/linux/kernel/events/ |
H A D | ring_buffer.c | 22 atomic_set(&handle->rb->poll, EPOLLIN); in perf_output_wakeup() 38 struct perf_buffer *rb = handle->rb; in perf_output_get_handle() local 46 (*(volatile unsigned int *)&rb->nest)++; in perf_output_get_handle() 47 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle() 52 struct perf_buffer *rb = handle->rb; in perf_output_put_handle() local 60 nest = READ_ONCE(rb->nest); in perf_output_put_handle() 62 WRITE_ONCE(rb->nest, nest - 1); in perf_output_put_handle() 76 head = local_read(&rb->head); in perf_output_put_handle() 110 WRITE_ONCE(rb->user_page->data_head, head); in perf_output_put_handle() 118 WRITE_ONCE(rb->nest, 0); in perf_output_put_handle() [all …]
|
H A D | internal.h | 62 extern void rb_free(struct perf_buffer *rb); 66 struct perf_buffer *rb; in rb_free_rcu() local 68 rb = container_of(rcu_head, struct perf_buffer, rcu_head); in rb_free_rcu() 69 rb_free(rb); in rb_free_rcu() 72 static inline void rb_toggle_paused(struct perf_buffer *rb, bool pause) in rb_toggle_paused() argument 74 if (!pause && rb->nr_pages) in rb_toggle_paused() 75 rb->paused = 0; in rb_toggle_paused() 77 rb->paused = 1; in rb_toggle_paused() 83 extern int rb_alloc_aux(struct perf_buffer *rb, struct perf_event *event, 85 extern void rb_free_aux(struct perf_buffer *rb); [all …]
|
/openbmc/obmc-console/ |
H A D | ringbuffer.c | 31 struct ringbuffer *rb; in ringbuffer_init() local 33 rb = malloc(sizeof(*rb) + size); in ringbuffer_init() 34 if (!rb) { in ringbuffer_init() 38 memset(rb, 0, sizeof(*rb)); in ringbuffer_init() 39 rb->size = size; in ringbuffer_init() 40 rb->buf = (void *)(rb + 1); in ringbuffer_init() 42 return rb; in ringbuffer_init() 45 void ringbuffer_fini(struct ringbuffer *rb) in ringbuffer_fini() argument 47 while (rb->n_consumers) { in ringbuffer_fini() 48 ringbuffer_consumer_unregister(rb->consumers[0]); in ringbuffer_fini() [all …]
|
/openbmc/linux/tools/lib/bpf/ |
H A D | ringbuf.c | 60 static void ringbuf_unmap_ring(struct ring_buffer *rb, struct ring *r) in ringbuf_unmap_ring() argument 63 munmap(r->consumer_pos, rb->page_size); in ringbuf_unmap_ring() 67 munmap(r->producer_pos, rb->page_size + 2 * (r->mask + 1)); in ringbuf_unmap_ring() 73 int ring_buffer__add(struct ring_buffer *rb, int map_fd, in ring_buffer__add() argument 100 tmp = libbpf_reallocarray(rb->rings, rb->ring_cnt + 1, sizeof(*rb->rings)); in ring_buffer__add() 103 rb->rings = tmp; in ring_buffer__add() 105 tmp = libbpf_reallocarray(rb->events, rb->ring_cnt + 1, sizeof(*rb->events)); in ring_buffer__add() 108 rb->events = tmp; in ring_buffer__add() 110 r = &rb->rings[rb->ring_cnt]; in ring_buffer__add() 119 tmp = mmap(NULL, rb->page_size, PROT_READ | PROT_WRITE, MAP_SHARED, map_fd, 0); in ring_buffer__add() [all …]
|
/openbmc/linux/drivers/scsi/bfa/ |
H A D | bfa_ioc_ct.c | 185 void __iomem *rb; in bfa_ioc_ct_reg_init() local 188 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 190 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 191 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 192 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 195 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 196 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 197 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() 198 ioc->ioc_regs.hfn_mbox_cmd = rb + ct_p0reg[pcifn].hfn; in bfa_ioc_ct_reg_init() 199 ioc->ioc_regs.lpu_mbox_cmd = rb + ct_p0reg[pcifn].lpu; in bfa_ioc_ct_reg_init() [all …]
|
H A D | bfa_ioc_cb.c | 138 void __iomem *rb; in bfa_ioc_cb_reg_init() local 141 rb = bfa_ioc_bar0(ioc); in bfa_ioc_cb_reg_init() 143 ioc->ioc_regs.hfn_mbox = rb + iocreg_fnreg[pcifn].hfn_mbox; in bfa_ioc_cb_reg_init() 144 ioc->ioc_regs.lpu_mbox = rb + iocreg_fnreg[pcifn].lpu_mbox; in bfa_ioc_cb_reg_init() 145 ioc->ioc_regs.host_page_num_fn = rb + iocreg_fnreg[pcifn].hfn_pgn; in bfa_ioc_cb_reg_init() 148 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_cb_reg_init() 149 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_cb_reg_init() 150 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_cb_reg_init() 152 ioc->ioc_regs.heartbeat = (rb + BFA_IOC1_HBEAT_REG); in bfa_ioc_cb_reg_init() 153 ioc->ioc_regs.ioc_fwstate = (rb + BFA_IOC1_STATE_REG); in bfa_ioc_cb_reg_init() [all …]
|
/openbmc/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc_ct.c | 49 static enum bfa_status bfa_ioc_ct_pll_init(void __iomem *rb, 51 static enum bfa_status bfa_ioc_ct2_pll_init(void __iomem *rb, 251 void __iomem *rb; in bfa_ioc_ct_reg_init() local 254 rb = bfa_ioc_bar0(ioc); in bfa_ioc_ct_reg_init() 256 ioc->ioc_regs.hfn_mbox = rb + ct_fnreg[pcifn].hfn_mbox; in bfa_ioc_ct_reg_init() 257 ioc->ioc_regs.lpu_mbox = rb + ct_fnreg[pcifn].lpu_mbox; in bfa_ioc_ct_reg_init() 258 ioc->ioc_regs.host_page_num_fn = rb + ct_fnreg[pcifn].hfn_pgn; in bfa_ioc_ct_reg_init() 261 ioc->ioc_regs.heartbeat = rb + BFA_IOC0_HBEAT_REG; in bfa_ioc_ct_reg_init() 262 ioc->ioc_regs.ioc_fwstate = rb + BFA_IOC0_STATE_REG; in bfa_ioc_ct_reg_init() 263 ioc->ioc_regs.alt_ioc_fwstate = rb + BFA_IOC1_STATE_REG; in bfa_ioc_ct_reg_init() [all …]
|
/openbmc/linux/kernel/bpf/ |
H A D | ringbuf.c | 80 struct bpf_ringbuf *rb; member 97 struct bpf_ringbuf *rb; in bpf_ringbuf_area_alloc() local 134 rb = vmap(pages, nr_meta_pages + 2 * nr_data_pages, in bpf_ringbuf_area_alloc() 136 if (rb) { in bpf_ringbuf_area_alloc() 138 rb->pages = pages; in bpf_ringbuf_area_alloc() 139 rb->nr_pages = nr_pages; in bpf_ringbuf_area_alloc() 140 return rb; in bpf_ringbuf_area_alloc() 152 struct bpf_ringbuf *rb = container_of(work, struct bpf_ringbuf, work); in bpf_ringbuf_notify() local 154 wake_up_all(&rb->waitq); in bpf_ringbuf_notify() 170 struct bpf_ringbuf *rb; in bpf_ringbuf_alloc() local [all …]
|
/openbmc/linux/drivers/hid/intel-ish-hid/ishtp/ |
H A D | client-buffers.c | 23 struct ishtp_cl_rb *rb; in ishtp_cl_alloc_rx_ring() local 28 rb = ishtp_io_rb_init(cl); in ishtp_cl_alloc_rx_ring() 29 if (!rb) { in ishtp_cl_alloc_rx_ring() 33 ret = ishtp_io_rb_alloc_buf(rb, len); in ishtp_cl_alloc_rx_ring() 37 list_add_tail(&rb->list, &cl->free_rb_list.list); in ishtp_cl_alloc_rx_ring() 99 struct ishtp_cl_rb *rb; in ishtp_cl_free_rx_ring() local 105 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, in ishtp_cl_free_rx_ring() 107 list_del(&rb->list); in ishtp_cl_free_rx_ring() 108 kfree(rb->buffer.data); in ishtp_cl_free_rx_ring() 109 kfree(rb); in ishtp_cl_free_rx_ring() [all …]
|
H A D | client.c | 44 struct ishtp_cl_rb *rb; in ishtp_read_list_flush() local 49 list_for_each_entry_safe(rb, next, &cl->dev->read_list.list, list) in ishtp_read_list_flush() 50 if (rb->cl && ishtp_cl_cmp_id(cl, rb->cl)) { in ishtp_read_list_flush() 51 list_del(&rb->list); in ishtp_read_list_flush() 52 ishtp_io_rb_free(rb); in ishtp_read_list_flush() 447 struct ishtp_cl_rb *rb; in ishtp_cl_read_start() local 477 rb = NULL; in ishtp_cl_read_start() 481 rb = list_entry(cl->free_rb_list.list.next, struct ishtp_cl_rb, list); in ishtp_cl_read_start() 482 list_del_init(&rb->list); in ishtp_cl_read_start() 485 rb->cl = cl; in ishtp_cl_read_start() [all …]
|
/openbmc/linux/lib/ |
H A D | rbtree_test.c | 20 struct rb_node rb; member 39 if (key < rb_entry(parent, struct test_node, rb)->key) in insert() 45 rb_link_node(&node->rb, parent, new); in insert() 46 rb_insert_color(&node->rb, &root->rb_root); in insert() 57 if (key < rb_entry(parent, struct test_node, rb)->key) in insert_cached() 65 rb_link_node(&node->rb, parent, new); in insert_cached() 66 rb_insert_color_cached(&node->rb, root, leftmost); in insert_cached() 71 rb_erase(&node->rb, &root->rb_root); in erase() 76 rb_erase_cached(&node->rb, root); in erase_cached() 83 struct test_node, rb, u32, augmented, NODE_VAL) in RB_DECLARE_CALLBACKS_MAX() argument [all …]
|
/openbmc/qemu/target/ppc/translate/ |
H A D | dfp-impl.c.inc | 13 TCGv_ptr rt, ra, rb; \ 18 rb = gen_fprp_ptr(a->rb); \ 19 gen_helper_##NAME(tcg_env, rt, ra, rb); \ 29 TCGv_ptr ra, rb; \ 33 rb = gen_fprp_ptr(a->rb); \ 35 tcg_env, ra, rb); \ 42 TCGv_ptr rb; \ 45 rb = gen_fprp_ptr(a->rb); \ 47 tcg_env, tcg_constant_i32(a->uim), rb);\ 66 TCGv_ptr rt, rb; \ [all …]
|
/openbmc/qemu/target/openrisc/ |
H A D | sys_helper.c | 42 void HELPER(mtspr)(CPUOpenRISCState *env, target_ulong spr, target_ulong rb) in HELPER() 54 cpu_set_fpcsr(env, rb); in HELPER() 65 env->evbar = rb; in HELPER() 72 if (env->pc != rb) { in HELPER() 73 env->pc = rb; in HELPER() 80 cpu_set_sr(env, rb); in HELPER() 84 env->epcr = rb; in HELPER() 88 env->eear = rb; in HELPER() 92 env->esr = rb; in HELPER() 97 env->shadow_gpr[idx / 32][idx % 32] = rb; in HELPER() [all …]
|
/openbmc/linux/kernel/printk/ |
H A D | printk_ringbuffer.c | 564 static bool data_make_reusable(struct printk_ringbuffer *rb, in data_make_reusable() argument 570 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_make_reusable() 571 struct prb_desc_ring *desc_ring = &rb->desc_ring; in data_make_reusable() 632 static bool data_push_tail(struct printk_ringbuffer *rb, unsigned long lpos) in data_push_tail() argument 634 struct prb_data_ring *data_ring = &rb->text_data_ring; in data_push_tail() 675 if (!data_make_reusable(rb, tail_lpos, lpos, &next_lpos)) { in data_push_tail() 767 static bool desc_push_tail(struct printk_ringbuffer *rb, in desc_push_tail() argument 770 struct prb_desc_ring *desc_ring = &rb->desc_ring; in desc_push_tail() 812 if (!data_push_tail(rb, desc.text_blk_lpos.next)) in desc_push_tail() 875 static bool desc_reserve(struct printk_ringbuffer *rb, unsigned long *id_out) in desc_reserve() argument [all …]
|
/openbmc/qemu/target/alpha/ |
H A D | translate.c | 314 static void gen_load_fp(DisasContext *ctx, int ra, int rb, int32_t disp16, in gen_load_fp() argument 320 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); in gen_load_fp() 325 static void gen_load_int(DisasContext *ctx, int ra, int rb, int32_t disp16, in gen_load_int() argument 338 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); in gen_load_int() 380 static void gen_store_fp(DisasContext *ctx, int ra, int rb, int32_t disp16, in gen_store_fp() argument 384 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); in gen_store_fp() 388 static void gen_store_int(DisasContext *ctx, int ra, int rb, int32_t disp16, in gen_store_int() argument 394 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); in gen_store_int() 405 static DisasJumpType gen_store_conditional(DisasContext *ctx, int ra, int rb, in gen_store_conditional() argument 413 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16); in gen_store_conditional() [all …]
|
/openbmc/qemu/migration/ |
H A D | postcopy-ram.c | 340 static int test_ramblock_postcopiable(RAMBlock *rb, Error **errp) in test_ramblock_postcopiable() argument 342 const char *block_name = qemu_ram_get_idstr(rb); in test_ramblock_postcopiable() 343 ram_addr_t length = qemu_ram_get_used_length(rb); in test_ramblock_postcopiable() 344 size_t pagesize = qemu_ram_pagesize(rb); in test_ramblock_postcopiable() 355 if (rb->fd >= 0) { in test_ramblock_postcopiable() 356 fs = qemu_fd_getfs(rb->fd); in test_ramblock_postcopiable() 488 static int init_range(RAMBlock *rb, void *opaque) in init_range() argument 490 const char *block_name = qemu_ram_get_idstr(rb); in init_range() 491 void *host_addr = qemu_ram_get_host_addr(rb); in init_range() 492 ram_addr_t offset = qemu_ram_get_offset(rb); in init_range() [all …]
|
/openbmc/qemu/include/exec/ |
H A D | cpu-common.h | 92 ram_addr_t qemu_ram_block_host_offset(RAMBlock *rb, void *host); 95 const char *qemu_ram_get_idstr(RAMBlock *rb); 96 void *qemu_ram_get_host_addr(RAMBlock *rb); 97 ram_addr_t qemu_ram_get_offset(RAMBlock *rb); 98 ram_addr_t qemu_ram_get_used_length(RAMBlock *rb); 99 ram_addr_t qemu_ram_get_max_length(RAMBlock *rb); 100 bool qemu_ram_is_shared(RAMBlock *rb); 101 bool qemu_ram_is_noreserve(RAMBlock *rb); 102 bool qemu_ram_is_uf_zeroable(RAMBlock *rb); 103 void qemu_ram_set_uf_zeroable(RAMBlock *rb); [all …]
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_mm.c | 152 INTERVAL_TREE_DEFINE(struct drm_mm_node, rb, in INTERVAL_TREE_DEFINE() argument 168 struct rb_node **link, *rb; in drm_mm_interval_tree_add_node() local 175 rb = &hole_node->rb; in drm_mm_interval_tree_add_node() 176 while (rb) { in drm_mm_interval_tree_add_node() 177 parent = rb_entry(rb, struct drm_mm_node, rb); in drm_mm_interval_tree_add_node() 182 rb = rb_parent(rb); in drm_mm_interval_tree_add_node() 185 rb = &hole_node->rb; in drm_mm_interval_tree_add_node() 186 link = &hole_node->rb.rb_right; in drm_mm_interval_tree_add_node() 189 rb = NULL; in drm_mm_interval_tree_add_node() 195 rb = *link; in drm_mm_interval_tree_add_node() [all …]
|
H A D | drm_prime.c | 100 struct rb_node **p, *rb; in drm_prime_add_buf_handle() local 110 rb = NULL; in drm_prime_add_buf_handle() 115 rb = *p; in drm_prime_add_buf_handle() 116 pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_add_buf_handle() 118 p = &rb->rb_right; in drm_prime_add_buf_handle() 120 p = &rb->rb_left; in drm_prime_add_buf_handle() 122 rb_link_node(&member->dmabuf_rb, rb, p); in drm_prime_add_buf_handle() 125 rb = NULL; in drm_prime_add_buf_handle() 130 rb = *p; in drm_prime_add_buf_handle() 131 pos = rb_entry(rb, struct drm_prime_member, handle_rb); in drm_prime_add_buf_handle() [all …]
|
/openbmc/linux/mm/ |
H A D | interval_tree.c | 23 INTERVAL_TREE_DEFINE(struct vm_area_struct, shared.rb, 38 if (!prev->shared.rb.rb_right) { in vma_interval_tree_insert_after() 40 link = &prev->shared.rb.rb_right; in vma_interval_tree_insert_after() 42 parent = rb_entry(prev->shared.rb.rb_right, in vma_interval_tree_insert_after() 43 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 46 while (parent->shared.rb.rb_left) { in vma_interval_tree_insert_after() 47 parent = rb_entry(parent->shared.rb.rb_left, in vma_interval_tree_insert_after() 48 struct vm_area_struct, shared.rb); in vma_interval_tree_insert_after() 52 link = &parent->shared.rb.rb_left; in vma_interval_tree_insert_after() 56 rb_link_node(&node->shared.rb, &parent->shared.rb, link); in vma_interval_tree_insert_after() [all …]
|
/openbmc/linux/drivers/misc/mchp_pci1xxxx/ |
H A D | mchp_pci1xxxx_otpe2p.c | 99 void __iomem *rb = priv->reg_base; in is_eeprom_responsive() local 104 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 106 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 111 true, rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in is_eeprom_responsive() 124 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_read() local 141 writel(EEPROM_CMD_EPC_BUSY_BIT | (off + byte), rb + in pci1xxxx_eeprom_read() 148 rb + MMAP_EEPROM_OFFSET(EEPROM_CMD_REG)); in pci1xxxx_eeprom_read() 154 buf[byte] = readl(rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_read() 165 void __iomem *rb = priv->reg_base; in pci1xxxx_eeprom_write() local 182 writel(*(value + byte), rb + MMAP_EEPROM_OFFSET(EEPROM_DATA_REG)); in pci1xxxx_eeprom_write() [all …]
|
/openbmc/linux/Documentation/translations/zh_CN/core-api/ |
H A D | rbtree.rst | 271 node = rb_entry(root->rb_node, struct interval_tree_node, rb); 274 if (node->rb.rb_left) { 276 rb_entry(node->rb.rb_left, 277 struct interval_tree_node, rb); 294 if (node->rb.rb_right) { 295 node = rb_entry(node->rb.rb_right, 296 struct interval_tree_node, rb); 311 if (node->rb.rb_left) { 312 subtree_last = rb_entry(node->rb.rb_left, 313 struct interval_tree_node, rb)->__subtree_last; [all …]
|
/openbmc/linux/arch/arm64/crypto/ |
H A D | sm3-neon-core.S | 42 #define rb w4 macro 356 ldp ra, rb, [RSTATE, #0] 401 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 0, 0, IW, _, 0) 402 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 1, 1, IW, _, 0) 403 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 2, 2, IW, _, 0) 404 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 3, 3, IW, _, 0) 407 R1(ra, rb, rc, rd, re, rf, rg, rh, k_even, KL, 4, 0, IW, _, 0) 408 R1(rd, ra, rb, rc, rh, re, rf, rg, k_odd, _, 5, 1, IW, _, 0) 409 R1(rc, rd, ra, rb, rg, rh, re, rf, k_even, KL, 6, 2, IW, SCHED_W_W0W1W2W3W4W5_1, 12) 410 R1(rb, rc, rd, ra, rf, rg, rh, re, k_odd, _, 7, 3, IW, SCHED_W_W0W1W2W3W4W5_2, 12) [all …]
|
/openbmc/linux/drivers/target/iscsi/ |
H A D | iscsi_target_configfs.c | 44 ssize_t rb; in lio_target_np_driver_show() local 48 rb = sysfs_emit(page, "1\n"); in lio_target_np_driver_show() 50 rb = sysfs_emit(page, "0\n"); in lio_target_np_driver_show() 52 return rb; in lio_target_np_driver_show() 474 ssize_t rb; \ 479 rb = snprintf(page, PAGE_SIZE, \ 483 rb = snprintf(page, PAGE_SIZE, "%u\n", \ 488 return rb; \ 530 ssize_t rb = 0; in lio_target_nacl_info_show() local 536 rb += sysfs_emit_at(page, rb, "No active iSCSI Session for Initiator" in lio_target_nacl_info_show() [all …]
|
/openbmc/linux/drivers/gpu/drm/amd/display/dmub/inc/ |
H A D | dmub_cmd.h | 4225 static inline bool dmub_rb_empty(struct dmub_rb *rb) in dmub_rb_empty() argument 4227 return (rb->wrpt == rb->rptr); in dmub_rb_empty() 4237 static inline bool dmub_rb_full(struct dmub_rb *rb) in dmub_rb_full() argument 4241 if (rb->wrpt >= rb->rptr) in dmub_rb_full() 4242 data_count = rb->wrpt - rb->rptr; in dmub_rb_full() 4244 data_count = rb->capacity - (rb->rptr - rb->wrpt); in dmub_rb_full() 4246 return (data_count == (rb->capacity - DMUB_RB_CMD_SIZE)); in dmub_rb_full() 4257 static inline bool dmub_rb_push_front(struct dmub_rb *rb, in dmub_rb_push_front() argument 4260 uint64_t volatile *dst = (uint64_t volatile *)((uint8_t *)(rb->base_address) + rb->wrpt); in dmub_rb_push_front() 4264 if (dmub_rb_full(rb)) in dmub_rb_push_front() [all …]
|