/openbmc/linux/lib/ |
H A D | sbitmap.c | 406 static unsigned int sbq_calc_wake_batch(struct sbitmap_queue *sbq, in sbq_calc_wake_batch() argument 428 shallow_depth = min(1U << sbq->sb.shift, sbq->min_shallow_depth); in sbq_calc_wake_batch() 429 depth = ((depth >> sbq->sb.shift) * shallow_depth + in sbq_calc_wake_batch() 430 min(depth & ((1U << sbq->sb.shift) - 1), shallow_depth)); in sbq_calc_wake_batch() 437 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, in sbitmap_queue_init_node() argument 443 ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node, in sbitmap_queue_init_node() 448 sbq->min_shallow_depth = UINT_MAX; in sbitmap_queue_init_node() 449 sbq->wake_batch = sbq_calc_wake_batch(sbq, depth); in sbitmap_queue_init_node() 450 atomic_set(&sbq->wake_index, 0); in sbitmap_queue_init_node() 451 atomic_set(&sbq->ws_active, 0); in sbitmap_queue_init_node() [all …]
|
/openbmc/linux/include/linux/ |
H A D | sbitmap.h | 418 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, 426 static inline void sbitmap_queue_free(struct sbitmap_queue *sbq) in sbitmap_queue_free() argument 428 kfree(sbq->ws); in sbitmap_queue_free() 429 sbitmap_free(&sbq->sb); in sbitmap_queue_free() 440 void sbitmap_queue_recalculate_wake_batch(struct sbitmap_queue *sbq, 452 void sbitmap_queue_resize(struct sbitmap_queue *sbq, unsigned int depth); 461 int __sbitmap_queue_get(struct sbitmap_queue *sbq); 473 unsigned long __sbitmap_queue_get_batch(struct sbitmap_queue *sbq, int nr_tags, 489 int sbitmap_queue_get_shallow(struct sbitmap_queue *sbq, 501 static inline int sbitmap_queue_get(struct sbitmap_queue *sbq, in sbitmap_queue_get() argument [all …]
|
/openbmc/linux/drivers/dma/idxd/ |
H A D | submit.c | 33 struct sbitmap_queue *sbq; in idxd_alloc_desc() local 38 sbq = &wq->sbq; in idxd_alloc_desc() 39 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 47 ws = &sbq->ws[0]; in idxd_alloc_desc() 49 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_INTERRUPTIBLE); in idxd_alloc_desc() 52 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 58 sbitmap_finish_wait(sbq, ws, &wait); in idxd_alloc_desc() 70 sbitmap_queue_clear(&wq->sbq, desc->id, cpu); in idxd_free_desc()
|
H A D | idxd.h | 221 struct sbitmap_queue sbq; member
|
H A D | device.c | 135 rc = sbitmap_queue_init_node(&wq->sbq, num_descs, -1, false, GFP_KERNEL, in idxd_wq_alloc_resources() 175 sbitmap_queue_free(&wq->sbq); in idxd_wq_free_resources()
|
/openbmc/linux/drivers/net/ethernet/fungible/funcore/ |
H A D | fun_dev.c | 339 struct sbitmap_queue *sbq = &fdev->admin_sbq; in fun_wait_for_tag() local 340 struct sbq_wait_state *ws = &sbq->ws[0]; in fun_wait_for_tag() 345 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_UNINTERRUPTIBLE); in fun_wait_for_tag() 350 tag = sbitmap_queue_get(sbq, cpup); in fun_wait_for_tag() 356 sbitmap_finish_wait(sbq, ws, &wait); in fun_wait_for_tag()
|
/openbmc/linux/drivers/staging/qlge/ |
H A D | qlge_main.c | 1145 sbq_fail = !!qlge_refill_bq(&rx_ring->sbq, gfp); in qlge_update_buffer_queues() 1154 if ((sbq_fail && QLGE_BQ_HW_OWNED(&rx_ring->sbq) < 2) || in qlge_update_buffer_queues() 1554 struct qlge_bq_desc *sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_process_mac_rx_skb() 1689 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1719 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1729 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1802 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 2742 struct qlge_bq_desc *sbq_desc = &rx_ring->sbq.queue[i]; in qlge_free_sbq_buffers() 2771 if (rx_ring->sbq.queue) in qlge_free_rx_buffers() 2818 if (rx_ring->sbq.base) { in qlge_free_rx_resources() [all …]
|
H A D | qlge.h | 1441 offsetof(struct rx_ring, sbq) : \ 1481 struct qlge_bq sbq; member
|
/openbmc/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_controlq.c | 62 struct ice_ctl_q_info *cq = &hw->sbq; in ice_sb_init_regs() 622 cq = &hw->sbq; in ice_init_ctrlq() 679 return &hw->sbq; in ice_get_sbq() 701 cq = &hw->sbq; in ice_shutdown_ctrlq() 813 ice_init_ctrlq_locks(&hw->sbq); in ice_create_all_ctrlq() 847 ice_destroy_ctrlq_locks(&hw->sbq); in ice_destroy_all_ctrlq()
|
H A D | ice_type.h | 871 struct ice_ctl_q_info sbq; 870 struct ice_ctl_q_info sbq; global() member
|
H A D | ice_main.c | 1425 cq = &hw->sbq; in __ice_clean_ctrlq() 1636 if (ice_ctrlq_pending(hw, &hw->sbq)) 2406 hw->sbq.num_rq_entries = ICE_SBQ_LEN; 2407 hw->sbq.num_sq_entries = ICE_SBQ_LEN; 2408 hw->sbq.rq_buf_size = ICE_SBQ_MAX_BUF_LEN; 2409 hw->sbq.sq_buf_size = ICE_SBQ_MAX_BUF_LEN; in ice_schedule_reset()
|
/openbmc/linux/drivers/target/iscsi/ |
H A D | iscsi_target_util.c | 130 struct sbitmap_queue *sbq; in iscsit_wait_for_tag() local 135 sbq = &se_sess->sess_tag_pool; in iscsit_wait_for_tag() 136 ws = &sbq->ws[0]; in iscsit_wait_for_tag() 138 sbitmap_prepare_to_wait(sbq, ws, &wait, state); in iscsit_wait_for_tag() 141 tag = sbitmap_queue_get(sbq, cpup); in iscsit_wait_for_tag() 147 sbitmap_finish_wait(sbq, ws, &wait); in iscsit_wait_for_tag()
|
/openbmc/linux/block/ |
H A D | blk-mq.c | 1815 struct sbitmap_queue *sbq; in blk_mq_dispatch_wake() local 1818 sbq = &hctx->tags->bitmap_tags; in blk_mq_dispatch_wake() 1819 atomic_dec(&sbq->ws_active); in blk_mq_dispatch_wake() 1836 struct sbitmap_queue *sbq; in blk_mq_mark_tag_wait() local 1861 sbq = &hctx->tags->breserved_tags; in blk_mq_mark_tag_wait() 1863 sbq = &hctx->tags->bitmap_tags; in blk_mq_mark_tag_wait() 1864 wq = &bt_wait_ptr(sbq, hctx)->wait; in blk_mq_mark_tag_wait() 1874 atomic_inc(&sbq->ws_active); in blk_mq_mark_tag_wait() 1911 atomic_dec(&sbq->ws_active); in blk_mq_mark_tag_wait()
|
H A D | kyber-iosched.c | 496 khd->domain_wait[i].sbq = NULL; in kyber_init_hctx()
|