/openbmc/linux/drivers/soc/ti/ |
H A D | knav_qmss_acc.c | 283 cmd->command, cmd->queue_mask, cmd->list_dma, in knav_acc_write() 289 writel_relaxed(cmd->queue_mask, &pdsp->acc_command->queue_mask); in knav_acc_write() 308 u32 queue_mask; in knav_acc_setup_cmd() local 313 queue_mask = BIT(range->num_queues) - 1; in knav_acc_setup_cmd() 317 queue_mask = 0; in knav_acc_setup_cmd() 322 cmd->queue_mask = queue_mask; in knav_acc_setup_cmd()
|
H A D | knav_qmss.h | 89 u32 queue_mask; member
|
/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_packet_manager_vi.c | 136 packet->queue_mask_lo = lower_32_bits(res->queue_mask); in pm_set_resources_vi() 137 packet->queue_mask_hi = upper_32_bits(res->queue_mask); in pm_set_resources_vi()
|
H A D | kfd_packet_manager_v9.c | 200 packet->queue_mask_lo = lower_32_bits(res->queue_mask); in pm_set_resources_v9() 201 packet->queue_mask_hi = upper_32_bits(res->queue_mask); in pm_set_resources_v9()
|
H A D | kfd_device_queue_manager.c | 1584 res.queue_mask = 0; in set_sched_resources() 1600 if (WARN_ON(i >= (sizeof(res.queue_mask)*8))) { in set_sched_resources() 1605 res.queue_mask |= 1ull in set_sched_resources() 1615 res.vmid_mask, res.queue_mask); in set_sched_resources()
|
H A D | kfd_priv.h | 625 uint64_t queue_mask; member
|
/openbmc/u-boot/drivers/net/ |
H A D | macb.c | 629 u32 queue_mask; in gmac_init_multi_queues() local 632 queue_mask = gem_readl(macb, DCFG6) & 0xff; in gmac_init_multi_queues() 633 queue_mask |= 0x1; in gmac_init_multi_queues() 636 if (queue_mask & (1 << i)) in gmac_init_multi_queues()
|
/openbmc/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gfx.c | 583 uint64_t queue_mask = 0; in amdgpu_gfx_enable_kcq() local 596 if (WARN_ON(i > (sizeof(queue_mask)*8))) { in amdgpu_gfx_enable_kcq() 601 queue_mask |= (1ull << amdgpu_queue_mask_bit_to_set_resource_bit(adev, i)); in amdgpu_gfx_enable_kcq() 619 queue_mask = ~0ULL; in amdgpu_gfx_enable_kcq() 621 kiq->pmf->kiq_set_resources(kiq_ring, queue_mask); in amdgpu_gfx_enable_kcq()
|
H A D | amdgpu_gfx.h | 132 uint64_t queue_mask);
|
H A D | gfx_v8_0.c | 4319 uint64_t queue_mask = 0; in gfx_v8_0_kiq_kcq_enable() local 4329 if (WARN_ON(i >= (sizeof(queue_mask)*8))) { in gfx_v8_0_kiq_kcq_enable() 4334 queue_mask |= (1ull << i); in gfx_v8_0_kiq_kcq_enable() 4345 amdgpu_ring_write(kiq_ring, lower_32_bits(queue_mask)); /* queue mask lo */ in gfx_v8_0_kiq_kcq_enable() 4346 amdgpu_ring_write(kiq_ring, upper_32_bits(queue_mask)); /* queue mask hi */ in gfx_v8_0_kiq_kcq_enable()
|
H A D | gfx_v9_4_3.c | 61 uint64_t queue_mask) in gfx_v9_4_3_kiq_set_resources() argument 69 lower_32_bits(queue_mask)); /* queue mask lo */ in gfx_v9_4_3_kiq_set_resources() 71 upper_32_bits(queue_mask)); /* queue mask hi */ in gfx_v9_4_3_kiq_set_resources()
|
H A D | gfx_v11_0.c | 135 static void gfx11_kiq_set_resources(struct amdgpu_ring *kiq_ring, uint64_t queue_mask) in gfx11_kiq_set_resources() argument 140 amdgpu_ring_write(kiq_ring, lower_32_bits(queue_mask)); /* queue mask lo */ in gfx11_kiq_set_resources() 141 amdgpu_ring_write(kiq_ring, upper_32_bits(queue_mask)); /* queue mask hi */ in gfx11_kiq_set_resources()
|
H A D | gfx_v9_0.c | 769 uint64_t queue_mask) in gfx_v9_0_kiq_set_resources() argument 777 lower_32_bits(queue_mask)); /* queue mask lo */ in gfx_v9_0_kiq_set_resources() 779 upper_32_bits(queue_mask)); /* queue mask hi */ in gfx_v9_0_kiq_set_resources()
|
H A D | gfx_v10_0.c | 3496 static void gfx10_kiq_set_resources(struct amdgpu_ring *kiq_ring, uint64_t queue_mask) in gfx10_kiq_set_resources() argument 3501 amdgpu_ring_write(kiq_ring, lower_32_bits(queue_mask)); /* queue mask lo */ in gfx10_kiq_set_resources() 3502 amdgpu_ring_write(kiq_ring, upper_32_bits(queue_mask)); /* queue mask hi */ in gfx10_kiq_set_resources()
|
/openbmc/linux/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 2259 u8 queue_mask; in mv643xx_eth_poll() local 2270 queue_mask = mp->work_tx | mp->work_tx_end | mp->work_rx; in mv643xx_eth_poll() 2272 queue_mask |= mp->work_rx_refill; in mv643xx_eth_poll() 2274 if (!queue_mask) { in mv643xx_eth_poll() 2280 queue = fls(queue_mask) - 1; in mv643xx_eth_poll() 2281 queue_mask = 1 << queue; in mv643xx_eth_poll() 2287 if (mp->work_tx_end & queue_mask) { in mv643xx_eth_poll() 2289 } else if (mp->work_tx & queue_mask) { in mv643xx_eth_poll() 2292 } else if (mp->work_rx & queue_mask) { in mv643xx_eth_poll() 2294 } else if (!mp->oom && (mp->work_rx_refill & queue_mask)) { in mv643xx_eth_poll()
|
/openbmc/linux/net/ethtool/ |
H A D | ioctl.c | 2530 DECLARE_BITMAP(queue_mask, MAX_NUM_QUEUE); in ethtool_get_per_queue_coalesce() 2537 bitmap_from_arr32(queue_mask, per_queue_opt->queue_mask, in ethtool_get_per_queue_coalesce() 2540 for_each_set_bit(bit, queue_mask, MAX_NUM_QUEUE) { in ethtool_get_per_queue_coalesce() 2563 DECLARE_BITMAP(queue_mask, MAX_NUM_QUEUE); in ethtool_set_per_queue_coalesce() 2571 bitmap_from_arr32(queue_mask, per_queue_opt->queue_mask, MAX_NUM_QUEUE); in ethtool_set_per_queue_coalesce() 2572 n_queue = bitmap_weight(queue_mask, MAX_NUM_QUEUE); in ethtool_set_per_queue_coalesce() 2577 for_each_set_bit(bit, queue_mask, MAX_NUM_QUEUE) { in ethtool_set_per_queue_coalesce() 2606 for_each_set_bit(i, queue_mask, bit) { in ethtool_set_per_queue_coalesce()
|
/openbmc/linux/drivers/net/ethernet/cadence/ |
H A D | macb_main.c | 3947 unsigned int *queue_mask, in macb_probe_queues() argument 3950 *queue_mask = 0x1; in macb_probe_queues() 3963 *queue_mask |= readl_relaxed(mem + GEM_DCFG6) & 0xff; in macb_probe_queues() 3964 *num_queues = hweight32(*queue_mask); in macb_probe_queues() 4083 if (!(bp->queue_mask & (1 << hw_q))) in macb_init() 4968 unsigned int queue_mask, num_queues; in macb_probe() local 5004 macb_probe_queues(mem, native_io, &queue_mask, &num_queues); in macb_probe() 5028 bp->queue_mask = queue_mask; in macb_probe()
|
H A D | macb.h | 1262 unsigned int queue_mask; member
|
/openbmc/linux/net/sched/ |
H A D | sch_taprio.c | 1433 u32 i, queue_mask = 0; in tc_map_to_queue_mask() local 1444 queue_mask |= GENMASK(offset + count - 1, offset); in tc_map_to_queue_mask() 1447 return queue_mask; in tc_map_to_queue_mask()
|
/openbmc/qemu/include/standard-headers/linux/ |
H A D | ethtool.h | 1776 uint32_t queue_mask[__KERNEL_DIV_ROUND_UP(MAX_NUM_QUEUE, 32)]; member
|
/openbmc/linux/include/uapi/linux/ |
H A D | ethtool.h | 1509 __u32 queue_mask[__KERNEL_DIV_ROUND_UP(MAX_NUM_QUEUE, 32)]; member
|
/openbmc/linux/Documentation/networking/device_drivers/ethernet/intel/ |
H A D | ice.rst | 961 # ethtool --per-queue <ethX> queue_mask 0xa --coalesce adaptive-rx off 966 # ethtool --per-queue <ethX> queue_mask 0xa --show-coalesce
|