/openbmc/linux/drivers/net/ethernet/huawei/hinic/ |
H A D | hinic_debugfs.c | 25 struct hinic_wq *wq = sq->wq; in hinic_dbg_get_sq_info() local 31 return atomic_read(&wq->prod_idx) & wq->mask; in hinic_dbg_get_sq_info() 33 return atomic_read(&wq->cons_idx) & wq->mask; in hinic_dbg_get_sq_info() 35 return be16_to_cpu(*(__be16 *)(sq->hw_ci_addr)) & wq->mask; in hinic_dbg_get_sq_info() 55 struct hinic_wq *wq = rq->wq; in hinic_dbg_get_rq_info() local 61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq->mask; in hinic_dbg_get_rq_info() 63 return atomic_read(&wq->cons_idx) & wq->mask; in hinic_dbg_get_rq_info() 65 return atomic_read(&wq->prod_idx) & wq->mask; in hinic_dbg_get_rq_info()
|
H A D | hinic_hw_cmdq.c | 343 struct hinic_wq *wq = cmdq->wq; in cmdq_sync_cmd_direct_resp() local 361 num_wqebbs = ALIGN(WQE_LCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; in cmdq_sync_cmd_direct_resp() 363 if (next_prod_idx >= wq->q_depth) { in cmdq_sync_cmd_direct_resp() 365 next_prod_idx -= wq->q_depth; in cmdq_sync_cmd_direct_resp() 422 struct hinic_wq *wq = cmdq->wq; in cmdq_set_arm_bit() local 440 num_wqebbs = ALIGN(WQE_SCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; in cmdq_set_arm_bit() 444 next_prod_idx -= wq->q_depth; in cmdq_set_arm_bit() 699 struct hinic_wq *wq = cmdq->wq; in cmdq_init_queue_ctxt() local 713 if (wq->num_q_pages != 1) { in cmdq_init_queue_ctxt() 743 cmdq->wq = wq; in init_cmdq() [all …]
|
H A D | hinic_hw_qp.h | 54 #define HINIC_MIN_TX_WQE_SIZE(wq) \ argument 55 ALIGN(HINIC_SQ_WQE_SIZE(1), (wq)->wqebb_size) 58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size) 82 struct hinic_wq *wq; member 101 struct hinic_wq *wq; member 139 struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr, 145 struct hinic_wq *wq, struct msix_entry *entry);
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_rx.c | 409 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_free_rx_wqes() local 427 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_free_rx_wqes() local 441 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_alloc_rx_wqes() local 587 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; in mlx5e_post_rx_mpwqe() local 764 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_alloc_rx_mpwqe() local 896 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_post_rx_wqes() local 1082 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; in mlx5e_post_rx_mpwqes() local 1824 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_handle_rx_cqe() local 1955 wq = &rq->mpwqe.wq; in mlx5e_handle_rx_cqe_mpwrq_rep() 2377 wq = &rq->mpwqe.wq; in mlx5e_handle_rx_cqe_mpwrq_shampo() [all …]
|
/openbmc/linux/io_uring/ |
H A D | io-wq.h | 47 void io_wq_exit_start(struct io_wq *wq); 48 void io_wq_put_and_exit(struct io_wq *wq); 50 void io_wq_enqueue(struct io_wq *wq, struct io_wq_work *work); 54 int io_wq_max_workers(struct io_wq *wq, int *new_count); 64 enum io_wq_cancel io_wq_cancel_cb(struct io_wq *wq, work_cancel_fn *cancel,
|
/openbmc/linux/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 422 struct lpfc_queue *wq; in lpfc_debug_dump_wq() local 429 wq = phba->sli4_hba.mbx_wq; in lpfc_debug_dump_wq() 432 wq = phba->sli4_hba.els_wq; in lpfc_debug_dump_wq() 435 wq = phba->sli4_hba.nvmels_wq; in lpfc_debug_dump_wq() 442 qtypestr, wqidx, wq->queue_id); in lpfc_debug_dump_wq() 445 qtypestr, wq->queue_id); in lpfc_debug_dump_wq() 447 lpfc_debug_dump_q(wq); in lpfc_debug_dump_wq() 462 struct lpfc_queue *wq, *cq, *eq; in lpfc_debug_dump_cq() local 474 wq = phba->sli4_hba.mbx_wq; in lpfc_debug_dump_cq() 478 wq = phba->sli4_hba.els_wq; in lpfc_debug_dump_cq() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mana/ |
H A D | qp.c | 111 struct mana_ib_wq *wq; in mana_ib_create_qp_rss() local 193 wq_spec.gdma_region = wq->gdma_region; in mana_ib_create_qp_rss() 194 wq_spec.queue_size = wq->wq_buf_size; in mana_ib_create_qp_rss() 202 &wq_spec, &cq_spec, &wq->rx_object); in mana_ib_create_qp_rss() 207 wq->gdma_region = GDMA_INVALID_DMA_REGION; in mana_ib_create_qp_rss() 210 wq->id = wq_spec.queue_index; in mana_ib_create_qp_rss() 215 ret, wq->rx_object, wq->id, cq->id); in mana_ib_create_qp_rss() 218 resp.entries[i].wqid = wq->id; in mana_ib_create_qp_rss() 220 mana_ind_table[i] = wq->rx_object; in mana_ib_create_qp_rss() 444 struct mana_ib_wq *wq; in mana_ib_destroy_qp_rss() local [all …]
|
/openbmc/linux/fs/xfs/ |
H A D | xfs_pwork.c | 74 pctl->wq = alloc_workqueue("%s-%d", in xfs_pwork_init() 77 if (!pctl->wq) in xfs_pwork_init() 97 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue() 105 destroy_workqueue(pctl->wq); in xfs_pwork_destroy() 106 pctl->wq = NULL; in xfs_pwork_destroy()
|
/openbmc/linux/tools/workqueue/ |
H A D | wq_monitor.py | 77 def __init__(self, wq): argument 78 self.name = wq.name.string_().decode() 79 self.unbound = wq.flags & WQ_UNBOUND != 0 80 self.mem_reclaim = wq.flags & WQ_MEM_RECLAIM != 0 82 for pwq in list_for_each_entry('struct pool_workqueue', wq.pwqs.address_of_(), 'pwqs_node'): 161 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'): 162 stats = WqStats(wq)
|
H A D | wq_dump.py | 156 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'): 158 if wq.flags & WQ_UNBOUND: 159 if wq.flags & WQ_ORDERED: 163 if wq.unbound_attrs.affn_strict: 171 pool_id = per_cpu_ptr(wq.cpu_pwq, cpu)[0].pool.id.value_() 175 if wq.flags & WQ_UNBOUND:
|
/openbmc/linux/fs/btrfs/ |
H A D | misc.h | 19 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument 25 if (wq_has_sleeper(wq)) in cond_wake_up() 26 wake_up(wq); in cond_wake_up() 29 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument 37 if (waitqueue_active(wq)) in cond_wake_up_nomb() 38 wake_up(wq); in cond_wake_up_nomb()
|
/openbmc/linux/net/core/ |
H A D | stream.c | 33 struct socket_wq *wq; in sk_stream_write_space() local 39 wq = rcu_dereference(sk->sk_wq); in sk_stream_write_space() 40 if (skwq_has_sleeper(wq)) in sk_stream_write_space() 41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | in sk_stream_write_space() 43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) in sk_stream_write_space() 44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); in sk_stream_write_space()
|
/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | cq.c | 196 wq = &qp->rq; in handle_responder() 197 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder() 198 ++wq->tail; in handle_responder() 403 cur = wq->head - wq->tail; in sw_comp() 412 idx = (is_send) ? wq->last_poll : wq->tail; in sw_comp() 417 wq->tail++; in sw_comp() 419 wq->last_poll = wq->w_list[idx].next; in sw_comp() 508 wq->tail = wq->wqe_head[idx] + 1; in mlx5_poll_one() 538 wq->tail = wq->wqe_head[idx] + 1; in mlx5_poll_one() 549 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one() [all …]
|
/openbmc/linux/drivers/power/supply/ |
H A D | ipaq_micro_battery.c | 40 struct workqueue_struct *wq; member 88 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); in micro_battery_work() 235 mb->wq = alloc_workqueue("ipaq-battery-wq", WQ_MEM_RECLAIM, 0); in micro_batt_probe() 236 if (!mb->wq) in micro_batt_probe() 241 queue_delayed_work(mb->wq, &mb->update, 1); in micro_batt_probe() 264 destroy_workqueue(mb->wq); in micro_batt_probe() 276 destroy_workqueue(mb->wq); in micro_batt_remove() 293 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); in micro_batt_resume()
|
/openbmc/linux/tools/testing/selftests/net/ |
H A D | srv6_end_dt46_l3vpn_test.sh | 269 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.accept_dad=0 270 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.default.accept_dad=0 276 ip netns exec ${nsname} sysctl -wq net.ipv4.ip_forward=1 277 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.forwarding=1 292 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.all.accept_dad=0 293 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.default.accept_dad=0 297 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.all.rp_filter=0 298 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.default.rp_filter=0 312 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.all.accept_dad=0 313 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.default.accept_dad=0 [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
H A D | rx.c | 23 struct mlx5_wq_cyc *wq = &icosq->wq; in mlx5e_xsk_alloc_rx_mpwqe() local 51 umr_wqe = mlx5_wq_cyc_get_wqe(wq, pi); in mlx5e_xsk_alloc_rx_mpwqe() 162 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_alloc_rx_wqes_batched() local 171 contig = mlx5_wq_cyc_get_size(wq) - ix; in mlx5e_xsk_alloc_rx_wqes_batched() 181 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes_batched() 186 wqe = mlx5_wq_cyc_get_wqe(wq, j); in mlx5e_xsk_alloc_rx_wqes_batched() 200 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_alloc_rx_wqes() local 204 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes() 209 wqe = mlx5_wq_cyc_get_wqe(wq, j); in mlx5e_xsk_alloc_rx_wqes()
|
/openbmc/linux/lib/raid6/ |
H A D | vpermxor.uc | 49 unative_t wp$$, wq$$, wd$$; 56 wp$$ = wq$$ = *(unative_t *)&dptr[z0][d+$$*NSIZE]; 64 asm(VPERMXOR(%0,%1,%2,%3):"=v"(wq$$):"v"(gf_high), "v"(gf_low), "v"(wq$$)); 65 wq$$ = vec_xor(wq$$, wd$$); 68 *(unative_t *)&q[d+NSIZE*$$] = wq$$;
|
/openbmc/linux/fs/jfs/ |
H A D | jfs_lock.h | 22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument 26 add_wait_queue(&wq, &__wait); \ 36 remove_wait_queue(&wq, &__wait); \
|
/openbmc/linux/net/sunrpc/ |
H A D | sched.c | 330 if (waitqueue_active(wq)) in rpc_complete_task() 1284 if (!wq) in rpciod_start() 1286 rpciod_workqueue = wq; in rpciod_start() 1288 if (!wq) in rpciod_start() 1290 xprtiod_workqueue = wq; in rpciod_start() 1293 wq = rpciod_workqueue; in rpciod_start() 1295 destroy_workqueue(wq); in rpciod_start() 1307 wq = rpciod_workqueue; in rpciod_stop() 1309 destroy_workqueue(wq); in rpciod_stop() 1310 wq = xprtiod_workqueue; in rpciod_stop() [all …]
|
/openbmc/openbmc/meta-phosphor/recipes-phosphor/users/phosphor-user-manager/ |
H A D | upgrade_hostconsole_group.sh | 7 if grep -wq hostconsole /etc/group; then 15 if id -nG root | grep -wq hostconsole; then 28 if id -nG "$usr" | grep -wq hostconsole; then
|
/openbmc/linux/drivers/platform/x86/ |
H A D | asus-wireless.c | 27 struct workqueue_struct *wq; member 108 queue_work(data->wq, &data->led_work); in led_state_set() 161 data->wq = create_singlethread_workqueue("asus_wireless_workqueue"); in asus_wireless_add() 162 if (!data->wq) in asus_wireless_add() 173 destroy_workqueue(data->wq); in asus_wireless_add() 182 if (data->wq) { in asus_wireless_remove() 184 destroy_workqueue(data->wq); in asus_wireless_remove()
|
/openbmc/linux/Documentation/core-api/ |
H A D | workqueue.rst | 14 is needed and the workqueue (wq) API is the most commonly used 31 In the original wq implementation, a multi threaded (MT) wq had one 41 MT wq albeit less severe on MT. Each wq maintained its own separate 152 ``alloc_workqueue()`` allocates a wq. The original 195 Work items of a highpri wq are queued to the highpri 217 This flag is meaningless for unbound wq. 244 time thus achieving the same ordering property as ST wq. 264 of possible sequences of events with the original wq. :: 328 reclaim, they should be queued to separate wq each with 337 * A wq serves as a domain for forward progress guarantee [all …]
|
/openbmc/linux/drivers/infiniband/hw/cxgb4/ |
H A D | device.c | 97 if (!wq->rdev->wr_log) in c4iw_log_wr_stats() 107 le.qid = wq->sq.qid; in c4iw_log_wr_stats() 109 le.post_host_time = wq->sq.sw_sq[wq->sq.cidx].host_time; in c4iw_log_wr_stats() 110 le.post_sge_ts = wq->sq.sw_sq[wq->sq.cidx].sge_ts; in c4iw_log_wr_stats() 113 le.qid = wq->rq.qid; in c4iw_log_wr_stats() 115 le.post_host_time = wq->rq.sw_rq[wq->rq.cidx].host_time; in c4iw_log_wr_stats() 116 le.post_sge_ts = wq->rq.sw_rq[wq->rq.cidx].sge_ts; in c4iw_log_wr_stats() 292 qp->wq.sq.qid, qp->wq.rq.qid, in dump_qp() 306 qp->wq.sq.qid, qp->wq.rq.qid, in dump_qp() 1288 t4_ring_sq_db(&qp->wq, qp->wq.sq.wq_pidx_inc, NULL); in resume_rc_qp() [all …]
|
/openbmc/linux/drivers/hid/ |
H A D | hid-elo.c | 34 static struct workqueue_struct *wq; variable 182 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); in elo_work() 258 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); in elo_probe() 296 wq = create_singlethread_workqueue("elousb"); in elo_driver_init() 297 if (!wq) in elo_driver_init() 302 destroy_workqueue(wq); in elo_driver_init() 311 destroy_workqueue(wq); in elo_driver_exit()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | xdp.c | 305 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdpsq_get_next_pi() local 308 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi() 322 mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_xdpsq_get_next_pi() 326 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi() 356 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdp_mpwqe_complete() local 360 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdp_mpwqe_complete() 487 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xmit_xdp_frame() local 541 wqe = mlx5_wq_cyc_get_wqe(wq, pi); in mlx5e_xmit_xdp_frame() 699 cqe = mlx5_cqwq_get_cqe(&cq->wq); in mlx5e_poll_xdpsq_cq() 714 mlx5_cqwq_pop(&cq->wq); in mlx5e_poll_xdpsq_cq() [all …]
|