Home
last modified time | relevance | path

Searched refs:wq (Results 51 – 75 of 692) sorted by relevance

12345678910>>...28

/openbmc/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_debugfs.c25 struct hinic_wq *wq = sq->wq; in hinic_dbg_get_sq_info() local
31 return atomic_read(&wq->prod_idx) & wq->mask; in hinic_dbg_get_sq_info()
33 return atomic_read(&wq->cons_idx) & wq->mask; in hinic_dbg_get_sq_info()
35 return be16_to_cpu(*(__be16 *)(sq->hw_ci_addr)) & wq->mask; in hinic_dbg_get_sq_info()
55 struct hinic_wq *wq = rq->wq; in hinic_dbg_get_rq_info() local
61 return be16_to_cpu(*(__be16 *)(rq->pi_virt_addr)) & wq->mask; in hinic_dbg_get_rq_info()
63 return atomic_read(&wq->cons_idx) & wq->mask; in hinic_dbg_get_rq_info()
65 return atomic_read(&wq->prod_idx) & wq->mask; in hinic_dbg_get_rq_info()
H A Dhinic_hw_cmdq.c343 struct hinic_wq *wq = cmdq->wq; in cmdq_sync_cmd_direct_resp() local
361 num_wqebbs = ALIGN(WQE_LCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; in cmdq_sync_cmd_direct_resp()
363 if (next_prod_idx >= wq->q_depth) { in cmdq_sync_cmd_direct_resp()
365 next_prod_idx -= wq->q_depth; in cmdq_sync_cmd_direct_resp()
422 struct hinic_wq *wq = cmdq->wq; in cmdq_set_arm_bit() local
440 num_wqebbs = ALIGN(WQE_SCMD_SIZE, wq->wqebb_size) / wq->wqebb_size; in cmdq_set_arm_bit()
444 next_prod_idx -= wq->q_depth; in cmdq_set_arm_bit()
699 struct hinic_wq *wq = cmdq->wq; in cmdq_init_queue_ctxt() local
713 if (wq->num_q_pages != 1) { in cmdq_init_queue_ctxt()
743 cmdq->wq = wq; in init_cmdq()
[all …]
H A Dhinic_hw_qp.h54 #define HINIC_MIN_TX_WQE_SIZE(wq) \ argument
55 ALIGN(HINIC_SQ_WQE_SIZE(1), (wq)->wqebb_size)
58 (HINIC_MIN_TX_WQE_SIZE((sq)->wq) / (sq)->wq->wqebb_size)
82 struct hinic_wq *wq; member
101 struct hinic_wq *wq; member
139 struct hinic_wq *wq, struct msix_entry *entry, void *ci_addr,
145 struct hinic_wq *wq, struct msix_entry *entry);
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_rx.c409 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_free_rx_wqes() local
427 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_free_rx_wqes() local
441 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_alloc_rx_wqes() local
587 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; in mlx5e_post_rx_mpwqe() local
764 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_alloc_rx_mpwqe() local
896 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_post_rx_wqes() local
1082 struct mlx5_wq_ll *wq = &rq->mpwqe.wq; in mlx5e_post_rx_mpwqes() local
1824 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_handle_rx_cqe() local
1955 wq = &rq->mpwqe.wq; in mlx5e_handle_rx_cqe_mpwrq_rep()
2377 wq = &rq->mpwqe.wq; in mlx5e_handle_rx_cqe_mpwrq_shampo()
[all …]
/openbmc/linux/io_uring/
H A Dio-wq.h47 void io_wq_exit_start(struct io_wq *wq);
48 void io_wq_put_and_exit(struct io_wq *wq);
50 void io_wq_enqueue(struct io_wq *wq, struct io_wq_work *work);
54 int io_wq_max_workers(struct io_wq *wq, int *new_count);
64 enum io_wq_cancel io_wq_cancel_cb(struct io_wq *wq, work_cancel_fn *cancel,
/openbmc/linux/drivers/scsi/lpfc/
H A Dlpfc_debugfs.h422 struct lpfc_queue *wq; in lpfc_debug_dump_wq() local
429 wq = phba->sli4_hba.mbx_wq; in lpfc_debug_dump_wq()
432 wq = phba->sli4_hba.els_wq; in lpfc_debug_dump_wq()
435 wq = phba->sli4_hba.nvmels_wq; in lpfc_debug_dump_wq()
442 qtypestr, wqidx, wq->queue_id); in lpfc_debug_dump_wq()
445 qtypestr, wq->queue_id); in lpfc_debug_dump_wq()
447 lpfc_debug_dump_q(wq); in lpfc_debug_dump_wq()
462 struct lpfc_queue *wq, *cq, *eq; in lpfc_debug_dump_cq() local
474 wq = phba->sli4_hba.mbx_wq; in lpfc_debug_dump_cq()
478 wq = phba->sli4_hba.els_wq; in lpfc_debug_dump_cq()
[all …]
/openbmc/linux/drivers/infiniband/hw/mana/
H A Dqp.c111 struct mana_ib_wq *wq; in mana_ib_create_qp_rss() local
193 wq_spec.gdma_region = wq->gdma_region; in mana_ib_create_qp_rss()
194 wq_spec.queue_size = wq->wq_buf_size; in mana_ib_create_qp_rss()
202 &wq_spec, &cq_spec, &wq->rx_object); in mana_ib_create_qp_rss()
207 wq->gdma_region = GDMA_INVALID_DMA_REGION; in mana_ib_create_qp_rss()
210 wq->id = wq_spec.queue_index; in mana_ib_create_qp_rss()
215 ret, wq->rx_object, wq->id, cq->id); in mana_ib_create_qp_rss()
218 resp.entries[i].wqid = wq->id; in mana_ib_create_qp_rss()
220 mana_ind_table[i] = wq->rx_object; in mana_ib_create_qp_rss()
444 struct mana_ib_wq *wq; in mana_ib_destroy_qp_rss() local
[all …]
/openbmc/linux/fs/xfs/
H A Dxfs_pwork.c74 pctl->wq = alloc_workqueue("%s-%d", in xfs_pwork_init()
77 if (!pctl->wq) in xfs_pwork_init()
97 queue_work(pctl->wq, &pwork->work); in xfs_pwork_queue()
105 destroy_workqueue(pctl->wq); in xfs_pwork_destroy()
106 pctl->wq = NULL; in xfs_pwork_destroy()
/openbmc/linux/tools/workqueue/
H A Dwq_monitor.py77 def __init__(self, wq): argument
78 self.name = wq.name.string_().decode()
79 self.unbound = wq.flags & WQ_UNBOUND != 0
80 self.mem_reclaim = wq.flags & WQ_MEM_RECLAIM != 0
82 for pwq in list_for_each_entry('struct pool_workqueue', wq.pwqs.address_of_(), 'pwqs_node'):
161 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'):
162 stats = WqStats(wq)
H A Dwq_dump.py156 for wq in list_for_each_entry('struct workqueue_struct', workqueues.address_of_(), 'list'):
158 if wq.flags & WQ_UNBOUND:
159 if wq.flags & WQ_ORDERED:
163 if wq.unbound_attrs.affn_strict:
171 pool_id = per_cpu_ptr(wq.cpu_pwq, cpu)[0].pool.id.value_()
175 if wq.flags & WQ_UNBOUND:
/openbmc/linux/fs/btrfs/
H A Dmisc.h19 static inline void cond_wake_up(struct wait_queue_head *wq) in cond_wake_up() argument
25 if (wq_has_sleeper(wq)) in cond_wake_up()
26 wake_up(wq); in cond_wake_up()
29 static inline void cond_wake_up_nomb(struct wait_queue_head *wq) in cond_wake_up_nomb() argument
37 if (waitqueue_active(wq)) in cond_wake_up_nomb()
38 wake_up(wq); in cond_wake_up_nomb()
/openbmc/linux/net/core/
H A Dstream.c33 struct socket_wq *wq; in sk_stream_write_space() local
39 wq = rcu_dereference(sk->sk_wq); in sk_stream_write_space()
40 if (skwq_has_sleeper(wq)) in sk_stream_write_space()
41 wake_up_interruptible_poll(&wq->wait, EPOLLOUT | in sk_stream_write_space()
43 if (wq && wq->fasync_list && !(sk->sk_shutdown & SEND_SHUTDOWN)) in sk_stream_write_space()
44 sock_wake_async(wq, SOCK_WAKE_SPACE, POLL_OUT); in sk_stream_write_space()
/openbmc/linux/drivers/infiniband/hw/mlx5/
H A Dcq.c196 wq = &qp->rq; in handle_responder()
197 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in handle_responder()
198 ++wq->tail; in handle_responder()
403 cur = wq->head - wq->tail; in sw_comp()
412 idx = (is_send) ? wq->last_poll : wq->tail; in sw_comp()
417 wq->tail++; in sw_comp()
419 wq->last_poll = wq->w_list[idx].next; in sw_comp()
508 wq->tail = wq->wqe_head[idx] + 1; in mlx5_poll_one()
538 wq->tail = wq->wqe_head[idx] + 1; in mlx5_poll_one()
549 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx5_poll_one()
[all …]
/openbmc/linux/drivers/power/supply/
H A Dipaq_micro_battery.c40 struct workqueue_struct *wq; member
88 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); in micro_battery_work()
235 mb->wq = alloc_workqueue("ipaq-battery-wq", WQ_MEM_RECLAIM, 0); in micro_batt_probe()
236 if (!mb->wq) in micro_batt_probe()
241 queue_delayed_work(mb->wq, &mb->update, 1); in micro_batt_probe()
264 destroy_workqueue(mb->wq); in micro_batt_probe()
276 destroy_workqueue(mb->wq); in micro_batt_remove()
293 queue_delayed_work(mb->wq, &mb->update, msecs_to_jiffies(BATT_PERIOD)); in micro_batt_resume()
/openbmc/linux/tools/testing/selftests/net/
H A Dsrv6_end_dt46_l3vpn_test.sh269 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.accept_dad=0
270 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.default.accept_dad=0
276 ip netns exec ${nsname} sysctl -wq net.ipv4.ip_forward=1
277 ip netns exec ${nsname} sysctl -wq net.ipv6.conf.all.forwarding=1
292 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.all.accept_dad=0
293 ip netns exec ${hsname} sysctl -wq net.ipv6.conf.default.accept_dad=0
297 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.all.rp_filter=0
298 ip netns exec ${rtname} sysctl -wq net.ipv4.conf.default.rp_filter=0
312 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.all.accept_dad=0
313 ip netns exec ${rtname} sysctl -wq net.ipv6.conf.default.accept_dad=0
[all …]
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
H A Drx.c23 struct mlx5_wq_cyc *wq = &icosq->wq; in mlx5e_xsk_alloc_rx_mpwqe() local
51 umr_wqe = mlx5_wq_cyc_get_wqe(wq, pi); in mlx5e_xsk_alloc_rx_mpwqe()
162 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_alloc_rx_wqes_batched() local
171 contig = mlx5_wq_cyc_get_size(wq) - ix; in mlx5e_xsk_alloc_rx_wqes_batched()
181 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes_batched()
186 wqe = mlx5_wq_cyc_get_wqe(wq, j); in mlx5e_xsk_alloc_rx_wqes_batched()
200 struct mlx5_wq_cyc *wq = &rq->wqe.wq; in mlx5e_xsk_alloc_rx_wqes() local
204 int j = mlx5_wq_cyc_ctr2ix(wq, ix + i); in mlx5e_xsk_alloc_rx_wqes()
209 wqe = mlx5_wq_cyc_get_wqe(wq, j); in mlx5e_xsk_alloc_rx_wqes()
/openbmc/linux/lib/raid6/
H A Dvpermxor.uc49 unative_t wp$$, wq$$, wd$$;
56 wp$$ = wq$$ = *(unative_t *)&dptr[z0][d+$$*NSIZE];
64 asm(VPERMXOR(%0,%1,%2,%3):"=v"(wq$$):"v"(gf_high), "v"(gf_low), "v"(wq$$));
65 wq$$ = vec_xor(wq$$, wd$$);
68 *(unative_t *)&q[d+NSIZE*$$] = wq$$;
/openbmc/linux/fs/jfs/
H A Djfs_lock.h22 #define __SLEEP_COND(wq, cond, lock_cmd, unlock_cmd) \ argument
26 add_wait_queue(&wq, &__wait); \
36 remove_wait_queue(&wq, &__wait); \
/openbmc/linux/net/sunrpc/
H A Dsched.c330 if (waitqueue_active(wq)) in rpc_complete_task()
1284 if (!wq) in rpciod_start()
1286 rpciod_workqueue = wq; in rpciod_start()
1288 if (!wq) in rpciod_start()
1290 xprtiod_workqueue = wq; in rpciod_start()
1293 wq = rpciod_workqueue; in rpciod_start()
1295 destroy_workqueue(wq); in rpciod_start()
1307 wq = rpciod_workqueue; in rpciod_stop()
1309 destroy_workqueue(wq); in rpciod_stop()
1310 wq = xprtiod_workqueue; in rpciod_stop()
[all …]
/openbmc/openbmc/meta-phosphor/recipes-phosphor/users/phosphor-user-manager/
H A Dupgrade_hostconsole_group.sh7 if grep -wq hostconsole /etc/group; then
15 if id -nG root | grep -wq hostconsole; then
28 if id -nG "$usr" | grep -wq hostconsole; then
/openbmc/linux/drivers/platform/x86/
H A Dasus-wireless.c27 struct workqueue_struct *wq; member
108 queue_work(data->wq, &data->led_work); in led_state_set()
161 data->wq = create_singlethread_workqueue("asus_wireless_workqueue"); in asus_wireless_add()
162 if (!data->wq) in asus_wireless_add()
173 destroy_workqueue(data->wq); in asus_wireless_add()
182 if (data->wq) { in asus_wireless_remove()
184 destroy_workqueue(data->wq); in asus_wireless_remove()
/openbmc/linux/Documentation/core-api/
H A Dworkqueue.rst14 is needed and the workqueue (wq) API is the most commonly used
31 In the original wq implementation, a multi threaded (MT) wq had one
41 MT wq albeit less severe on MT. Each wq maintained its own separate
152 ``alloc_workqueue()`` allocates a wq. The original
195 Work items of a highpri wq are queued to the highpri
217 This flag is meaningless for unbound wq.
244 time thus achieving the same ordering property as ST wq.
264 of possible sequences of events with the original wq. ::
328 reclaim, they should be queued to separate wq each with
337 * A wq serves as a domain for forward progress guarantee
[all …]
/openbmc/linux/drivers/infiniband/hw/cxgb4/
H A Ddevice.c97 if (!wq->rdev->wr_log) in c4iw_log_wr_stats()
107 le.qid = wq->sq.qid; in c4iw_log_wr_stats()
109 le.post_host_time = wq->sq.sw_sq[wq->sq.cidx].host_time; in c4iw_log_wr_stats()
110 le.post_sge_ts = wq->sq.sw_sq[wq->sq.cidx].sge_ts; in c4iw_log_wr_stats()
113 le.qid = wq->rq.qid; in c4iw_log_wr_stats()
115 le.post_host_time = wq->rq.sw_rq[wq->rq.cidx].host_time; in c4iw_log_wr_stats()
116 le.post_sge_ts = wq->rq.sw_rq[wq->rq.cidx].sge_ts; in c4iw_log_wr_stats()
292 qp->wq.sq.qid, qp->wq.rq.qid, in dump_qp()
306 qp->wq.sq.qid, qp->wq.rq.qid, in dump_qp()
1288 t4_ring_sq_db(&qp->wq, qp->wq.sq.wq_pidx_inc, NULL); in resume_rc_qp()
[all …]
/openbmc/linux/drivers/hid/
H A Dhid-elo.c34 static struct workqueue_struct *wq; variable
182 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); in elo_work()
258 queue_delayed_work(wq, &priv->work, ELO_PERIODIC_READ_INTERVAL); in elo_probe()
296 wq = create_singlethread_workqueue("elousb"); in elo_driver_init()
297 if (!wq) in elo_driver_init()
302 destroy_workqueue(wq); in elo_driver_init()
311 destroy_workqueue(wq); in elo_driver_exit()
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dxdp.c305 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdpsq_get_next_pi() local
308 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi()
322 mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_xdpsq_get_next_pi()
326 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdpsq_get_next_pi()
356 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xdp_mpwqe_complete() local
360 u16 pi = mlx5_wq_cyc_ctr2ix(wq, sq->pc); in mlx5e_xdp_mpwqe_complete()
487 struct mlx5_wq_cyc *wq = &sq->wq; in mlx5e_xmit_xdp_frame() local
541 wqe = mlx5_wq_cyc_get_wqe(wq, pi); in mlx5e_xmit_xdp_frame()
699 cqe = mlx5_cqwq_get_cqe(&cq->wq); in mlx5e_poll_xdpsq_cq()
714 mlx5_cqwq_pop(&cq->wq); in mlx5e_poll_xdpsq_cq()
[all …]

12345678910>>...28