/openbmc/linux/fs/smb/server/ |
H A D | ksmbd_work.c | 23 if (work) { in ksmbd_alloc_work_struct() 32 work->iov = kcalloc(work->iov_alloc_cnt, sizeof(struct kvec), in ksmbd_alloc_work_struct() 36 work = NULL; in ksmbd_alloc_work_struct() 39 return work; in ksmbd_alloc_work_struct() 63 ksmbd_release_id(&work->conn->async_ida, work->async_id); in ksmbd_free_work_struct() 98 return queue_work(ksmbd_wq, &work->work); in ksmbd_queue_work() 104 work->iov[++work->iov_idx].iov_base = ib; in __ksmbd_iov_pin() 105 work->iov[work->iov_idx].iov_len = ib_len; in __ksmbd_iov_pin() 122 if (work->iov_alloc_cnt < work->iov_cnt + need_iov_cnt) { in __ksmbd_iov_pin_rsp() 139 work->iov[work->iov_idx].iov_base = work->response_buf; in __ksmbd_iov_pin_rsp() [all …]
|
H A D | server.c | 139 if (work->sess && conn->ops->is_sign_req(work, command)) { in __process_request() 147 ret = cmds->proc(work); in __process_request() 234 if (work->sess && in __handle_ksmbd_work() 235 (work->sess->sign || smb3_11_final_sess_setup_resp(work) || in __handle_ksmbd_work() 241 if (work->tcon) in __handle_ksmbd_work() 244 if (work->sess && work->sess->enc && work->encrypted && in __handle_ksmbd_work() 251 ksmbd_conn_write(work); in __handle_ksmbd_work() 262 struct ksmbd_work *work = container_of(wk, struct ksmbd_work, work); in handle_ksmbd_work() local 293 if (!work) { in queue_ksmbd_work() 298 work->conn = conn; in queue_ksmbd_work() [all …]
|
H A D | smb2pdu.h | 436 bool is_smb2_rsp(struct ksmbd_work *work); 479 int smb2_open(struct ksmbd_work *work); 482 int smb2_close(struct ksmbd_work *work); 483 int smb2_echo(struct ksmbd_work *work); 484 int smb2_set_info(struct ksmbd_work *work); 485 int smb2_read(struct ksmbd_work *work); 486 int smb2_write(struct ksmbd_work *work); 487 int smb2_flush(struct ksmbd_work *work); 488 int smb2_cancel(struct ksmbd_work *work); 489 int smb2_lock(struct ksmbd_work *work); [all …]
|
H A D | ksmbd_work.h | 86 struct work_struct work; member 99 static inline void *ksmbd_resp_buf_next(struct ksmbd_work *work) in ksmbd_resp_buf_next() argument 101 return work->response_buf + work->next_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_next() 108 static inline void *ksmbd_resp_buf_curr(struct ksmbd_work *work) in ksmbd_resp_buf_curr() argument 110 return work->response_buf + work->curr_smb2_rsp_hdr_off + 4; in ksmbd_resp_buf_curr() 117 static inline void *ksmbd_req_buf_next(struct ksmbd_work *work) in ksmbd_req_buf_next() argument 119 return work->request_buf + work->next_smb2_rcv_hdr_off + 4; in ksmbd_req_buf_next() 123 void ksmbd_free_work_struct(struct ksmbd_work *work); 130 bool ksmbd_queue_work(struct ksmbd_work *work); 133 int ksmbd_iov_pin_rsp(struct ksmbd_work *work, void *ib, int len); [all …]
|
H A D | smb2pdu.c | 125 work->tcon = ksmbd_tree_conn_lookup(work->sess, tree_id); in smb2_get_ksmbd_tcon() 399 work->iov[work->iov_idx].iov_len += (new_len - len); in init_chained_smb2_rsp() 404 work->curr_smb2_rsp_hdr_off = work->next_smb2_rsp_hdr_off; in init_chained_smb2_rsp() 480 work->iov[work->iov_idx].iov_len += len; in is_chained_smb2_message() 483 work->curr_smb2_rsp_hdr_off = work->next_smb2_rsp_hdr_off; in is_chained_smb2_message() 4362 query_dir_private.work = work; in smb2_query_dir() 8696 iov = &work->iov[work->iov_idx - 1]; in smb2_set_sign_rsp() 8699 iov = &work->iov[work->iov_idx]; in smb2_set_sign_rsp() 8788 chann = lookup_chann_list(work->sess, work->conn); in smb3_set_sign_rsp() 8802 iov = &work->iov[work->iov_idx - 1]; in smb3_set_sign_rsp() [all …]
|
H A D | smb_common.c | 143 hdr = work->request_buf; in ksmbd_verify_smb_message() 146 work->conn->outstanding_credits++; in ksmbd_verify_smb_message() 344 unsigned int cmd = work->conn->ops->get_cmd_val(work); in smb1_check_user_session() 364 if (!work->response_buf) { in smb1_allocate_rsp_buf() 380 work->send_no_response = 1; in set_smb1_rsp_status() 603 init_smb2_neg_rsp(work); in ksmbd_smb_negotiate_common() 769 WARN_ON(work->saved_cred); in ksmbd_override_fsids() 771 if (!work->saved_cred) { in ksmbd_override_fsids() 782 WARN_ON(!work->saved_cred); in ksmbd_revert_fsids() 785 revert_creds(work->saved_cred); in ksmbd_revert_fsids() [all …]
|
/openbmc/linux/virt/kvm/ |
H A D | async_pf.c | 48 container_of(work, struct kvm_async_pf, work); in async_pf_execute() 107 WARN_ON_ONCE(work->work.func); in kvm_flush_and_free_async_pf_work() 109 flush_work(&work->work); in kvm_flush_and_free_async_pf_work() 133 flush_work(&work->work); in kvm_clear_async_pf_completion_queue() 135 if (cancel_work_sync(&work->work)) { in kvm_clear_async_pf_completion_queue() 165 work = list_first_entry(&vcpu->async_pf.done, typeof(*work), in kvm_check_async_pf_completion() 201 if (!work) in kvm_setup_async_pf() 210 mmget(work->mm); in kvm_setup_async_pf() 212 INIT_WORK(&work->work, async_pf_execute); in kvm_setup_async_pf() 218 schedule_work(&work->work); in kvm_setup_async_pf() [all …]
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_flip_work.c | 86 work->func(work, val); in drm_flip_work_queue() 107 list_splice_tail(&work->queued, &work->commited); in drm_flip_work_commit() 108 INIT_LIST_HEAD(&work->queued); in drm_flip_work_commit() 110 queue_work(wq, &work->worker); in drm_flip_work_commit() 133 work->func(work, task->data); in flip_worker() 150 work->name = name; in drm_flip_work_init() 151 INIT_LIST_HEAD(&work->queued); in drm_flip_work_init() 152 INIT_LIST_HEAD(&work->commited); in drm_flip_work_init() 153 spin_lock_init(&work->lock); in drm_flip_work_init() 154 work->func = func; in drm_flip_work_init() [all …]
|
H A D | drm_vblank_work.c | 58 list_del_init(&work->node); in drm_handle_vblank_works() 77 list_del_init(&work->node); in drm_vblank_cancel_pending_works() 117 if (work->cancelling) in drm_vblank_work_schedule() 126 if (list_empty(&work->node)) { in drm_vblank_work_schedule() 137 work->count = count; in drm_vblank_work_schedule() 150 list_del_init(&work->node); in drm_vblank_work_schedule() 189 if (!list_empty(&work->node)) { in drm_vblank_work_cancel_sync() 190 list_del_init(&work->node); in drm_vblank_work_cancel_sync() 195 work->cancelling++; in drm_vblank_work_cancel_sync() 204 work->cancelling--; in drm_vblank_work_cancel_sync() [all …]
|
/openbmc/linux/include/trace/events/ |
H A D | workqueue.h | 31 __field( void *, work ) 39 __entry->work = work; 63 TP_ARGS(work), 66 __field( void *, work ) 70 __entry->work = work; 86 TP_ARGS(work), 89 __field( void *, work ) 94 __entry->work = work; 112 TP_ARGS(work, function), 115 __field( void *, work ) [all …]
|
/openbmc/linux/kernel/ |
H A D | irq_work.c | 112 irq_work_raise(work); in __irq_work_queue_local() 119 if (!irq_work_claim(work)) in irq_work_queue() 124 __irq_work_queue_local(work); in irq_work_queue() 140 return irq_work_queue(work); in irq_work_queue_on() 147 if (!irq_work_claim(work)) in irq_work_queue_on() 169 if (!irq_work_claim(work)) in irq_work_queue_on() 203 struct irq_work *work = arg; in irq_work_single() local 221 work->func(work); in irq_work_single() 237 struct irq_work *work, *tmp; in irq_work_run_list() local 252 irq_work_single(work); in irq_work_run_list() [all …]
|
H A D | task_work.c | 54 work->next = head; in task_work_add() 105 while (work) { in task_work_cancel_match() 107 pprev = &work->next; in task_work_cancel_match() 109 } else if (try_cmpxchg(pprev, &work, work->next)) in task_work_cancel_match() 114 return work; in task_work_cancel_match() 160 if (!work) { in task_work_run() 168 if (!work) in task_work_run() 179 next = work->next; in task_work_run() 180 work->func(work); in task_work_run() 181 work = next; in task_work_run() [all …]
|
H A D | kthread.c | 827 work = NULL; in kthread_worker_fn() 837 if (work) { in kthread_worker_fn() 841 work->func(work); in kthread_worker_fn() 974 return !list_empty(&work->node) || work->canceling; in queuing_blocked() 983 WARN_ON_ONCE(work->worker && work->worker != worker); in kthread_insert_work_sanity_check() 1040 struct kthread_work *work = &dwork->work; in kthread_delayed_work_timer_fn() local 1070 struct kthread_work *work = &dwork->work; in __kthread_queue_delayed_work() local 1113 struct kthread_work *work = &dwork->work; in kthread_queue_delayed_work() local 1137 container_of(work, struct kthread_flush_work, work); in kthread_flush_work_fn() 1190 container_of(work, struct kthread_delayed_work, work); in kthread_cancel_delayed_work_timer() [all …]
|
/openbmc/linux/include/linux/ |
H A D | completion.h | 35 #define COMPLETION_INITIALIZER(work) \ argument 39 (*({ init_completion_map(&(work), &(map)); &(work); })) 41 #define COMPLETION_INITIALIZER_ONSTACK(work) \ argument 42 (*({ init_completion(&work); &work; })) 52 #define DECLARE_COMPLETION(work) \ argument 53 struct completion work = COMPLETION_INITIALIZER(work) 68 # define DECLARE_COMPLETION_ONSTACK(work) \ argument 69 struct completion work = COMPLETION_INITIALIZER_ONSTACK(work) 71 struct completion work = COMPLETION_INITIALIZER_ONSTACK_MAP(work, map) 73 # define DECLARE_COMPLETION_ONSTACK(work) DECLARE_COMPLETION(work) argument [all …]
|
H A D | workqueue.h | 28 #define work_data_bits(work) ((unsigned long *)(&(work)->data)) argument 112 struct work_struct work; member 121 struct work_struct work; member 208 return container_of(work, struct delayed_work, work); in to_delayed_work() 213 return container_of(work, struct rcu_work, work); in to_rcu_work() 217 struct work_struct work; member 240 .work = __WORK_INITIALIZER((n).work, (f)), \ 349 #define work_pending(work) \ argument 358 work_pending(&(w)->work) 497 struct work_struct *work); [all …]
|
H A D | jump_label_ratelimit.h | 12 struct delayed_work work; member 18 struct delayed_work work; member 24 struct delayed_work work; member 28 __static_key_slow_dec_deferred(&(x)->key, &(x)->work, (x)->timeout) 30 __static_key_slow_dec_deferred(&(x)->key.key, &(x)->work, (x)->timeout) 33 __static_key_deferred_flush((x), &(x)->work) 37 struct delayed_work *work, 39 extern void __static_key_deferred_flush(void *key, struct delayed_work *work); 43 extern void jump_label_update_timeout(struct work_struct *work); 49 .work = __DELAYED_WORK_INITIALIZER((name).work, \ [all …]
|
H A D | irq_work.h | 37 void init_irq_work(struct irq_work *work, void (*func)(struct irq_work *)) in init_irq_work() argument 39 *work = IRQ_WORK_INIT(func); in init_irq_work() 42 static inline bool irq_work_is_pending(struct irq_work *work) in irq_work_is_pending() argument 44 return atomic_read(&work->node.a_flags) & IRQ_WORK_PENDING; in irq_work_is_pending() 47 static inline bool irq_work_is_busy(struct irq_work *work) in irq_work_is_busy() argument 49 return atomic_read(&work->node.a_flags) & IRQ_WORK_BUSY; in irq_work_is_busy() 52 static inline bool irq_work_is_hard(struct irq_work *work) in irq_work_is_hard() argument 54 return atomic_read(&work->node.a_flags) & IRQ_WORK_HARD_IRQ; in irq_work_is_hard() 57 bool irq_work_queue(struct irq_work *work); 58 bool irq_work_queue_on(struct irq_work *work, int cpu); [all …]
|
H A D | kthread.h | 141 struct kthread_work work; member 145 #define KTHREAD_WORK_INIT(work, fn) { \ argument 146 .node = LIST_HEAD_INIT((work).node), \ 151 .work = KTHREAD_WORK_INIT((dwork).work, (fn)), \ 156 #define DEFINE_KTHREAD_WORK(work, fn) \ argument 157 struct kthread_work work = KTHREAD_WORK_INIT(work, fn) 172 #define kthread_init_work(work, fn) \ argument 175 INIT_LIST_HEAD(&(work)->node); \ 176 (work)->func = (fn); \ 181 kthread_init_work(&(dwork)->work, (fn)); \ [all …]
|
/openbmc/linux/fs/btrfs/ |
H A D | async-thread.c | 57 return work->wq->fs_info; in btrfs_work_owner() 214 struct btrfs_work *work; in run_ordered_work() local 245 work->ordered_func(work); in run_ordered_work() 252 if (work == self) { in run_ordered_work() 280 work->ordered_free(work); in run_ordered_work() 309 if (work->ordered_func) in btrfs_work_helper() 314 work->func(work); in btrfs_work_helper() 334 work->func = func; in btrfs_init_work() 339 work->flags = 0; in btrfs_init_work() 346 work->wq = wq; in btrfs_queue_work() [all …]
|
/openbmc/linux/drivers/staging/octeon/ |
H A D | ethernet-rx.c | 72 if ((work->word2.snoip.err_code == 10) && (work->word1.len <= 64)) in cvm_oct_check_rcv_error() 112 work->word1.len -= i + 5; in cvm_oct_check_rcv_error() 119 work->word1.len -= i + 4; in cvm_oct_check_rcv_error() 131 cvm_oct_free_work(work); in cvm_oct_check_rcv_error() 138 cvm_oct_free_work(work); in cvm_oct_check_rcv_error() 146 int len = work->word1.len; in copy_segments_to_skb() 222 struct cvmx_wqe *work; in cvm_oct_poll() local 230 prefetch(work); in cvm_oct_poll() 232 if (!work) { in cvm_oct_poll() 299 cvm_oct_free_work(work); in cvm_oct_poll() [all …]
|
H A D | ethernet-tx.c | 516 if (unlikely(!work)) { in cvm_oct_xmit_pow() 560 work->word1.len = skb->len; in cvm_oct_xmit_pow() 567 work->word2.u64 = 0; in cvm_oct_xmit_pow() 568 work->word2.s.bufs = 1; in cvm_oct_xmit_pow() 569 work->packet_ptr.u64 = 0; in cvm_oct_xmit_pow() 590 work->word2.s.is_v6 = 0; in cvm_oct_xmit_pow() 601 work->word2.s.IP_exc = 0; in cvm_oct_xmit_pow() 607 work->word2.s.not_IP = 0; in cvm_oct_xmit_pow() 641 memcpy(work->packet_data, skb->data, sizeof(work->packet_data)); in cvm_oct_xmit_pow() 645 cvmx_pow_work_submit(work, work->word1.tag, work->word1.tag_type, in cvm_oct_xmit_pow() [all …]
|
/openbmc/linux/tools/perf/ |
H A D | builtin-kwork.c | 284 return work; in work_search() 318 struct kwork_work *work = zalloc(sizeof(*work)); in work_new() local 332 return work; in work_new() 342 return work; in work_findnew() 345 if (work) in work_findnew() 348 return work; in work_findnew() 666 if (work->class && work->class->work_name) { in timehist_print_event() 667 work->class->work_name(work, kwork_name, in timehist_print_event() 1092 if (work->class && work->class->work_name) { in report_print_work() 1093 work->class->work_name(work, kwork_name, in report_print_work() [all …]
|
/openbmc/linux/drivers/accessibility/speakup/ |
H A D | selection.c | 20 struct work_struct work; member 25 static void __speakup_set_selection(struct work_struct *work) in __speakup_set_selection() argument 28 container_of(work, struct speakup_selection_work, work); in __speakup_set_selection() 58 .work = __WORK_INITIALIZER(speakup_sel_work.work, 87 schedule_work_on(WORK_CPU_UNBOUND, &speakup_sel_work.work); in speakup_set_selection() 96 cancel_work_sync(&speakup_sel_work.work); in speakup_cancel_selection() 106 static void __speakup_paste_selection(struct work_struct *work) in __speakup_paste_selection() argument 109 container_of(work, struct speakup_selection_work, work); in __speakup_paste_selection() 117 .work = __WORK_INITIALIZER(speakup_paste_work.work, 129 schedule_work_on(WORK_CPU_UNBOUND, &speakup_paste_work.work); in speakup_paste_selection() [all …]
|
/openbmc/linux/drivers/net/wireless/st/cw1200/ |
H A D | sta.h | 59 void cw1200_event_handler(struct work_struct *work); 60 void cw1200_bss_loss_work(struct work_struct *work); 61 void cw1200_bss_params_work(struct work_struct *work); 62 void cw1200_keep_alive_work(struct work_struct *work); 63 void cw1200_tx_failure_work(struct work_struct *work); 79 void cw1200_join_timeout(struct work_struct *work); 80 void cw1200_unjoin_work(struct work_struct *work); 82 void cw1200_wep_key_work(struct work_struct *work); 91 void cw1200_ba_work(struct work_struct *work); 114 void cw1200_set_tim_work(struct work_struct *work); [all …]
|
/openbmc/linux/io_uring/ |
H A D | io-wq.c | 492 return work; in io_get_next_work() 503 return work; in io_get_next_work() 535 if (work) { in io_assign_current_work() 567 if (work) { in io_worker_handle_work() 582 if (!work) in io_worker_handle_work() 621 } while (work); in io_worker_handle_work() 898 work = wq->free_work(work); in io_run_cancel() 899 } while (work); in io_run_cancel() 934 .data = work, in io_wq_enqueue() 994 if (work && match->fn(work, match->data)) { in __io_wq_worker_cancel() [all …]
|