1 /* Kernel thread helper functions. 2 * Copyright (C) 2004 IBM Corporation, Rusty Russell. 3 * 4 * Creation is done via kthreadd, so that we get a clean environment 5 * even if we're invoked from userspace (think modprobe, hotplug cpu, 6 * etc.). 7 */ 8 #include <uapi/linux/sched/types.h> 9 #include <linux/sched.h> 10 #include <linux/sched/task.h> 11 #include <linux/kthread.h> 12 #include <linux/completion.h> 13 #include <linux/err.h> 14 #include <linux/cpuset.h> 15 #include <linux/unistd.h> 16 #include <linux/file.h> 17 #include <linux/export.h> 18 #include <linux/mutex.h> 19 #include <linux/slab.h> 20 #include <linux/freezer.h> 21 #include <linux/ptrace.h> 22 #include <linux/uaccess.h> 23 #include <trace/events/sched.h> 24 25 static DEFINE_SPINLOCK(kthread_create_lock); 26 static LIST_HEAD(kthread_create_list); 27 struct task_struct *kthreadd_task; 28 29 struct kthread_create_info 30 { 31 /* Information passed to kthread() from kthreadd. */ 32 int (*threadfn)(void *data); 33 void *data; 34 int node; 35 36 /* Result passed back to kthread_create() from kthreadd. */ 37 struct task_struct *result; 38 struct completion *done; 39 40 struct list_head list; 41 }; 42 43 struct kthread { 44 unsigned long flags; 45 unsigned int cpu; 46 void *data; 47 struct completion parked; 48 struct completion exited; 49 #ifdef CONFIG_BLK_CGROUP 50 struct cgroup_subsys_state *blkcg_css; 51 #endif 52 }; 53 54 enum KTHREAD_BITS { 55 KTHREAD_IS_PER_CPU = 0, 56 KTHREAD_SHOULD_STOP, 57 KTHREAD_SHOULD_PARK, 58 }; 59 60 static inline void set_kthread_struct(void *kthread) 61 { 62 /* 63 * We abuse ->set_child_tid to avoid the new member and because it 64 * can't be wrongly copied by copy_process(). We also rely on fact 65 * that the caller can't exec, so PF_KTHREAD can't be cleared. 66 */ 67 current->set_child_tid = (__force void __user *)kthread; 68 } 69 70 static inline struct kthread *to_kthread(struct task_struct *k) 71 { 72 WARN_ON(!(k->flags & PF_KTHREAD)); 73 return (__force void *)k->set_child_tid; 74 } 75 76 void free_kthread_struct(struct task_struct *k) 77 { 78 struct kthread *kthread; 79 80 /* 81 * Can be NULL if this kthread was created by kernel_thread() 82 * or if kmalloc() in kthread() failed. 83 */ 84 kthread = to_kthread(k); 85 #ifdef CONFIG_BLK_CGROUP 86 WARN_ON_ONCE(kthread && kthread->blkcg_css); 87 #endif 88 kfree(kthread); 89 } 90 91 /** 92 * kthread_should_stop - should this kthread return now? 93 * 94 * When someone calls kthread_stop() on your kthread, it will be woken 95 * and this will return true. You should then return, and your return 96 * value will be passed through to kthread_stop(). 97 */ 98 bool kthread_should_stop(void) 99 { 100 return test_bit(KTHREAD_SHOULD_STOP, &to_kthread(current)->flags); 101 } 102 EXPORT_SYMBOL(kthread_should_stop); 103 104 /** 105 * kthread_should_park - should this kthread park now? 106 * 107 * When someone calls kthread_park() on your kthread, it will be woken 108 * and this will return true. You should then do the necessary 109 * cleanup and call kthread_parkme() 110 * 111 * Similar to kthread_should_stop(), but this keeps the thread alive 112 * and in a park position. kthread_unpark() "restarts" the thread and 113 * calls the thread function again. 114 */ 115 bool kthread_should_park(void) 116 { 117 return test_bit(KTHREAD_SHOULD_PARK, &to_kthread(current)->flags); 118 } 119 EXPORT_SYMBOL_GPL(kthread_should_park); 120 121 /** 122 * kthread_freezable_should_stop - should this freezable kthread return now? 123 * @was_frozen: optional out parameter, indicates whether %current was frozen 124 * 125 * kthread_should_stop() for freezable kthreads, which will enter 126 * refrigerator if necessary. This function is safe from kthread_stop() / 127 * freezer deadlock and freezable kthreads should use this function instead 128 * of calling try_to_freeze() directly. 129 */ 130 bool kthread_freezable_should_stop(bool *was_frozen) 131 { 132 bool frozen = false; 133 134 might_sleep(); 135 136 if (unlikely(freezing(current))) 137 frozen = __refrigerator(true); 138 139 if (was_frozen) 140 *was_frozen = frozen; 141 142 return kthread_should_stop(); 143 } 144 EXPORT_SYMBOL_GPL(kthread_freezable_should_stop); 145 146 /** 147 * kthread_data - return data value specified on kthread creation 148 * @task: kthread task in question 149 * 150 * Return the data value specified when kthread @task was created. 151 * The caller is responsible for ensuring the validity of @task when 152 * calling this function. 153 */ 154 void *kthread_data(struct task_struct *task) 155 { 156 return to_kthread(task)->data; 157 } 158 159 /** 160 * kthread_probe_data - speculative version of kthread_data() 161 * @task: possible kthread task in question 162 * 163 * @task could be a kthread task. Return the data value specified when it 164 * was created if accessible. If @task isn't a kthread task or its data is 165 * inaccessible for any reason, %NULL is returned. This function requires 166 * that @task itself is safe to dereference. 167 */ 168 void *kthread_probe_data(struct task_struct *task) 169 { 170 struct kthread *kthread = to_kthread(task); 171 void *data = NULL; 172 173 probe_kernel_read(&data, &kthread->data, sizeof(data)); 174 return data; 175 } 176 177 static void __kthread_parkme(struct kthread *self) 178 { 179 for (;;) { 180 set_current_state(TASK_PARKED); 181 if (!test_bit(KTHREAD_SHOULD_PARK, &self->flags)) 182 break; 183 schedule(); 184 } 185 __set_current_state(TASK_RUNNING); 186 } 187 188 void kthread_parkme(void) 189 { 190 __kthread_parkme(to_kthread(current)); 191 } 192 EXPORT_SYMBOL_GPL(kthread_parkme); 193 194 void kthread_park_complete(struct task_struct *k) 195 { 196 complete(&to_kthread(k)->parked); 197 } 198 199 static int kthread(void *_create) 200 { 201 /* Copy data: it's on kthread's stack */ 202 struct kthread_create_info *create = _create; 203 int (*threadfn)(void *data) = create->threadfn; 204 void *data = create->data; 205 struct completion *done; 206 struct kthread *self; 207 int ret; 208 209 self = kzalloc(sizeof(*self), GFP_KERNEL); 210 set_kthread_struct(self); 211 212 /* If user was SIGKILLed, I release the structure. */ 213 done = xchg(&create->done, NULL); 214 if (!done) { 215 kfree(create); 216 do_exit(-EINTR); 217 } 218 219 if (!self) { 220 create->result = ERR_PTR(-ENOMEM); 221 complete(done); 222 do_exit(-ENOMEM); 223 } 224 225 self->data = data; 226 init_completion(&self->exited); 227 init_completion(&self->parked); 228 current->vfork_done = &self->exited; 229 230 /* OK, tell user we're spawned, wait for stop or wakeup */ 231 __set_current_state(TASK_UNINTERRUPTIBLE); 232 create->result = current; 233 complete(done); 234 schedule(); 235 236 ret = -EINTR; 237 if (!test_bit(KTHREAD_SHOULD_STOP, &self->flags)) { 238 cgroup_kthread_ready(); 239 __kthread_parkme(self); 240 ret = threadfn(data); 241 } 242 do_exit(ret); 243 } 244 245 /* called from do_fork() to get node information for about to be created task */ 246 int tsk_fork_get_node(struct task_struct *tsk) 247 { 248 #ifdef CONFIG_NUMA 249 if (tsk == kthreadd_task) 250 return tsk->pref_node_fork; 251 #endif 252 return NUMA_NO_NODE; 253 } 254 255 static void create_kthread(struct kthread_create_info *create) 256 { 257 int pid; 258 259 #ifdef CONFIG_NUMA 260 current->pref_node_fork = create->node; 261 #endif 262 /* We want our own signal handler (we take no signals by default). */ 263 pid = kernel_thread(kthread, create, CLONE_FS | CLONE_FILES | SIGCHLD); 264 if (pid < 0) { 265 /* If user was SIGKILLed, I release the structure. */ 266 struct completion *done = xchg(&create->done, NULL); 267 268 if (!done) { 269 kfree(create); 270 return; 271 } 272 create->result = ERR_PTR(pid); 273 complete(done); 274 } 275 } 276 277 static __printf(4, 0) 278 struct task_struct *__kthread_create_on_node(int (*threadfn)(void *data), 279 void *data, int node, 280 const char namefmt[], 281 va_list args) 282 { 283 DECLARE_COMPLETION_ONSTACK(done); 284 struct task_struct *task; 285 struct kthread_create_info *create = kmalloc(sizeof(*create), 286 GFP_KERNEL); 287 288 if (!create) 289 return ERR_PTR(-ENOMEM); 290 create->threadfn = threadfn; 291 create->data = data; 292 create->node = node; 293 create->done = &done; 294 295 spin_lock(&kthread_create_lock); 296 list_add_tail(&create->list, &kthread_create_list); 297 spin_unlock(&kthread_create_lock); 298 299 wake_up_process(kthreadd_task); 300 /* 301 * Wait for completion in killable state, for I might be chosen by 302 * the OOM killer while kthreadd is trying to allocate memory for 303 * new kernel thread. 304 */ 305 if (unlikely(wait_for_completion_killable(&done))) { 306 /* 307 * If I was SIGKILLed before kthreadd (or new kernel thread) 308 * calls complete(), leave the cleanup of this structure to 309 * that thread. 310 */ 311 if (xchg(&create->done, NULL)) 312 return ERR_PTR(-EINTR); 313 /* 314 * kthreadd (or new kernel thread) will call complete() 315 * shortly. 316 */ 317 wait_for_completion(&done); 318 } 319 task = create->result; 320 if (!IS_ERR(task)) { 321 static const struct sched_param param = { .sched_priority = 0 }; 322 323 vsnprintf(task->comm, sizeof(task->comm), namefmt, args); 324 /* 325 * root may have changed our (kthreadd's) priority or CPU mask. 326 * The kernel thread should not inherit these properties. 327 */ 328 sched_setscheduler_nocheck(task, SCHED_NORMAL, ¶m); 329 set_cpus_allowed_ptr(task, cpu_all_mask); 330 } 331 kfree(create); 332 return task; 333 } 334 335 /** 336 * kthread_create_on_node - create a kthread. 337 * @threadfn: the function to run until signal_pending(current). 338 * @data: data ptr for @threadfn. 339 * @node: task and thread structures for the thread are allocated on this node 340 * @namefmt: printf-style name for the thread. 341 * 342 * Description: This helper function creates and names a kernel 343 * thread. The thread will be stopped: use wake_up_process() to start 344 * it. See also kthread_run(). The new thread has SCHED_NORMAL policy and 345 * is affine to all CPUs. 346 * 347 * If thread is going to be bound on a particular cpu, give its node 348 * in @node, to get NUMA affinity for kthread stack, or else give NUMA_NO_NODE. 349 * When woken, the thread will run @threadfn() with @data as its 350 * argument. @threadfn() can either call do_exit() directly if it is a 351 * standalone thread for which no one will call kthread_stop(), or 352 * return when 'kthread_should_stop()' is true (which means 353 * kthread_stop() has been called). The return value should be zero 354 * or a negative error number; it will be passed to kthread_stop(). 355 * 356 * Returns a task_struct or ERR_PTR(-ENOMEM) or ERR_PTR(-EINTR). 357 */ 358 struct task_struct *kthread_create_on_node(int (*threadfn)(void *data), 359 void *data, int node, 360 const char namefmt[], 361 ...) 362 { 363 struct task_struct *task; 364 va_list args; 365 366 va_start(args, namefmt); 367 task = __kthread_create_on_node(threadfn, data, node, namefmt, args); 368 va_end(args); 369 370 return task; 371 } 372 EXPORT_SYMBOL(kthread_create_on_node); 373 374 static void __kthread_bind_mask(struct task_struct *p, const struct cpumask *mask, long state) 375 { 376 unsigned long flags; 377 378 if (!wait_task_inactive(p, state)) { 379 WARN_ON(1); 380 return; 381 } 382 383 /* It's safe because the task is inactive. */ 384 raw_spin_lock_irqsave(&p->pi_lock, flags); 385 do_set_cpus_allowed(p, mask); 386 p->flags |= PF_NO_SETAFFINITY; 387 raw_spin_unlock_irqrestore(&p->pi_lock, flags); 388 } 389 390 static void __kthread_bind(struct task_struct *p, unsigned int cpu, long state) 391 { 392 __kthread_bind_mask(p, cpumask_of(cpu), state); 393 } 394 395 void kthread_bind_mask(struct task_struct *p, const struct cpumask *mask) 396 { 397 __kthread_bind_mask(p, mask, TASK_UNINTERRUPTIBLE); 398 } 399 400 /** 401 * kthread_bind - bind a just-created kthread to a cpu. 402 * @p: thread created by kthread_create(). 403 * @cpu: cpu (might not be online, must be possible) for @k to run on. 404 * 405 * Description: This function is equivalent to set_cpus_allowed(), 406 * except that @cpu doesn't need to be online, and the thread must be 407 * stopped (i.e., just returned from kthread_create()). 408 */ 409 void kthread_bind(struct task_struct *p, unsigned int cpu) 410 { 411 __kthread_bind(p, cpu, TASK_UNINTERRUPTIBLE); 412 } 413 EXPORT_SYMBOL(kthread_bind); 414 415 /** 416 * kthread_create_on_cpu - Create a cpu bound kthread 417 * @threadfn: the function to run until signal_pending(current). 418 * @data: data ptr for @threadfn. 419 * @cpu: The cpu on which the thread should be bound, 420 * @namefmt: printf-style name for the thread. Format is restricted 421 * to "name.*%u". Code fills in cpu number. 422 * 423 * Description: This helper function creates and names a kernel thread 424 * The thread will be woken and put into park mode. 425 */ 426 struct task_struct *kthread_create_on_cpu(int (*threadfn)(void *data), 427 void *data, unsigned int cpu, 428 const char *namefmt) 429 { 430 struct task_struct *p; 431 432 p = kthread_create_on_node(threadfn, data, cpu_to_node(cpu), namefmt, 433 cpu); 434 if (IS_ERR(p)) 435 return p; 436 kthread_bind(p, cpu); 437 /* CPU hotplug need to bind once again when unparking the thread. */ 438 set_bit(KTHREAD_IS_PER_CPU, &to_kthread(p)->flags); 439 to_kthread(p)->cpu = cpu; 440 return p; 441 } 442 443 /** 444 * kthread_unpark - unpark a thread created by kthread_create(). 445 * @k: thread created by kthread_create(). 446 * 447 * Sets kthread_should_park() for @k to return false, wakes it, and 448 * waits for it to return. If the thread is marked percpu then its 449 * bound to the cpu again. 450 */ 451 void kthread_unpark(struct task_struct *k) 452 { 453 struct kthread *kthread = to_kthread(k); 454 455 /* 456 * Newly created kthread was parked when the CPU was offline. 457 * The binding was lost and we need to set it again. 458 */ 459 if (test_bit(KTHREAD_IS_PER_CPU, &kthread->flags)) 460 __kthread_bind(k, kthread->cpu, TASK_PARKED); 461 462 clear_bit(KTHREAD_SHOULD_PARK, &kthread->flags); 463 wake_up_state(k, TASK_PARKED); 464 } 465 EXPORT_SYMBOL_GPL(kthread_unpark); 466 467 /** 468 * kthread_park - park a thread created by kthread_create(). 469 * @k: thread created by kthread_create(). 470 * 471 * Sets kthread_should_park() for @k to return true, wakes it, and 472 * waits for it to return. This can also be called after kthread_create() 473 * instead of calling wake_up_process(): the thread will park without 474 * calling threadfn(). 475 * 476 * Returns 0 if the thread is parked, -ENOSYS if the thread exited. 477 * If called by the kthread itself just the park bit is set. 478 */ 479 int kthread_park(struct task_struct *k) 480 { 481 struct kthread *kthread = to_kthread(k); 482 483 if (WARN_ON(k->flags & PF_EXITING)) 484 return -ENOSYS; 485 486 if (WARN_ON_ONCE(test_bit(KTHREAD_SHOULD_PARK, &kthread->flags))) 487 return -EBUSY; 488 489 set_bit(KTHREAD_SHOULD_PARK, &kthread->flags); 490 if (k != current) { 491 wake_up_process(k); 492 wait_for_completion(&kthread->parked); 493 } 494 495 return 0; 496 } 497 EXPORT_SYMBOL_GPL(kthread_park); 498 499 /** 500 * kthread_stop - stop a thread created by kthread_create(). 501 * @k: thread created by kthread_create(). 502 * 503 * Sets kthread_should_stop() for @k to return true, wakes it, and 504 * waits for it to exit. This can also be called after kthread_create() 505 * instead of calling wake_up_process(): the thread will exit without 506 * calling threadfn(). 507 * 508 * If threadfn() may call do_exit() itself, the caller must ensure 509 * task_struct can't go away. 510 * 511 * Returns the result of threadfn(), or %-EINTR if wake_up_process() 512 * was never called. 513 */ 514 int kthread_stop(struct task_struct *k) 515 { 516 struct kthread *kthread; 517 int ret; 518 519 trace_sched_kthread_stop(k); 520 521 get_task_struct(k); 522 kthread = to_kthread(k); 523 set_bit(KTHREAD_SHOULD_STOP, &kthread->flags); 524 kthread_unpark(k); 525 wake_up_process(k); 526 wait_for_completion(&kthread->exited); 527 ret = k->exit_code; 528 put_task_struct(k); 529 530 trace_sched_kthread_stop_ret(ret); 531 return ret; 532 } 533 EXPORT_SYMBOL(kthread_stop); 534 535 int kthreadd(void *unused) 536 { 537 struct task_struct *tsk = current; 538 539 /* Setup a clean context for our children to inherit. */ 540 set_task_comm(tsk, "kthreadd"); 541 ignore_signals(tsk); 542 set_cpus_allowed_ptr(tsk, cpu_all_mask); 543 set_mems_allowed(node_states[N_MEMORY]); 544 545 current->flags |= PF_NOFREEZE; 546 cgroup_init_kthreadd(); 547 548 for (;;) { 549 set_current_state(TASK_INTERRUPTIBLE); 550 if (list_empty(&kthread_create_list)) 551 schedule(); 552 __set_current_state(TASK_RUNNING); 553 554 spin_lock(&kthread_create_lock); 555 while (!list_empty(&kthread_create_list)) { 556 struct kthread_create_info *create; 557 558 create = list_entry(kthread_create_list.next, 559 struct kthread_create_info, list); 560 list_del_init(&create->list); 561 spin_unlock(&kthread_create_lock); 562 563 create_kthread(create); 564 565 spin_lock(&kthread_create_lock); 566 } 567 spin_unlock(&kthread_create_lock); 568 } 569 570 return 0; 571 } 572 573 void __kthread_init_worker(struct kthread_worker *worker, 574 const char *name, 575 struct lock_class_key *key) 576 { 577 memset(worker, 0, sizeof(struct kthread_worker)); 578 spin_lock_init(&worker->lock); 579 lockdep_set_class_and_name(&worker->lock, key, name); 580 INIT_LIST_HEAD(&worker->work_list); 581 INIT_LIST_HEAD(&worker->delayed_work_list); 582 } 583 EXPORT_SYMBOL_GPL(__kthread_init_worker); 584 585 /** 586 * kthread_worker_fn - kthread function to process kthread_worker 587 * @worker_ptr: pointer to initialized kthread_worker 588 * 589 * This function implements the main cycle of kthread worker. It processes 590 * work_list until it is stopped with kthread_stop(). It sleeps when the queue 591 * is empty. 592 * 593 * The works are not allowed to keep any locks, disable preemption or interrupts 594 * when they finish. There is defined a safe point for freezing when one work 595 * finishes and before a new one is started. 596 * 597 * Also the works must not be handled by more than one worker at the same time, 598 * see also kthread_queue_work(). 599 */ 600 int kthread_worker_fn(void *worker_ptr) 601 { 602 struct kthread_worker *worker = worker_ptr; 603 struct kthread_work *work; 604 605 /* 606 * FIXME: Update the check and remove the assignment when all kthread 607 * worker users are created using kthread_create_worker*() functions. 608 */ 609 WARN_ON(worker->task && worker->task != current); 610 worker->task = current; 611 612 if (worker->flags & KTW_FREEZABLE) 613 set_freezable(); 614 615 repeat: 616 set_current_state(TASK_INTERRUPTIBLE); /* mb paired w/ kthread_stop */ 617 618 if (kthread_should_stop()) { 619 __set_current_state(TASK_RUNNING); 620 spin_lock_irq(&worker->lock); 621 worker->task = NULL; 622 spin_unlock_irq(&worker->lock); 623 return 0; 624 } 625 626 work = NULL; 627 spin_lock_irq(&worker->lock); 628 if (!list_empty(&worker->work_list)) { 629 work = list_first_entry(&worker->work_list, 630 struct kthread_work, node); 631 list_del_init(&work->node); 632 } 633 worker->current_work = work; 634 spin_unlock_irq(&worker->lock); 635 636 if (work) { 637 __set_current_state(TASK_RUNNING); 638 work->func(work); 639 } else if (!freezing(current)) 640 schedule(); 641 642 try_to_freeze(); 643 cond_resched(); 644 goto repeat; 645 } 646 EXPORT_SYMBOL_GPL(kthread_worker_fn); 647 648 static __printf(3, 0) struct kthread_worker * 649 __kthread_create_worker(int cpu, unsigned int flags, 650 const char namefmt[], va_list args) 651 { 652 struct kthread_worker *worker; 653 struct task_struct *task; 654 int node = -1; 655 656 worker = kzalloc(sizeof(*worker), GFP_KERNEL); 657 if (!worker) 658 return ERR_PTR(-ENOMEM); 659 660 kthread_init_worker(worker); 661 662 if (cpu >= 0) 663 node = cpu_to_node(cpu); 664 665 task = __kthread_create_on_node(kthread_worker_fn, worker, 666 node, namefmt, args); 667 if (IS_ERR(task)) 668 goto fail_task; 669 670 if (cpu >= 0) 671 kthread_bind(task, cpu); 672 673 worker->flags = flags; 674 worker->task = task; 675 wake_up_process(task); 676 return worker; 677 678 fail_task: 679 kfree(worker); 680 return ERR_CAST(task); 681 } 682 683 /** 684 * kthread_create_worker - create a kthread worker 685 * @flags: flags modifying the default behavior of the worker 686 * @namefmt: printf-style name for the kthread worker (task). 687 * 688 * Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM) 689 * when the needed structures could not get allocated, and ERR_PTR(-EINTR) 690 * when the worker was SIGKILLed. 691 */ 692 struct kthread_worker * 693 kthread_create_worker(unsigned int flags, const char namefmt[], ...) 694 { 695 struct kthread_worker *worker; 696 va_list args; 697 698 va_start(args, namefmt); 699 worker = __kthread_create_worker(-1, flags, namefmt, args); 700 va_end(args); 701 702 return worker; 703 } 704 EXPORT_SYMBOL(kthread_create_worker); 705 706 /** 707 * kthread_create_worker_on_cpu - create a kthread worker and bind it 708 * it to a given CPU and the associated NUMA node. 709 * @cpu: CPU number 710 * @flags: flags modifying the default behavior of the worker 711 * @namefmt: printf-style name for the kthread worker (task). 712 * 713 * Use a valid CPU number if you want to bind the kthread worker 714 * to the given CPU and the associated NUMA node. 715 * 716 * A good practice is to add the cpu number also into the worker name. 717 * For example, use kthread_create_worker_on_cpu(cpu, "helper/%d", cpu). 718 * 719 * Returns a pointer to the allocated worker on success, ERR_PTR(-ENOMEM) 720 * when the needed structures could not get allocated, and ERR_PTR(-EINTR) 721 * when the worker was SIGKILLed. 722 */ 723 struct kthread_worker * 724 kthread_create_worker_on_cpu(int cpu, unsigned int flags, 725 const char namefmt[], ...) 726 { 727 struct kthread_worker *worker; 728 va_list args; 729 730 va_start(args, namefmt); 731 worker = __kthread_create_worker(cpu, flags, namefmt, args); 732 va_end(args); 733 734 return worker; 735 } 736 EXPORT_SYMBOL(kthread_create_worker_on_cpu); 737 738 /* 739 * Returns true when the work could not be queued at the moment. 740 * It happens when it is already pending in a worker list 741 * or when it is being cancelled. 742 */ 743 static inline bool queuing_blocked(struct kthread_worker *worker, 744 struct kthread_work *work) 745 { 746 lockdep_assert_held(&worker->lock); 747 748 return !list_empty(&work->node) || work->canceling; 749 } 750 751 static void kthread_insert_work_sanity_check(struct kthread_worker *worker, 752 struct kthread_work *work) 753 { 754 lockdep_assert_held(&worker->lock); 755 WARN_ON_ONCE(!list_empty(&work->node)); 756 /* Do not use a work with >1 worker, see kthread_queue_work() */ 757 WARN_ON_ONCE(work->worker && work->worker != worker); 758 } 759 760 /* insert @work before @pos in @worker */ 761 static void kthread_insert_work(struct kthread_worker *worker, 762 struct kthread_work *work, 763 struct list_head *pos) 764 { 765 kthread_insert_work_sanity_check(worker, work); 766 767 list_add_tail(&work->node, pos); 768 work->worker = worker; 769 if (!worker->current_work && likely(worker->task)) 770 wake_up_process(worker->task); 771 } 772 773 /** 774 * kthread_queue_work - queue a kthread_work 775 * @worker: target kthread_worker 776 * @work: kthread_work to queue 777 * 778 * Queue @work to work processor @task for async execution. @task 779 * must have been created with kthread_worker_create(). Returns %true 780 * if @work was successfully queued, %false if it was already pending. 781 * 782 * Reinitialize the work if it needs to be used by another worker. 783 * For example, when the worker was stopped and started again. 784 */ 785 bool kthread_queue_work(struct kthread_worker *worker, 786 struct kthread_work *work) 787 { 788 bool ret = false; 789 unsigned long flags; 790 791 spin_lock_irqsave(&worker->lock, flags); 792 if (!queuing_blocked(worker, work)) { 793 kthread_insert_work(worker, work, &worker->work_list); 794 ret = true; 795 } 796 spin_unlock_irqrestore(&worker->lock, flags); 797 return ret; 798 } 799 EXPORT_SYMBOL_GPL(kthread_queue_work); 800 801 /** 802 * kthread_delayed_work_timer_fn - callback that queues the associated kthread 803 * delayed work when the timer expires. 804 * @t: pointer to the expired timer 805 * 806 * The format of the function is defined by struct timer_list. 807 * It should have been called from irqsafe timer with irq already off. 808 */ 809 void kthread_delayed_work_timer_fn(struct timer_list *t) 810 { 811 struct kthread_delayed_work *dwork = from_timer(dwork, t, timer); 812 struct kthread_work *work = &dwork->work; 813 struct kthread_worker *worker = work->worker; 814 815 /* 816 * This might happen when a pending work is reinitialized. 817 * It means that it is used a wrong way. 818 */ 819 if (WARN_ON_ONCE(!worker)) 820 return; 821 822 spin_lock(&worker->lock); 823 /* Work must not be used with >1 worker, see kthread_queue_work(). */ 824 WARN_ON_ONCE(work->worker != worker); 825 826 /* Move the work from worker->delayed_work_list. */ 827 WARN_ON_ONCE(list_empty(&work->node)); 828 list_del_init(&work->node); 829 kthread_insert_work(worker, work, &worker->work_list); 830 831 spin_unlock(&worker->lock); 832 } 833 EXPORT_SYMBOL(kthread_delayed_work_timer_fn); 834 835 void __kthread_queue_delayed_work(struct kthread_worker *worker, 836 struct kthread_delayed_work *dwork, 837 unsigned long delay) 838 { 839 struct timer_list *timer = &dwork->timer; 840 struct kthread_work *work = &dwork->work; 841 842 WARN_ON_ONCE(timer->function != kthread_delayed_work_timer_fn); 843 844 /* 845 * If @delay is 0, queue @dwork->work immediately. This is for 846 * both optimization and correctness. The earliest @timer can 847 * expire is on the closest next tick and delayed_work users depend 848 * on that there's no such delay when @delay is 0. 849 */ 850 if (!delay) { 851 kthread_insert_work(worker, work, &worker->work_list); 852 return; 853 } 854 855 /* Be paranoid and try to detect possible races already now. */ 856 kthread_insert_work_sanity_check(worker, work); 857 858 list_add(&work->node, &worker->delayed_work_list); 859 work->worker = worker; 860 timer->expires = jiffies + delay; 861 add_timer(timer); 862 } 863 864 /** 865 * kthread_queue_delayed_work - queue the associated kthread work 866 * after a delay. 867 * @worker: target kthread_worker 868 * @dwork: kthread_delayed_work to queue 869 * @delay: number of jiffies to wait before queuing 870 * 871 * If the work has not been pending it starts a timer that will queue 872 * the work after the given @delay. If @delay is zero, it queues the 873 * work immediately. 874 * 875 * Return: %false if the @work has already been pending. It means that 876 * either the timer was running or the work was queued. It returns %true 877 * otherwise. 878 */ 879 bool kthread_queue_delayed_work(struct kthread_worker *worker, 880 struct kthread_delayed_work *dwork, 881 unsigned long delay) 882 { 883 struct kthread_work *work = &dwork->work; 884 unsigned long flags; 885 bool ret = false; 886 887 spin_lock_irqsave(&worker->lock, flags); 888 889 if (!queuing_blocked(worker, work)) { 890 __kthread_queue_delayed_work(worker, dwork, delay); 891 ret = true; 892 } 893 894 spin_unlock_irqrestore(&worker->lock, flags); 895 return ret; 896 } 897 EXPORT_SYMBOL_GPL(kthread_queue_delayed_work); 898 899 struct kthread_flush_work { 900 struct kthread_work work; 901 struct completion done; 902 }; 903 904 static void kthread_flush_work_fn(struct kthread_work *work) 905 { 906 struct kthread_flush_work *fwork = 907 container_of(work, struct kthread_flush_work, work); 908 complete(&fwork->done); 909 } 910 911 /** 912 * kthread_flush_work - flush a kthread_work 913 * @work: work to flush 914 * 915 * If @work is queued or executing, wait for it to finish execution. 916 */ 917 void kthread_flush_work(struct kthread_work *work) 918 { 919 struct kthread_flush_work fwork = { 920 KTHREAD_WORK_INIT(fwork.work, kthread_flush_work_fn), 921 COMPLETION_INITIALIZER_ONSTACK(fwork.done), 922 }; 923 struct kthread_worker *worker; 924 bool noop = false; 925 926 worker = work->worker; 927 if (!worker) 928 return; 929 930 spin_lock_irq(&worker->lock); 931 /* Work must not be used with >1 worker, see kthread_queue_work(). */ 932 WARN_ON_ONCE(work->worker != worker); 933 934 if (!list_empty(&work->node)) 935 kthread_insert_work(worker, &fwork.work, work->node.next); 936 else if (worker->current_work == work) 937 kthread_insert_work(worker, &fwork.work, 938 worker->work_list.next); 939 else 940 noop = true; 941 942 spin_unlock_irq(&worker->lock); 943 944 if (!noop) 945 wait_for_completion(&fwork.done); 946 } 947 EXPORT_SYMBOL_GPL(kthread_flush_work); 948 949 /* 950 * This function removes the work from the worker queue. Also it makes sure 951 * that it won't get queued later via the delayed work's timer. 952 * 953 * The work might still be in use when this function finishes. See the 954 * current_work proceed by the worker. 955 * 956 * Return: %true if @work was pending and successfully canceled, 957 * %false if @work was not pending 958 */ 959 static bool __kthread_cancel_work(struct kthread_work *work, bool is_dwork, 960 unsigned long *flags) 961 { 962 /* Try to cancel the timer if exists. */ 963 if (is_dwork) { 964 struct kthread_delayed_work *dwork = 965 container_of(work, struct kthread_delayed_work, work); 966 struct kthread_worker *worker = work->worker; 967 968 /* 969 * del_timer_sync() must be called to make sure that the timer 970 * callback is not running. The lock must be temporary released 971 * to avoid a deadlock with the callback. In the meantime, 972 * any queuing is blocked by setting the canceling counter. 973 */ 974 work->canceling++; 975 spin_unlock_irqrestore(&worker->lock, *flags); 976 del_timer_sync(&dwork->timer); 977 spin_lock_irqsave(&worker->lock, *flags); 978 work->canceling--; 979 } 980 981 /* 982 * Try to remove the work from a worker list. It might either 983 * be from worker->work_list or from worker->delayed_work_list. 984 */ 985 if (!list_empty(&work->node)) { 986 list_del_init(&work->node); 987 return true; 988 } 989 990 return false; 991 } 992 993 /** 994 * kthread_mod_delayed_work - modify delay of or queue a kthread delayed work 995 * @worker: kthread worker to use 996 * @dwork: kthread delayed work to queue 997 * @delay: number of jiffies to wait before queuing 998 * 999 * If @dwork is idle, equivalent to kthread_queue_delayed_work(). Otherwise, 1000 * modify @dwork's timer so that it expires after @delay. If @delay is zero, 1001 * @work is guaranteed to be queued immediately. 1002 * 1003 * Return: %true if @dwork was pending and its timer was modified, 1004 * %false otherwise. 1005 * 1006 * A special case is when the work is being canceled in parallel. 1007 * It might be caused either by the real kthread_cancel_delayed_work_sync() 1008 * or yet another kthread_mod_delayed_work() call. We let the other command 1009 * win and return %false here. The caller is supposed to synchronize these 1010 * operations a reasonable way. 1011 * 1012 * This function is safe to call from any context including IRQ handler. 1013 * See __kthread_cancel_work() and kthread_delayed_work_timer_fn() 1014 * for details. 1015 */ 1016 bool kthread_mod_delayed_work(struct kthread_worker *worker, 1017 struct kthread_delayed_work *dwork, 1018 unsigned long delay) 1019 { 1020 struct kthread_work *work = &dwork->work; 1021 unsigned long flags; 1022 int ret = false; 1023 1024 spin_lock_irqsave(&worker->lock, flags); 1025 1026 /* Do not bother with canceling when never queued. */ 1027 if (!work->worker) 1028 goto fast_queue; 1029 1030 /* Work must not be used with >1 worker, see kthread_queue_work() */ 1031 WARN_ON_ONCE(work->worker != worker); 1032 1033 /* Do not fight with another command that is canceling this work. */ 1034 if (work->canceling) 1035 goto out; 1036 1037 ret = __kthread_cancel_work(work, true, &flags); 1038 fast_queue: 1039 __kthread_queue_delayed_work(worker, dwork, delay); 1040 out: 1041 spin_unlock_irqrestore(&worker->lock, flags); 1042 return ret; 1043 } 1044 EXPORT_SYMBOL_GPL(kthread_mod_delayed_work); 1045 1046 static bool __kthread_cancel_work_sync(struct kthread_work *work, bool is_dwork) 1047 { 1048 struct kthread_worker *worker = work->worker; 1049 unsigned long flags; 1050 int ret = false; 1051 1052 if (!worker) 1053 goto out; 1054 1055 spin_lock_irqsave(&worker->lock, flags); 1056 /* Work must not be used with >1 worker, see kthread_queue_work(). */ 1057 WARN_ON_ONCE(work->worker != worker); 1058 1059 ret = __kthread_cancel_work(work, is_dwork, &flags); 1060 1061 if (worker->current_work != work) 1062 goto out_fast; 1063 1064 /* 1065 * The work is in progress and we need to wait with the lock released. 1066 * In the meantime, block any queuing by setting the canceling counter. 1067 */ 1068 work->canceling++; 1069 spin_unlock_irqrestore(&worker->lock, flags); 1070 kthread_flush_work(work); 1071 spin_lock_irqsave(&worker->lock, flags); 1072 work->canceling--; 1073 1074 out_fast: 1075 spin_unlock_irqrestore(&worker->lock, flags); 1076 out: 1077 return ret; 1078 } 1079 1080 /** 1081 * kthread_cancel_work_sync - cancel a kthread work and wait for it to finish 1082 * @work: the kthread work to cancel 1083 * 1084 * Cancel @work and wait for its execution to finish. This function 1085 * can be used even if the work re-queues itself. On return from this 1086 * function, @work is guaranteed to be not pending or executing on any CPU. 1087 * 1088 * kthread_cancel_work_sync(&delayed_work->work) must not be used for 1089 * delayed_work's. Use kthread_cancel_delayed_work_sync() instead. 1090 * 1091 * The caller must ensure that the worker on which @work was last 1092 * queued can't be destroyed before this function returns. 1093 * 1094 * Return: %true if @work was pending, %false otherwise. 1095 */ 1096 bool kthread_cancel_work_sync(struct kthread_work *work) 1097 { 1098 return __kthread_cancel_work_sync(work, false); 1099 } 1100 EXPORT_SYMBOL_GPL(kthread_cancel_work_sync); 1101 1102 /** 1103 * kthread_cancel_delayed_work_sync - cancel a kthread delayed work and 1104 * wait for it to finish. 1105 * @dwork: the kthread delayed work to cancel 1106 * 1107 * This is kthread_cancel_work_sync() for delayed works. 1108 * 1109 * Return: %true if @dwork was pending, %false otherwise. 1110 */ 1111 bool kthread_cancel_delayed_work_sync(struct kthread_delayed_work *dwork) 1112 { 1113 return __kthread_cancel_work_sync(&dwork->work, true); 1114 } 1115 EXPORT_SYMBOL_GPL(kthread_cancel_delayed_work_sync); 1116 1117 /** 1118 * kthread_flush_worker - flush all current works on a kthread_worker 1119 * @worker: worker to flush 1120 * 1121 * Wait until all currently executing or pending works on @worker are 1122 * finished. 1123 */ 1124 void kthread_flush_worker(struct kthread_worker *worker) 1125 { 1126 struct kthread_flush_work fwork = { 1127 KTHREAD_WORK_INIT(fwork.work, kthread_flush_work_fn), 1128 COMPLETION_INITIALIZER_ONSTACK(fwork.done), 1129 }; 1130 1131 kthread_queue_work(worker, &fwork.work); 1132 wait_for_completion(&fwork.done); 1133 } 1134 EXPORT_SYMBOL_GPL(kthread_flush_worker); 1135 1136 /** 1137 * kthread_destroy_worker - destroy a kthread worker 1138 * @worker: worker to be destroyed 1139 * 1140 * Flush and destroy @worker. The simple flush is enough because the kthread 1141 * worker API is used only in trivial scenarios. There are no multi-step state 1142 * machines needed. 1143 */ 1144 void kthread_destroy_worker(struct kthread_worker *worker) 1145 { 1146 struct task_struct *task; 1147 1148 task = worker->task; 1149 if (WARN_ON(!task)) 1150 return; 1151 1152 kthread_flush_worker(worker); 1153 kthread_stop(task); 1154 WARN_ON(!list_empty(&worker->work_list)); 1155 kfree(worker); 1156 } 1157 EXPORT_SYMBOL(kthread_destroy_worker); 1158 1159 #ifdef CONFIG_BLK_CGROUP 1160 /** 1161 * kthread_associate_blkcg - associate blkcg to current kthread 1162 * @css: the cgroup info 1163 * 1164 * Current thread must be a kthread. The thread is running jobs on behalf of 1165 * other threads. In some cases, we expect the jobs attach cgroup info of 1166 * original threads instead of that of current thread. This function stores 1167 * original thread's cgroup info in current kthread context for later 1168 * retrieval. 1169 */ 1170 void kthread_associate_blkcg(struct cgroup_subsys_state *css) 1171 { 1172 struct kthread *kthread; 1173 1174 if (!(current->flags & PF_KTHREAD)) 1175 return; 1176 kthread = to_kthread(current); 1177 if (!kthread) 1178 return; 1179 1180 if (kthread->blkcg_css) { 1181 css_put(kthread->blkcg_css); 1182 kthread->blkcg_css = NULL; 1183 } 1184 if (css) { 1185 css_get(css); 1186 kthread->blkcg_css = css; 1187 } 1188 } 1189 EXPORT_SYMBOL(kthread_associate_blkcg); 1190 1191 /** 1192 * kthread_blkcg - get associated blkcg css of current kthread 1193 * 1194 * Current thread must be a kthread. 1195 */ 1196 struct cgroup_subsys_state *kthread_blkcg(void) 1197 { 1198 struct kthread *kthread; 1199 1200 if (current->flags & PF_KTHREAD) { 1201 kthread = to_kthread(current); 1202 if (kthread) 1203 return kthread->blkcg_css; 1204 } 1205 return NULL; 1206 } 1207 EXPORT_SYMBOL(kthread_blkcg); 1208 #endif 1209