1 /* 2 * Wrappers around mutex/cond/thread functions 3 * 4 * Copyright Red Hat, Inc. 2009 5 * 6 * Author: 7 * Marcelo Tosatti <mtosatti@redhat.com> 8 * 9 * This work is licensed under the terms of the GNU GPL, version 2 or later. 10 * See the COPYING file in the top-level directory. 11 * 12 */ 13 #include "qemu/osdep.h" 14 #include "qemu/thread.h" 15 #include "qemu/atomic.h" 16 #include "qemu/notify.h" 17 #include "qemu-thread-common.h" 18 #include "qemu/tsan.h" 19 #include "qemu/bitmap.h" 20 21 static bool name_threads; 22 23 void qemu_thread_naming(bool enable) 24 { 25 name_threads = enable; 26 27 #if !defined CONFIG_PTHREAD_SETNAME_NP_W_TID && \ 28 !defined CONFIG_PTHREAD_SETNAME_NP_WO_TID 29 /* This is a debugging option, not fatal */ 30 if (enable) { 31 fprintf(stderr, "qemu: thread naming not supported on this host\n"); 32 } 33 #endif 34 } 35 36 static void error_exit(int err, const char *msg) 37 { 38 fprintf(stderr, "qemu: %s: %s\n", msg, strerror(err)); 39 abort(); 40 } 41 42 static inline clockid_t qemu_timedwait_clockid(void) 43 { 44 #ifdef CONFIG_PTHREAD_CONDATTR_SETCLOCK 45 return CLOCK_MONOTONIC; 46 #else 47 return CLOCK_REALTIME; 48 #endif 49 } 50 51 static void compute_abs_deadline(struct timespec *ts, int ms) 52 { 53 clock_gettime(qemu_timedwait_clockid(), ts); 54 ts->tv_nsec += (ms % 1000) * 1000000; 55 ts->tv_sec += ms / 1000; 56 if (ts->tv_nsec >= 1000000000) { 57 ts->tv_sec++; 58 ts->tv_nsec -= 1000000000; 59 } 60 } 61 62 void qemu_mutex_init(QemuMutex *mutex) 63 { 64 int err; 65 66 err = pthread_mutex_init(&mutex->lock, NULL); 67 if (err) 68 error_exit(err, __func__); 69 qemu_mutex_post_init(mutex); 70 } 71 72 void qemu_mutex_destroy(QemuMutex *mutex) 73 { 74 int err; 75 76 assert(mutex->initialized); 77 mutex->initialized = false; 78 err = pthread_mutex_destroy(&mutex->lock); 79 if (err) 80 error_exit(err, __func__); 81 } 82 83 void qemu_mutex_lock_impl(QemuMutex *mutex, const char *file, const int line) 84 { 85 int err; 86 87 assert(mutex->initialized); 88 qemu_mutex_pre_lock(mutex, file, line); 89 err = pthread_mutex_lock(&mutex->lock); 90 if (err) 91 error_exit(err, __func__); 92 qemu_mutex_post_lock(mutex, file, line); 93 } 94 95 int qemu_mutex_trylock_impl(QemuMutex *mutex, const char *file, const int line) 96 { 97 int err; 98 99 assert(mutex->initialized); 100 err = pthread_mutex_trylock(&mutex->lock); 101 if (err == 0) { 102 qemu_mutex_post_lock(mutex, file, line); 103 return 0; 104 } 105 if (err != EBUSY) { 106 error_exit(err, __func__); 107 } 108 return -EBUSY; 109 } 110 111 void qemu_mutex_unlock_impl(QemuMutex *mutex, const char *file, const int line) 112 { 113 int err; 114 115 assert(mutex->initialized); 116 qemu_mutex_pre_unlock(mutex, file, line); 117 err = pthread_mutex_unlock(&mutex->lock); 118 if (err) 119 error_exit(err, __func__); 120 } 121 122 void qemu_rec_mutex_init(QemuRecMutex *mutex) 123 { 124 int err; 125 pthread_mutexattr_t attr; 126 127 pthread_mutexattr_init(&attr); 128 pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE); 129 err = pthread_mutex_init(&mutex->m.lock, &attr); 130 pthread_mutexattr_destroy(&attr); 131 if (err) { 132 error_exit(err, __func__); 133 } 134 mutex->m.initialized = true; 135 } 136 137 void qemu_rec_mutex_destroy(QemuRecMutex *mutex) 138 { 139 qemu_mutex_destroy(&mutex->m); 140 } 141 142 void qemu_rec_mutex_lock_impl(QemuRecMutex *mutex, const char *file, int line) 143 { 144 qemu_mutex_lock_impl(&mutex->m, file, line); 145 } 146 147 int qemu_rec_mutex_trylock_impl(QemuRecMutex *mutex, const char *file, int line) 148 { 149 return qemu_mutex_trylock_impl(&mutex->m, file, line); 150 } 151 152 void qemu_rec_mutex_unlock_impl(QemuRecMutex *mutex, const char *file, int line) 153 { 154 qemu_mutex_unlock_impl(&mutex->m, file, line); 155 } 156 157 void qemu_cond_init(QemuCond *cond) 158 { 159 pthread_condattr_t attr; 160 int err; 161 162 err = pthread_condattr_init(&attr); 163 if (err) { 164 error_exit(err, __func__); 165 } 166 #ifdef CONFIG_PTHREAD_CONDATTR_SETCLOCK 167 err = pthread_condattr_setclock(&attr, qemu_timedwait_clockid()); 168 if (err) { 169 error_exit(err, __func__); 170 } 171 #endif 172 err = pthread_cond_init(&cond->cond, &attr); 173 if (err) { 174 error_exit(err, __func__); 175 } 176 err = pthread_condattr_destroy(&attr); 177 if (err) { 178 error_exit(err, __func__); 179 } 180 cond->initialized = true; 181 } 182 183 void qemu_cond_destroy(QemuCond *cond) 184 { 185 int err; 186 187 assert(cond->initialized); 188 cond->initialized = false; 189 err = pthread_cond_destroy(&cond->cond); 190 if (err) 191 error_exit(err, __func__); 192 } 193 194 void qemu_cond_signal(QemuCond *cond) 195 { 196 int err; 197 198 assert(cond->initialized); 199 err = pthread_cond_signal(&cond->cond); 200 if (err) 201 error_exit(err, __func__); 202 } 203 204 void qemu_cond_broadcast(QemuCond *cond) 205 { 206 int err; 207 208 assert(cond->initialized); 209 err = pthread_cond_broadcast(&cond->cond); 210 if (err) 211 error_exit(err, __func__); 212 } 213 214 void qemu_cond_wait_impl(QemuCond *cond, QemuMutex *mutex, const char *file, const int line) 215 { 216 int err; 217 218 assert(cond->initialized); 219 qemu_mutex_pre_unlock(mutex, file, line); 220 err = pthread_cond_wait(&cond->cond, &mutex->lock); 221 qemu_mutex_post_lock(mutex, file, line); 222 if (err) 223 error_exit(err, __func__); 224 } 225 226 static bool TSA_NO_TSA 227 qemu_cond_timedwait_ts(QemuCond *cond, QemuMutex *mutex, struct timespec *ts, 228 const char *file, const int line) 229 { 230 int err; 231 232 assert(cond->initialized); 233 trace_qemu_mutex_unlock(mutex, file, line); 234 err = pthread_cond_timedwait(&cond->cond, &mutex->lock, ts); 235 trace_qemu_mutex_locked(mutex, file, line); 236 if (err && err != ETIMEDOUT) { 237 error_exit(err, __func__); 238 } 239 return err != ETIMEDOUT; 240 } 241 242 bool qemu_cond_timedwait_impl(QemuCond *cond, QemuMutex *mutex, int ms, 243 const char *file, const int line) 244 { 245 struct timespec ts; 246 247 compute_abs_deadline(&ts, ms); 248 return qemu_cond_timedwait_ts(cond, mutex, &ts, file, line); 249 } 250 251 void qemu_sem_init(QemuSemaphore *sem, int init) 252 { 253 qemu_mutex_init(&sem->mutex); 254 qemu_cond_init(&sem->cond); 255 256 if (init < 0) { 257 error_exit(EINVAL, __func__); 258 } 259 sem->count = init; 260 } 261 262 void qemu_sem_destroy(QemuSemaphore *sem) 263 { 264 qemu_cond_destroy(&sem->cond); 265 qemu_mutex_destroy(&sem->mutex); 266 } 267 268 void qemu_sem_post(QemuSemaphore *sem) 269 { 270 qemu_mutex_lock(&sem->mutex); 271 if (sem->count == UINT_MAX) { 272 error_exit(EINVAL, __func__); 273 } else { 274 sem->count++; 275 qemu_cond_signal(&sem->cond); 276 } 277 qemu_mutex_unlock(&sem->mutex); 278 } 279 280 int qemu_sem_timedwait(QemuSemaphore *sem, int ms) 281 { 282 bool rc = true; 283 struct timespec ts; 284 285 compute_abs_deadline(&ts, ms); 286 qemu_mutex_lock(&sem->mutex); 287 while (sem->count == 0) { 288 if (ms == 0) { 289 rc = false; 290 } else { 291 rc = qemu_cond_timedwait_ts(&sem->cond, &sem->mutex, &ts, 292 __FILE__, __LINE__); 293 } 294 if (!rc) { /* timeout */ 295 break; 296 } 297 } 298 if (rc) { 299 --sem->count; 300 } 301 qemu_mutex_unlock(&sem->mutex); 302 return (rc ? 0 : -1); 303 } 304 305 void qemu_sem_wait(QemuSemaphore *sem) 306 { 307 qemu_mutex_lock(&sem->mutex); 308 while (sem->count == 0) { 309 qemu_cond_wait(&sem->cond, &sem->mutex); 310 } 311 --sem->count; 312 qemu_mutex_unlock(&sem->mutex); 313 } 314 315 #ifdef __linux__ 316 #include "qemu/futex.h" 317 #else 318 static inline void qemu_futex_wake(QemuEvent *ev, int n) 319 { 320 assert(ev->initialized); 321 pthread_mutex_lock(&ev->lock); 322 if (n == 1) { 323 pthread_cond_signal(&ev->cond); 324 } else { 325 pthread_cond_broadcast(&ev->cond); 326 } 327 pthread_mutex_unlock(&ev->lock); 328 } 329 330 static inline void qemu_futex_wait(QemuEvent *ev, unsigned val) 331 { 332 assert(ev->initialized); 333 pthread_mutex_lock(&ev->lock); 334 if (ev->value == val) { 335 pthread_cond_wait(&ev->cond, &ev->lock); 336 } 337 pthread_mutex_unlock(&ev->lock); 338 } 339 #endif 340 341 /* Valid transitions: 342 * - free->set, when setting the event 343 * - busy->set, when setting the event, followed by qemu_futex_wake 344 * - set->free, when resetting the event 345 * - free->busy, when waiting 346 * 347 * set->busy does not happen (it can be observed from the outside but 348 * it really is set->free->busy). 349 * 350 * busy->free provably cannot happen; to enforce it, the set->free transition 351 * is done with an OR, which becomes a no-op if the event has concurrently 352 * transitioned to free or busy. 353 */ 354 355 #define EV_SET 0 356 #define EV_FREE 1 357 #define EV_BUSY -1 358 359 void qemu_event_init(QemuEvent *ev, bool init) 360 { 361 #ifndef __linux__ 362 pthread_mutex_init(&ev->lock, NULL); 363 pthread_cond_init(&ev->cond, NULL); 364 #endif 365 366 ev->value = (init ? EV_SET : EV_FREE); 367 ev->initialized = true; 368 } 369 370 void qemu_event_destroy(QemuEvent *ev) 371 { 372 assert(ev->initialized); 373 ev->initialized = false; 374 #ifndef __linux__ 375 pthread_mutex_destroy(&ev->lock); 376 pthread_cond_destroy(&ev->cond); 377 #endif 378 } 379 380 void qemu_event_set(QemuEvent *ev) 381 { 382 /* qemu_event_set has release semantics, but because it *loads* 383 * ev->value we need a full memory barrier here. 384 */ 385 assert(ev->initialized); 386 smp_mb(); 387 if (qatomic_read(&ev->value) != EV_SET) { 388 if (qatomic_xchg(&ev->value, EV_SET) == EV_BUSY) { 389 /* There were waiters, wake them up. */ 390 qemu_futex_wake(ev, INT_MAX); 391 } 392 } 393 } 394 395 void qemu_event_reset(QemuEvent *ev) 396 { 397 unsigned value; 398 399 assert(ev->initialized); 400 value = qatomic_read(&ev->value); 401 smp_mb_acquire(); 402 if (value == EV_SET) { 403 /* 404 * If there was a concurrent reset (or even reset+wait), 405 * do nothing. Otherwise change EV_SET->EV_FREE. 406 */ 407 qatomic_or(&ev->value, EV_FREE); 408 } 409 } 410 411 void qemu_event_wait(QemuEvent *ev) 412 { 413 unsigned value; 414 415 assert(ev->initialized); 416 value = qatomic_read(&ev->value); 417 smp_mb_acquire(); 418 if (value != EV_SET) { 419 if (value == EV_FREE) { 420 /* 421 * Leave the event reset and tell qemu_event_set that there 422 * are waiters. No need to retry, because there cannot be 423 * a concurrent busy->free transition. After the CAS, the 424 * event will be either set or busy. 425 */ 426 if (qatomic_cmpxchg(&ev->value, EV_FREE, EV_BUSY) == EV_SET) { 427 return; 428 } 429 } 430 qemu_futex_wait(ev, EV_BUSY); 431 } 432 } 433 434 static __thread NotifierList thread_exit; 435 436 /* 437 * Note that in this implementation you can register a thread-exit 438 * notifier for the main thread, but it will never be called. 439 * This is OK because main thread exit can only happen when the 440 * entire process is exiting, and the API allows notifiers to not 441 * be called on process exit. 442 */ 443 void qemu_thread_atexit_add(Notifier *notifier) 444 { 445 notifier_list_add(&thread_exit, notifier); 446 } 447 448 void qemu_thread_atexit_remove(Notifier *notifier) 449 { 450 notifier_remove(notifier); 451 } 452 453 static void qemu_thread_atexit_notify(void *arg) 454 { 455 /* 456 * Called when non-main thread exits (via qemu_thread_exit() 457 * or by returning from its start routine.) 458 */ 459 notifier_list_notify(&thread_exit, NULL); 460 } 461 462 typedef struct { 463 void *(*start_routine)(void *); 464 void *arg; 465 char *name; 466 } QemuThreadArgs; 467 468 static void *qemu_thread_start(void *args) 469 { 470 QemuThreadArgs *qemu_thread_args = args; 471 void *(*start_routine)(void *) = qemu_thread_args->start_routine; 472 void *arg = qemu_thread_args->arg; 473 void *r; 474 475 /* Attempt to set the threads name; note that this is for debug, so 476 * we're not going to fail if we can't set it. 477 */ 478 if (name_threads && qemu_thread_args->name) { 479 # if defined(CONFIG_PTHREAD_SETNAME_NP_W_TID) 480 pthread_setname_np(pthread_self(), qemu_thread_args->name); 481 # elif defined(CONFIG_PTHREAD_SETNAME_NP_WO_TID) 482 pthread_setname_np(qemu_thread_args->name); 483 # endif 484 } 485 QEMU_TSAN_ANNOTATE_THREAD_NAME(qemu_thread_args->name); 486 g_free(qemu_thread_args->name); 487 g_free(qemu_thread_args); 488 489 /* 490 * GCC 11 with glibc 2.17 on PowerPC reports 491 * 492 * qemu-thread-posix.c:540:5: error: ‘__sigsetjmp’ accessing 656 bytes 493 * in a region of size 528 [-Werror=stringop-overflow=] 494 * 540 | pthread_cleanup_push(qemu_thread_atexit_notify, NULL); 495 * | ^~~~~~~~~~~~~~~~~~~~ 496 * 497 * which is clearly nonsense. 498 */ 499 #pragma GCC diagnostic push 500 #ifndef __clang__ 501 #pragma GCC diagnostic ignored "-Wstringop-overflow" 502 #endif 503 504 pthread_cleanup_push(qemu_thread_atexit_notify, NULL); 505 r = start_routine(arg); 506 pthread_cleanup_pop(1); 507 508 #pragma GCC diagnostic pop 509 510 return r; 511 } 512 513 void qemu_thread_create(QemuThread *thread, const char *name, 514 void *(*start_routine)(void*), 515 void *arg, int mode) 516 { 517 sigset_t set, oldset; 518 int err; 519 pthread_attr_t attr; 520 QemuThreadArgs *qemu_thread_args; 521 522 err = pthread_attr_init(&attr); 523 if (err) { 524 error_exit(err, __func__); 525 } 526 527 if (mode == QEMU_THREAD_DETACHED) { 528 pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); 529 } 530 531 /* Leave signal handling to the iothread. */ 532 sigfillset(&set); 533 /* Blocking the signals can result in undefined behaviour. */ 534 sigdelset(&set, SIGSEGV); 535 sigdelset(&set, SIGFPE); 536 sigdelset(&set, SIGILL); 537 /* TODO avoid SIGBUS loss on macOS */ 538 pthread_sigmask(SIG_SETMASK, &set, &oldset); 539 540 qemu_thread_args = g_new0(QemuThreadArgs, 1); 541 qemu_thread_args->name = g_strdup(name); 542 qemu_thread_args->start_routine = start_routine; 543 qemu_thread_args->arg = arg; 544 545 err = pthread_create(&thread->thread, &attr, 546 qemu_thread_start, qemu_thread_args); 547 548 if (err) 549 error_exit(err, __func__); 550 551 pthread_sigmask(SIG_SETMASK, &oldset, NULL); 552 553 pthread_attr_destroy(&attr); 554 } 555 556 int qemu_thread_set_affinity(QemuThread *thread, unsigned long *host_cpus, 557 unsigned long nbits) 558 { 559 #if defined(CONFIG_PTHREAD_AFFINITY_NP) 560 const size_t setsize = CPU_ALLOC_SIZE(nbits); 561 unsigned long value; 562 cpu_set_t *cpuset; 563 int err; 564 565 cpuset = CPU_ALLOC(nbits); 566 g_assert(cpuset); 567 568 CPU_ZERO_S(setsize, cpuset); 569 value = find_first_bit(host_cpus, nbits); 570 while (value < nbits) { 571 CPU_SET_S(value, setsize, cpuset); 572 value = find_next_bit(host_cpus, nbits, value + 1); 573 } 574 575 err = pthread_setaffinity_np(thread->thread, setsize, cpuset); 576 CPU_FREE(cpuset); 577 return err; 578 #else 579 return -ENOSYS; 580 #endif 581 } 582 583 int qemu_thread_get_affinity(QemuThread *thread, unsigned long **host_cpus, 584 unsigned long *nbits) 585 { 586 #if defined(CONFIG_PTHREAD_AFFINITY_NP) 587 unsigned long tmpbits; 588 cpu_set_t *cpuset; 589 size_t setsize; 590 int i, err; 591 592 tmpbits = CPU_SETSIZE; 593 while (true) { 594 setsize = CPU_ALLOC_SIZE(tmpbits); 595 cpuset = CPU_ALLOC(tmpbits); 596 g_assert(cpuset); 597 598 err = pthread_getaffinity_np(thread->thread, setsize, cpuset); 599 if (err) { 600 CPU_FREE(cpuset); 601 if (err != -EINVAL) { 602 return err; 603 } 604 tmpbits *= 2; 605 } else { 606 break; 607 } 608 } 609 610 /* Convert the result into a proper bitmap. */ 611 *nbits = tmpbits; 612 *host_cpus = bitmap_new(tmpbits); 613 for (i = 0; i < tmpbits; i++) { 614 if (CPU_ISSET(i, cpuset)) { 615 set_bit(i, *host_cpus); 616 } 617 } 618 CPU_FREE(cpuset); 619 return 0; 620 #else 621 return -ENOSYS; 622 #endif 623 } 624 625 void qemu_thread_get_self(QemuThread *thread) 626 { 627 thread->thread = pthread_self(); 628 } 629 630 bool qemu_thread_is_self(QemuThread *thread) 631 { 632 return pthread_equal(pthread_self(), thread->thread); 633 } 634 635 void qemu_thread_exit(void *retval) 636 { 637 pthread_exit(retval); 638 } 639 640 void *qemu_thread_join(QemuThread *thread) 641 { 642 int err; 643 void *ret; 644 645 err = pthread_join(thread->thread, &ret); 646 if (err) { 647 error_exit(err, __func__); 648 } 649 return ret; 650 } 651