|  | /* | 
|  | * Wrappers around mutex/cond/thread functions | 
|  | * | 
|  | * Copyright Red Hat, Inc. 2009 | 
|  | * | 
|  | * Author: | 
|  | *  Marcelo Tosatti <mtosatti@redhat.com> | 
|  | * | 
|  | * This work is licensed under the terms of the GNU GPL, version 2 or later. | 
|  | * See the COPYING file in the top-level directory. | 
|  | * | 
|  | */ | 
|  | #include "qemu/osdep.h" | 
|  | #include "qemu/thread.h" | 
|  | #include "qemu/atomic.h" | 
|  | #include "qemu/notify.h" | 
|  | #include "qemu-thread-common.h" | 
|  | #include "qemu/tsan.h" | 
|  | #include "qemu/bitmap.h" | 
|  |  | 
|  | #ifdef CONFIG_PTHREAD_SET_NAME_NP | 
|  | #include <pthread_np.h> | 
|  | #endif | 
|  |  | 
|  | static bool name_threads; | 
|  |  | 
|  | void qemu_thread_naming(bool enable) | 
|  | { | 
|  | name_threads = enable; | 
|  |  | 
|  | #if !defined CONFIG_PTHREAD_SETNAME_NP_W_TID && \ | 
|  | !defined CONFIG_PTHREAD_SETNAME_NP_WO_TID && \ | 
|  | !defined CONFIG_PTHREAD_SET_NAME_NP | 
|  | /* This is a debugging option, not fatal */ | 
|  | if (enable) { | 
|  | fprintf(stderr, "qemu: thread naming not supported on this host\n"); | 
|  | } | 
|  | #endif | 
|  | } | 
|  |  | 
|  | static void error_exit(int err, const char *msg) | 
|  | { | 
|  | fprintf(stderr, "qemu: %s: %s\n", msg, strerror(err)); | 
|  | abort(); | 
|  | } | 
|  |  | 
|  | static inline clockid_t qemu_timedwait_clockid(void) | 
|  | { | 
|  | #ifdef CONFIG_PTHREAD_CONDATTR_SETCLOCK | 
|  | return CLOCK_MONOTONIC; | 
|  | #else | 
|  | return CLOCK_REALTIME; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | static void compute_abs_deadline(struct timespec *ts, int ms) | 
|  | { | 
|  | clock_gettime(qemu_timedwait_clockid(), ts); | 
|  | ts->tv_nsec += (ms % 1000) * 1000000; | 
|  | ts->tv_sec += ms / 1000; | 
|  | if (ts->tv_nsec >= 1000000000) { | 
|  | ts->tv_sec++; | 
|  | ts->tv_nsec -= 1000000000; | 
|  | } | 
|  | } | 
|  |  | 
|  | void qemu_mutex_init(QemuMutex *mutex) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | err = pthread_mutex_init(&mutex->lock, NULL); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | qemu_mutex_post_init(mutex); | 
|  | } | 
|  |  | 
|  | void qemu_mutex_destroy(QemuMutex *mutex) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(mutex->initialized); | 
|  | mutex->initialized = false; | 
|  | err = pthread_mutex_destroy(&mutex->lock); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | void qemu_mutex_lock_impl(QemuMutex *mutex, const char *file, const int line) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(mutex->initialized); | 
|  | qemu_mutex_pre_lock(mutex, file, line); | 
|  | err = pthread_mutex_lock(&mutex->lock); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | qemu_mutex_post_lock(mutex, file, line); | 
|  | } | 
|  |  | 
|  | int qemu_mutex_trylock_impl(QemuMutex *mutex, const char *file, const int line) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(mutex->initialized); | 
|  | err = pthread_mutex_trylock(&mutex->lock); | 
|  | if (err == 0) { | 
|  | qemu_mutex_post_lock(mutex, file, line); | 
|  | return 0; | 
|  | } | 
|  | if (err != EBUSY) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | void qemu_mutex_unlock_impl(QemuMutex *mutex, const char *file, const int line) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(mutex->initialized); | 
|  | qemu_mutex_pre_unlock(mutex, file, line); | 
|  | err = pthread_mutex_unlock(&mutex->lock); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | void qemu_rec_mutex_init(QemuRecMutex *mutex) | 
|  | { | 
|  | int err; | 
|  | pthread_mutexattr_t attr; | 
|  |  | 
|  | pthread_mutexattr_init(&attr); | 
|  | pthread_mutexattr_settype(&attr, PTHREAD_MUTEX_RECURSIVE); | 
|  | err = pthread_mutex_init(&mutex->m.lock, &attr); | 
|  | pthread_mutexattr_destroy(&attr); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | mutex->m.initialized = true; | 
|  | } | 
|  |  | 
|  | void qemu_rec_mutex_destroy(QemuRecMutex *mutex) | 
|  | { | 
|  | qemu_mutex_destroy(&mutex->m); | 
|  | } | 
|  |  | 
|  | void qemu_rec_mutex_lock_impl(QemuRecMutex *mutex, const char *file, int line) | 
|  | { | 
|  | qemu_mutex_lock_impl(&mutex->m, file, line); | 
|  | } | 
|  |  | 
|  | int qemu_rec_mutex_trylock_impl(QemuRecMutex *mutex, const char *file, int line) | 
|  | { | 
|  | return qemu_mutex_trylock_impl(&mutex->m, file, line); | 
|  | } | 
|  |  | 
|  | void qemu_rec_mutex_unlock_impl(QemuRecMutex *mutex, const char *file, int line) | 
|  | { | 
|  | qemu_mutex_unlock_impl(&mutex->m, file, line); | 
|  | } | 
|  |  | 
|  | void qemu_cond_init(QemuCond *cond) | 
|  | { | 
|  | pthread_condattr_t attr; | 
|  | int err; | 
|  |  | 
|  | err = pthread_condattr_init(&attr); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | #ifdef CONFIG_PTHREAD_CONDATTR_SETCLOCK | 
|  | err = pthread_condattr_setclock(&attr, qemu_timedwait_clockid()); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | #endif | 
|  | err = pthread_cond_init(&cond->cond, &attr); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | err = pthread_condattr_destroy(&attr); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | cond->initialized = true; | 
|  | } | 
|  |  | 
|  | void qemu_cond_destroy(QemuCond *cond) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(cond->initialized); | 
|  | cond->initialized = false; | 
|  | err = pthread_cond_destroy(&cond->cond); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | void qemu_cond_signal(QemuCond *cond) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(cond->initialized); | 
|  | err = pthread_cond_signal(&cond->cond); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | void qemu_cond_broadcast(QemuCond *cond) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(cond->initialized); | 
|  | err = pthread_cond_broadcast(&cond->cond); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | void qemu_cond_wait_impl(QemuCond *cond, QemuMutex *mutex, const char *file, const int line) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(cond->initialized); | 
|  | qemu_mutex_pre_unlock(mutex, file, line); | 
|  | err = pthread_cond_wait(&cond->cond, &mutex->lock); | 
|  | qemu_mutex_post_lock(mutex, file, line); | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | static bool TSA_NO_TSA | 
|  | qemu_cond_timedwait_ts(QemuCond *cond, QemuMutex *mutex, struct timespec *ts, | 
|  | const char *file, const int line) | 
|  | { | 
|  | int err; | 
|  |  | 
|  | assert(cond->initialized); | 
|  | trace_qemu_mutex_unlock(mutex, file, line); | 
|  | err = pthread_cond_timedwait(&cond->cond, &mutex->lock, ts); | 
|  | trace_qemu_mutex_locked(mutex, file, line); | 
|  | if (err && err != ETIMEDOUT) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | return err != ETIMEDOUT; | 
|  | } | 
|  |  | 
|  | bool qemu_cond_timedwait_impl(QemuCond *cond, QemuMutex *mutex, int ms, | 
|  | const char *file, const int line) | 
|  | { | 
|  | struct timespec ts; | 
|  |  | 
|  | compute_abs_deadline(&ts, ms); | 
|  | return qemu_cond_timedwait_ts(cond, mutex, &ts, file, line); | 
|  | } | 
|  |  | 
|  | void qemu_sem_init(QemuSemaphore *sem, int init) | 
|  | { | 
|  | qemu_mutex_init(&sem->mutex); | 
|  | qemu_cond_init(&sem->cond); | 
|  |  | 
|  | if (init < 0) { | 
|  | error_exit(EINVAL, __func__); | 
|  | } | 
|  | sem->count = init; | 
|  | } | 
|  |  | 
|  | void qemu_sem_destroy(QemuSemaphore *sem) | 
|  | { | 
|  | qemu_cond_destroy(&sem->cond); | 
|  | qemu_mutex_destroy(&sem->mutex); | 
|  | } | 
|  |  | 
|  | void qemu_sem_post(QemuSemaphore *sem) | 
|  | { | 
|  | qemu_mutex_lock(&sem->mutex); | 
|  | if (sem->count == UINT_MAX) { | 
|  | error_exit(EINVAL, __func__); | 
|  | } else { | 
|  | sem->count++; | 
|  | qemu_cond_signal(&sem->cond); | 
|  | } | 
|  | qemu_mutex_unlock(&sem->mutex); | 
|  | } | 
|  |  | 
|  | int qemu_sem_timedwait(QemuSemaphore *sem, int ms) | 
|  | { | 
|  | bool rc = true; | 
|  | struct timespec ts; | 
|  |  | 
|  | compute_abs_deadline(&ts, ms); | 
|  | qemu_mutex_lock(&sem->mutex); | 
|  | while (sem->count == 0) { | 
|  | if (ms == 0) { | 
|  | rc = false; | 
|  | } else { | 
|  | rc = qemu_cond_timedwait_ts(&sem->cond, &sem->mutex, &ts, | 
|  | __FILE__, __LINE__); | 
|  | } | 
|  | if (!rc) { /* timeout */ | 
|  | break; | 
|  | } | 
|  | } | 
|  | if (rc) { | 
|  | --sem->count; | 
|  | } | 
|  | qemu_mutex_unlock(&sem->mutex); | 
|  | return (rc ? 0 : -1); | 
|  | } | 
|  |  | 
|  | void qemu_sem_wait(QemuSemaphore *sem) | 
|  | { | 
|  | qemu_mutex_lock(&sem->mutex); | 
|  | while (sem->count == 0) { | 
|  | qemu_cond_wait(&sem->cond, &sem->mutex); | 
|  | } | 
|  | --sem->count; | 
|  | qemu_mutex_unlock(&sem->mutex); | 
|  | } | 
|  |  | 
|  | #ifdef __linux__ | 
|  | #include "qemu/futex.h" | 
|  | #else | 
|  | static inline void qemu_futex_wake(QemuEvent *ev, int n) | 
|  | { | 
|  | assert(ev->initialized); | 
|  | pthread_mutex_lock(&ev->lock); | 
|  | if (n == 1) { | 
|  | pthread_cond_signal(&ev->cond); | 
|  | } else { | 
|  | pthread_cond_broadcast(&ev->cond); | 
|  | } | 
|  | pthread_mutex_unlock(&ev->lock); | 
|  | } | 
|  |  | 
|  | static inline void qemu_futex_wait(QemuEvent *ev, unsigned val) | 
|  | { | 
|  | assert(ev->initialized); | 
|  | pthread_mutex_lock(&ev->lock); | 
|  | if (ev->value == val) { | 
|  | pthread_cond_wait(&ev->cond, &ev->lock); | 
|  | } | 
|  | pthread_mutex_unlock(&ev->lock); | 
|  | } | 
|  | #endif | 
|  |  | 
|  | /* Valid transitions: | 
|  | * - free->set, when setting the event | 
|  | * - busy->set, when setting the event, followed by qemu_futex_wake | 
|  | * - set->free, when resetting the event | 
|  | * - free->busy, when waiting | 
|  | * | 
|  | * set->busy does not happen (it can be observed from the outside but | 
|  | * it really is set->free->busy). | 
|  | * | 
|  | * busy->free provably cannot happen; to enforce it, the set->free transition | 
|  | * is done with an OR, which becomes a no-op if the event has concurrently | 
|  | * transitioned to free or busy. | 
|  | */ | 
|  |  | 
|  | #define EV_SET         0 | 
|  | #define EV_FREE        1 | 
|  | #define EV_BUSY       -1 | 
|  |  | 
|  | void qemu_event_init(QemuEvent *ev, bool init) | 
|  | { | 
|  | #ifndef __linux__ | 
|  | pthread_mutex_init(&ev->lock, NULL); | 
|  | pthread_cond_init(&ev->cond, NULL); | 
|  | #endif | 
|  |  | 
|  | ev->value = (init ? EV_SET : EV_FREE); | 
|  | ev->initialized = true; | 
|  | } | 
|  |  | 
|  | void qemu_event_destroy(QemuEvent *ev) | 
|  | { | 
|  | assert(ev->initialized); | 
|  | ev->initialized = false; | 
|  | #ifndef __linux__ | 
|  | pthread_mutex_destroy(&ev->lock); | 
|  | pthread_cond_destroy(&ev->cond); | 
|  | #endif | 
|  | } | 
|  |  | 
|  | void qemu_event_set(QemuEvent *ev) | 
|  | { | 
|  | assert(ev->initialized); | 
|  |  | 
|  | /* | 
|  | * Pairs with both qemu_event_reset() and qemu_event_wait(). | 
|  | * | 
|  | * qemu_event_set has release semantics, but because it *loads* | 
|  | * ev->value we need a full memory barrier here. | 
|  | */ | 
|  | smp_mb(); | 
|  | if (qatomic_read(&ev->value) != EV_SET) { | 
|  | int old = qatomic_xchg(&ev->value, EV_SET); | 
|  |  | 
|  | /* Pairs with memory barrier in kernel futex_wait system call.  */ | 
|  | smp_mb__after_rmw(); | 
|  | if (old == EV_BUSY) { | 
|  | /* There were waiters, wake them up.  */ | 
|  | qemu_futex_wake(ev, INT_MAX); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void qemu_event_reset(QemuEvent *ev) | 
|  | { | 
|  | assert(ev->initialized); | 
|  |  | 
|  | /* | 
|  | * If there was a concurrent reset (or even reset+wait), | 
|  | * do nothing.  Otherwise change EV_SET->EV_FREE. | 
|  | */ | 
|  | qatomic_or(&ev->value, EV_FREE); | 
|  |  | 
|  | /* | 
|  | * Order reset before checking the condition in the caller. | 
|  | * Pairs with the first memory barrier in qemu_event_set(). | 
|  | */ | 
|  | smp_mb__after_rmw(); | 
|  | } | 
|  |  | 
|  | void qemu_event_wait(QemuEvent *ev) | 
|  | { | 
|  | unsigned value; | 
|  |  | 
|  | assert(ev->initialized); | 
|  |  | 
|  | /* | 
|  | * qemu_event_wait must synchronize with qemu_event_set even if it does | 
|  | * not go down the slow path, so this load-acquire is needed that | 
|  | * synchronizes with the first memory barrier in qemu_event_set(). | 
|  | * | 
|  | * If we do go down the slow path, there is no requirement at all: we | 
|  | * might miss a qemu_event_set() here but ultimately the memory barrier in | 
|  | * qemu_futex_wait() will ensure the check is done correctly. | 
|  | */ | 
|  | value = qatomic_load_acquire(&ev->value); | 
|  | if (value != EV_SET) { | 
|  | if (value == EV_FREE) { | 
|  | /* | 
|  | * Leave the event reset and tell qemu_event_set that there are | 
|  | * waiters.  No need to retry, because there cannot be a concurrent | 
|  | * busy->free transition.  After the CAS, the event will be either | 
|  | * set or busy. | 
|  | * | 
|  | * This cmpxchg doesn't have particular ordering requirements if it | 
|  | * succeeds (moving the store earlier can only cause qemu_event_set() | 
|  | * to issue _more_ wakeups), the failing case needs acquire semantics | 
|  | * like the load above. | 
|  | */ | 
|  | if (qatomic_cmpxchg(&ev->value, EV_FREE, EV_BUSY) == EV_SET) { | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* | 
|  | * This is the final check for a concurrent set, so it does need | 
|  | * a smp_mb() pairing with the second barrier of qemu_event_set(). | 
|  | * The barrier is inside the FUTEX_WAIT system call. | 
|  | */ | 
|  | qemu_futex_wait(ev, EV_BUSY); | 
|  | } | 
|  | } | 
|  |  | 
|  | static __thread NotifierList thread_exit; | 
|  |  | 
|  | /* | 
|  | * Note that in this implementation you can register a thread-exit | 
|  | * notifier for the main thread, but it will never be called. | 
|  | * This is OK because main thread exit can only happen when the | 
|  | * entire process is exiting, and the API allows notifiers to not | 
|  | * be called on process exit. | 
|  | */ | 
|  | void qemu_thread_atexit_add(Notifier *notifier) | 
|  | { | 
|  | notifier_list_add(&thread_exit, notifier); | 
|  | } | 
|  |  | 
|  | void qemu_thread_atexit_remove(Notifier *notifier) | 
|  | { | 
|  | notifier_remove(notifier); | 
|  | } | 
|  |  | 
|  | static void qemu_thread_atexit_notify(void *arg) | 
|  | { | 
|  | /* | 
|  | * Called when non-main thread exits (via qemu_thread_exit() | 
|  | * or by returning from its start routine.) | 
|  | */ | 
|  | notifier_list_notify(&thread_exit, NULL); | 
|  | } | 
|  |  | 
|  | typedef struct { | 
|  | void *(*start_routine)(void *); | 
|  | void *arg; | 
|  | char *name; | 
|  | } QemuThreadArgs; | 
|  |  | 
|  | static void *qemu_thread_start(void *args) | 
|  | { | 
|  | QemuThreadArgs *qemu_thread_args = args; | 
|  | void *(*start_routine)(void *) = qemu_thread_args->start_routine; | 
|  | void *arg = qemu_thread_args->arg; | 
|  | void *r; | 
|  |  | 
|  | /* Attempt to set the threads name; note that this is for debug, so | 
|  | * we're not going to fail if we can't set it. | 
|  | */ | 
|  | if (name_threads && qemu_thread_args->name) { | 
|  | # if defined(CONFIG_PTHREAD_SETNAME_NP_W_TID) | 
|  | pthread_setname_np(pthread_self(), qemu_thread_args->name); | 
|  | # elif defined(CONFIG_PTHREAD_SETNAME_NP_WO_TID) | 
|  | pthread_setname_np(qemu_thread_args->name); | 
|  | # elif defined(CONFIG_PTHREAD_SET_NAME_NP) | 
|  | pthread_set_name_np(pthread_self(), qemu_thread_args->name); | 
|  | # endif | 
|  | } | 
|  | QEMU_TSAN_ANNOTATE_THREAD_NAME(qemu_thread_args->name); | 
|  | g_free(qemu_thread_args->name); | 
|  | g_free(qemu_thread_args); | 
|  |  | 
|  | /* | 
|  | * GCC 11 with glibc 2.17 on PowerPC reports | 
|  | * | 
|  | * qemu-thread-posix.c:540:5: error: ‘__sigsetjmp’ accessing 656 bytes | 
|  | *   in a region of size 528 [-Werror=stringop-overflow=] | 
|  | * 540 |     pthread_cleanup_push(qemu_thread_atexit_notify, NULL); | 
|  | *     |     ^~~~~~~~~~~~~~~~~~~~ | 
|  | * | 
|  | * which is clearly nonsense. | 
|  | */ | 
|  | #pragma GCC diagnostic push | 
|  | #ifndef __clang__ | 
|  | #pragma GCC diagnostic ignored "-Wstringop-overflow" | 
|  | #endif | 
|  |  | 
|  | pthread_cleanup_push(qemu_thread_atexit_notify, NULL); | 
|  | r = start_routine(arg); | 
|  | pthread_cleanup_pop(1); | 
|  |  | 
|  | #pragma GCC diagnostic pop | 
|  |  | 
|  | return r; | 
|  | } | 
|  |  | 
|  | void qemu_thread_create(QemuThread *thread, const char *name, | 
|  | void *(*start_routine)(void*), | 
|  | void *arg, int mode) | 
|  | { | 
|  | sigset_t set, oldset; | 
|  | int err; | 
|  | pthread_attr_t attr; | 
|  | QemuThreadArgs *qemu_thread_args; | 
|  |  | 
|  | err = pthread_attr_init(&attr); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  |  | 
|  | if (mode == QEMU_THREAD_DETACHED) { | 
|  | pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_DETACHED); | 
|  | } | 
|  |  | 
|  | /* Leave signal handling to the iothread.  */ | 
|  | sigfillset(&set); | 
|  | /* Blocking the signals can result in undefined behaviour. */ | 
|  | sigdelset(&set, SIGSEGV); | 
|  | sigdelset(&set, SIGFPE); | 
|  | sigdelset(&set, SIGILL); | 
|  | /* TODO avoid SIGBUS loss on macOS */ | 
|  | pthread_sigmask(SIG_SETMASK, &set, &oldset); | 
|  |  | 
|  | qemu_thread_args = g_new0(QemuThreadArgs, 1); | 
|  | qemu_thread_args->name = g_strdup(name); | 
|  | qemu_thread_args->start_routine = start_routine; | 
|  | qemu_thread_args->arg = arg; | 
|  |  | 
|  | err = pthread_create(&thread->thread, &attr, | 
|  | qemu_thread_start, qemu_thread_args); | 
|  |  | 
|  | if (err) | 
|  | error_exit(err, __func__); | 
|  |  | 
|  | pthread_sigmask(SIG_SETMASK, &oldset, NULL); | 
|  |  | 
|  | pthread_attr_destroy(&attr); | 
|  | } | 
|  |  | 
|  | int qemu_thread_set_affinity(QemuThread *thread, unsigned long *host_cpus, | 
|  | unsigned long nbits) | 
|  | { | 
|  | #if defined(CONFIG_PTHREAD_AFFINITY_NP) | 
|  | const size_t setsize = CPU_ALLOC_SIZE(nbits); | 
|  | unsigned long value; | 
|  | cpu_set_t *cpuset; | 
|  | int err; | 
|  |  | 
|  | cpuset = CPU_ALLOC(nbits); | 
|  | g_assert(cpuset); | 
|  |  | 
|  | CPU_ZERO_S(setsize, cpuset); | 
|  | value = find_first_bit(host_cpus, nbits); | 
|  | while (value < nbits) { | 
|  | CPU_SET_S(value, setsize, cpuset); | 
|  | value = find_next_bit(host_cpus, nbits, value + 1); | 
|  | } | 
|  |  | 
|  | err = pthread_setaffinity_np(thread->thread, setsize, cpuset); | 
|  | CPU_FREE(cpuset); | 
|  | return err; | 
|  | #else | 
|  | return -ENOSYS; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | int qemu_thread_get_affinity(QemuThread *thread, unsigned long **host_cpus, | 
|  | unsigned long *nbits) | 
|  | { | 
|  | #if defined(CONFIG_PTHREAD_AFFINITY_NP) | 
|  | unsigned long tmpbits; | 
|  | cpu_set_t *cpuset; | 
|  | size_t setsize; | 
|  | int i, err; | 
|  |  | 
|  | tmpbits = CPU_SETSIZE; | 
|  | while (true) { | 
|  | setsize = CPU_ALLOC_SIZE(tmpbits); | 
|  | cpuset = CPU_ALLOC(tmpbits); | 
|  | g_assert(cpuset); | 
|  |  | 
|  | err = pthread_getaffinity_np(thread->thread, setsize, cpuset); | 
|  | if (err) { | 
|  | CPU_FREE(cpuset); | 
|  | if (err != -EINVAL) { | 
|  | return err; | 
|  | } | 
|  | tmpbits *= 2; | 
|  | } else { | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | /* Convert the result into a proper bitmap. */ | 
|  | *nbits = tmpbits; | 
|  | *host_cpus = bitmap_new(tmpbits); | 
|  | for (i = 0; i < tmpbits; i++) { | 
|  | if (CPU_ISSET(i, cpuset)) { | 
|  | set_bit(i, *host_cpus); | 
|  | } | 
|  | } | 
|  | CPU_FREE(cpuset); | 
|  | return 0; | 
|  | #else | 
|  | return -ENOSYS; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | void qemu_thread_get_self(QemuThread *thread) | 
|  | { | 
|  | thread->thread = pthread_self(); | 
|  | } | 
|  |  | 
|  | bool qemu_thread_is_self(QemuThread *thread) | 
|  | { | 
|  | return pthread_equal(pthread_self(), thread->thread); | 
|  | } | 
|  |  | 
|  | void qemu_thread_exit(void *retval) | 
|  | { | 
|  | pthread_exit(retval); | 
|  | } | 
|  |  | 
|  | void *qemu_thread_join(QemuThread *thread) | 
|  | { | 
|  | int err; | 
|  | void *ret; | 
|  |  | 
|  | err = pthread_join(thread->thread, &ret); | 
|  | if (err) { | 
|  | error_exit(err, __func__); | 
|  | } | 
|  | return ret; | 
|  | } |