| /* |
| * Copyright (C) 2008 The Android Open Source Project |
| * All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in |
| * the documentation and/or other materials provided with the |
| * distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS |
| * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE |
| * COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, |
| * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, |
| * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS |
| * OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED |
| * AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, |
| * OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT |
| * OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF |
| * SUCH DAMAGE. |
| */ |
| |
| #include <pthread.h> |
| |
| #include <errno.h> |
| #include <limits.h> |
| #include <stdatomic.h> |
| #include <sys/mman.h> |
| #include <time.h> |
| #include <unistd.h> |
| |
| #include "pthread_internal.h" |
| |
| #include "private/bionic_futex.h" |
| #include "private/bionic_time_conversions.h" |
| #include "private/bionic_tls.h" |
| |
| // We use one bit in pthread_condattr_t (long) values as the 'shared' flag |
| // and one bit for the clock type (CLOCK_REALTIME is ((clockid_t) 1), and |
| // CLOCK_MONOTONIC is ((clockid_t) 0).). The rest of the bits are a counter. |
| // |
| // The 'value' field pthread_cond_t has the same layout. |
| |
| #define COND_SHARED_MASK 0x0001 |
| #define COND_CLOCK_MASK 0x0002 |
| #define COND_COUNTER_STEP 0x0004 |
| #define COND_FLAGS_MASK (COND_SHARED_MASK | COND_CLOCK_MASK) |
| #define COND_COUNTER_MASK (~COND_FLAGS_MASK) |
| |
| #define COND_IS_SHARED(c) (((c) & COND_SHARED_MASK) != 0) |
| #define COND_GET_CLOCK(c) (((c) & COND_CLOCK_MASK) >> 1) |
| #define COND_SET_CLOCK(attr, c) ((attr) | (c << 1)) |
| |
| |
| int pthread_condattr_init(pthread_condattr_t* attr) { |
| *attr = 0; |
| *attr |= PTHREAD_PROCESS_PRIVATE; |
| *attr |= (CLOCK_REALTIME << 1); |
| return 0; |
| } |
| |
| int pthread_condattr_getpshared(const pthread_condattr_t* attr, int* pshared) { |
| *pshared = static_cast<int>(COND_IS_SHARED(*attr)); |
| return 0; |
| } |
| |
| int pthread_condattr_setpshared(pthread_condattr_t* attr, int pshared) { |
| if (pshared != PTHREAD_PROCESS_SHARED && pshared != PTHREAD_PROCESS_PRIVATE) { |
| return EINVAL; |
| } |
| |
| *attr |= pshared; |
| return 0; |
| } |
| |
| int pthread_condattr_getclock(const pthread_condattr_t* attr, clockid_t* clock) { |
| *clock = COND_GET_CLOCK(*attr); |
| return 0; |
| } |
| |
| int pthread_condattr_setclock(pthread_condattr_t* attr, clockid_t clock) { |
| if (clock != CLOCK_MONOTONIC && clock != CLOCK_REALTIME) { |
| return EINVAL; |
| } |
| |
| *attr = COND_SET_CLOCK(*attr, clock); |
| return 0; |
| } |
| |
| int pthread_condattr_destroy(pthread_condattr_t* attr) { |
| *attr = 0xdeada11d; |
| return 0; |
| } |
| |
| static inline atomic_uint* COND_TO_ATOMIC_POINTER(pthread_cond_t* cond) { |
| static_assert(sizeof(atomic_uint) == sizeof(cond->value), |
| "cond->value should actually be atomic_uint in implementation."); |
| |
| // We prefer casting to atomic_uint instead of declaring cond->value to be atomic_uint directly. |
| // Because using the second method pollutes pthread.h, and causes an error when compiling libcxx. |
| return reinterpret_cast<atomic_uint*>(&cond->value); |
| } |
| |
| // XXX *technically* there is a race condition that could allow |
| // XXX a signal to be missed. If thread A is preempted in _wait() |
| // XXX after unlocking the mutex and before waiting, and if other |
| // XXX threads call signal or broadcast UINT_MAX/2 times (exactly), |
| // XXX before thread A is scheduled again and calls futex_wait(), |
| // XXX then the signal will be lost. |
| |
| int pthread_cond_init(pthread_cond_t* cond, const pthread_condattr_t* attr) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| unsigned int init_value = 0; |
| |
| if (attr != NULL) { |
| init_value = (*attr & COND_FLAGS_MASK); |
| } |
| atomic_init(cond_value_ptr, init_value); |
| |
| return 0; |
| } |
| |
| int pthread_cond_destroy(pthread_cond_t* cond) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| atomic_store_explicit(cond_value_ptr, 0xdeadc04d, memory_order_relaxed); |
| return 0; |
| } |
| |
| // This function is used by pthread_cond_broadcast and |
| // pthread_cond_signal to atomically decrement the counter |
| // then wake up thread_count threads. |
| static int __pthread_cond_pulse(atomic_uint* cond_value_ptr, int thread_count) { |
| unsigned int old_value = atomic_load_explicit(cond_value_ptr, memory_order_relaxed); |
| bool shared = COND_IS_SHARED(old_value); |
| |
| // We don't use a release/seq_cst fence here. Because pthread_cond_wait/signal can't be |
| // used as a method for memory synchronization by itself. It should always be used with |
| // pthread mutexes. Note that Spurious wakeups from pthread_cond_wait/timedwait may occur, |
| // so when using condition variables there is always a boolean predicate involving shared |
| // variables associated with each condition wait that is true if the thread should proceed. |
| // If the predicate is seen true before a condition wait, pthread_cond_wait/timedwait will |
| // not be called. That's why pthread_wait/signal pair can't be used as a method for memory |
| // synchronization. And it doesn't help even if we use any fence here. |
| |
| // The increase of value should leave flags alone, even if the value can overflows. |
| atomic_fetch_add_explicit(cond_value_ptr, COND_COUNTER_STEP, memory_order_relaxed); |
| |
| __futex_wake_ex(cond_value_ptr, shared, thread_count); |
| return 0; |
| } |
| |
| __LIBC_HIDDEN__ |
| int __pthread_cond_timedwait_relative(atomic_uint* cond_value_ptr, pthread_mutex_t* mutex, |
| const timespec* reltime) { |
| unsigned int old_value = atomic_load_explicit(cond_value_ptr, memory_order_relaxed); |
| bool shared = COND_IS_SHARED(old_value); |
| |
| pthread_mutex_unlock(mutex); |
| int status = __futex_wait_ex(cond_value_ptr, shared, old_value, reltime); |
| pthread_mutex_lock(mutex); |
| |
| if (status == -ETIMEDOUT) { |
| return ETIMEDOUT; |
| } |
| return 0; |
| } |
| |
| __LIBC_HIDDEN__ |
| int __pthread_cond_timedwait(atomic_uint* cond_value_ptr, pthread_mutex_t* mutex, |
| const timespec* abs_ts, clockid_t clock) { |
| timespec ts; |
| timespec* tsp; |
| |
| if (abs_ts != NULL) { |
| if (!timespec_from_absolute_timespec(ts, *abs_ts, clock)) { |
| return ETIMEDOUT; |
| } |
| tsp = &ts; |
| } else { |
| tsp = NULL; |
| } |
| |
| return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, tsp); |
| } |
| |
| int pthread_cond_broadcast(pthread_cond_t* cond) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_pulse(cond_value_ptr, INT_MAX); |
| } |
| |
| int pthread_cond_signal(pthread_cond_t* cond) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_pulse(cond_value_ptr, 1); |
| } |
| |
| int pthread_cond_wait(pthread_cond_t* cond, pthread_mutex_t* mutex) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait(cond_value_ptr, mutex, NULL, |
| COND_GET_CLOCK(atomic_load_explicit(cond_value_ptr, memory_order_relaxed))); |
| } |
| |
| int pthread_cond_timedwait(pthread_cond_t *cond, pthread_mutex_t * mutex, const timespec *abstime) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime, |
| COND_GET_CLOCK(atomic_load_explicit(cond_value_ptr, memory_order_relaxed))); |
| } |
| |
| #if !defined(__LP64__) |
| // TODO: this exists only for backward binary compatibility on 32 bit platforms. |
| extern "C" int pthread_cond_timedwait_monotonic(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime, CLOCK_MONOTONIC); |
| } |
| |
| extern "C" int pthread_cond_timedwait_monotonic_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* abstime) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait(cond_value_ptr, mutex, abstime, CLOCK_MONOTONIC); |
| } |
| |
| extern "C" int pthread_cond_timedwait_relative_np(pthread_cond_t* cond, pthread_mutex_t* mutex, const timespec* reltime) { |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, reltime); |
| } |
| |
| extern "C" int pthread_cond_timeout_np(pthread_cond_t* cond, pthread_mutex_t* mutex, unsigned ms) { |
| timespec ts; |
| timespec_from_ms(ts, ms); |
| atomic_uint* cond_value_ptr = COND_TO_ATOMIC_POINTER(cond); |
| return __pthread_cond_timedwait_relative(cond_value_ptr, mutex, &ts); |
| } |
| #endif // !defined(__LP64__) |