|  | /* | 
|  | * Copyright (C) 2007 The Android Open Source Project | 
|  | * | 
|  | * Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | * you may not use this file except in compliance with the License. | 
|  | * You may obtain a copy of the License at | 
|  | * | 
|  | *      http://www.apache.org/licenses/LICENSE-2.0 | 
|  | * | 
|  | * Unless required by applicable law or agreed to in writing, software | 
|  | * distributed under the License is distributed on an "AS IS" BASIS, | 
|  | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | * See the License for the specific language governing permissions and | 
|  | * limitations under the License. | 
|  | */ | 
|  |  | 
|  | #ifndef ANDROID_CUTILS_ATOMIC_H | 
|  | #define ANDROID_CUTILS_ATOMIC_H | 
|  |  | 
|  | #include <stdint.h> | 
|  | #include <sys/types.h> | 
|  | #include <stdatomic.h> | 
|  |  | 
|  | #ifndef ANDROID_ATOMIC_INLINE | 
|  | #define ANDROID_ATOMIC_INLINE static inline | 
|  | #endif | 
|  |  | 
|  | /* | 
|  | * A handful of basic atomic operations. | 
|  | * THESE ARE HERE FOR LEGACY REASONS ONLY.  AVOID. | 
|  | * | 
|  | * PREFERRED ALTERNATIVES: | 
|  | * - Use C++/C/pthread locks/mutexes whenever there is not a | 
|  | *   convincing reason to do otherwise.  Note that very clever and | 
|  | *   complicated, but correct, lock-free code is often slower than | 
|  | *   using locks, especially where nontrivial data structures | 
|  | *   are involved. | 
|  | * - C11 stdatomic.h. | 
|  | * - Where supported, C++11 std::atomic<T> . | 
|  | * | 
|  | * PLEASE STOP READING HERE UNLESS YOU ARE TRYING TO UNDERSTAND | 
|  | * OR UPDATE OLD CODE. | 
|  | * | 
|  | * The "acquire" and "release" terms can be defined intuitively in terms | 
|  | * of the placement of memory barriers in a simple lock implementation: | 
|  | *   - wait until compare-and-swap(lock-is-free --> lock-is-held) succeeds | 
|  | *   - barrier | 
|  | *   - [do work] | 
|  | *   - barrier | 
|  | *   - store(lock-is-free) | 
|  | * In very crude terms, the initial (acquire) barrier prevents any of the | 
|  | * "work" from happening before the lock is held, and the later (release) | 
|  | * barrier ensures that all of the work happens before the lock is released. | 
|  | * (Think of cached writes, cache read-ahead, and instruction reordering | 
|  | * around the CAS and store instructions.) | 
|  | * | 
|  | * The barriers must apply to both the compiler and the CPU.  Note it is | 
|  | * legal for instructions that occur before an "acquire" barrier to be | 
|  | * moved down below it, and for instructions that occur after a "release" | 
|  | * barrier to be moved up above it. | 
|  | * | 
|  | * The ARM-driven implementation we use here is short on subtlety, | 
|  | * and actually requests a full barrier from the compiler and the CPU. | 
|  | * The only difference between acquire and release is in whether they | 
|  | * are issued before or after the atomic operation with which they | 
|  | * are associated.  To ease the transition to C/C++ atomic intrinsics, | 
|  | * you should not rely on this, and instead assume that only the minimal | 
|  | * acquire/release protection is provided. | 
|  | * | 
|  | * NOTE: all int32_t* values are expected to be aligned on 32-bit boundaries. | 
|  | * If they are not, atomicity is not guaranteed. | 
|  | */ | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | volatile atomic_int_least32_t* to_atomic_int_least32_t(volatile const int32_t* addr) { | 
|  | #ifdef __cplusplus | 
|  | return reinterpret_cast<volatile atomic_int_least32_t*>(const_cast<volatile int32_t*>(addr)); | 
|  | #else | 
|  | return (volatile atomic_int_least32_t*)addr; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Basic arithmetic and bitwise operations.  These all provide a | 
|  | * barrier with "release" ordering, and return the previous value. | 
|  | * | 
|  | * These have the same characteristics (e.g. what happens on overflow) | 
|  | * as the equivalent non-atomic C operations. | 
|  | */ | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_inc(volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | /* Int32_t, if it exists, is the same as int_least32_t. */ | 
|  | return atomic_fetch_add_explicit(a, 1, memory_order_release); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_dec(volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return atomic_fetch_sub_explicit(a, 1, memory_order_release); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_add(int32_t value, volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return atomic_fetch_add_explicit(a, value, memory_order_release); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_and(int32_t value, volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return atomic_fetch_and_explicit(a, value, memory_order_release); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_or(int32_t value, volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return atomic_fetch_or_explicit(a, value, memory_order_release); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Perform an atomic load with "acquire" or "release" ordering. | 
|  | * | 
|  | * Note that the notion of a "release" ordering for a load does not | 
|  | * really fit into the C11 or C++11 memory model.  The extra ordering | 
|  | * is normally observable only by code using memory_order_relaxed | 
|  | * atomics, or data races.  In the rare cases in which such ordering | 
|  | * is called for, use memory_order_relaxed atomics and a leading | 
|  | * atomic_thread_fence (typically with memory_order_acquire, | 
|  | * not memory_order_release!) instead.  If you do not understand | 
|  | * this comment, you are in the vast majority, and should not be | 
|  | * using release loads or replacing them with anything other than | 
|  | * locks or default sequentially consistent atomics. | 
|  | */ | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_acquire_load(volatile const int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return atomic_load_explicit(a, memory_order_acquire); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int32_t android_atomic_release_load(volatile const int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | atomic_thread_fence(memory_order_seq_cst); | 
|  | /* Any reasonable clients of this interface would probably prefer   */ | 
|  | /* something weaker.  But some remaining clients seem to be         */ | 
|  | /* abusing this API in strange ways, e.g. by using it as a fence.   */ | 
|  | /* Thus we are conservative until we can get rid of remaining       */ | 
|  | /* clients (and this function).                                     */ | 
|  | return atomic_load_explicit(a, memory_order_relaxed); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Perform an atomic store with "acquire" or "release" ordering. | 
|  | * | 
|  | * Note that the notion of an "acquire" ordering for a store does not | 
|  | * really fit into the C11 or C++11 memory model.  The extra ordering | 
|  | * is normally observable only by code using memory_order_relaxed | 
|  | * atomics, or data races.  In the rare cases in which such ordering | 
|  | * is called for, use memory_order_relaxed atomics and a trailing | 
|  | * atomic_thread_fence (typically with memory_order_release, | 
|  | * not memory_order_acquire!) instead. | 
|  | */ | 
|  | ANDROID_ATOMIC_INLINE | 
|  | void android_atomic_acquire_store(int32_t value, volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | atomic_store_explicit(a, value, memory_order_relaxed); | 
|  | atomic_thread_fence(memory_order_seq_cst); | 
|  | /* Again overly conservative to accomodate weird clients.   */ | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | void android_atomic_release_store(int32_t value, volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | atomic_store_explicit(a, value, memory_order_release); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Compare-and-set operation with "acquire" or "release" ordering. | 
|  | * | 
|  | * This returns zero if the new value was successfully stored, which will | 
|  | * only happen when *addr == oldvalue. | 
|  | * | 
|  | * (The return value is inverted from implementations on other platforms, | 
|  | * but matches the ARM ldrex/strex result.) | 
|  | * | 
|  | * Implementations that use the release CAS in a loop may be less efficient | 
|  | * than possible, because we re-issue the memory barrier on each iteration. | 
|  | */ | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int android_atomic_acquire_cas(int32_t oldvalue, int32_t newvalue, | 
|  | volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return !atomic_compare_exchange_strong_explicit( | 
|  | a, &oldvalue, newvalue, | 
|  | memory_order_acquire, | 
|  | memory_order_acquire); | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | int android_atomic_release_cas(int32_t oldvalue, int32_t newvalue, | 
|  | volatile int32_t* addr) | 
|  | { | 
|  | volatile atomic_int_least32_t* a = to_atomic_int_least32_t(addr); | 
|  | return !atomic_compare_exchange_strong_explicit( | 
|  | a, &oldvalue, newvalue, | 
|  | memory_order_release, | 
|  | memory_order_relaxed); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Fence primitives. | 
|  | */ | 
|  | ANDROID_ATOMIC_INLINE | 
|  | void android_compiler_barrier(void) | 
|  | { | 
|  | __asm__ __volatile__ ("" : : : "memory"); | 
|  | /* Could probably also be:                          */ | 
|  | /* atomic_signal_fence(memory_order_seq_cst);       */ | 
|  | } | 
|  |  | 
|  | ANDROID_ATOMIC_INLINE | 
|  | void android_memory_barrier(void) | 
|  | { | 
|  | atomic_thread_fence(memory_order_seq_cst); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Aliases for code using an older version of this header.  These are now | 
|  | * deprecated and should not be used.  The definitions will be removed | 
|  | * in a future release. | 
|  | */ | 
|  | #define android_atomic_write android_atomic_release_store | 
|  | #define android_atomic_cmpxchg android_atomic_release_cas | 
|  |  | 
|  | #endif // ANDROID_CUTILS_ATOMIC_H |