| /* |
| * Copyright (C) 2007, 2008, 2010, 2012 Apple Inc. All rights reserved. |
| * Copyright (C) 2007 Justin Haygood (jhaygood@reaktix.com) |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of |
| * its contributors may be used to endorse or promote products derived |
| * from this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY |
| * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
| * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
| * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
| * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #ifndef SKY_ENGINE_WTF_ATOMICS_H_ |
| #define SKY_ENGINE_WTF_ATOMICS_H_ |
| |
| #include "sky/engine/wtf/Assertions.h" |
| #include "sky/engine/wtf/CPU.h" |
| |
| #include <stdint.h> |
| |
| #if defined(THREAD_SANITIZER) |
| #include <sanitizer/tsan_interface_atomic.h> |
| #endif |
| |
| namespace WTF { |
| |
| // atomicAdd returns the result of the addition. |
| ALWAYS_INLINE int atomicAdd(int volatile* addend, int increment) { return __sync_add_and_fetch(addend, increment); } |
| // atomicSubtract returns the result of the subtraction. |
| ALWAYS_INLINE int atomicSubtract(int volatile* addend, int decrement) { return __sync_sub_and_fetch(addend, decrement); } |
| |
| ALWAYS_INLINE int atomicIncrement(int volatile* addend) { return atomicAdd(addend, 1); } |
| ALWAYS_INLINE int atomicDecrement(int volatile* addend) { return atomicSubtract(addend, 1); } |
| |
| ALWAYS_INLINE int64_t atomicIncrement(int64_t volatile* addend) { return __sync_add_and_fetch(addend, 1); } |
| ALWAYS_INLINE int64_t atomicDecrement(int64_t volatile* addend) { return __sync_sub_and_fetch(addend, 1); } |
| |
| ALWAYS_INLINE int atomicTestAndSetToOne(int volatile* ptr) |
| { |
| int ret = __sync_lock_test_and_set(ptr, 1); |
| ASSERT(!ret || ret == 1); |
| return ret; |
| } |
| |
| ALWAYS_INLINE void atomicSetOneToZero(int volatile* ptr) |
| { |
| ASSERT(*ptr == 1); |
| __sync_lock_release(ptr); |
| } |
| |
| #if defined(THREAD_SANITIZER) |
| ALWAYS_INLINE void releaseStore(volatile int* ptr, int value) |
| { |
| __tsan_atomic32_store(ptr, value, __tsan_memory_order_release); |
| } |
| |
| ALWAYS_INLINE int acquireLoad(volatile const int* ptr) |
| { |
| return __tsan_atomic32_load(ptr, __tsan_memory_order_acquire); |
| } |
| |
| ALWAYS_INLINE void releaseStore(volatile unsigned* ptr, unsigned value) |
| { |
| __tsan_atomic32_store(reinterpret_cast<volatile int*>(ptr), static_cast<int>(value), __tsan_memory_order_release); |
| } |
| |
| ALWAYS_INLINE unsigned acquireLoad(volatile const unsigned* ptr) |
| { |
| return static_cast<unsigned>(__tsan_atomic32_load(reinterpret_cast<volatile const int*>(ptr), __tsan_memory_order_acquire)); |
| } |
| #else |
| |
| #if CPU(X86) || CPU(X86_64) |
| // Only compiler barrier is needed. |
| #define MEMORY_BARRIER() __asm__ __volatile__("" : : : "memory") |
| #elif CPU(ARM) && (OS(LINUX) || OS(ANDROID)) |
| // On ARM __sync_synchronize generates dmb which is very expensive on single |
| // core devices which don't actually need it. Avoid the cost by calling into |
| // kuser_memory_barrier helper. |
| inline void memoryBarrier() |
| { |
| // Note: This is a function call, which is also an implicit compiler barrier. |
| typedef void (*KernelMemoryBarrierFunc)(); |
| ((KernelMemoryBarrierFunc)0xffff0fa0)(); |
| } |
| #define MEMORY_BARRIER() memoryBarrier() |
| #else |
| // Fallback to the compiler intrinsic on all other platforms. |
| #define MEMORY_BARRIER() __sync_synchronize() |
| #endif |
| |
| ALWAYS_INLINE void releaseStore(volatile int* ptr, int value) |
| { |
| MEMORY_BARRIER(); |
| *ptr = value; |
| } |
| |
| ALWAYS_INLINE int acquireLoad(volatile const int* ptr) |
| { |
| int value = *ptr; |
| MEMORY_BARRIER(); |
| return value; |
| } |
| |
| ALWAYS_INLINE void releaseStore(volatile unsigned* ptr, unsigned value) |
| { |
| MEMORY_BARRIER(); |
| *ptr = value; |
| } |
| |
| ALWAYS_INLINE unsigned acquireLoad(volatile const unsigned* ptr) |
| { |
| unsigned value = *ptr; |
| MEMORY_BARRIER(); |
| return value; |
| } |
| |
| #undef MEMORY_BARRIER |
| |
| #endif |
| |
| } // namespace WTF |
| |
| using WTF::atomicAdd; |
| using WTF::atomicSubtract; |
| using WTF::atomicDecrement; |
| using WTF::atomicIncrement; |
| using WTF::atomicTestAndSetToOne; |
| using WTF::atomicSetOneToZero; |
| using WTF::acquireLoad; |
| using WTF::releaseStore; |
| |
| #endif // SKY_ENGINE_WTF_ATOMICS_H_ |