/* * Copyright 2011 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "SkThread.h" #include <pthread.h> #include <errno.h> #ifndef SK_BUILD_FOR_ANDROID /** We prefer the GCC intrinsic implementation of the atomic operations over the SkMutex-based implementation. The SkMutex version suffers from static destructor ordering problems. Note clang also defines the GCC version macros and implements the intrinsics. TODO: Verify that gcc-style __sync_* intrinsics work on ARM According to this the intrinsics are supported on ARM in LLVM 2.7+ http://llvm.org/releases/2.7/docs/ReleaseNotes.html */ #if (__GNUC__ == 4 && __GNUC_MINOR__ >= 1) || __GNUC__ > 4 #if (defined(__x86_64) || defined(__i386__)) #define GCC_INTRINSIC #endif #endif #if defined(GCC_INTRINSIC) int32_t sk_atomic_inc(int32_t* addr) { return __sync_fetch_and_add(addr, 1); } int32_t sk_atomic_add(int32_t* addr, int32_t inc) { return __sync_fetch_and_add(addr, inc); } int32_t sk_atomic_dec(int32_t* addr) { return __sync_fetch_and_add(addr, -1); } void sk_membar_aquire__after_atomic_dec() { } int32_t sk_atomic_conditional_inc(int32_t* addr) { int32_t value = *addr; while (true) { if (value == 0) { return 0; } int32_t before = __sync_val_compare_and_swap(addr, value, value + 1); if (before == value) { return value; } else { value = before; } } } void sk_membar_aquire__after_atomic_conditional_inc() { } #else SkMutex gAtomicMutex; int32_t sk_atomic_inc(int32_t* addr) { SkAutoMutexAcquire ac(gAtomicMutex); int32_t value = *addr; *addr = value + 1; return value; } int32_t sk_atomic_add(int32_t* addr, int32_t inc) { SkAutoMutexAcquire ac(gAtomicMutex); int32_t value = *addr; *addr = value + inc; return value; } int32_t sk_atomic_dec(int32_t* addr) { SkAutoMutexAcquire ac(gAtomicMutex); int32_t value = *addr; *addr = value - 1; return value; } void sk_membar_aquire__after_atomic_dec() { } int32_t sk_atomic_conditional_inc(int32_t* addr) { SkAutoMutexAcquire ac(gAtomicMutex); int32_t value = *addr; if (value != 0) ++*addr; return value; } void sk_membar_aquire__after_atomic_conditional_inc() { } #endif #endif // SK_BUILD_FOR_ANDROID ////////////////////////////////////////////////////////////////////////////// static void print_pthread_error(int status) { switch (status) { case 0: // success break; case EINVAL: SkDebugf("pthread error [%d] EINVAL\n", status); break; case EBUSY: SkDebugf("pthread error [%d] EBUSY\n", status); break; default: SkDebugf("pthread error [%d] unknown\n", status); break; } } #ifdef SK_USE_POSIX_THREADS SkMutex::SkMutex() { int status; status = pthread_mutex_init(&fMutex, NULL); if (status != 0) { print_pthread_error(status); SkASSERT(0 == status); } } SkMutex::~SkMutex() { int status = pthread_mutex_destroy(&fMutex); // only report errors on non-global mutexes if (status != 0) { print_pthread_error(status); SkASSERT(0 == status); } } #else // !SK_USE_POSIX_THREADS SkMutex::SkMutex() { if (sizeof(pthread_mutex_t) > sizeof(fStorage)) { SkDEBUGF(("pthread mutex size = %d\n", sizeof(pthread_mutex_t))); SkDEBUGFAIL("mutex storage is too small"); } int status; pthread_mutexattr_t attr; status = pthread_mutexattr_init(&attr); print_pthread_error(status); SkASSERT(0 == status); status = pthread_mutex_init((pthread_mutex_t*)fStorage, &attr); print_pthread_error(status); SkASSERT(0 == status); } SkMutex::~SkMutex() { int status = pthread_mutex_destroy((pthread_mutex_t*)fStorage); #if 0 // only report errors on non-global mutexes if (!fIsGlobal) { print_pthread_error(status); SkASSERT(0 == status); } #endif } void SkMutex::acquire() { int status = pthread_mutex_lock((pthread_mutex_t*)fStorage); print_pthread_error(status); SkASSERT(0 == status); } void SkMutex::release() { int status = pthread_mutex_unlock((pthread_mutex_t*)fStorage); print_pthread_error(status); SkASSERT(0 == status); } #endif // !SK_USE_POSIX_THREADS