| // Copyright (c) 2017 Facebook Inc. |
| // Copyright (c) 2015-2017 Georgia Institute of Technology |
| // All rights reserved. |
| // |
| // Copyright 2019 Google LLC |
| // |
| // This source code is licensed under the BSD-style license found in the |
| // LICENSE file in the root directory of this source tree. |
| |
| #ifndef __PTHREADPOOL_SRC_THREADPOOL_ATOMICS_H_ |
| #define __PTHREADPOOL_SRC_THREADPOOL_ATOMICS_H_ |
| |
| /* Standard C headers */ |
| #include <stdatomic.h> |
| #include <stdbool.h> |
| #include <stddef.h> |
| #include <stdint.h> |
| #include <stdio.h> |
| |
| /* Windows-specific headers */ |
| #ifdef _WIN32 |
| #ifndef WIN32_LEAN_AND_MEAN |
| #define WIN32_LEAN_AND_MEAN |
| #endif |
| #include <windows.h> |
| #else |
| #include <sched.h> |
| #endif |
| |
| /* SSE-specific headers */ |
| #if defined(__i386__) || defined(__i686__) || defined(__x86_64__) || \ |
| defined(_M_IX86) || defined(_M_X64) && !defined(_M_ARM64EC) |
| #include <xmmintrin.h> |
| #endif |
| |
| /* ARM-specific headers */ |
| #if defined(__ARM_ACLE) |
| #include <arm_acle.h> |
| #endif |
| |
| /* MSVC-specific headers */ |
| #ifdef _MSC_VER |
| #include <intrin.h> |
| #endif |
| |
| /* Configuration header */ |
| #include "threadpool-common.h" |
| |
| typedef atomic_uint_fast32_t pthreadpool_atomic_uint32_t; |
| typedef atomic_size_t pthreadpool_atomic_size_t; |
| typedef atomic_uintptr_t pthreadpool_atomic_void_p; |
| |
| static inline uint32_t pthreadpool_load_relaxed_uint32_t( |
| pthreadpool_atomic_uint32_t* address) { |
| return atomic_load_explicit(address, memory_order_relaxed); |
| } |
| |
| static inline size_t pthreadpool_load_relaxed_size_t( |
| pthreadpool_atomic_size_t* address) { |
| return atomic_load_explicit(address, memory_order_relaxed); |
| } |
| |
| static inline void* pthreadpool_load_relaxed_void_p( |
| pthreadpool_atomic_void_p* address) { |
| return (void*)atomic_load_explicit(address, memory_order_relaxed); |
| } |
| |
| static inline uint32_t pthreadpool_load_acquire_uint32_t( |
| pthreadpool_atomic_uint32_t* address) { |
| return atomic_load_explicit(address, memory_order_acquire); |
| } |
| |
| static inline size_t pthreadpool_load_acquire_size_t( |
| pthreadpool_atomic_size_t* address) { |
| return atomic_load_explicit(address, memory_order_acquire); |
| } |
| |
| static inline void pthreadpool_store_relaxed_uint32_t( |
| pthreadpool_atomic_uint32_t* address, uint32_t value) { |
| atomic_store_explicit(address, value, memory_order_relaxed); |
| } |
| |
| static inline void pthreadpool_store_relaxed_size_t( |
| pthreadpool_atomic_size_t* address, size_t value) { |
| atomic_store_explicit(address, value, memory_order_relaxed); |
| } |
| |
| static inline void pthreadpool_store_relaxed_void_p( |
| pthreadpool_atomic_void_p* address, void* value) { |
| atomic_store_explicit(address, (uintptr_t)value, memory_order_relaxed); |
| } |
| |
| static inline void pthreadpool_store_release_uint32_t( |
| pthreadpool_atomic_uint32_t* address, uint32_t value) { |
| atomic_store_explicit(address, value, memory_order_release); |
| } |
| |
| static inline void pthreadpool_store_release_size_t( |
| pthreadpool_atomic_size_t* address, size_t value) { |
| atomic_store_explicit(address, value, memory_order_release); |
| } |
| |
| static inline size_t pthreadpool_decrement_fetch_relaxed_size_t( |
| pthreadpool_atomic_size_t* address) { |
| return atomic_fetch_sub_explicit(address, 1, memory_order_relaxed) - 1; |
| } |
| |
| static inline size_t pthreadpool_decrement_fetch_release_size_t( |
| pthreadpool_atomic_size_t* address) { |
| return atomic_fetch_sub_explicit(address, 1, memory_order_release) - 1; |
| } |
| |
| static inline size_t pthreadpool_decrement_fetch_acquire_release_size_t( |
| pthreadpool_atomic_size_t* address) { |
| return atomic_fetch_sub_explicit(address, 1, memory_order_acq_rel) - 1; |
| } |
| |
| static inline bool pthreadpool_try_decrement_relaxed_size_t( |
| pthreadpool_atomic_size_t* value) { |
| size_t actual_value = atomic_load_explicit(value, memory_order_acquire); |
| while (actual_value != 0) { |
| if (atomic_compare_exchange_weak_explicit( |
| value, &actual_value, actual_value - 1, memory_order_relaxed, |
| memory_order_relaxed)) { |
| return true; |
| } |
| } |
| return false; |
| } |
| |
| static inline size_t pthreadpool_fetch_add_relaxed_size_t( |
| pthreadpool_atomic_size_t* address, size_t value) { |
| return atomic_fetch_add_explicit(address, value, memory_order_relaxed); |
| } |
| |
| static inline void pthreadpool_fence_acquire() { |
| atomic_thread_fence(memory_order_acquire); |
| } |
| |
| static inline void pthreadpool_fence_release() { |
| atomic_thread_fence(memory_order_release); |
| } |
| |
| static inline void pthreadpool_yield(uint32_t step) { |
| if (step < PTHREADPOOL_SPIN_PAUSE_ITERATIONS) { |
| #if defined(__ARM_ACLE) || \ |
| defined(_MSC_VER) && \ |
| (defined(_M_ARM) || defined(_M_ARM64) || defined(_M_ARM64EC)) |
| __yield(); |
| #elif defined(__GNUC__) && \ |
| (defined(__ARM_ARCH) && (__ARM_ARCH >= 7) || \ |
| (defined(__ARM_ARCH_6K__) || defined(__ARM_ARCH_6KZ__)) && \ |
| !defined(__thumb__)) |
| __asm__ __volatile__("yield"); |
| #elif defined(__i386__) || defined(__i686__) || defined(__x86_64__) || \ |
| defined(_M_IX86) || defined(_M_X64) |
| _mm_pause(); |
| #else |
| pthreadpool_fence_acquire(); |
| #endif |
| } else { |
| #ifdef _WIN32 |
| Sleep(0); |
| #else |
| sched_yield(); |
| #endif |
| } |
| } |
| |
| #endif // __PTHREADPOOL_SRC_THREADPOOL_ATOMICS_H_ |