aboutsummaryrefslogtreecommitdiff
path: root/include/EASTL/internal/atomic/atomic_pointer.h
diff options
context:
space:
mode:
Diffstat (limited to 'include/EASTL/internal/atomic/atomic_pointer.h')
-rw-r--r--include/EASTL/internal/atomic/atomic_pointer.h281
1 files changed, 0 insertions, 281 deletions
diff --git a/include/EASTL/internal/atomic/atomic_pointer.h b/include/EASTL/internal/atomic/atomic_pointer.h
deleted file mode 100644
index c0b19e6..0000000
--- a/include/EASTL/internal/atomic/atomic_pointer.h
+++ /dev/null
@@ -1,281 +0,0 @@
-/////////////////////////////////////////////////////////////////////////////////
-// Copyright (c) Electronic Arts Inc. All rights reserved.
-/////////////////////////////////////////////////////////////////////////////////
-
-
-#ifndef EASTL_ATOMIC_INTERNAL_POINTER_H
-#define EASTL_ATOMIC_INTERNAL_POINTER_H
-
-#if defined(EA_PRAGMA_ONCE_SUPPORTED)
- #pragma once
-#endif
-
-
-#include "atomic_push_compiler_options.h"
-
-
-namespace eastl
-{
-
-
-namespace internal
-{
-
-
- template <typename T, unsigned width = sizeof(T)>
- struct atomic_pointer_base;
-
-#define EASTL_ATOMIC_POINTER_STATIC_ASSERT_FUNCS_IMPL(funcName) \
- template <typename Order> \
- T* funcName(ptrdiff_t /*arg*/, Order /*order*/) EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_INVALID_MEMORY_ORDER(T); \
- } \
- \
- template <typename Order> \
- T* funcName(ptrdiff_t /*arg*/, Order /*order*/) volatile EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_VOLATILE_MEM_FN(T); \
- } \
- \
- T* funcName(ptrdiff_t /*arg*/) volatile EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_VOLATILE_MEM_FN(T); \
- }
-
-#define EASTL_ATOMIC_POINTER_STATIC_ASSERT_INC_DEC_OPERATOR_IMPL(operatorOp) \
- T* operator operatorOp() volatile EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_VOLATILE_MEM_FN(T); \
- } \
- \
- T* operator operatorOp(int) volatile EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_VOLATILE_MEM_FN(T); \
- }
-
-#define EASTL_ATOMIC_POINTER_STATIC_ASSERT_ASSIGNMENT_OPERATOR_IMPL(operatorOp) \
- T* operator operatorOp(ptrdiff_t /*arg*/) volatile EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_VOLATILE_MEM_FN(T); \
- }
-
-
- template <typename T, unsigned width>
- struct atomic_pointer_base<T*, width> : public atomic_base_width<T*, width>
- {
- private:
-
- using Base = atomic_base_width<T*, width>;
-
- public: /* ctors */
-
- EA_CONSTEXPR atomic_pointer_base(T* desired) EA_NOEXCEPT
- : Base{ desired }
- {
- }
-
- EA_CONSTEXPR atomic_pointer_base() EA_NOEXCEPT = default;
-
- atomic_pointer_base(const atomic_pointer_base&) EA_NOEXCEPT = delete;
-
- public: /* assignment operators */
-
- using Base::operator=;
-
- atomic_pointer_base& operator=(const atomic_pointer_base&) EA_NOEXCEPT = delete;
- atomic_pointer_base& operator=(const atomic_pointer_base&) volatile EA_NOEXCEPT = delete;
-
- public: /* fetch_add */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_FUNCS_IMPL(fetch_add)
-
- public: /* add_fetch */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_FUNCS_IMPL(add_fetch)
-
- public: /* fetch_sub */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_FUNCS_IMPL(fetch_sub)
-
- public: /* sub_fetch */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_FUNCS_IMPL(sub_fetch)
-
- public: /* operator++ && operator-- */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_INC_DEC_OPERATOR_IMPL(++)
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_INC_DEC_OPERATOR_IMPL(--)
-
- public: /* operator+= && operator-= */
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_ASSIGNMENT_OPERATOR_IMPL(+=)
-
- EASTL_ATOMIC_POINTER_STATIC_ASSERT_ASSIGNMENT_OPERATOR_IMPL(-=)
-
- };
-
-
- template <typename T, unsigned width = sizeof(T)>
- struct atomic_pointer_width;
-
-#define EASTL_ATOMIC_POINTER_FUNC_IMPL(op, bits) \
- T* retVal; \
- { \
- ptr_integral_type retType; \
- ptr_integral_type addend = static_cast<ptr_integral_type>(arg) * static_cast<ptr_integral_type>(sizeof(T)); \
- \
- EA_PREPROCESSOR_JOIN(op, bits)(ptr_integral_type, retType, EASTL_ATOMIC_INTEGRAL_CAST(ptr_integral_type, this->GetAtomicAddress()), addend); \
- \
- retVal = reinterpret_cast<T*>(retType); \
- } \
- return retVal;
-
-#define EASTL_ATOMIC_POINTER_FETCH_IMPL(funcName, op, bits) \
- T* funcName(ptrdiff_t arg) EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_TYPE_IS_OBJECT(T); \
- EASTL_ATOMIC_POINTER_FUNC_IMPL(op, bits); \
- }
-
-#define EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, orderType, op, bits) \
- T* funcName(ptrdiff_t arg, orderType) EA_NOEXCEPT \
- { \
- EASTL_ATOMIC_STATIC_ASSERT_TYPE_IS_OBJECT(T); \
- EASTL_ATOMIC_POINTER_FUNC_IMPL(op, bits); \
- }
-
-#define EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, Order) \
- EA_PREPROCESSOR_JOIN(EA_PREPROCESSOR_JOIN(EASTL_ATOMIC_, fetchOp), Order)
-
-#define EASTL_ATOMIC_POINTER_FETCH_FUNCS_IMPL(funcName, fetchOp, bits) \
- using Base::funcName; \
- \
- EASTL_ATOMIC_POINTER_FETCH_IMPL(funcName, EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _SEQ_CST_), bits) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, eastl::internal::memory_order_relaxed_s, \
- EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _RELAXED_), bits) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, eastl::internal::memory_order_acquire_s, \
- EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _ACQUIRE_), bits) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, eastl::internal::memory_order_release_s, \
- EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _RELEASE_), bits) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, eastl::internal::memory_order_acq_rel_s, \
- EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _ACQ_REL_), bits) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ORDER_IMPL(funcName, eastl::internal::memory_order_seq_cst_s, \
- EASTL_ATOMIC_POINTER_FETCH_OP_JOIN(fetchOp, _SEQ_CST_), bits)
-
-#define EASTL_ATOMIC_POINTER_FETCH_INC_DEC_OPERATOR_IMPL(operatorOp, preFuncName, postFuncName) \
- using Base::operator operatorOp; \
- \
- T* operator operatorOp() EA_NOEXCEPT \
- { \
- return preFuncName(1, eastl::memory_order_seq_cst); \
- } \
- \
- T* operator operatorOp(int) EA_NOEXCEPT \
- { \
- return postFuncName(1, eastl::memory_order_seq_cst); \
- }
-
-#define EASTL_ATOMIC_POINTER_FETCH_ASSIGNMENT_OPERATOR_IMPL(operatorOp, funcName) \
- using Base::operator operatorOp; \
- \
- T* operator operatorOp(ptrdiff_t arg) EA_NOEXCEPT \
- { \
- return funcName(arg, eastl::memory_order_seq_cst); \
- }
-
-
-#define EASTL_ATOMIC_POINTER_WIDTH_SPECIALIZE(bytes, bits) \
- template <typename T> \
- struct atomic_pointer_width<T*, bytes> : public atomic_pointer_base<T*, bytes> \
- { \
- private: \
- \
- using Base = atomic_pointer_base<T*, bytes>; \
- using u_ptr_integral_type = EA_PREPROCESSOR_JOIN(EA_PREPROCESSOR_JOIN(uint, bits), _t); \
- using ptr_integral_type = EA_PREPROCESSOR_JOIN(EA_PREPROCESSOR_JOIN(int, bits), _t); \
- \
- public: /* ctors */ \
- \
- EA_CONSTEXPR atomic_pointer_width(T* desired) EA_NOEXCEPT \
- : Base{ desired } \
- { \
- } \
- \
- EA_CONSTEXPR atomic_pointer_width() EA_NOEXCEPT = default; \
- \
- atomic_pointer_width(const atomic_pointer_width&) EA_NOEXCEPT = delete; \
- \
- public: /* assignment operators */ \
- \
- using Base::operator=; \
- \
- atomic_pointer_width& operator=(const atomic_pointer_width&) EA_NOEXCEPT = delete; \
- atomic_pointer_width& operator=(const atomic_pointer_width&) volatile EA_NOEXCEPT = delete; \
- \
- public: /* fetch_add */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_FUNCS_IMPL(fetch_add, FETCH_ADD, bits) \
- \
- public: /* add_fetch */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_FUNCS_IMPL(add_fetch, ADD_FETCH, bits) \
- \
- public: /* fetch_sub */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_FUNCS_IMPL(fetch_sub, FETCH_SUB, bits) \
- \
- public: /* sub_fetch */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_FUNCS_IMPL(sub_fetch, SUB_FETCH, bits) \
- \
- public: /* operator++ && operator-- */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_INC_DEC_OPERATOR_IMPL(++, add_fetch, fetch_add) \
- \
- EASTL_ATOMIC_POINTER_FETCH_INC_DEC_OPERATOR_IMPL(--, sub_fetch, fetch_sub) \
- \
- public: /* operator+= && operator-= */ \
- \
- EASTL_ATOMIC_POINTER_FETCH_ASSIGNMENT_OPERATOR_IMPL(+=, add_fetch) \
- \
- EASTL_ATOMIC_POINTER_FETCH_ASSIGNMENT_OPERATOR_IMPL(-=, sub_fetch) \
- \
- public: \
- \
- using Base::load; \
- \
- T* load(eastl::internal::memory_order_read_depends_s) EA_NOEXCEPT \
- { \
- T* retPointer; \
- EA_PREPROCESSOR_JOIN(EASTL_ATOMIC_LOAD_READ_DEPENDS_, bits)(T*, retPointer, this->GetAtomicAddress()); \
- return retPointer; \
- } \
- };
-
-
-#if defined(EASTL_ATOMIC_HAS_32BIT) && EA_PLATFORM_PTR_SIZE == 4
- EASTL_ATOMIC_POINTER_WIDTH_SPECIALIZE(4, 32)
-#endif
-
-#if defined(EASTL_ATOMIC_HAS_64BIT) && EA_PLATFORM_PTR_SIZE == 8
- EASTL_ATOMIC_POINTER_WIDTH_SPECIALIZE(8, 64)
-#endif
-
-
-} // namespace internal
-
-
-} // namespace eastl
-
-
-#include "atomic_pop_compiler_options.h"
-
-
-#endif /* EASTL_ATOMIC_INTERNAL_POINTER_H */