Intel(R) Threading Building Blocks Doxygen Documentation  version 4.2.3
gcc_generic.h
Go to the documentation of this file.
1 /*
2  Copyright (c) 2005-2019 Intel Corporation
3 
4  Licensed under the Apache License, Version 2.0 (the "License");
5  you may not use this file except in compliance with the License.
6  You may obtain a copy of the License at
7 
8  http://www.apache.org/licenses/LICENSE-2.0
9 
10  Unless required by applicable law or agreed to in writing, software
11  distributed under the License is distributed on an "AS IS" BASIS,
12  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  See the License for the specific language governing permissions and
14  limitations under the License.
15 
16 
17 
18 
19 */
20 
21 #if !defined(__TBB_machine_H) || defined(__TBB_machine_gcc_generic_H)
22 #error Do not #include this internal file directly; use public TBB headers instead.
23 #endif
24 
25 #define __TBB_machine_gcc_generic_H
26 
27 #include <stdint.h>
28 #include <unistd.h>
29 
30 #define __TBB_WORDSIZE __SIZEOF_POINTER__
31 
32 #if __TBB_GCC_64BIT_ATOMIC_BUILTINS_BROKEN
33  #define __TBB_64BIT_ATOMICS 0
34 #endif
35 
37 #if __ANDROID__ && __TBB_generic_arch
38  #define __TBB_CPU_CTL_ENV_PRESENT 0
39 #endif
40 
41 // __BYTE_ORDER__ is used in accordance with http://gcc.gnu.org/onlinedocs/cpp/Common-Predefined-Macros.html,
42 // but __BIG_ENDIAN__ or __LITTLE_ENDIAN__ may be more commonly found instead.
43 #if __BIG_ENDIAN__ || (defined(__BYTE_ORDER__) && __BYTE_ORDER__==__ORDER_BIG_ENDIAN__)
44  #define __TBB_ENDIANNESS __TBB_ENDIAN_BIG
45 #elif __LITTLE_ENDIAN__ || (defined(__BYTE_ORDER__) && __BYTE_ORDER__==__ORDER_LITTLE_ENDIAN__)
46  #define __TBB_ENDIANNESS __TBB_ENDIAN_LITTLE
47 #elif defined(__BYTE_ORDER__)
48  #define __TBB_ENDIANNESS __TBB_ENDIAN_UNSUPPORTED
49 #else
50  #define __TBB_ENDIANNESS __TBB_ENDIAN_DETECT
51 #endif
52 
53 #if __TBB_GCC_VERSION < 40700
54 // Use __sync_* builtins
55 
61 #define __TBB_acquire_consistency_helper() __sync_synchronize()
62 #define __TBB_release_consistency_helper() __sync_synchronize()
63 #define __TBB_full_memory_fence() __sync_synchronize()
64 #define __TBB_control_consistency_helper() __sync_synchronize()
65 
66 #define __TBB_MACHINE_DEFINE_ATOMICS(S,T) \
67 inline T __TBB_machine_cmpswp##S( volatile void *ptr, T value, T comparand ) { \
68  return __sync_val_compare_and_swap(reinterpret_cast<volatile T *>(ptr),comparand,value); \
69 } \
70 inline T __TBB_machine_fetchadd##S( volatile void *ptr, T value ) { \
71  return __sync_fetch_and_add(reinterpret_cast<volatile T *>(ptr),value); \
72 }
73 
74 #define __TBB_USE_GENERIC_FETCH_STORE 1
75 
76 #else
77 // __TBB_GCC_VERSION >= 40700; use __atomic_* builtins available since gcc 4.7
78 
79 #define __TBB_compiler_fence() __asm__ __volatile__("": : :"memory")
80 // Acquire and release fence intrinsics in GCC might miss compiler fence.
81 // Adding it at both sides of an intrinsic, as we do not know what reordering can be made.
82 #define __TBB_acquire_consistency_helper() __TBB_compiler_fence(); __atomic_thread_fence(__ATOMIC_ACQUIRE); __TBB_compiler_fence()
83 #define __TBB_release_consistency_helper() __TBB_compiler_fence(); __atomic_thread_fence(__ATOMIC_RELEASE); __TBB_compiler_fence()
84 #define __TBB_full_memory_fence() __atomic_thread_fence(__ATOMIC_SEQ_CST)
85 #define __TBB_control_consistency_helper() __TBB_acquire_consistency_helper()
86 
87 #define __TBB_MACHINE_DEFINE_ATOMICS(S,T) \
88 inline T __TBB_machine_cmpswp##S( volatile void *ptr, T value, T comparand ) { \
89  (void)__atomic_compare_exchange_n(reinterpret_cast<volatile T *>(ptr), &comparand, value, \
90  false, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST); \
91  return comparand; \
92 } \
93 inline T __TBB_machine_fetchadd##S( volatile void *ptr, T value ) { \
94  return __atomic_fetch_add(reinterpret_cast<volatile T *>(ptr), value, __ATOMIC_SEQ_CST); \
95 } \
96 inline T __TBB_machine_fetchstore##S( volatile void *ptr, T value ) { \
97  return __atomic_exchange_n(reinterpret_cast<volatile T *>(ptr), value, __ATOMIC_SEQ_CST); \
98 }
99 
100 #endif // __TBB_GCC_VERSION < 40700
101 
106 
107 #undef __TBB_MACHINE_DEFINE_ATOMICS
108 
109 typedef unsigned char __TBB_Flag;
111 
112 #if __TBB_GCC_VERSION < 40700
113 // Use __sync_* builtins
114 
115 // Use generic machine_load_store functions if there are no builtin atomics
116 #define __TBB_USE_GENERIC_HALF_FENCED_LOAD_STORE 1
117 #define __TBB_USE_GENERIC_RELAXED_LOAD_STORE 1
118 #define __TBB_USE_GENERIC_SEQUENTIAL_CONSISTENCY_LOAD_STORE 1
119 
120 static inline void __TBB_machine_or( volatile void *ptr, uintptr_t addend ) {
121  __sync_fetch_and_or(reinterpret_cast<volatile uintptr_t *>(ptr),addend);
122 }
123 
124 static inline void __TBB_machine_and( volatile void *ptr, uintptr_t addend ) {
125  __sync_fetch_and_and(reinterpret_cast<volatile uintptr_t *>(ptr),addend);
126 }
127 
129  return __sync_lock_test_and_set(&flag,1)==0;
130 }
131 
133  __sync_lock_release(&flag);
134 }
135 
136 #else
137 // __TBB_GCC_VERSION >= 40700; use __atomic_* builtins available since gcc 4.7
138 
139 static inline void __TBB_machine_or( volatile void *ptr, uintptr_t addend ) {
140  __atomic_fetch_or(reinterpret_cast<volatile uintptr_t *>(ptr),addend,__ATOMIC_SEQ_CST);
141 }
142 
143 static inline void __TBB_machine_and( volatile void *ptr, uintptr_t addend ) {
144  __atomic_fetch_and(reinterpret_cast<volatile uintptr_t *>(ptr),addend,__ATOMIC_SEQ_CST);
145 }
146 
147 inline bool __TBB_machine_try_lock_byte( __TBB_atomic_flag &flag ) {
148  return !__atomic_test_and_set(&flag,__ATOMIC_ACQUIRE);
149 }
150 
151 inline void __TBB_machine_unlock_byte( __TBB_atomic_flag &flag ) {
152  __atomic_clear(&flag,__ATOMIC_RELEASE);
153 }
154 
155 namespace tbb { namespace internal {
156 
161 template <typename T, int MemOrder>
162 inline T __TBB_machine_atomic_load( const volatile T& location) {
163  if (MemOrder == __ATOMIC_SEQ_CST) __TBB_compiler_fence();
164  T value = __atomic_load_n(&location, MemOrder);
165  if (MemOrder != __ATOMIC_RELAXED) __TBB_compiler_fence();
166  return value;
167 }
168 
169 template <typename T, int MemOrder>
170 inline void __TBB_machine_atomic_store( volatile T& location, T value) {
171  if (MemOrder != __ATOMIC_RELAXED) __TBB_compiler_fence();
172  __atomic_store_n(&location, value, MemOrder);
173  if (MemOrder == __ATOMIC_SEQ_CST) __TBB_compiler_fence();
174 }
175 
176 template <typename T, size_t S>
177 struct machine_load_store {
178  static T load_with_acquire ( const volatile T& location ) {
179  return __TBB_machine_atomic_load<T, __ATOMIC_ACQUIRE>(location);
180  }
181  static void store_with_release ( volatile T &location, T value ) {
182  __TBB_machine_atomic_store<T, __ATOMIC_RELEASE>(location, value);
183  }
184 };
185 
186 template <typename T, size_t S>
187 struct machine_load_store_relaxed {
188  static inline T load ( const volatile T& location ) {
189  return __TBB_machine_atomic_load<T, __ATOMIC_RELAXED>(location);
190  }
191  static inline void store ( volatile T& location, T value ) {
192  __TBB_machine_atomic_store<T, __ATOMIC_RELAXED>(location, value);
193  }
194 };
195 
196 template <typename T, size_t S>
197 struct machine_load_store_seq_cst {
198  static T load ( const volatile T& location ) {
199  return __TBB_machine_atomic_load<T, __ATOMIC_SEQ_CST>(location);
200  }
201  static void store ( volatile T &location, T value ) {
202  __TBB_machine_atomic_store<T, __ATOMIC_SEQ_CST>(location, value);
203  }
204 };
205 
206 }} // namespace tbb::internal
207 
208 #endif // __TBB_GCC_VERSION < 40700
209 
210 // Machine specific atomic operations
211 #define __TBB_AtomicOR(P,V) __TBB_machine_or(P,V)
212 #define __TBB_AtomicAND(P,V) __TBB_machine_and(P,V)
213 
214 #define __TBB_TryLockByte __TBB_machine_try_lock_byte
215 #define __TBB_UnlockByte __TBB_machine_unlock_byte
216 
217 // __builtin_clz counts the number of leading zeroes
218 namespace tbb{ namespace internal { namespace gcc_builtins {
219  inline int clz(unsigned int x){ return __builtin_clz(x); };
220  inline int clz(unsigned long int x){ return __builtin_clzl(x); };
221  inline int clz(unsigned long long int x){ return __builtin_clzll(x); };
222 }}}
223 // logarithm is the index of the most significant non-zero bit
224 static inline intptr_t __TBB_machine_lg( uintptr_t x ) {
225  // If P is a power of 2 and x<P, then (P-1)-x == (P-1) XOR x
226  return (sizeof(x)*8 - 1) ^ tbb::internal::gcc_builtins::clz(x);
227 }
228 
229 #define __TBB_Log2(V) __TBB_machine_lg(V)
230 
231 #if __TBB_WORDSIZE==4
232  #define __TBB_USE_GENERIC_DWORD_LOAD_STORE 1
233 #endif
234 
235 #if __TBB_x86_32 || __TBB_x86_64
236 #include "gcc_ia32_common.h"
237 #endif
static void store_with_release(volatile T &location, T value)
Definition: icc_generic.h:102
static void __TBB_machine_and(volatile void *ptr, uintptr_t addend)
Definition: gcc_generic.h:124
void __TBB_machine_unlock_byte(__TBB_atomic_flag &flag)
Definition: gcc_generic.h:132
bool __TBB_machine_try_lock_byte(__TBB_atomic_flag &flag)
Definition: gcc_generic.h:128
__TBB_atomic __TBB_Flag __TBB_atomic_flag
Definition: gcc_generic.h:110
static void store(volatile T &location, T value)
Definition: icc_generic.h:123
unsigned char __TBB_Flag
Definition: gcc_generic.h:109
static void __TBB_machine_or(volatile void *ptr, uintptr_t addend)
Definition: gcc_generic.h:120
static T load_with_acquire(const volatile T &location)
Definition: icc_generic.h:99
static intptr_t __TBB_machine_lg(uintptr_t x)
Definition: gcc_generic.h:224
static T load(const volatile T &location)
Definition: icc_generic.h:119
#define __TBB_compiler_fence()
Definition: icc_generic.h:55
The graph class.
void const char const char int ITT_FORMAT __itt_group_sync x void const char ITT_FORMAT __itt_group_sync s void ITT_FORMAT __itt_group_sync p void ITT_FORMAT p void ITT_FORMAT p no args __itt_suppress_mode_t unsigned int void size_t ITT_FORMAT d void ITT_FORMAT p void ITT_FORMAT p __itt_model_site __itt_model_site_instance ITT_FORMAT p __itt_model_task __itt_model_task_instance ITT_FORMAT p void ITT_FORMAT p void ITT_FORMAT p void size_t ITT_FORMAT d void ITT_FORMAT p const wchar_t ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s const char ITT_FORMAT s no args void ITT_FORMAT p size_t ITT_FORMAT d no args const wchar_t const wchar_t ITT_FORMAT s __itt_heap_function void size_t int ITT_FORMAT d __itt_heap_function void ITT_FORMAT p __itt_heap_function void void size_t int ITT_FORMAT d no args no args unsigned int ITT_FORMAT u const __itt_domain __itt_id ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain __itt_id ITT_FORMAT p const __itt_domain __itt_id __itt_timestamp __itt_timestamp ITT_FORMAT lu const __itt_domain __itt_id __itt_id __itt_string_handle ITT_FORMAT p const __itt_domain ITT_FORMAT p const __itt_domain __itt_string_handle unsigned long long value
int clz(unsigned int x)
Definition: gcc_generic.h:219
#define __TBB_atomic
Definition: tbb_stddef.h:241
static void store(T &location, T value)
Definition: icc_generic.h:112
#define __TBB_MACHINE_DEFINE_ATOMICS(S, T)
Definition: gcc_generic.h:66
static T load(const T &location)
Definition: icc_generic.h:109

Copyright © 2005-2019 Intel Corporation. All Rights Reserved.

Intel, Pentium, Intel Xeon, Itanium, Intel XScale and VTune are registered trademarks or trademarks of Intel Corporation or its subsidiaries in the United States and other countries.

* Other names and brands may be claimed as the property of others.