OLD | NEW |
1 // Copyright 2010 the V8 project authors. All rights reserved. | 1 // Copyright 2010 the V8 project authors. All rights reserved. |
2 // Redistribution and use in source and binary forms, with or without | 2 // Redistribution and use in source and binary forms, with or without |
3 // modification, are permitted provided that the following conditions are | 3 // modification, are permitted provided that the following conditions are |
4 // met: | 4 // met: |
5 // | 5 // |
6 // * Redistributions of source code must retain the above copyright | 6 // * Redistributions of source code must retain the above copyright |
7 // notice, this list of conditions and the following disclaimer. | 7 // notice, this list of conditions and the following disclaimer. |
8 // * Redistributions in binary form must reproduce the above | 8 // * Redistributions in binary form must reproduce the above |
9 // copyright notice, this list of conditions and the following | 9 // copyright notice, this list of conditions and the following |
10 // disclaimer in the documentation and/or other materials provided | 10 // disclaimer in the documentation and/or other materials provided |
(...skipping 133 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
144 void Acquire_Store(volatile Atomic64* ptr, Atomic64 value); | 144 void Acquire_Store(volatile Atomic64* ptr, Atomic64 value); |
145 void Release_Store(volatile Atomic64* ptr, Atomic64 value); | 145 void Release_Store(volatile Atomic64* ptr, Atomic64 value); |
146 Atomic64 NoBarrier_Load(volatile const Atomic64* ptr); | 146 Atomic64 NoBarrier_Load(volatile const Atomic64* ptr); |
147 Atomic64 Acquire_Load(volatile const Atomic64* ptr); | 147 Atomic64 Acquire_Load(volatile const Atomic64* ptr); |
148 Atomic64 Release_Load(volatile const Atomic64* ptr); | 148 Atomic64 Release_Load(volatile const Atomic64* ptr); |
149 #endif // V8_HOST_ARCH_64_BIT | 149 #endif // V8_HOST_ARCH_64_BIT |
150 | 150 |
151 } } // namespace v8::internal | 151 } } // namespace v8::internal |
152 | 152 |
153 // Include our platform specific implementation. | 153 // Include our platform specific implementation. |
154 #if defined(_MSC_VER) && \ | 154 #if defined(THREAD_SANITIZER) |
| 155 #include "atomicops_internals_tsan.h" |
| 156 #elif defined(_MSC_VER) && \ |
155 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) | 157 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) |
156 #include "atomicops_internals_x86_msvc.h" | 158 #include "atomicops_internals_x86_msvc.h" |
157 #elif defined(__APPLE__) && \ | 159 #elif defined(__APPLE__) && \ |
158 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) | 160 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) |
159 #include "atomicops_internals_x86_macosx.h" | 161 #include "atomicops_internals_x86_macosx.h" |
160 #elif defined(__GNUC__) && \ | 162 #elif defined(__GNUC__) && \ |
161 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) | 163 (defined(V8_HOST_ARCH_IA32) || defined(V8_HOST_ARCH_X64)) |
162 #include "atomicops_internals_x86_gcc.h" | 164 #include "atomicops_internals_x86_gcc.h" |
163 #elif defined(__GNUC__) && defined(V8_HOST_ARCH_ARM) | 165 #elif defined(__GNUC__) && defined(V8_HOST_ARCH_ARM) |
164 #include "atomicops_internals_arm_gcc.h" | 166 #include "atomicops_internals_arm_gcc.h" |
165 #elif defined(__GNUC__) && defined(V8_HOST_ARCH_MIPS) | 167 #elif defined(__GNUC__) && defined(V8_HOST_ARCH_MIPS) |
166 #include "atomicops_internals_mips_gcc.h" | 168 #include "atomicops_internals_mips_gcc.h" |
167 #else | 169 #else |
168 #error "Atomic operations are not supported on your platform" | 170 #error "Atomic operations are not supported on your platform" |
169 #endif | 171 #endif |
170 | 172 |
171 #endif // V8_ATOMICOPS_H_ | 173 #endif // V8_ATOMICOPS_H_ |
OLD | NEW |