OLD | NEW |
---|---|
1 // Copyright 2013 The Chromium Authors. All rights reserved. | 1 // Copyright 2013 The Chromium Authors. All rights reserved. |
2 // Use of this source code is governed by a BSD-style license that can be | 2 // Use of this source code is governed by a BSD-style license that can be |
3 // found in the LICENSE file. | 3 // found in the LICENSE file. |
4 | 4 |
5 // This file is an internal atomic implementation, use base/atomicops.h instead. | 5 // This file is an internal atomic implementation, use base/atomicops.h instead. |
6 // | 6 // |
7 // LinuxKernelCmpxchg and Barrier_AtomicIncrement are from Google Gears. | 7 // LinuxKernelCmpxchg and Barrier_AtomicIncrement are from Google Gears. |
8 | 8 |
9 #ifndef BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ | 9 #ifndef BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ |
10 #define BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ | 10 #define BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ |
(...skipping 52 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
63 int reloop; | 63 int reloop; |
64 do { | 64 do { |
65 // The following is equivalent to: | 65 // The following is equivalent to: |
66 // | 66 // |
67 // prev_value = LDREX(ptr) | 67 // prev_value = LDREX(ptr) |
68 // reloop = 0 | 68 // reloop = 0 |
69 // if (prev_value != old_value) | 69 // if (prev_value != old_value) |
70 // reloop = STREX(ptr, new_value) | 70 // reloop = STREX(ptr, new_value) |
71 __asm__ __volatile__(" ldrex %0, [%3]\n" | 71 __asm__ __volatile__(" ldrex %0, [%3]\n" |
72 " mov %1, #0\n" | 72 " mov %1, #0\n" |
73 " teq %0, %4\n" | 73 " cmp %0, %4\n" |
74 #ifdef __thumb2__ | 74 #ifdef __thumb2__ |
75 " it eq\n" | 75 " it eq\n" |
76 #endif | 76 #endif |
77 " strexeq %1, %5, [%3]\n" | 77 " strexeq %1, %5, [%3]\n" |
78 : "=&r"(prev_value), "=&r"(reloop), "+m"(*ptr) | 78 : "=&r"(prev_value), "=&r"(reloop), "+m"(*ptr) |
79 : "r"(ptr), "r"(old_value), "r"(new_value) | 79 : "r"(ptr), "r"(old_value), "r"(new_value) |
80 : "cc", "memory"); | 80 : "cc", "memory"); |
81 } while (reloop != 0); | 81 } while (reloop != 0); |
82 return prev_value; | 82 return prev_value; |
83 } | 83 } |
(...skipping 44 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
128 MemoryBarrier(); | 128 MemoryBarrier(); |
129 return result; | 129 return result; |
130 } | 130 } |
131 | 131 |
132 inline Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, | 132 inline Atomic32 NoBarrier_AtomicExchange(volatile Atomic32* ptr, |
133 Atomic32 new_value) { | 133 Atomic32 new_value) { |
134 Atomic32 old_value; | 134 Atomic32 old_value; |
135 int reloop; | 135 int reloop; |
136 do { | 136 do { |
137 // old_value = LDREX(ptr) | 137 // old_value = LDREX(ptr) |
138 // fail = STREX(ptr, new_value) | 138 // reloop = STREX(ptr, new_value) |
139 __asm__ __volatile__(" ldrex %0, [%3]\n" | 139 __asm__ __volatile__(" ldrex %0, [%3]\n" |
140 " strex %1, %4, [%3]\n" | 140 " strex %1, %4, [%3]\n" |
141 : "=&r"(old_value), "=&r"(reloop), "+m"(*ptr) | 141 : "=&r"(old_value), "=&r"(reloop), "+m"(*ptr) |
142 : "r"(ptr), "r"(new_value) | 142 : "r"(ptr), "r"(new_value) |
143 : "cc", "memory"); | 143 : "cc", "memory"); |
144 } while (reloop != 0); | 144 } while (reloop != 0); |
145 return old_value; | 145 return old_value; |
146 } | 146 } |
147 | 147 |
148 // This tests against any known ARMv5 variant. | 148 // This tests against any known ARMv5 variant. |
(...skipping 76 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... | |
225 return prev_value; | 225 return prev_value; |
226 } | 226 } |
227 if (!LinuxKernelCmpxchg(old_value, new_value, ptr)) | 227 if (!LinuxKernelCmpxchg(old_value, new_value, ptr)) |
228 return old_value; | 228 return old_value; |
229 } | 229 } |
230 } | 230 } |
231 | 231 |
232 inline Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr, | 232 inline Atomic32 Release_CompareAndSwap(volatile Atomic32* ptr, |
233 Atomic32 old_value, | 233 Atomic32 old_value, |
234 Atomic32 new_value) { | 234 Atomic32 new_value) { |
235 // Use NoBarrier_CompareAndSwap(), because its implementation | 235 // This could be implemented as: |
236 // ensures that all stores happen through the kernel helper | 236 // MemoryBarrier(); |
237 // which always implement a full barrier. | 237 // return NoBarrier_CompareAndSwap(); |
238 return NoBarrier_CompareAndSwap(ptr, old_value, new_value); | 238 // |
239 // But would use 3 barriers per succesful CAS. To save performance, | |
240 // use Acquire_CompareAndSwap(). Its implementation guarantees that: | |
241 // - A succesful swap uses only 2 barriers (in the kernel helper). | |
242 // - An early return due to (prev_value != old_value) performs | |
243 // a memory barrier with no store, which is equivalent to the | |
244 // generic implementation above. | |
245 return Acquire_CompareAndSwap(ptr, old_value, new_value); | |
Dmitry Vyukov
2013/07/15 14:42:32
A failing Release_CompareAndSwap() does not need a
| |
239 } | 246 } |
240 | 247 |
241 #else | 248 #else |
242 # error "Your CPU's ARM architecture is not supported yet" | 249 # error "Your CPU's ARM architecture is not supported yet" |
243 #endif | 250 #endif |
244 | 251 |
245 // NOTE: Atomicity of the following load and store operations is only | 252 // NOTE: Atomicity of the following load and store operations is only |
246 // guaranteed in case of 32-bit alignement of |ptr| values. | 253 // guaranteed in case of 32-bit alignement of |ptr| values. |
247 | 254 |
248 inline void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value) { | 255 inline void NoBarrier_Store(volatile Atomic32* ptr, Atomic32 value) { |
(...skipping 20 matching lines...) Expand all Loading... | |
269 | 276 |
270 inline Atomic32 Release_Load(volatile const Atomic32* ptr) { | 277 inline Atomic32 Release_Load(volatile const Atomic32* ptr) { |
271 MemoryBarrier(); | 278 MemoryBarrier(); |
272 return *ptr; | 279 return *ptr; |
273 } | 280 } |
274 | 281 |
275 } // namespace base::subtle | 282 } // namespace base::subtle |
276 } // namespace base | 283 } // namespace base |
277 | 284 |
278 #endif // BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ | 285 #endif // BASE_ATOMICOPS_INTERNALS_ARM_GCC_H_ |
OLD | NEW |