diff options
author | Russell King <rmk@dyn-67.arm.linux.org.uk> | 2005-07-26 14:44:26 -0400 |
---|---|---|
committer | Russell King <rmk+kernel@arm.linux.org.uk> | 2005-07-26 14:44:26 -0400 |
commit | 6d9b37a3a80195d317887ff81aad6a58a66954b5 (patch) | |
tree | 98d1c96416b660070b1fd37dbbc0c80d39c55b98 /include | |
parent | 9560782f9a68a5de3e72bc3ba71317f373844549 (diff) |
[PATCH] ARM SMP: Add ARMv6 memory barriers
Convert explicit gcc asm-based memory barriers into smp_mb() calls.
These change between barrier() and the ARMv6 data memory barrier
instruction depending on whether ARMv6 SMP is enabled.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
Diffstat (limited to 'include')
-rw-r--r-- | include/asm-arm/bitops.h | 4 | ||||
-rw-r--r-- | include/asm-arm/locks.h | 36 | ||||
-rw-r--r-- | include/asm-arm/spinlock.h | 53 | ||||
-rw-r--r-- | include/asm-arm/system.h | 5 |
4 files changed, 69 insertions, 29 deletions
diff --git a/include/asm-arm/bitops.h b/include/asm-arm/bitops.h index 4edd4dc40c5b..c1adc6b3e86d 100644 --- a/include/asm-arm/bitops.h +++ b/include/asm-arm/bitops.h | |||
@@ -21,8 +21,8 @@ | |||
21 | 21 | ||
22 | #include <asm/system.h> | 22 | #include <asm/system.h> |
23 | 23 | ||
24 | #define smp_mb__before_clear_bit() do { } while (0) | 24 | #define smp_mb__before_clear_bit() mb() |
25 | #define smp_mb__after_clear_bit() do { } while (0) | 25 | #define smp_mb__after_clear_bit() mb() |
26 | 26 | ||
27 | /* | 27 | /* |
28 | * These functions are the basis of our bit ops. | 28 | * These functions are the basis of our bit ops. |
diff --git a/include/asm-arm/locks.h b/include/asm-arm/locks.h index 9cb33fcc06c1..f08dc8447913 100644 --- a/include/asm-arm/locks.h +++ b/include/asm-arm/locks.h | |||
@@ -28,7 +28,8 @@ | |||
28 | " blmi " #fail \ | 28 | " blmi " #fail \ |
29 | : \ | 29 | : \ |
30 | : "r" (ptr), "I" (1) \ | 30 | : "r" (ptr), "I" (1) \ |
31 | : "ip", "lr", "cc", "memory"); \ | 31 | : "ip", "lr", "cc"); \ |
32 | smp_mb(); \ | ||
32 | }) | 33 | }) |
33 | 34 | ||
34 | #define __down_op_ret(ptr,fail) \ | 35 | #define __down_op_ret(ptr,fail) \ |
@@ -48,12 +49,14 @@ | |||
48 | " mov %0, ip" \ | 49 | " mov %0, ip" \ |
49 | : "=&r" (ret) \ | 50 | : "=&r" (ret) \ |
50 | : "r" (ptr), "I" (1) \ | 51 | : "r" (ptr), "I" (1) \ |
51 | : "ip", "lr", "cc", "memory"); \ | 52 | : "ip", "lr", "cc"); \ |
53 | smp_mb(); \ | ||
52 | ret; \ | 54 | ret; \ |
53 | }) | 55 | }) |
54 | 56 | ||
55 | #define __up_op(ptr,wake) \ | 57 | #define __up_op(ptr,wake) \ |
56 | ({ \ | 58 | ({ \ |
59 | smp_mb(); \ | ||
57 | __asm__ __volatile__( \ | 60 | __asm__ __volatile__( \ |
58 | "@ up_op\n" \ | 61 | "@ up_op\n" \ |
59 | "1: ldrex lr, [%0]\n" \ | 62 | "1: ldrex lr, [%0]\n" \ |
@@ -66,7 +69,7 @@ | |||
66 | " blle " #wake \ | 69 | " blle " #wake \ |
67 | : \ | 70 | : \ |
68 | : "r" (ptr), "I" (1) \ | 71 | : "r" (ptr), "I" (1) \ |
69 | : "ip", "lr", "cc", "memory"); \ | 72 | : "ip", "lr", "cc"); \ |
70 | }) | 73 | }) |
71 | 74 | ||
72 | /* | 75 | /* |
@@ -92,11 +95,13 @@ | |||
92 | " blne " #fail \ | 95 | " blne " #fail \ |
93 | : \ | 96 | : \ |
94 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ | 97 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ |
95 | : "ip", "lr", "cc", "memory"); \ | 98 | : "ip", "lr", "cc"); \ |
99 | smp_mb(); \ | ||
96 | }) | 100 | }) |
97 | 101 | ||
98 | #define __up_op_write(ptr,wake) \ | 102 | #define __up_op_write(ptr,wake) \ |
99 | ({ \ | 103 | ({ \ |
104 | smp_mb(); \ | ||
100 | __asm__ __volatile__( \ | 105 | __asm__ __volatile__( \ |
101 | "@ up_op_read\n" \ | 106 | "@ up_op_read\n" \ |
102 | "1: ldrex lr, [%0]\n" \ | 107 | "1: ldrex lr, [%0]\n" \ |
@@ -108,7 +113,7 @@ | |||
108 | " blcs " #wake \ | 113 | " blcs " #wake \ |
109 | : \ | 114 | : \ |
110 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ | 115 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ |
111 | : "ip", "lr", "cc", "memory"); \ | 116 | : "ip", "lr", "cc"); \ |
112 | }) | 117 | }) |
113 | 118 | ||
114 | #define __down_op_read(ptr,fail) \ | 119 | #define __down_op_read(ptr,fail) \ |
@@ -116,6 +121,7 @@ | |||
116 | 121 | ||
117 | #define __up_op_read(ptr,wake) \ | 122 | #define __up_op_read(ptr,wake) \ |
118 | ({ \ | 123 | ({ \ |
124 | smp_mb(); \ | ||
119 | __asm__ __volatile__( \ | 125 | __asm__ __volatile__( \ |
120 | "@ up_op_read\n" \ | 126 | "@ up_op_read\n" \ |
121 | "1: ldrex lr, [%0]\n" \ | 127 | "1: ldrex lr, [%0]\n" \ |
@@ -128,7 +134,7 @@ | |||
128 | " bleq " #wake \ | 134 | " bleq " #wake \ |
129 | : \ | 135 | : \ |
130 | : "r" (ptr), "I" (1) \ | 136 | : "r" (ptr), "I" (1) \ |
131 | : "ip", "lr", "cc", "memory"); \ | 137 | : "ip", "lr", "cc"); \ |
132 | }) | 138 | }) |
133 | 139 | ||
134 | #else | 140 | #else |
@@ -148,7 +154,8 @@ | |||
148 | " blmi " #fail \ | 154 | " blmi " #fail \ |
149 | : \ | 155 | : \ |
150 | : "r" (ptr), "I" (1) \ | 156 | : "r" (ptr), "I" (1) \ |
151 | : "ip", "lr", "cc", "memory"); \ | 157 | : "ip", "lr", "cc"); \ |
158 | smp_mb(); \ | ||
152 | }) | 159 | }) |
153 | 160 | ||
154 | #define __down_op_ret(ptr,fail) \ | 161 | #define __down_op_ret(ptr,fail) \ |
@@ -169,12 +176,14 @@ | |||
169 | " mov %0, ip" \ | 176 | " mov %0, ip" \ |
170 | : "=&r" (ret) \ | 177 | : "=&r" (ret) \ |
171 | : "r" (ptr), "I" (1) \ | 178 | : "r" (ptr), "I" (1) \ |
172 | : "ip", "lr", "cc", "memory"); \ | 179 | : "ip", "lr", "cc"); \ |
180 | smp_mb(); \ | ||
173 | ret; \ | 181 | ret; \ |
174 | }) | 182 | }) |
175 | 183 | ||
176 | #define __up_op(ptr,wake) \ | 184 | #define __up_op(ptr,wake) \ |
177 | ({ \ | 185 | ({ \ |
186 | smp_mb(); \ | ||
178 | __asm__ __volatile__( \ | 187 | __asm__ __volatile__( \ |
179 | "@ up_op\n" \ | 188 | "@ up_op\n" \ |
180 | " mrs ip, cpsr\n" \ | 189 | " mrs ip, cpsr\n" \ |
@@ -188,7 +197,7 @@ | |||
188 | " blle " #wake \ | 197 | " blle " #wake \ |
189 | : \ | 198 | : \ |
190 | : "r" (ptr), "I" (1) \ | 199 | : "r" (ptr), "I" (1) \ |
191 | : "ip", "lr", "cc", "memory"); \ | 200 | : "ip", "lr", "cc"); \ |
192 | }) | 201 | }) |
193 | 202 | ||
194 | /* | 203 | /* |
@@ -215,7 +224,8 @@ | |||
215 | " blne " #fail \ | 224 | " blne " #fail \ |
216 | : \ | 225 | : \ |
217 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ | 226 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ |
218 | : "ip", "lr", "cc", "memory"); \ | 227 | : "ip", "lr", "cc"); \ |
228 | smp_mb(); \ | ||
219 | }) | 229 | }) |
220 | 230 | ||
221 | #define __up_op_write(ptr,wake) \ | 231 | #define __up_op_write(ptr,wake) \ |
@@ -233,7 +243,8 @@ | |||
233 | " blcs " #wake \ | 243 | " blcs " #wake \ |
234 | : \ | 244 | : \ |
235 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ | 245 | : "r" (ptr), "I" (RW_LOCK_BIAS) \ |
236 | : "ip", "lr", "cc", "memory"); \ | 246 | : "ip", "lr", "cc"); \ |
247 | smp_mb(); \ | ||
237 | }) | 248 | }) |
238 | 249 | ||
239 | #define __down_op_read(ptr,fail) \ | 250 | #define __down_op_read(ptr,fail) \ |
@@ -241,6 +252,7 @@ | |||
241 | 252 | ||
242 | #define __up_op_read(ptr,wake) \ | 253 | #define __up_op_read(ptr,wake) \ |
243 | ({ \ | 254 | ({ \ |
255 | smp_mb(); \ | ||
244 | __asm__ __volatile__( \ | 256 | __asm__ __volatile__( \ |
245 | "@ up_op_read\n" \ | 257 | "@ up_op_read\n" \ |
246 | " mrs ip, cpsr\n" \ | 258 | " mrs ip, cpsr\n" \ |
@@ -254,7 +266,7 @@ | |||
254 | " bleq " #wake \ | 266 | " bleq " #wake \ |
255 | : \ | 267 | : \ |
256 | : "r" (ptr), "I" (1) \ | 268 | : "r" (ptr), "I" (1) \ |
257 | : "ip", "lr", "cc", "memory"); \ | 269 | : "ip", "lr", "cc"); \ |
258 | }) | 270 | }) |
259 | 271 | ||
260 | #endif | 272 | #endif |
diff --git a/include/asm-arm/spinlock.h b/include/asm-arm/spinlock.h index 9705d5eec94c..1f906d09b688 100644 --- a/include/asm-arm/spinlock.h +++ b/include/asm-arm/spinlock.h | |||
@@ -8,9 +8,10 @@ | |||
8 | /* | 8 | /* |
9 | * ARMv6 Spin-locking. | 9 | * ARMv6 Spin-locking. |
10 | * | 10 | * |
11 | * We (exclusively) read the old value, and decrement it. If it | 11 | * We exclusively read the old value. If it is zero, we may have |
12 | * hits zero, we may have won the lock, so we try (exclusively) | 12 | * won the lock, so we try exclusively storing it. A memory barrier |
13 | * storing it. | 13 | * is required after we get a lock, and before we release it, because |
14 | * V6 CPUs are assumed to have weakly ordered memory. | ||
14 | * | 15 | * |
15 | * Unlocked value: 0 | 16 | * Unlocked value: 0 |
16 | * Locked value: 1 | 17 | * Locked value: 1 |
@@ -41,7 +42,9 @@ static inline void _raw_spin_lock(spinlock_t *lock) | |||
41 | " bne 1b" | 42 | " bne 1b" |
42 | : "=&r" (tmp) | 43 | : "=&r" (tmp) |
43 | : "r" (&lock->lock), "r" (1) | 44 | : "r" (&lock->lock), "r" (1) |
44 | : "cc", "memory"); | 45 | : "cc"); |
46 | |||
47 | smp_mb(); | ||
45 | } | 48 | } |
46 | 49 | ||
47 | static inline int _raw_spin_trylock(spinlock_t *lock) | 50 | static inline int _raw_spin_trylock(spinlock_t *lock) |
@@ -54,18 +57,25 @@ static inline int _raw_spin_trylock(spinlock_t *lock) | |||
54 | " strexeq %0, %2, [%1]" | 57 | " strexeq %0, %2, [%1]" |
55 | : "=&r" (tmp) | 58 | : "=&r" (tmp) |
56 | : "r" (&lock->lock), "r" (1) | 59 | : "r" (&lock->lock), "r" (1) |
57 | : "cc", "memory"); | 60 | : "cc"); |
58 | 61 | ||
59 | return tmp == 0; | 62 | if (tmp == 0) { |
63 | smp_mb(); | ||
64 | return 1; | ||
65 | } else { | ||
66 | return 0; | ||
67 | } | ||
60 | } | 68 | } |
61 | 69 | ||
62 | static inline void _raw_spin_unlock(spinlock_t *lock) | 70 | static inline void _raw_spin_unlock(spinlock_t *lock) |
63 | { | 71 | { |
72 | smp_mb(); | ||
73 | |||
64 | __asm__ __volatile__( | 74 | __asm__ __volatile__( |
65 | " str %1, [%0]" | 75 | " str %1, [%0]" |
66 | : | 76 | : |
67 | : "r" (&lock->lock), "r" (0) | 77 | : "r" (&lock->lock), "r" (0) |
68 | : "cc", "memory"); | 78 | : "cc"); |
69 | } | 79 | } |
70 | 80 | ||
71 | /* | 81 | /* |
@@ -98,7 +108,9 @@ static inline void _raw_write_lock(rwlock_t *rw) | |||
98 | " bne 1b" | 108 | " bne 1b" |
99 | : "=&r" (tmp) | 109 | : "=&r" (tmp) |
100 | : "r" (&rw->lock), "r" (0x80000000) | 110 | : "r" (&rw->lock), "r" (0x80000000) |
101 | : "cc", "memory"); | 111 | : "cc"); |
112 | |||
113 | smp_mb(); | ||
102 | } | 114 | } |
103 | 115 | ||
104 | static inline int _raw_write_trylock(rwlock_t *rw) | 116 | static inline int _raw_write_trylock(rwlock_t *rw) |
@@ -111,18 +123,25 @@ static inline int _raw_write_trylock(rwlock_t *rw) | |||
111 | " strexeq %0, %2, [%1]" | 123 | " strexeq %0, %2, [%1]" |
112 | : "=&r" (tmp) | 124 | : "=&r" (tmp) |
113 | : "r" (&rw->lock), "r" (0x80000000) | 125 | : "r" (&rw->lock), "r" (0x80000000) |
114 | : "cc", "memory"); | 126 | : "cc"); |
115 | 127 | ||
116 | return tmp == 0; | 128 | if (tmp == 0) { |
129 | smp_mb(); | ||
130 | return 1; | ||
131 | } else { | ||
132 | return 0; | ||
133 | } | ||
117 | } | 134 | } |
118 | 135 | ||
119 | static inline void _raw_write_unlock(rwlock_t *rw) | 136 | static inline void _raw_write_unlock(rwlock_t *rw) |
120 | { | 137 | { |
138 | smp_mb(); | ||
139 | |||
121 | __asm__ __volatile__( | 140 | __asm__ __volatile__( |
122 | "str %1, [%0]" | 141 | "str %1, [%0]" |
123 | : | 142 | : |
124 | : "r" (&rw->lock), "r" (0) | 143 | : "r" (&rw->lock), "r" (0) |
125 | : "cc", "memory"); | 144 | : "cc"); |
126 | } | 145 | } |
127 | 146 | ||
128 | /* | 147 | /* |
@@ -149,13 +168,17 @@ static inline void _raw_read_lock(rwlock_t *rw) | |||
149 | " bmi 1b" | 168 | " bmi 1b" |
150 | : "=&r" (tmp), "=&r" (tmp2) | 169 | : "=&r" (tmp), "=&r" (tmp2) |
151 | : "r" (&rw->lock) | 170 | : "r" (&rw->lock) |
152 | : "cc", "memory"); | 171 | : "cc"); |
172 | |||
173 | smp_mb(); | ||
153 | } | 174 | } |
154 | 175 | ||
155 | static inline void _raw_read_unlock(rwlock_t *rw) | 176 | static inline void _raw_read_unlock(rwlock_t *rw) |
156 | { | 177 | { |
157 | unsigned long tmp, tmp2; | 178 | unsigned long tmp, tmp2; |
158 | 179 | ||
180 | smp_mb(); | ||
181 | |||
159 | __asm__ __volatile__( | 182 | __asm__ __volatile__( |
160 | "1: ldrex %0, [%2]\n" | 183 | "1: ldrex %0, [%2]\n" |
161 | " sub %0, %0, #1\n" | 184 | " sub %0, %0, #1\n" |
@@ -164,7 +187,7 @@ static inline void _raw_read_unlock(rwlock_t *rw) | |||
164 | " bne 1b" | 187 | " bne 1b" |
165 | : "=&r" (tmp), "=&r" (tmp2) | 188 | : "=&r" (tmp), "=&r" (tmp2) |
166 | : "r" (&rw->lock) | 189 | : "r" (&rw->lock) |
167 | : "cc", "memory"); | 190 | : "cc"); |
168 | } | 191 | } |
169 | 192 | ||
170 | #define _raw_read_trylock(lock) generic_raw_read_trylock(lock) | 193 | #define _raw_read_trylock(lock) generic_raw_read_trylock(lock) |
diff --git a/include/asm-arm/system.h b/include/asm-arm/system.h index 3f9c3626a73c..8efa4ebdcacb 100644 --- a/include/asm-arm/system.h +++ b/include/asm-arm/system.h | |||
@@ -139,7 +139,12 @@ extern unsigned int user_debug; | |||
139 | #define vectors_high() (0) | 139 | #define vectors_high() (0) |
140 | #endif | 140 | #endif |
141 | 141 | ||
142 | #if __LINUX_ARM_ARCH__ >= 6 | ||
143 | #define mb() __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" \ | ||
144 | : : "r" (0) : "memory") | ||
145 | #else | ||
142 | #define mb() __asm__ __volatile__ ("" : : : "memory") | 146 | #define mb() __asm__ __volatile__ ("" : : : "memory") |
147 | #endif | ||
143 | #define rmb() mb() | 148 | #define rmb() mb() |
144 | #define wmb() mb() | 149 | #define wmb() mb() |
145 | #define read_barrier_depends() do { } while(0) | 150 | #define read_barrier_depends() do { } while(0) |