diff options
author | Peter Zijlstra <peterz@infradead.org> | 2014-03-13 14:00:37 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2014-04-18 05:40:33 -0400 |
commit | 8715466b602729061394df18864ea64b97951589 (patch) | |
tree | 4ff3109c25f89b45b041bd9ef1360644eb2e7f77 | |
parent | 030d0178bdbd237c1f0577f03bbc1d7140a75044 (diff) |
arch,arm64: Convert smp_mb__*()
AARGH64 uses ll/sc primitives that do not imply any barriers for the
normal atomics, therefore smp_mb__{before,after} should be a full
barrier.
Since AARGH64 doesn't use asm-generic/barrier.h, add the required
definitions to its asm/barrier.h.
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Link: http://lkml.kernel.org/n/tip-8p5iclqgy78al33kck3ht7nr@git.kernel.org
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: Chen Gang <gang.chen@asianux.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Russell King <rmk+kernel@arm.linux.org.uk>
Cc: Will Deacon <will.deacon@arm.com>
Cc: linux-arm-kernel@lists.infradead.org
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
-rw-r--r-- | arch/arm64/include/asm/atomic.h | 5 | ||||
-rw-r--r-- | arch/arm64/include/asm/barrier.h | 3 | ||||
-rw-r--r-- | arch/arm64/include/asm/bitops.h | 9 |
3 files changed, 3 insertions, 14 deletions
diff --git a/arch/arm64/include/asm/atomic.h b/arch/arm64/include/asm/atomic.h index 0237f0867e37..57e8cb49824c 100644 --- a/arch/arm64/include/asm/atomic.h +++ b/arch/arm64/include/asm/atomic.h | |||
@@ -152,11 +152,6 @@ static inline int __atomic_add_unless(atomic_t *v, int a, int u) | |||
152 | 152 | ||
153 | #define atomic_add_negative(i,v) (atomic_add_return(i, v) < 0) | 153 | #define atomic_add_negative(i,v) (atomic_add_return(i, v) < 0) |
154 | 154 | ||
155 | #define smp_mb__before_atomic_dec() smp_mb() | ||
156 | #define smp_mb__after_atomic_dec() smp_mb() | ||
157 | #define smp_mb__before_atomic_inc() smp_mb() | ||
158 | #define smp_mb__after_atomic_inc() smp_mb() | ||
159 | |||
160 | /* | 155 | /* |
161 | * 64-bit atomic operations. | 156 | * 64-bit atomic operations. |
162 | */ | 157 | */ |
diff --git a/arch/arm64/include/asm/barrier.h b/arch/arm64/include/asm/barrier.h index 66eb7648043b..48b9e704af7c 100644 --- a/arch/arm64/include/asm/barrier.h +++ b/arch/arm64/include/asm/barrier.h | |||
@@ -98,6 +98,9 @@ do { \ | |||
98 | #define set_mb(var, value) do { var = value; smp_mb(); } while (0) | 98 | #define set_mb(var, value) do { var = value; smp_mb(); } while (0) |
99 | #define nop() asm volatile("nop"); | 99 | #define nop() asm volatile("nop"); |
100 | 100 | ||
101 | #define smp_mb__before_atomic() smp_mb() | ||
102 | #define smp_mb__after_atomic() smp_mb() | ||
103 | |||
101 | #endif /* __ASSEMBLY__ */ | 104 | #endif /* __ASSEMBLY__ */ |
102 | 105 | ||
103 | #endif /* __ASM_BARRIER_H */ | 106 | #endif /* __ASM_BARRIER_H */ |
diff --git a/arch/arm64/include/asm/bitops.h b/arch/arm64/include/asm/bitops.h index aa5b59d6ba43..9c19594ce7cb 100644 --- a/arch/arm64/include/asm/bitops.h +++ b/arch/arm64/include/asm/bitops.h | |||
@@ -17,17 +17,8 @@ | |||
17 | #define __ASM_BITOPS_H | 17 | #define __ASM_BITOPS_H |
18 | 18 | ||
19 | #include <linux/compiler.h> | 19 | #include <linux/compiler.h> |
20 | |||
21 | #include <asm/barrier.h> | 20 | #include <asm/barrier.h> |
22 | 21 | ||
23 | /* | ||
24 | * clear_bit may not imply a memory barrier | ||
25 | */ | ||
26 | #ifndef smp_mb__before_clear_bit | ||
27 | #define smp_mb__before_clear_bit() smp_mb() | ||
28 | #define smp_mb__after_clear_bit() smp_mb() | ||
29 | #endif | ||
30 | |||
31 | #ifndef _LINUX_BITOPS_H | 22 | #ifndef _LINUX_BITOPS_H |
32 | #error only <linux/bitops.h> can be included directly | 23 | #error only <linux/bitops.h> can be included directly |
33 | #endif | 24 | #endif |