aboutsummaryrefslogtreecommitdiffstats
path: root/include/asm-s390/percpu.h
diff options
context:
space:
mode:
Diffstat (limited to 'include/asm-s390/percpu.h')
-rw-r--r--include/asm-s390/percpu.h62
1 files changed, 9 insertions, 53 deletions
diff --git a/include/asm-s390/percpu.h b/include/asm-s390/percpu.h
index 545857e64443..408d60b4f75b 100644
--- a/include/asm-s390/percpu.h
+++ b/include/asm-s390/percpu.h
@@ -4,8 +4,6 @@
4#include <linux/compiler.h> 4#include <linux/compiler.h>
5#include <asm/lowcore.h> 5#include <asm/lowcore.h>
6 6
7#define __GENERIC_PER_CPU
8
9/* 7/*
10 * s390 uses its own implementation for per cpu data, the offset of 8 * s390 uses its own implementation for per cpu data, the offset of
11 * the cpu local data area is cached in the cpu's lowcore memory. 9 * the cpu local data area is cached in the cpu's lowcore memory.
@@ -15,67 +13,25 @@
15 */ 13 */
16#if defined(__s390x__) && defined(MODULE) 14#if defined(__s390x__) && defined(MODULE)
17 15
18#define __reloc_hide(var,offset) (*({ \ 16#define SHIFT_PERCPU_PTR(ptr,offset) (({ \
19 extern int simple_identifier_##var(void); \ 17 extern int simple_identifier_##var(void); \
20 unsigned long *__ptr; \ 18 unsigned long *__ptr; \
21 asm ( "larl %0,per_cpu__"#var"@GOTENT" \ 19 asm ( "larl %0, %1@GOTENT" \
22 : "=a" (__ptr) : "X" (per_cpu__##var) ); \ 20 : "=a" (__ptr) : "X" (ptr) ); \
23 (typeof(&per_cpu__##var))((*__ptr) + (offset)); })) 21 (typeof(ptr))((*__ptr) + (offset)); }))
24 22
25#else 23#else
26 24
27#define __reloc_hide(var, offset) (*({ \ 25#define SHIFT_PERCPU_PTR(ptr, offset) (({ \
28 extern int simple_identifier_##var(void); \ 26 extern int simple_identifier_##var(void); \
29 unsigned long __ptr; \ 27 unsigned long __ptr; \
30 asm ( "" : "=a" (__ptr) : "0" (&per_cpu__##var) ); \ 28 asm ( "" : "=a" (__ptr) : "0" (ptr) ); \
31 (typeof(&per_cpu__##var)) (__ptr + (offset)); })) 29 (typeof(ptr)) (__ptr + (offset)); }))
32 30
33#endif 31#endif
34 32
35#ifdef CONFIG_SMP 33#define __my_cpu_offset S390_lowcore.percpu_offset
36
37extern unsigned long __per_cpu_offset[NR_CPUS];
38
39/* Separate out the type, so (int[3], foo) works. */
40#define DEFINE_PER_CPU(type, name) \
41 __attribute__((__section__(".data.percpu"))) \
42 __typeof__(type) per_cpu__##name
43
44#define DEFINE_PER_CPU_SHARED_ALIGNED(type, name) \
45 __attribute__((__section__(".data.percpu.shared_aligned"))) \
46 __typeof__(type) per_cpu__##name \
47 ____cacheline_aligned_in_smp
48
49#define __get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset)
50#define __raw_get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset)
51#define per_cpu(var,cpu) __reloc_hide(var,__per_cpu_offset[cpu])
52#define per_cpu_offset(x) (__per_cpu_offset[x])
53
54/* A macro to avoid #include hell... */
55#define percpu_modcopy(pcpudst, src, size) \
56do { \
57 unsigned int __i; \
58 for_each_possible_cpu(__i) \
59 memcpy((pcpudst)+__per_cpu_offset[__i], \
60 (src), (size)); \
61} while (0)
62
63#else /* ! SMP */
64
65#define DEFINE_PER_CPU(type, name) \
66 __typeof__(type) per_cpu__##name
67#define DEFINE_PER_CPU_SHARED_ALIGNED(type, name) \
68 DEFINE_PER_CPU(type, name)
69
70#define __get_cpu_var(var) __reloc_hide(var,0)
71#define __raw_get_cpu_var(var) __reloc_hide(var,0)
72#define per_cpu(var,cpu) __reloc_hide(var,0)
73
74#endif /* SMP */
75
76#define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name
77 34
78#define EXPORT_PER_CPU_SYMBOL(var) EXPORT_SYMBOL(per_cpu__##var) 35#include <asm-generic/percpu.h>
79#define EXPORT_PER_CPU_SYMBOL_GPL(var) EXPORT_SYMBOL_GPL(per_cpu__##var)
80 36
81#endif /* __ARCH_S390_PERCPU__ */ 37#endif /* __ARCH_S390_PERCPU__ */