diff options
author | David S. Miller <davem@davemloft.net> | 2009-04-01 04:47:10 -0400 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2009-06-16 07:56:11 -0400 |
commit | 5a5488d3bb9a23d9884572e5d85dfeefe8749d3d (patch) | |
tree | afa8db75cdf771257cd5541ed80a606df60f9cf8 /arch/sparc/include | |
parent | 19f0fa3fb3499d8c5fb861933959f546d05fc202 (diff) |
sparc64: Store per-cpu offset in trap_block[]
Surprisingly this actually makes LOAD_PER_CPU_BASE() a little
more efficient.
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/sparc/include')
-rw-r--r-- | arch/sparc/include/asm/percpu_64.h | 6 | ||||
-rw-r--r-- | arch/sparc/include/asm/trap_block.h | 14 |
2 files changed, 10 insertions, 10 deletions
diff --git a/arch/sparc/include/asm/percpu_64.h b/arch/sparc/include/asm/percpu_64.h index bee64593023e..c0ab102d11f6 100644 --- a/arch/sparc/include/asm/percpu_64.h +++ b/arch/sparc/include/asm/percpu_64.h | |||
@@ -7,12 +7,12 @@ register unsigned long __local_per_cpu_offset asm("g5"); | |||
7 | 7 | ||
8 | #ifdef CONFIG_SMP | 8 | #ifdef CONFIG_SMP |
9 | 9 | ||
10 | #include <asm/trap_block.h> | ||
11 | |||
10 | extern void real_setup_per_cpu_areas(void); | 12 | extern void real_setup_per_cpu_areas(void); |
11 | 13 | ||
12 | extern unsigned long __per_cpu_base; | ||
13 | extern unsigned long __per_cpu_shift; | ||
14 | #define __per_cpu_offset(__cpu) \ | 14 | #define __per_cpu_offset(__cpu) \ |
15 | (__per_cpu_base + ((unsigned long)(__cpu) << __per_cpu_shift)) | 15 | (trap_block[(__cpu)].__per_cpu_base) |
16 | #define per_cpu_offset(x) (__per_cpu_offset(x)) | 16 | #define per_cpu_offset(x) (__per_cpu_offset(x)) |
17 | 17 | ||
18 | #define __my_cpu_offset __local_per_cpu_offset | 18 | #define __my_cpu_offset __local_per_cpu_offset |
diff --git a/arch/sparc/include/asm/trap_block.h b/arch/sparc/include/asm/trap_block.h index 68fd9ee3e8ae..7e26b2db6211 100644 --- a/arch/sparc/include/asm/trap_block.h +++ b/arch/sparc/include/asm/trap_block.h | |||
@@ -48,7 +48,7 @@ struct trap_per_cpu { | |||
48 | unsigned int dev_mondo_qmask; | 48 | unsigned int dev_mondo_qmask; |
49 | unsigned int resum_qmask; | 49 | unsigned int resum_qmask; |
50 | unsigned int nonresum_qmask; | 50 | unsigned int nonresum_qmask; |
51 | unsigned long __unused; | 51 | unsigned long __per_cpu_base; |
52 | } __attribute__((aligned(64))); | 52 | } __attribute__((aligned(64))); |
53 | extern struct trap_per_cpu trap_block[NR_CPUS]; | 53 | extern struct trap_per_cpu trap_block[NR_CPUS]; |
54 | extern void init_cur_cpu_trap(struct thread_info *); | 54 | extern void init_cur_cpu_trap(struct thread_info *); |
@@ -101,6 +101,7 @@ extern struct sun4v_2insn_patch_entry __sun4v_2insn_patch, | |||
101 | #define TRAP_PER_CPU_DEV_MONDO_QMASK 0xec | 101 | #define TRAP_PER_CPU_DEV_MONDO_QMASK 0xec |
102 | #define TRAP_PER_CPU_RESUM_QMASK 0xf0 | 102 | #define TRAP_PER_CPU_RESUM_QMASK 0xf0 |
103 | #define TRAP_PER_CPU_NONRESUM_QMASK 0xf4 | 103 | #define TRAP_PER_CPU_NONRESUM_QMASK 0xf4 |
104 | #define TRAP_PER_CPU_PER_CPU_BASE 0xf8 | ||
104 | 105 | ||
105 | #define TRAP_BLOCK_SZ_SHIFT 8 | 106 | #define TRAP_BLOCK_SZ_SHIFT 8 |
106 | 107 | ||
@@ -172,12 +173,11 @@ extern struct sun4v_2insn_patch_entry __sun4v_2insn_patch, | |||
172 | */ | 173 | */ |
173 | #define LOAD_PER_CPU_BASE(DEST, THR, REG1, REG2, REG3) \ | 174 | #define LOAD_PER_CPU_BASE(DEST, THR, REG1, REG2, REG3) \ |
174 | lduh [THR + TI_CPU], REG1; \ | 175 | lduh [THR + TI_CPU], REG1; \ |
175 | sethi %hi(__per_cpu_shift), REG3; \ | 176 | sethi %hi(trap_block), REG2; \ |
176 | sethi %hi(__per_cpu_base), REG2; \ | 177 | sllx REG1, TRAP_BLOCK_SZ_SHIFT, REG1; \ |
177 | ldx [REG3 + %lo(__per_cpu_shift)], REG3; \ | 178 | or REG2, %lo(trap_block), REG2; \ |
178 | ldx [REG2 + %lo(__per_cpu_base)], REG2; \ | 179 | add REG2, REG1, REG2; \ |
179 | sllx REG1, REG3, REG3; \ | 180 | ldx [REG2 + TRAP_PER_CPU_PER_CPU_BASE], DEST; |
180 | add REG3, REG2, DEST; | ||
181 | 181 | ||
182 | #else | 182 | #else |
183 | 183 | ||