aboutsummaryrefslogtreecommitdiffstats
path: root/arch/sparc/include
diff options
context:
space:
mode:
authorDavid S. Miller <davem@davemloft.net>2009-04-01 04:47:10 -0400
committerDavid S. Miller <davem@davemloft.net>2009-06-16 07:56:11 -0400
commit5a5488d3bb9a23d9884572e5d85dfeefe8749d3d (patch)
treeafa8db75cdf771257cd5541ed80a606df60f9cf8 /arch/sparc/include
parent19f0fa3fb3499d8c5fb861933959f546d05fc202 (diff)
sparc64: Store per-cpu offset in trap_block[]
Surprisingly this actually makes LOAD_PER_CPU_BASE() a little more efficient. Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/sparc/include')
-rw-r--r--arch/sparc/include/asm/percpu_64.h6
-rw-r--r--arch/sparc/include/asm/trap_block.h14
2 files changed, 10 insertions, 10 deletions
diff --git a/arch/sparc/include/asm/percpu_64.h b/arch/sparc/include/asm/percpu_64.h
index bee64593023e..c0ab102d11f6 100644
--- a/arch/sparc/include/asm/percpu_64.h
+++ b/arch/sparc/include/asm/percpu_64.h
@@ -7,12 +7,12 @@ register unsigned long __local_per_cpu_offset asm("g5");
7 7
8#ifdef CONFIG_SMP 8#ifdef CONFIG_SMP
9 9
10#include <asm/trap_block.h>
11
10extern void real_setup_per_cpu_areas(void); 12extern void real_setup_per_cpu_areas(void);
11 13
12extern unsigned long __per_cpu_base;
13extern unsigned long __per_cpu_shift;
14#define __per_cpu_offset(__cpu) \ 14#define __per_cpu_offset(__cpu) \
15 (__per_cpu_base + ((unsigned long)(__cpu) << __per_cpu_shift)) 15 (trap_block[(__cpu)].__per_cpu_base)
16#define per_cpu_offset(x) (__per_cpu_offset(x)) 16#define per_cpu_offset(x) (__per_cpu_offset(x))
17 17
18#define __my_cpu_offset __local_per_cpu_offset 18#define __my_cpu_offset __local_per_cpu_offset
diff --git a/arch/sparc/include/asm/trap_block.h b/arch/sparc/include/asm/trap_block.h
index 68fd9ee3e8ae..7e26b2db6211 100644
--- a/arch/sparc/include/asm/trap_block.h
+++ b/arch/sparc/include/asm/trap_block.h
@@ -48,7 +48,7 @@ struct trap_per_cpu {
48 unsigned int dev_mondo_qmask; 48 unsigned int dev_mondo_qmask;
49 unsigned int resum_qmask; 49 unsigned int resum_qmask;
50 unsigned int nonresum_qmask; 50 unsigned int nonresum_qmask;
51 unsigned long __unused; 51 unsigned long __per_cpu_base;
52} __attribute__((aligned(64))); 52} __attribute__((aligned(64)));
53extern struct trap_per_cpu trap_block[NR_CPUS]; 53extern struct trap_per_cpu trap_block[NR_CPUS];
54extern void init_cur_cpu_trap(struct thread_info *); 54extern void init_cur_cpu_trap(struct thread_info *);
@@ -101,6 +101,7 @@ extern struct sun4v_2insn_patch_entry __sun4v_2insn_patch,
101#define TRAP_PER_CPU_DEV_MONDO_QMASK 0xec 101#define TRAP_PER_CPU_DEV_MONDO_QMASK 0xec
102#define TRAP_PER_CPU_RESUM_QMASK 0xf0 102#define TRAP_PER_CPU_RESUM_QMASK 0xf0
103#define TRAP_PER_CPU_NONRESUM_QMASK 0xf4 103#define TRAP_PER_CPU_NONRESUM_QMASK 0xf4
104#define TRAP_PER_CPU_PER_CPU_BASE 0xf8
104 105
105#define TRAP_BLOCK_SZ_SHIFT 8 106#define TRAP_BLOCK_SZ_SHIFT 8
106 107
@@ -172,12 +173,11 @@ extern struct sun4v_2insn_patch_entry __sun4v_2insn_patch,
172 */ 173 */
173#define LOAD_PER_CPU_BASE(DEST, THR, REG1, REG2, REG3) \ 174#define LOAD_PER_CPU_BASE(DEST, THR, REG1, REG2, REG3) \
174 lduh [THR + TI_CPU], REG1; \ 175 lduh [THR + TI_CPU], REG1; \
175 sethi %hi(__per_cpu_shift), REG3; \ 176 sethi %hi(trap_block), REG2; \
176 sethi %hi(__per_cpu_base), REG2; \ 177 sllx REG1, TRAP_BLOCK_SZ_SHIFT, REG1; \
177 ldx [REG3 + %lo(__per_cpu_shift)], REG3; \ 178 or REG2, %lo(trap_block), REG2; \
178 ldx [REG2 + %lo(__per_cpu_base)], REG2; \ 179 add REG2, REG1, REG2; \
179 sllx REG1, REG3, REG3; \ 180 ldx [REG2 + TRAP_PER_CPU_PER_CPU_BASE], DEST;
180 add REG3, REG2, DEST;
181 181
182#else 182#else
183 183