aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/sparc64/kernel/binfmt_aout32.c3
-rw-r--r--arch/sparc64/kernel/process.c3
-rw-r--r--arch/sparc64/kernel/tsb.S55
-rw-r--r--arch/sparc64/mm/tsb.c109
-rw-r--r--include/asm-sparc64/mmu.h13
-rw-r--r--include/asm-sparc64/mmu_context.h15
-rw-r--r--include/asm-sparc64/tsb.h2
7 files changed, 142 insertions, 58 deletions
diff --git a/arch/sparc64/kernel/binfmt_aout32.c b/arch/sparc64/kernel/binfmt_aout32.c
index a57d7f2b6f13..181c8cdf9541 100644
--- a/arch/sparc64/kernel/binfmt_aout32.c
+++ b/arch/sparc64/kernel/binfmt_aout32.c
@@ -330,8 +330,7 @@ beyond_if:
330 330
331 current->mm->start_stack = 331 current->mm->start_stack =
332 (unsigned long) create_aout32_tables((char __user *)bprm->p, bprm); 332 (unsigned long) create_aout32_tables((char __user *)bprm->p, bprm);
333 tsb_context_switch(__pa(current->mm->pgd), 333 tsb_context_switch(mm);
334 current->mm->context.sparc64_tsb);
335 334
336 start_thread32(regs, ex.a_entry, current->mm->start_stack); 335 start_thread32(regs, ex.a_entry, current->mm->start_stack);
337 if (current->ptrace & PT_PTRACED) 336 if (current->ptrace & PT_PTRACED)
diff --git a/arch/sparc64/kernel/process.c b/arch/sparc64/kernel/process.c
index 2784aab0d3e5..26548fc604b6 100644
--- a/arch/sparc64/kernel/process.c
+++ b/arch/sparc64/kernel/process.c
@@ -441,8 +441,7 @@ void flush_thread(void)
441 441
442 mm = t->task->mm; 442 mm = t->task->mm;
443 if (mm) 443 if (mm)
444 tsb_context_switch(__pa(mm->pgd), 444 tsb_context_switch(mm);
445 mm->context.sparc64_tsb);
446 445
447 set_thread_wsaved(0); 446 set_thread_wsaved(0);
448 447
diff --git a/arch/sparc64/kernel/tsb.S b/arch/sparc64/kernel/tsb.S
index 76f2c0b01f36..fe266bad0a28 100644
--- a/arch/sparc64/kernel/tsb.S
+++ b/arch/sparc64/kernel/tsb.S
@@ -130,48 +130,36 @@ winfix_trampoline:
130 * schedule() time. 130 * schedule() time.
131 * 131 *
132 * %o0: page table physical address 132 * %o0: page table physical address
133 * %o1: TSB address 133 * %o1: TSB register value
134 * %o2: TSB virtual address
135 * %o3: TSB mapping locked PTE
136 *
137 * We have to run this whole thing with interrupts
138 * disabled so that the current cpu doesn't change
139 * due to preemption.
134 */ 140 */
135 .align 32 141 .align 32
136 .globl tsb_context_switch 142 .globl __tsb_context_switch
137tsb_context_switch: 143__tsb_context_switch:
138 rdpr %pstate, %o5 144 rdpr %pstate, %o5
139 wrpr %o5, PSTATE_IE, %pstate 145 wrpr %o5, PSTATE_IE, %pstate
140 146
141 ldub [%g6 + TI_CPU], %o3 147 ldub [%g6 + TI_CPU], %g1
142 sethi %hi(trap_block), %o4 148 sethi %hi(trap_block), %g2
143 sllx %o3, TRAP_BLOCK_SZ_SHIFT, %o3 149 sllx %g1, TRAP_BLOCK_SZ_SHIFT, %g1
144 or %o4, %lo(trap_block), %o4 150 or %g2, %lo(trap_block), %g2
145 add %o4, %o3, %o4 151 add %g2, %g1, %g2
146 stx %o0, [%o4 + TRAP_PER_CPU_PGD_PADDR] 152 stx %o0, [%g2 + TRAP_PER_CPU_PGD_PADDR]
147
148 brgez %o1, 9f
149 nop
150
151 /* Lock TSB into D-TLB. */
152 sethi %hi(PAGE_SIZE), %o3
153 and %o3, %o1, %o3
154 sethi %hi(TSBMAP_BASE), %o2
155 add %o2, %o3, %o2
156 153
157 /* XXX handle PAGE_SIZE != 8K correctly... */
158 mov TSB_REG, %g1 154 mov TSB_REG, %g1
159 stxa %o2, [%g1] ASI_DMMU 155 stxa %o1, [%g1] ASI_DMMU
160 membar #Sync 156 membar #Sync
161 157
162 stxa %o2, [%g1] ASI_IMMU 158 stxa %o1, [%g1] ASI_IMMU
163 membar #Sync 159 membar #Sync
164 160
165#define KERN_HIGHBITS ((_PAGE_VALID|_PAGE_SZBITS)^0xfffff80000000000) 161 brz %o2, 9f
166#define KERN_LOWBITS (_PAGE_CP | _PAGE_CV | _PAGE_P | _PAGE_W | _PAGE_L) 162 nop
167 sethi %uhi(KERN_HIGHBITS), %g2
168 or %g2, %ulo(KERN_HIGHBITS), %g2
169 sllx %g2, 32, %g2
170 or %g2, KERN_LOWBITS, %g2
171#undef KERN_HIGHBITS
172#undef KERN_LOWBITS
173
174 xor %o1, %g2, %o1
175 163
176 /* We use entry 61 for this locked entry. This is the spitfire 164 /* We use entry 61 for this locked entry. This is the spitfire
177 * TLB entry number, and luckily cheetah masks the value with 165 * TLB entry number, and luckily cheetah masks the value with
@@ -184,11 +172,10 @@ tsb_context_switch:
184 stxa %o2, [%g1] ASI_DMMU 172 stxa %o2, [%g1] ASI_DMMU
185 membar #Sync 173 membar #Sync
186 mov (61 << 3), %g1 174 mov (61 << 3), %g1
187 stxa %o1, [%g1] ASI_DTLB_DATA_ACCESS 175 stxa %o3, [%g1] ASI_DTLB_DATA_ACCESS
188 membar #Sync 176 membar #Sync
189
1909: 1779:
191 wrpr %o5, %pstate 178 wrpr %o5, %pstate
192 179
193 retl 180 retl
194 mov %o2, %o0 181 nop
diff --git a/arch/sparc64/mm/tsb.c b/arch/sparc64/mm/tsb.c
index 2f84cef6c1b5..dfe7144fcdf6 100644
--- a/arch/sparc64/mm/tsb.c
+++ b/arch/sparc64/mm/tsb.c
@@ -9,13 +9,7 @@
9#include <asm/tlbflush.h> 9#include <asm/tlbflush.h>
10#include <asm/tlb.h> 10#include <asm/tlb.h>
11#include <asm/mmu_context.h> 11#include <asm/mmu_context.h>
12 12#include <asm/pgtable.h>
13#define TSB_ENTRY_ALIGNMENT 16
14
15struct tsb {
16 unsigned long tag;
17 unsigned long pte;
18} __attribute__((aligned(TSB_ENTRY_ALIGNMENT)));
19 13
20/* We use an 8K TSB for the whole kernel, this allows to 14/* We use an 8K TSB for the whole kernel, this allows to
21 * handle about 4MB of modules and vmalloc mappings without 15 * handle about 4MB of modules and vmalloc mappings without
@@ -27,10 +21,10 @@ struct tsb {
27 21
28extern struct tsb swapper_tsb[KERNEL_TSB_NENTRIES]; 22extern struct tsb swapper_tsb[KERNEL_TSB_NENTRIES];
29 23
30static inline unsigned long tsb_hash(unsigned long vaddr) 24static inline unsigned long tsb_hash(unsigned long vaddr, unsigned long nentries)
31{ 25{
32 vaddr >>= PAGE_SHIFT; 26 vaddr >>= PAGE_SHIFT;
33 return vaddr & (KERNEL_TSB_NENTRIES - 1); 27 return vaddr & (nentries - 1);
34} 28}
35 29
36static inline int tag_compare(struct tsb *entry, unsigned long vaddr, unsigned long context) 30static inline int tag_compare(struct tsb *entry, unsigned long vaddr, unsigned long context)
@@ -51,7 +45,8 @@ void flush_tsb_kernel_range(unsigned long start, unsigned long end)
51 unsigned long v; 45 unsigned long v;
52 46
53 for (v = start; v < end; v += PAGE_SIZE) { 47 for (v = start; v < end; v += PAGE_SIZE) {
54 struct tsb *ent = &swapper_tsb[tsb_hash(v)]; 48 unsigned long hash = tsb_hash(v, KERNEL_TSB_NENTRIES);
49 struct tsb *ent = &swapper_tsb[hash];
55 50
56 if (tag_compare(ent, v, 0)) { 51 if (tag_compare(ent, v, 0)) {
57 ent->tag = 0UL; 52 ent->tag = 0UL;
@@ -63,8 +58,9 @@ void flush_tsb_kernel_range(unsigned long start, unsigned long end)
63void flush_tsb_user(struct mmu_gather *mp) 58void flush_tsb_user(struct mmu_gather *mp)
64{ 59{
65 struct mm_struct *mm = mp->mm; 60 struct mm_struct *mm = mp->mm;
66 struct tsb *tsb = (struct tsb *) mm->context.sparc64_tsb; 61 struct tsb *tsb = mm->context.tsb;
67 unsigned long ctx = ~0UL; 62 unsigned long ctx = ~0UL;
63 unsigned long nentries = mm->context.tsb_nentries;
68 int i; 64 int i;
69 65
70 if (CTX_VALID(mm->context)) 66 if (CTX_VALID(mm->context))
@@ -76,7 +72,7 @@ void flush_tsb_user(struct mmu_gather *mp)
76 72
77 v &= ~0x1UL; 73 v &= ~0x1UL;
78 74
79 ent = &tsb[tsb_hash(v)]; 75 ent = &tsb[tsb_hash(v, nentries)];
80 if (tag_compare(ent, v, ctx)) { 76 if (tag_compare(ent, v, ctx)) {
81 ent->tag = 0UL; 77 ent->tag = 0UL;
82 membar_storeload_storestore(); 78 membar_storeload_storestore();
@@ -84,6 +80,83 @@ void flush_tsb_user(struct mmu_gather *mp)
84 } 80 }
85} 81}
86 82
83static void setup_tsb_params(struct mm_struct *mm, unsigned long tsb_bytes)
84{
85 unsigned long tsb_reg, base, tsb_paddr;
86 unsigned long page_sz, tte;
87
88 mm->context.tsb_nentries = tsb_bytes / sizeof(struct tsb);
89
90 base = TSBMAP_BASE;
91 tte = (_PAGE_VALID | _PAGE_L | _PAGE_CP |
92 _PAGE_CV | _PAGE_P | _PAGE_W);
93 tsb_paddr = __pa(mm->context.tsb);
94
95 /* Use the smallest page size that can map the whole TSB
96 * in one TLB entry.
97 */
98 switch (tsb_bytes) {
99 case 8192 << 0:
100 tsb_reg = 0x0UL;
101#ifdef DCACHE_ALIASING_POSSIBLE
102 base += (tsb_paddr & 8192);
103#endif
104 tte |= _PAGE_SZ8K;
105 page_sz = 8192;
106 break;
107
108 case 8192 << 1:
109 tsb_reg = 0x1UL;
110 tte |= _PAGE_SZ64K;
111 page_sz = 64 * 1024;
112 break;
113
114 case 8192 << 2:
115 tsb_reg = 0x2UL;
116 tte |= _PAGE_SZ64K;
117 page_sz = 64 * 1024;
118 break;
119
120 case 8192 << 3:
121 tsb_reg = 0x3UL;
122 tte |= _PAGE_SZ64K;
123 page_sz = 64 * 1024;
124 break;
125
126 case 8192 << 4:
127 tsb_reg = 0x4UL;
128 tte |= _PAGE_SZ512K;
129 page_sz = 512 * 1024;
130 break;
131
132 case 8192 << 5:
133 tsb_reg = 0x5UL;
134 tte |= _PAGE_SZ512K;
135 page_sz = 512 * 1024;
136 break;
137
138 case 8192 << 6:
139 tsb_reg = 0x6UL;
140 tte |= _PAGE_SZ512K;
141 page_sz = 512 * 1024;
142 break;
143
144 case 8192 << 7:
145 tsb_reg = 0x7UL;
146 tte |= _PAGE_SZ4MB;
147 page_sz = 4 * 1024 * 1024;
148 break;
149 };
150
151 tsb_reg |= base;
152 tsb_reg |= (tsb_paddr & (page_sz - 1UL));
153 tte |= (tsb_paddr & ~(page_sz - 1UL));
154
155 mm->context.tsb_reg_val = tsb_reg;
156 mm->context.tsb_map_vaddr = base;
157 mm->context.tsb_map_pte = tte;
158}
159
87int init_new_context(struct task_struct *tsk, struct mm_struct *mm) 160int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
88{ 161{
89 unsigned long page = get_zeroed_page(GFP_KERNEL); 162 unsigned long page = get_zeroed_page(GFP_KERNEL);
@@ -92,14 +165,22 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
92 if (unlikely(!page)) 165 if (unlikely(!page))
93 return -ENOMEM; 166 return -ENOMEM;
94 167
95 mm->context.sparc64_tsb = (unsigned long *) page; 168 mm->context.tsb = (struct tsb *) page;
169 setup_tsb_params(mm, PAGE_SIZE);
96 170
97 return 0; 171 return 0;
98} 172}
99 173
100void destroy_context(struct mm_struct *mm) 174void destroy_context(struct mm_struct *mm)
101{ 175{
102 free_page((unsigned long) mm->context.sparc64_tsb); 176 free_page((unsigned long) mm->context.tsb);
177
178 /* We can remove these later, but for now it's useful
179 * to catch any bogus post-destroy_context() references
180 * to the TSB.
181 */
182 mm->context.tsb = NULL;
183 mm->context.tsb_reg_val = 0UL;
103 184
104 spin_lock(&ctx_alloc_lock); 185 spin_lock(&ctx_alloc_lock);
105 186
diff --git a/include/asm-sparc64/mmu.h b/include/asm-sparc64/mmu.h
index 36384cf7faa6..2effeba2476c 100644
--- a/include/asm-sparc64/mmu.h
+++ b/include/asm-sparc64/mmu.h
@@ -90,9 +90,20 @@
90 90
91#ifndef __ASSEMBLY__ 91#ifndef __ASSEMBLY__
92 92
93#define TSB_ENTRY_ALIGNMENT 16
94
95struct tsb {
96 unsigned long tag;
97 unsigned long pte;
98} __attribute__((aligned(TSB_ENTRY_ALIGNMENT)));
99
93typedef struct { 100typedef struct {
94 unsigned long sparc64_ctx_val; 101 unsigned long sparc64_ctx_val;
95 unsigned long *sparc64_tsb; 102 struct tsb *tsb;
103 unsigned long tsb_nentries;
104 unsigned long tsb_reg_val;
105 unsigned long tsb_map_vaddr;
106 unsigned long tsb_map_pte;
96} mm_context_t; 107} mm_context_t;
97 108
98#endif /* !__ASSEMBLY__ */ 109#endif /* !__ASSEMBLY__ */
diff --git a/include/asm-sparc64/mmu_context.h b/include/asm-sparc64/mmu_context.h
index 0dffb4ce8a1d..0a950f151d2b 100644
--- a/include/asm-sparc64/mmu_context.h
+++ b/include/asm-sparc64/mmu_context.h
@@ -22,7 +22,15 @@ extern void get_new_mmu_context(struct mm_struct *mm);
22extern int init_new_context(struct task_struct *tsk, struct mm_struct *mm); 22extern int init_new_context(struct task_struct *tsk, struct mm_struct *mm);
23extern void destroy_context(struct mm_struct *mm); 23extern void destroy_context(struct mm_struct *mm);
24 24
25extern unsigned long tsb_context_switch(unsigned long pgd_pa, unsigned long *tsb); 25extern void __tsb_context_switch(unsigned long pgd_pa, unsigned long tsb_reg,
26 unsigned long tsb_vaddr, unsigned long tsb_pte);
27
28static inline void tsb_context_switch(struct mm_struct *mm)
29{
30 __tsb_context_switch(__pa(mm->pgd), mm->context.tsb_reg_val,
31 mm->context.tsb_map_vaddr,
32 mm->context.tsb_map_pte);
33}
26 34
27/* Set MMU context in the actual hardware. */ 35/* Set MMU context in the actual hardware. */
28#define load_secondary_context(__mm) \ 36#define load_secondary_context(__mm) \
@@ -52,8 +60,7 @@ static inline void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, str
52 60
53 if (!ctx_valid || (old_mm != mm)) { 61 if (!ctx_valid || (old_mm != mm)) {
54 load_secondary_context(mm); 62 load_secondary_context(mm);
55 tsb_context_switch(__pa(mm->pgd), 63 tsb_context_switch(mm);
56 mm->context.sparc64_tsb);
57 } 64 }
58 65
59 /* Even if (mm == old_mm) we _must_ check 66 /* Even if (mm == old_mm) we _must_ check
@@ -91,7 +98,7 @@ static inline void activate_mm(struct mm_struct *active_mm, struct mm_struct *mm
91 98
92 load_secondary_context(mm); 99 load_secondary_context(mm);
93 __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT); 100 __flush_tlb_mm(CTX_HWBITS(mm->context), SECONDARY_CONTEXT);
94 tsb_context_switch(__pa(mm->pgd), mm->context.sparc64_tsb); 101 tsb_context_switch(mm);
95} 102}
96 103
97#endif /* !(__ASSEMBLY__) */ 104#endif /* !(__ASSEMBLY__) */
diff --git a/include/asm-sparc64/tsb.h b/include/asm-sparc64/tsb.h
index 03d272e0e477..1f93b7d8cdbc 100644
--- a/include/asm-sparc64/tsb.h
+++ b/include/asm-sparc64/tsb.h
@@ -19,7 +19,7 @@
19 * stxa %g5, [%g0] ASI_{D,I}TLB_DATA_IN 19 * stxa %g5, [%g0] ASI_{D,I}TLB_DATA_IN
20 * retry 20 * retry
21 * 21 *
22 22 *
23 * Each 16-byte slot of the TSB is the 8-byte tag and then the 8-byte 23 * Each 16-byte slot of the TSB is the 8-byte tag and then the 8-byte
24 * PTE. The TAG is of the same layout as the TLB TAG TARGET mmu 24 * PTE. The TAG is of the same layout as the TLB TAG TARGET mmu
25 * register which is: 25 * register which is: