diff options
author | Xiantao Zhang <xiantao.zhang@intel.com> | 2008-04-03 14:02:58 -0400 |
---|---|---|
committer | Tony Luck <tony.luck@intel.com> | 2008-04-03 14:02:58 -0400 |
commit | 96651896b8d9ad4244a1c3ed9691faa3e382f503 (patch) | |
tree | d1982cf832e019deb8619fd0eae9133e5b98bc99 /arch/ia64/mm | |
parent | 9597362d354f8655ece324b01d0c640a0e99c077 (diff) |
[IA64] Add API for allocating Dynamic TR resource.
Dynamic TR resource should be managed in the uniform way.
Add two interfaces for kernel:
ia64_itr_entry: Allocate a (pair of) TR for caller.
ia64_ptr_entry: Purge a (pair of ) TR by caller.
Signed-off-by: Xiantao Zhang <xiantao.zhang@intel.com>
Signed-off-by: Anthony Xu <anthony.xu@intel.com>
Signed-off-by: Tony Luck <tony.luck@intel.com>
Diffstat (limited to 'arch/ia64/mm')
-rw-r--r-- | arch/ia64/mm/tlb.c | 196 |
1 files changed, 196 insertions, 0 deletions
diff --git a/arch/ia64/mm/tlb.c b/arch/ia64/mm/tlb.c index 655da240d13c..3d8903f936a5 100644 --- a/arch/ia64/mm/tlb.c +++ b/arch/ia64/mm/tlb.c | |||
@@ -26,6 +26,8 @@ | |||
26 | #include <asm/pal.h> | 26 | #include <asm/pal.h> |
27 | #include <asm/tlbflush.h> | 27 | #include <asm/tlbflush.h> |
28 | #include <asm/dma.h> | 28 | #include <asm/dma.h> |
29 | #include <asm/processor.h> | ||
30 | #include <asm/tlb.h> | ||
29 | 31 | ||
30 | static struct { | 32 | static struct { |
31 | unsigned long mask; /* mask of supported purge page-sizes */ | 33 | unsigned long mask; /* mask of supported purge page-sizes */ |
@@ -39,6 +41,10 @@ struct ia64_ctx ia64_ctx = { | |||
39 | }; | 41 | }; |
40 | 42 | ||
41 | DEFINE_PER_CPU(u8, ia64_need_tlb_flush); | 43 | DEFINE_PER_CPU(u8, ia64_need_tlb_flush); |
44 | DEFINE_PER_CPU(u8, ia64_tr_num); /*Number of TR slots in current processor*/ | ||
45 | DEFINE_PER_CPU(u8, ia64_tr_used); /*Max Slot number used by kernel*/ | ||
46 | |||
47 | struct ia64_tr_entry __per_cpu_idtrs[NR_CPUS][2][IA64_TR_ALLOC_MAX]; | ||
42 | 48 | ||
43 | /* | 49 | /* |
44 | * Initializes the ia64_ctx.bitmap array based on max_ctx+1. | 50 | * Initializes the ia64_ctx.bitmap array based on max_ctx+1. |
@@ -190,6 +196,9 @@ ia64_tlb_init (void) | |||
190 | ia64_ptce_info_t uninitialized_var(ptce_info); /* GCC be quiet */ | 196 | ia64_ptce_info_t uninitialized_var(ptce_info); /* GCC be quiet */ |
191 | unsigned long tr_pgbits; | 197 | unsigned long tr_pgbits; |
192 | long status; | 198 | long status; |
199 | pal_vm_info_1_u_t vm_info_1; | ||
200 | pal_vm_info_2_u_t vm_info_2; | ||
201 | int cpu = smp_processor_id(); | ||
193 | 202 | ||
194 | if ((status = ia64_pal_vm_page_size(&tr_pgbits, &purge.mask)) != 0) { | 203 | if ((status = ia64_pal_vm_page_size(&tr_pgbits, &purge.mask)) != 0) { |
195 | printk(KERN_ERR "PAL_VM_PAGE_SIZE failed with status=%ld; " | 204 | printk(KERN_ERR "PAL_VM_PAGE_SIZE failed with status=%ld; " |
@@ -206,4 +215,191 @@ ia64_tlb_init (void) | |||
206 | local_cpu_data->ptce_stride[1] = ptce_info.stride[1]; | 215 | local_cpu_data->ptce_stride[1] = ptce_info.stride[1]; |
207 | 216 | ||
208 | local_flush_tlb_all(); /* nuke left overs from bootstrapping... */ | 217 | local_flush_tlb_all(); /* nuke left overs from bootstrapping... */ |
218 | status = ia64_pal_vm_summary(&vm_info_1, &vm_info_2); | ||
219 | |||
220 | if (status) { | ||
221 | printk(KERN_ERR "ia64_pal_vm_summary=%ld\n", status); | ||
222 | per_cpu(ia64_tr_num, cpu) = 8; | ||
223 | return; | ||
224 | } | ||
225 | per_cpu(ia64_tr_num, cpu) = vm_info_1.pal_vm_info_1_s.max_itr_entry+1; | ||
226 | if (per_cpu(ia64_tr_num, cpu) > | ||
227 | (vm_info_1.pal_vm_info_1_s.max_dtr_entry+1)) | ||
228 | per_cpu(ia64_tr_num, cpu) = | ||
229 | vm_info_1.pal_vm_info_1_s.max_dtr_entry+1; | ||
230 | if (per_cpu(ia64_tr_num, cpu) > IA64_TR_ALLOC_MAX) { | ||
231 | per_cpu(ia64_tr_num, cpu) = IA64_TR_ALLOC_MAX; | ||
232 | printk(KERN_DEBUG "TR register number exceeds IA64_TR_ALLOC_MAX!" | ||
233 | "IA64_TR_ALLOC_MAX should be extended\n"); | ||
234 | } | ||
235 | } | ||
236 | |||
237 | /* | ||
238 | * is_tr_overlap | ||
239 | * | ||
240 | * Check overlap with inserted TRs. | ||
241 | */ | ||
242 | static int is_tr_overlap(struct ia64_tr_entry *p, u64 va, u64 log_size) | ||
243 | { | ||
244 | u64 tr_log_size; | ||
245 | u64 tr_end; | ||
246 | u64 va_rr = ia64_get_rr(va); | ||
247 | u64 va_rid = RR_TO_RID(va_rr); | ||
248 | u64 va_end = va + (1<<log_size) - 1; | ||
249 | |||
250 | if (va_rid != RR_TO_RID(p->rr)) | ||
251 | return 0; | ||
252 | tr_log_size = (p->itir & 0xff) >> 2; | ||
253 | tr_end = p->ifa + (1<<tr_log_size) - 1; | ||
254 | |||
255 | if (va > tr_end || p->ifa > va_end) | ||
256 | return 0; | ||
257 | return 1; | ||
258 | |||
259 | } | ||
260 | |||
261 | /* | ||
262 | * ia64_insert_tr in virtual mode. Allocate a TR slot | ||
263 | * | ||
264 | * target_mask : 0x1 : itr, 0x2 : dtr, 0x3 : idtr | ||
265 | * | ||
266 | * va : virtual address. | ||
267 | * pte : pte entries inserted. | ||
268 | * log_size: range to be covered. | ||
269 | * | ||
270 | * Return value: <0 : error No. | ||
271 | * | ||
272 | * >=0 : slot number allocated for TR. | ||
273 | * Must be called with preemption disabled. | ||
274 | */ | ||
275 | int ia64_itr_entry(u64 target_mask, u64 va, u64 pte, u64 log_size) | ||
276 | { | ||
277 | int i, r; | ||
278 | unsigned long psr; | ||
279 | struct ia64_tr_entry *p; | ||
280 | int cpu = smp_processor_id(); | ||
281 | |||
282 | r = -EINVAL; | ||
283 | /*Check overlap with existing TR entries*/ | ||
284 | if (target_mask & 0x1) { | ||
285 | p = &__per_cpu_idtrs[cpu][0][0]; | ||
286 | for (i = IA64_TR_ALLOC_BASE; i <= per_cpu(ia64_tr_used, cpu); | ||
287 | i++, p++) { | ||
288 | if (p->pte & 0x1) | ||
289 | if (is_tr_overlap(p, va, log_size)) { | ||
290 | printk(KERN_DEBUG "Overlapped Entry" | ||
291 | "Inserted for TR Reigster!!\n"); | ||
292 | goto out; | ||
293 | } | ||
294 | } | ||
295 | } | ||
296 | if (target_mask & 0x2) { | ||
297 | p = &__per_cpu_idtrs[cpu][1][0]; | ||
298 | for (i = IA64_TR_ALLOC_BASE; i <= per_cpu(ia64_tr_used, cpu); | ||
299 | i++, p++) { | ||
300 | if (p->pte & 0x1) | ||
301 | if (is_tr_overlap(p, va, log_size)) { | ||
302 | printk(KERN_DEBUG "Overlapped Entry" | ||
303 | "Inserted for TR Reigster!!\n"); | ||
304 | goto out; | ||
305 | } | ||
306 | } | ||
307 | } | ||
308 | |||
309 | for (i = IA64_TR_ALLOC_BASE; i < per_cpu(ia64_tr_num, cpu); i++) { | ||
310 | switch (target_mask & 0x3) { | ||
311 | case 1: | ||
312 | if (!(__per_cpu_idtrs[cpu][0][i].pte & 0x1)) | ||
313 | goto found; | ||
314 | continue; | ||
315 | case 2: | ||
316 | if (!(__per_cpu_idtrs[cpu][1][i].pte & 0x1)) | ||
317 | goto found; | ||
318 | continue; | ||
319 | case 3: | ||
320 | if (!(__per_cpu_idtrs[cpu][0][i].pte & 0x1) && | ||
321 | !(__per_cpu_idtrs[cpu][1][i].pte & 0x1)) | ||
322 | goto found; | ||
323 | continue; | ||
324 | default: | ||
325 | r = -EINVAL; | ||
326 | goto out; | ||
327 | } | ||
328 | } | ||
329 | found: | ||
330 | if (i >= per_cpu(ia64_tr_num, cpu)) | ||
331 | return -EBUSY; | ||
332 | |||
333 | /*Record tr info for mca hander use!*/ | ||
334 | if (i > per_cpu(ia64_tr_used, cpu)) | ||
335 | per_cpu(ia64_tr_used, cpu) = i; | ||
336 | |||
337 | psr = ia64_clear_ic(); | ||
338 | if (target_mask & 0x1) { | ||
339 | ia64_itr(0x1, i, va, pte, log_size); | ||
340 | ia64_srlz_i(); | ||
341 | p = &__per_cpu_idtrs[cpu][0][i]; | ||
342 | p->ifa = va; | ||
343 | p->pte = pte; | ||
344 | p->itir = log_size << 2; | ||
345 | p->rr = ia64_get_rr(va); | ||
346 | } | ||
347 | if (target_mask & 0x2) { | ||
348 | ia64_itr(0x2, i, va, pte, log_size); | ||
349 | ia64_srlz_i(); | ||
350 | p = &__per_cpu_idtrs[cpu][1][i]; | ||
351 | p->ifa = va; | ||
352 | p->pte = pte; | ||
353 | p->itir = log_size << 2; | ||
354 | p->rr = ia64_get_rr(va); | ||
355 | } | ||
356 | ia64_set_psr(psr); | ||
357 | r = i; | ||
358 | out: | ||
359 | return r; | ||
360 | } | ||
361 | EXPORT_SYMBOL_GPL(ia64_itr_entry); | ||
362 | |||
363 | /* | ||
364 | * ia64_purge_tr | ||
365 | * | ||
366 | * target_mask: 0x1: purge itr, 0x2 : purge dtr, 0x3 purge idtr. | ||
367 | * slot: slot number to be freed. | ||
368 | * | ||
369 | * Must be called with preemption disabled. | ||
370 | */ | ||
371 | void ia64_ptr_entry(u64 target_mask, int slot) | ||
372 | { | ||
373 | int cpu = smp_processor_id(); | ||
374 | int i; | ||
375 | struct ia64_tr_entry *p; | ||
376 | |||
377 | if (slot < IA64_TR_ALLOC_BASE || slot >= per_cpu(ia64_tr_num, cpu)) | ||
378 | return; | ||
379 | |||
380 | if (target_mask & 0x1) { | ||
381 | p = &__per_cpu_idtrs[cpu][0][slot]; | ||
382 | if ((p->pte&0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) { | ||
383 | p->pte = 0; | ||
384 | ia64_ptr(0x1, p->ifa, p->itir>>2); | ||
385 | ia64_srlz_i(); | ||
386 | } | ||
387 | } | ||
388 | |||
389 | if (target_mask & 0x2) { | ||
390 | p = &__per_cpu_idtrs[cpu][1][slot]; | ||
391 | if ((p->pte & 0x1) && is_tr_overlap(p, p->ifa, p->itir>>2)) { | ||
392 | p->pte = 0; | ||
393 | ia64_ptr(0x2, p->ifa, p->itir>>2); | ||
394 | ia64_srlz_i(); | ||
395 | } | ||
396 | } | ||
397 | |||
398 | for (i = per_cpu(ia64_tr_used, cpu); i >= IA64_TR_ALLOC_BASE; i--) { | ||
399 | if ((__per_cpu_idtrs[cpu][0][i].pte & 0x1) || | ||
400 | (__per_cpu_idtrs[cpu][1][i].pte & 0x1)) | ||
401 | break; | ||
402 | } | ||
403 | per_cpu(ia64_tr_used, cpu) = i; | ||
209 | } | 404 | } |
405 | EXPORT_SYMBOL_GPL(ia64_ptr_entry); | ||