diff options
author | Benjamin Herrenschmidt <benh@kernel.crashing.org> | 2008-12-18 14:13:42 -0500 |
---|---|---|
committer | Paul Mackerras <paulus@samba.org> | 2008-12-20 22:21:16 -0500 |
commit | 2a4aca1144394653269720ffbb5a325a77abd5fa (patch) | |
tree | 553bbcbb294ac5923f72430b7317b5c80a27141c /arch/powerpc/kernel/misc_32.S | |
parent | f048aace29e007f2b642097e2da8231e0e9cce2d (diff) |
powerpc/mm: Split low level tlb invalidate for nohash processors
Currently, the various forms of low level TLB invalidations are all
implemented in misc_32.S for 32-bit processors, in a fairly scary
mess of #ifdef's and with interesting duplication such as a whole
bunch of code for FSL _tlbie and _tlbia which are no longer used.
This moves things around such that _tlbie is now defined in
hash_low_32.S and is only used by the 32-bit hash code, and all
nohash CPUs use the various _tlbil_* forms that are now moved to
a new file, tlb_nohash_low.S.
I moved all the definitions for that stuff out of
include/asm/tlbflush.h as they are really internal mm stuff, into
mm/mmu_decl.h
The code should have no functional changes. I kept some variants
inline for trivial forms on things like 40x and 8xx.
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Acked-by: Kumar Gala <galak@kernel.crashing.org>
Signed-off-by: Paul Mackerras <paulus@samba.org>
Diffstat (limited to 'arch/powerpc/kernel/misc_32.S')
-rw-r--r-- | arch/powerpc/kernel/misc_32.S | 233 |
1 files changed, 0 insertions, 233 deletions
diff --git a/arch/powerpc/kernel/misc_32.S b/arch/powerpc/kernel/misc_32.S index 2c2ab89f0b64..ae0d084b6a24 100644 --- a/arch/powerpc/kernel/misc_32.S +++ b/arch/powerpc/kernel/misc_32.S | |||
@@ -272,239 +272,6 @@ _GLOBAL(real_writeb) | |||
272 | 272 | ||
273 | #endif /* CONFIG_40x */ | 273 | #endif /* CONFIG_40x */ |
274 | 274 | ||
275 | /* | ||
276 | * Flush MMU TLB | ||
277 | */ | ||
278 | #ifndef CONFIG_FSL_BOOKE | ||
279 | _GLOBAL(_tlbil_all) | ||
280 | _GLOBAL(_tlbil_pid) | ||
281 | #endif | ||
282 | _GLOBAL(_tlbia) | ||
283 | #if defined(CONFIG_40x) | ||
284 | sync /* Flush to memory before changing mapping */ | ||
285 | tlbia | ||
286 | isync /* Flush shadow TLB */ | ||
287 | #elif defined(CONFIG_44x) | ||
288 | li r3,0 | ||
289 | sync | ||
290 | |||
291 | /* Load high watermark */ | ||
292 | lis r4,tlb_44x_hwater@ha | ||
293 | lwz r5,tlb_44x_hwater@l(r4) | ||
294 | |||
295 | 1: tlbwe r3,r3,PPC44x_TLB_PAGEID | ||
296 | addi r3,r3,1 | ||
297 | cmpw 0,r3,r5 | ||
298 | ble 1b | ||
299 | |||
300 | isync | ||
301 | #elif defined(CONFIG_FSL_BOOKE) | ||
302 | /* Invalidate all entries in TLB0 */ | ||
303 | li r3, 0x04 | ||
304 | tlbivax 0,3 | ||
305 | /* Invalidate all entries in TLB1 */ | ||
306 | li r3, 0x0c | ||
307 | tlbivax 0,3 | ||
308 | msync | ||
309 | #ifdef CONFIG_SMP | ||
310 | tlbsync | ||
311 | #endif /* CONFIG_SMP */ | ||
312 | #else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */ | ||
313 | #if defined(CONFIG_SMP) | ||
314 | rlwinm r8,r1,0,0,(31-THREAD_SHIFT) | ||
315 | lwz r8,TI_CPU(r8) | ||
316 | oris r8,r8,10 | ||
317 | mfmsr r10 | ||
318 | SYNC | ||
319 | rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */ | ||
320 | rlwinm r0,r0,0,28,26 /* clear DR */ | ||
321 | mtmsr r0 | ||
322 | SYNC_601 | ||
323 | isync | ||
324 | lis r9,mmu_hash_lock@h | ||
325 | ori r9,r9,mmu_hash_lock@l | ||
326 | tophys(r9,r9) | ||
327 | 10: lwarx r7,0,r9 | ||
328 | cmpwi 0,r7,0 | ||
329 | bne- 10b | ||
330 | stwcx. r8,0,r9 | ||
331 | bne- 10b | ||
332 | sync | ||
333 | tlbia | ||
334 | sync | ||
335 | TLBSYNC | ||
336 | li r0,0 | ||
337 | stw r0,0(r9) /* clear mmu_hash_lock */ | ||
338 | mtmsr r10 | ||
339 | SYNC_601 | ||
340 | isync | ||
341 | #else /* CONFIG_SMP */ | ||
342 | sync | ||
343 | tlbia | ||
344 | sync | ||
345 | #endif /* CONFIG_SMP */ | ||
346 | #endif /* ! defined(CONFIG_40x) */ | ||
347 | blr | ||
348 | |||
349 | /* | ||
350 | * Flush MMU TLB for a particular address | ||
351 | */ | ||
352 | #ifndef CONFIG_FSL_BOOKE | ||
353 | _GLOBAL(_tlbil_va) | ||
354 | #endif | ||
355 | _GLOBAL(_tlbie) | ||
356 | #if defined(CONFIG_40x) | ||
357 | /* We run the search with interrupts disabled because we have to change | ||
358 | * the PID and I don't want to preempt when that happens. | ||
359 | */ | ||
360 | mfmsr r5 | ||
361 | mfspr r6,SPRN_PID | ||
362 | wrteei 0 | ||
363 | mtspr SPRN_PID,r4 | ||
364 | tlbsx. r3, 0, r3 | ||
365 | mtspr SPRN_PID,r6 | ||
366 | wrtee r5 | ||
367 | bne 10f | ||
368 | sync | ||
369 | /* There are only 64 TLB entries, so r3 < 64, which means bit 25 is clear. | ||
370 | * Since 25 is the V bit in the TLB_TAG, loading this value will invalidate | ||
371 | * the TLB entry. */ | ||
372 | tlbwe r3, r3, TLB_TAG | ||
373 | isync | ||
374 | 10: | ||
375 | |||
376 | #elif defined(CONFIG_44x) | ||
377 | mfspr r5,SPRN_MMUCR | ||
378 | rlwimi r5,r4,0,24,31 /* Set TID */ | ||
379 | |||
380 | /* We have to run the search with interrupts disabled, even critical | ||
381 | * and debug interrupts (in fact the only critical exceptions we have | ||
382 | * are debug and machine check). Otherwise an interrupt which causes | ||
383 | * a TLB miss can clobber the MMUCR between the mtspr and the tlbsx. */ | ||
384 | mfmsr r4 | ||
385 | lis r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@ha | ||
386 | addi r6,r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@l | ||
387 | andc r6,r4,r6 | ||
388 | mtmsr r6 | ||
389 | mtspr SPRN_MMUCR,r5 | ||
390 | tlbsx. r3, 0, r3 | ||
391 | mtmsr r4 | ||
392 | bne 10f | ||
393 | sync | ||
394 | /* There are only 64 TLB entries, so r3 < 64, | ||
395 | * which means bit 22, is clear. Since 22 is | ||
396 | * the V bit in the TLB_PAGEID, loading this | ||
397 | * value will invalidate the TLB entry. | ||
398 | */ | ||
399 | tlbwe r3, r3, PPC44x_TLB_PAGEID | ||
400 | isync | ||
401 | 10: | ||
402 | #elif defined(CONFIG_FSL_BOOKE) | ||
403 | rlwinm r4, r3, 0, 0, 19 | ||
404 | ori r5, r4, 0x08 /* TLBSEL = 1 */ | ||
405 | tlbivax 0, r4 | ||
406 | tlbivax 0, r5 | ||
407 | msync | ||
408 | #if defined(CONFIG_SMP) | ||
409 | tlbsync | ||
410 | #endif /* CONFIG_SMP */ | ||
411 | #else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */ | ||
412 | #if defined(CONFIG_SMP) | ||
413 | rlwinm r8,r1,0,0,(31-THREAD_SHIFT) | ||
414 | lwz r8,TI_CPU(r8) | ||
415 | oris r8,r8,11 | ||
416 | mfmsr r10 | ||
417 | SYNC | ||
418 | rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */ | ||
419 | rlwinm r0,r0,0,28,26 /* clear DR */ | ||
420 | mtmsr r0 | ||
421 | SYNC_601 | ||
422 | isync | ||
423 | lis r9,mmu_hash_lock@h | ||
424 | ori r9,r9,mmu_hash_lock@l | ||
425 | tophys(r9,r9) | ||
426 | 10: lwarx r7,0,r9 | ||
427 | cmpwi 0,r7,0 | ||
428 | bne- 10b | ||
429 | stwcx. r8,0,r9 | ||
430 | bne- 10b | ||
431 | eieio | ||
432 | tlbie r3 | ||
433 | sync | ||
434 | TLBSYNC | ||
435 | li r0,0 | ||
436 | stw r0,0(r9) /* clear mmu_hash_lock */ | ||
437 | mtmsr r10 | ||
438 | SYNC_601 | ||
439 | isync | ||
440 | #else /* CONFIG_SMP */ | ||
441 | tlbie r3 | ||
442 | sync | ||
443 | #endif /* CONFIG_SMP */ | ||
444 | #endif /* ! CONFIG_40x */ | ||
445 | blr | ||
446 | |||
447 | #if defined(CONFIG_FSL_BOOKE) | ||
448 | /* | ||
449 | * Flush MMU TLB, but only on the local processor (no broadcast) | ||
450 | */ | ||
451 | _GLOBAL(_tlbil_all) | ||
452 | #define MMUCSR0_TLBFI (MMUCSR0_TLB0FI | MMUCSR0_TLB1FI | \ | ||
453 | MMUCSR0_TLB2FI | MMUCSR0_TLB3FI) | ||
454 | li r3,(MMUCSR0_TLBFI)@l | ||
455 | mtspr SPRN_MMUCSR0, r3 | ||
456 | 1: | ||
457 | mfspr r3,SPRN_MMUCSR0 | ||
458 | andi. r3,r3,MMUCSR0_TLBFI@l | ||
459 | bne 1b | ||
460 | blr | ||
461 | |||
462 | /* | ||
463 | * Flush MMU TLB for a particular process id, but only on the local processor | ||
464 | * (no broadcast) | ||
465 | */ | ||
466 | _GLOBAL(_tlbil_pid) | ||
467 | /* we currently do an invalidate all since we don't have per pid invalidate */ | ||
468 | li r3,(MMUCSR0_TLBFI)@l | ||
469 | mtspr SPRN_MMUCSR0, r3 | ||
470 | 1: | ||
471 | mfspr r3,SPRN_MMUCSR0 | ||
472 | andi. r3,r3,MMUCSR0_TLBFI@l | ||
473 | bne 1b | ||
474 | msync | ||
475 | isync | ||
476 | blr | ||
477 | |||
478 | /* | ||
479 | * Flush MMU TLB for a particular address, but only on the local processor | ||
480 | * (no broadcast) | ||
481 | */ | ||
482 | _GLOBAL(_tlbil_va) | ||
483 | mfmsr r10 | ||
484 | wrteei 0 | ||
485 | slwi r4,r4,16 | ||
486 | mtspr SPRN_MAS6,r4 /* assume AS=0 for now */ | ||
487 | tlbsx 0,r3 | ||
488 | mfspr r4,SPRN_MAS1 /* check valid */ | ||
489 | andis. r3,r4,MAS1_VALID@h | ||
490 | beq 1f | ||
491 | rlwinm r4,r4,0,1,31 | ||
492 | mtspr SPRN_MAS1,r4 | ||
493 | tlbwe | ||
494 | msync | ||
495 | isync | ||
496 | 1: wrtee r10 | ||
497 | blr | ||
498 | #endif /* CONFIG_FSL_BOOKE */ | ||
499 | |||
500 | /* | ||
501 | * Nobody implements this yet | ||
502 | */ | ||
503 | _GLOBAL(_tlbivax_bcast) | ||
504 | 1: trap | ||
505 | EMIT_BUG_ENTRY 1b,__FILE__,__LINE__,0; | ||
506 | blr | ||
507 | |||
508 | 275 | ||
509 | /* | 276 | /* |
510 | * Flush instruction cache. | 277 | * Flush instruction cache. |