aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/kernel/misc_32.S
diff options
context:
space:
mode:
authorBenjamin Herrenschmidt <benh@kernel.crashing.org>2008-12-18 14:13:42 -0500
committerPaul Mackerras <paulus@samba.org>2008-12-20 22:21:16 -0500
commit2a4aca1144394653269720ffbb5a325a77abd5fa (patch)
tree553bbcbb294ac5923f72430b7317b5c80a27141c /arch/powerpc/kernel/misc_32.S
parentf048aace29e007f2b642097e2da8231e0e9cce2d (diff)
powerpc/mm: Split low level tlb invalidate for nohash processors
Currently, the various forms of low level TLB invalidations are all implemented in misc_32.S for 32-bit processors, in a fairly scary mess of #ifdef's and with interesting duplication such as a whole bunch of code for FSL _tlbie and _tlbia which are no longer used. This moves things around such that _tlbie is now defined in hash_low_32.S and is only used by the 32-bit hash code, and all nohash CPUs use the various _tlbil_* forms that are now moved to a new file, tlb_nohash_low.S. I moved all the definitions for that stuff out of include/asm/tlbflush.h as they are really internal mm stuff, into mm/mmu_decl.h The code should have no functional changes. I kept some variants inline for trivial forms on things like 40x and 8xx. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Acked-by: Kumar Gala <galak@kernel.crashing.org> Signed-off-by: Paul Mackerras <paulus@samba.org>
Diffstat (limited to 'arch/powerpc/kernel/misc_32.S')
-rw-r--r--arch/powerpc/kernel/misc_32.S233
1 files changed, 0 insertions, 233 deletions
diff --git a/arch/powerpc/kernel/misc_32.S b/arch/powerpc/kernel/misc_32.S
index 2c2ab89f0b64..ae0d084b6a24 100644
--- a/arch/powerpc/kernel/misc_32.S
+++ b/arch/powerpc/kernel/misc_32.S
@@ -272,239 +272,6 @@ _GLOBAL(real_writeb)
272 272
273#endif /* CONFIG_40x */ 273#endif /* CONFIG_40x */
274 274
275/*
276 * Flush MMU TLB
277 */
278#ifndef CONFIG_FSL_BOOKE
279_GLOBAL(_tlbil_all)
280_GLOBAL(_tlbil_pid)
281#endif
282_GLOBAL(_tlbia)
283#if defined(CONFIG_40x)
284 sync /* Flush to memory before changing mapping */
285 tlbia
286 isync /* Flush shadow TLB */
287#elif defined(CONFIG_44x)
288 li r3,0
289 sync
290
291 /* Load high watermark */
292 lis r4,tlb_44x_hwater@ha
293 lwz r5,tlb_44x_hwater@l(r4)
294
2951: tlbwe r3,r3,PPC44x_TLB_PAGEID
296 addi r3,r3,1
297 cmpw 0,r3,r5
298 ble 1b
299
300 isync
301#elif defined(CONFIG_FSL_BOOKE)
302 /* Invalidate all entries in TLB0 */
303 li r3, 0x04
304 tlbivax 0,3
305 /* Invalidate all entries in TLB1 */
306 li r3, 0x0c
307 tlbivax 0,3
308 msync
309#ifdef CONFIG_SMP
310 tlbsync
311#endif /* CONFIG_SMP */
312#else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */
313#if defined(CONFIG_SMP)
314 rlwinm r8,r1,0,0,(31-THREAD_SHIFT)
315 lwz r8,TI_CPU(r8)
316 oris r8,r8,10
317 mfmsr r10
318 SYNC
319 rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */
320 rlwinm r0,r0,0,28,26 /* clear DR */
321 mtmsr r0
322 SYNC_601
323 isync
324 lis r9,mmu_hash_lock@h
325 ori r9,r9,mmu_hash_lock@l
326 tophys(r9,r9)
32710: lwarx r7,0,r9
328 cmpwi 0,r7,0
329 bne- 10b
330 stwcx. r8,0,r9
331 bne- 10b
332 sync
333 tlbia
334 sync
335 TLBSYNC
336 li r0,0
337 stw r0,0(r9) /* clear mmu_hash_lock */
338 mtmsr r10
339 SYNC_601
340 isync
341#else /* CONFIG_SMP */
342 sync
343 tlbia
344 sync
345#endif /* CONFIG_SMP */
346#endif /* ! defined(CONFIG_40x) */
347 blr
348
349/*
350 * Flush MMU TLB for a particular address
351 */
352#ifndef CONFIG_FSL_BOOKE
353_GLOBAL(_tlbil_va)
354#endif
355_GLOBAL(_tlbie)
356#if defined(CONFIG_40x)
357 /* We run the search with interrupts disabled because we have to change
358 * the PID and I don't want to preempt when that happens.
359 */
360 mfmsr r5
361 mfspr r6,SPRN_PID
362 wrteei 0
363 mtspr SPRN_PID,r4
364 tlbsx. r3, 0, r3
365 mtspr SPRN_PID,r6
366 wrtee r5
367 bne 10f
368 sync
369 /* There are only 64 TLB entries, so r3 < 64, which means bit 25 is clear.
370 * Since 25 is the V bit in the TLB_TAG, loading this value will invalidate
371 * the TLB entry. */
372 tlbwe r3, r3, TLB_TAG
373 isync
37410:
375
376#elif defined(CONFIG_44x)
377 mfspr r5,SPRN_MMUCR
378 rlwimi r5,r4,0,24,31 /* Set TID */
379
380 /* We have to run the search with interrupts disabled, even critical
381 * and debug interrupts (in fact the only critical exceptions we have
382 * are debug and machine check). Otherwise an interrupt which causes
383 * a TLB miss can clobber the MMUCR between the mtspr and the tlbsx. */
384 mfmsr r4
385 lis r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@ha
386 addi r6,r6,(MSR_EE|MSR_CE|MSR_ME|MSR_DE)@l
387 andc r6,r4,r6
388 mtmsr r6
389 mtspr SPRN_MMUCR,r5
390 tlbsx. r3, 0, r3
391 mtmsr r4
392 bne 10f
393 sync
394 /* There are only 64 TLB entries, so r3 < 64,
395 * which means bit 22, is clear. Since 22 is
396 * the V bit in the TLB_PAGEID, loading this
397 * value will invalidate the TLB entry.
398 */
399 tlbwe r3, r3, PPC44x_TLB_PAGEID
400 isync
40110:
402#elif defined(CONFIG_FSL_BOOKE)
403 rlwinm r4, r3, 0, 0, 19
404 ori r5, r4, 0x08 /* TLBSEL = 1 */
405 tlbivax 0, r4
406 tlbivax 0, r5
407 msync
408#if defined(CONFIG_SMP)
409 tlbsync
410#endif /* CONFIG_SMP */
411#else /* !(CONFIG_40x || CONFIG_44x || CONFIG_FSL_BOOKE) */
412#if defined(CONFIG_SMP)
413 rlwinm r8,r1,0,0,(31-THREAD_SHIFT)
414 lwz r8,TI_CPU(r8)
415 oris r8,r8,11
416 mfmsr r10
417 SYNC
418 rlwinm r0,r10,0,17,15 /* clear bit 16 (MSR_EE) */
419 rlwinm r0,r0,0,28,26 /* clear DR */
420 mtmsr r0
421 SYNC_601
422 isync
423 lis r9,mmu_hash_lock@h
424 ori r9,r9,mmu_hash_lock@l
425 tophys(r9,r9)
42610: lwarx r7,0,r9
427 cmpwi 0,r7,0
428 bne- 10b
429 stwcx. r8,0,r9
430 bne- 10b
431 eieio
432 tlbie r3
433 sync
434 TLBSYNC
435 li r0,0
436 stw r0,0(r9) /* clear mmu_hash_lock */
437 mtmsr r10
438 SYNC_601
439 isync
440#else /* CONFIG_SMP */
441 tlbie r3
442 sync
443#endif /* CONFIG_SMP */
444#endif /* ! CONFIG_40x */
445 blr
446
447#if defined(CONFIG_FSL_BOOKE)
448/*
449 * Flush MMU TLB, but only on the local processor (no broadcast)
450 */
451_GLOBAL(_tlbil_all)
452#define MMUCSR0_TLBFI (MMUCSR0_TLB0FI | MMUCSR0_TLB1FI | \
453 MMUCSR0_TLB2FI | MMUCSR0_TLB3FI)
454 li r3,(MMUCSR0_TLBFI)@l
455 mtspr SPRN_MMUCSR0, r3
4561:
457 mfspr r3,SPRN_MMUCSR0
458 andi. r3,r3,MMUCSR0_TLBFI@l
459 bne 1b
460 blr
461
462/*
463 * Flush MMU TLB for a particular process id, but only on the local processor
464 * (no broadcast)
465 */
466_GLOBAL(_tlbil_pid)
467/* we currently do an invalidate all since we don't have per pid invalidate */
468 li r3,(MMUCSR0_TLBFI)@l
469 mtspr SPRN_MMUCSR0, r3
4701:
471 mfspr r3,SPRN_MMUCSR0
472 andi. r3,r3,MMUCSR0_TLBFI@l
473 bne 1b
474 msync
475 isync
476 blr
477
478/*
479 * Flush MMU TLB for a particular address, but only on the local processor
480 * (no broadcast)
481 */
482_GLOBAL(_tlbil_va)
483 mfmsr r10
484 wrteei 0
485 slwi r4,r4,16
486 mtspr SPRN_MAS6,r4 /* assume AS=0 for now */
487 tlbsx 0,r3
488 mfspr r4,SPRN_MAS1 /* check valid */
489 andis. r3,r4,MAS1_VALID@h
490 beq 1f
491 rlwinm r4,r4,0,1,31
492 mtspr SPRN_MAS1,r4
493 tlbwe
494 msync
495 isync
4961: wrtee r10
497 blr
498#endif /* CONFIG_FSL_BOOKE */
499
500/*
501 * Nobody implements this yet
502 */
503_GLOBAL(_tlbivax_bcast)
5041: trap
505 EMIT_BUG_ENTRY 1b,__FILE__,__LINE__,0;
506 blr
507
508 275
509/* 276/*
510 * Flush instruction cache. 277 * Flush instruction cache.