aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/include/asm/mmu-44x.h
diff options
context:
space:
mode:
authorBenjamin Herrenschmidt <benh@kernel.crashing.org>2008-12-18 14:13:29 -0500
committerPaul Mackerras <paulus@samba.org>2008-12-20 22:21:15 -0500
commit2ca8cf738907180e7fbda90f25f32b86feda609f (patch)
tree60d8af9b53a78ae9300ef7e68f222b02fe3be542 /arch/powerpc/include/asm/mmu-44x.h
parent5e696617c425eb97bd943d781f3941fb1e8f0e5b (diff)
powerpc/mm: Rework context management for CPUs with no hash table
This reworks the context management code used by 4xx,8xx and freescale BookE. It adds support for SMP by implementing a concept of stale context map to lazily flush the TLB on processors where a context may have been invalidated. This also contains the ground work for generalizing such lazy TLB flushing by just picking up a new PID and marking the old one stale. This will be implemented later. This is a first implementation that uses a global spinlock. Ideally, we should try to get at least the fast path (context ID already assigned) lockless or limited to a per context lock, but for now this will do. I tried to keep the UP case reasonably simple to avoid adding too much overhead to 8xx which does a lot of context stealing since it effectively has only 16 PIDs available. Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org> Acked-by: Kumar Gala <galak@kernel.crashing.org> Signed-off-by: Paul Mackerras <paulus@samba.org>
Diffstat (limited to 'arch/powerpc/include/asm/mmu-44x.h')
-rw-r--r--arch/powerpc/include/asm/mmu-44x.h5
1 files changed, 3 insertions, 2 deletions
diff --git a/arch/powerpc/include/asm/mmu-44x.h b/arch/powerpc/include/asm/mmu-44x.h
index a825524c981a..b21af32ac6d6 100644
--- a/arch/powerpc/include/asm/mmu-44x.h
+++ b/arch/powerpc/include/asm/mmu-44x.h
@@ -56,8 +56,9 @@
56extern unsigned int tlb_44x_hwater; 56extern unsigned int tlb_44x_hwater;
57 57
58typedef struct { 58typedef struct {
59 unsigned long id; 59 unsigned int id;
60 unsigned long vdso_base; 60 unsigned int active;
61 unsigned long vdso_base;
61} mm_context_t; 62} mm_context_t;
62 63
63#endif /* !__ASSEMBLY__ */ 64#endif /* !__ASSEMBLY__ */