diff options
author | Akira Takeuchi <takeuchi.akr@jp.panasonic.com> | 2010-10-27 12:28:49 -0400 |
---|---|---|
committer | David Howells <dhowells@redhat.com> | 2010-10-27 12:28:49 -0400 |
commit | a9bc60ebfd5766ce5f6095d0fed3d9978990122f (patch) | |
tree | d6044bbb56bbb06fb6f13fab9f079a20938d0960 /arch | |
parent | 492e675116003b99dfcf0fa70084027e86bc0161 (diff) |
MN10300: Make the use of PIDR to mark TLB entries controllable
Make controllable the use of the PIDR register to mark TLB entries as belonging
to particular processes.
Signed-off-by: Akira Takeuchi <takeuchi.akr@jp.panasonic.com>
Signed-off-by: Kiyoshi Owada <owada.kiyoshi@jp.panasonic.com>
Signed-off-by: David Howells <dhowells@redhat.com>
Diffstat (limited to 'arch')
-rw-r--r-- | arch/mn10300/Kconfig | 3 | ||||
-rw-r--r-- | arch/mn10300/include/asm/mmu_context.h | 59 | ||||
-rw-r--r-- | arch/mn10300/include/asm/tlbflush.h | 43 | ||||
-rw-r--r-- | arch/mn10300/mm/mmu-context.c | 41 |
4 files changed, 84 insertions, 62 deletions
diff --git a/arch/mn10300/Kconfig b/arch/mn10300/Kconfig index dd7b5700358b..7bd920b1c06f 100644 --- a/arch/mn10300/Kconfig +++ b/arch/mn10300/Kconfig | |||
@@ -142,6 +142,9 @@ config FPU | |||
142 | 142 | ||
143 | source "arch/mn10300/mm/Kconfig.cache" | 143 | source "arch/mn10300/mm/Kconfig.cache" |
144 | 144 | ||
145 | config MN10300_TLB_USE_PIDR | ||
146 | def_bool y | ||
147 | |||
145 | menu "Memory layout options" | 148 | menu "Memory layout options" |
146 | 149 | ||
147 | config KERNEL_RAM_BASE_ADDRESS | 150 | config KERNEL_RAM_BASE_ADDRESS |
diff --git a/arch/mn10300/include/asm/mmu_context.h b/arch/mn10300/include/asm/mmu_context.h index 24d63f0f7377..5fb3648968ae 100644 --- a/arch/mn10300/include/asm/mmu_context.h +++ b/arch/mn10300/include/asm/mmu_context.h | |||
@@ -27,28 +27,22 @@ | |||
27 | #include <asm/tlbflush.h> | 27 | #include <asm/tlbflush.h> |
28 | #include <asm-generic/mm_hooks.h> | 28 | #include <asm-generic/mm_hooks.h> |
29 | 29 | ||
30 | #define MMU_CONTEXT_TLBPID_NR 256 | ||
30 | #define MMU_CONTEXT_TLBPID_MASK 0x000000ffUL | 31 | #define MMU_CONTEXT_TLBPID_MASK 0x000000ffUL |
31 | #define MMU_CONTEXT_VERSION_MASK 0xffffff00UL | 32 | #define MMU_CONTEXT_VERSION_MASK 0xffffff00UL |
32 | #define MMU_CONTEXT_FIRST_VERSION 0x00000100UL | 33 | #define MMU_CONTEXT_FIRST_VERSION 0x00000100UL |
33 | #define MMU_NO_CONTEXT 0x00000000UL | 34 | #define MMU_NO_CONTEXT 0x00000000UL |
34 | 35 | #define MMU_CONTEXT_TLBPID_LOCK_NR 0 | |
35 | extern unsigned long mmu_context_cache[NR_CPUS]; | ||
36 | #define mm_context(mm) (mm->context.tlbpid[smp_processor_id()]) | ||
37 | 36 | ||
38 | #define enter_lazy_tlb(mm, tsk) do {} while (0) | 37 | #define enter_lazy_tlb(mm, tsk) do {} while (0) |
39 | 38 | ||
40 | #ifdef CONFIG_SMP | 39 | #ifdef CONFIG_MN10300_TLB_USE_PIDR |
41 | #define cpu_ran_vm(cpu, mm) \ | 40 | extern unsigned long mmu_context_cache[NR_CPUS]; |
42 | cpumask_set_cpu((cpu), mm_cpumask(mm)) | 41 | #define mm_context(mm) (mm->context.tlbpid[smp_processor_id()]) |
43 | #define cpu_maybe_ran_vm(cpu, mm) \ | ||
44 | cpumask_test_and_set_cpu((cpu), mm_cpumask(mm)) | ||
45 | #else | ||
46 | #define cpu_ran_vm(cpu, mm) do {} while (0) | ||
47 | #define cpu_maybe_ran_vm(cpu, mm) true | ||
48 | #endif /* CONFIG_SMP */ | ||
49 | 42 | ||
50 | /* | 43 | /** |
51 | * allocate an MMU context | 44 | * allocate_mmu_context - Allocate storage for the arch-specific MMU data |
45 | * @mm: The userspace VM context being set up | ||
52 | */ | 46 | */ |
53 | static inline unsigned long allocate_mmu_context(struct mm_struct *mm) | 47 | static inline unsigned long allocate_mmu_context(struct mm_struct *mm) |
54 | { | 48 | { |
@@ -101,34 +95,41 @@ static inline int init_new_context(struct task_struct *tsk, | |||
101 | } | 95 | } |
102 | 96 | ||
103 | /* | 97 | /* |
104 | * destroy context related info for an mm_struct that is about to be put to | ||
105 | * rest | ||
106 | */ | ||
107 | #define destroy_context(mm) do { } while (0) | ||
108 | |||
109 | /* | ||
110 | * after we have set current->mm to a new value, this activates the context for | 98 | * after we have set current->mm to a new value, this activates the context for |
111 | * the new mm so we see the new mappings. | 99 | * the new mm so we see the new mappings. |
112 | */ | 100 | */ |
113 | static inline void activate_context(struct mm_struct *mm, int cpu) | 101 | static inline void activate_context(struct mm_struct *mm) |
114 | { | 102 | { |
115 | PIDR = get_mmu_context(mm) & MMU_CONTEXT_TLBPID_MASK; | 103 | PIDR = get_mmu_context(mm) & MMU_CONTEXT_TLBPID_MASK; |
116 | } | 104 | } |
105 | #else /* CONFIG_MN10300_TLB_USE_PIDR */ | ||
117 | 106 | ||
118 | /* | 107 | #define init_new_context(tsk, mm) (0) |
119 | * change between virtual memory sets | 108 | #define activate_context(mm) local_flush_tlb() |
109 | |||
110 | #endif /* CONFIG_MN10300_TLB_USE_PIDR */ | ||
111 | |||
112 | /** | ||
113 | * destroy_context - Destroy mm context information | ||
114 | * @mm: The MM being destroyed. | ||
115 | * | ||
116 | * Destroy context related info for an mm_struct that is about to be put to | ||
117 | * rest | ||
118 | */ | ||
119 | #define destroy_context(mm) do {} while (0) | ||
120 | |||
121 | /** | ||
122 | * switch_mm - Change between userspace virtual memory contexts | ||
123 | * @prev: The outgoing MM context. | ||
124 | * @next: The incoming MM context. | ||
125 | * @tsk: The incoming task. | ||
120 | */ | 126 | */ |
121 | static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, | 127 | static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, |
122 | struct task_struct *tsk) | 128 | struct task_struct *tsk) |
123 | { | 129 | { |
124 | int cpu = smp_processor_id(); | ||
125 | |||
126 | if (prev != next) { | 130 | if (prev != next) { |
127 | cpu_ran_vm(cpu, next); | ||
128 | activate_context(next, cpu); | ||
129 | PTBR = (unsigned long) next->pgd; | 131 | PTBR = (unsigned long) next->pgd; |
130 | } else if (!cpu_maybe_ran_vm(cpu, next)) { | 132 | activate_context(next); |
131 | activate_context(next, cpu); | ||
132 | } | 133 | } |
133 | } | 134 | } |
134 | 135 | ||
diff --git a/arch/mn10300/include/asm/tlbflush.h b/arch/mn10300/include/asm/tlbflush.h index 5d54bf57e6c3..c3c194d4031e 100644 --- a/arch/mn10300/include/asm/tlbflush.h +++ b/arch/mn10300/include/asm/tlbflush.h | |||
@@ -13,6 +13,12 @@ | |||
13 | 13 | ||
14 | #include <asm/processor.h> | 14 | #include <asm/processor.h> |
15 | 15 | ||
16 | struct tlb_state { | ||
17 | struct mm_struct *active_mm; | ||
18 | int state; | ||
19 | }; | ||
20 | DECLARE_PER_CPU_SHARED_ALIGNED(struct tlb_state, cpu_tlbstate); | ||
21 | |||
16 | /** | 22 | /** |
17 | * local_flush_tlb - Flush the current MM's entries from the local CPU's TLBs | 23 | * local_flush_tlb - Flush the current MM's entries from the local CPU's TLBs |
18 | */ | 24 | */ |
@@ -31,20 +37,51 @@ static inline void local_flush_tlb(void) | |||
31 | /** | 37 | /** |
32 | * local_flush_tlb_all - Flush all entries from the local CPU's TLBs | 38 | * local_flush_tlb_all - Flush all entries from the local CPU's TLBs |
33 | */ | 39 | */ |
34 | #define local_flush_tlb_all() local_flush_tlb() | 40 | static inline void local_flush_tlb_all(void) |
41 | { | ||
42 | local_flush_tlb(); | ||
43 | } | ||
35 | 44 | ||
36 | /** | 45 | /** |
37 | * local_flush_tlb_one - Flush one entry from the local CPU's TLBs | 46 | * local_flush_tlb_one - Flush one entry from the local CPU's TLBs |
38 | */ | 47 | */ |
39 | #define local_flush_tlb_one(addr) local_flush_tlb() | 48 | static inline void local_flush_tlb_one(unsigned long addr) |
49 | { | ||
50 | local_flush_tlb(); | ||
51 | } | ||
40 | 52 | ||
41 | /** | 53 | /** |
42 | * local_flush_tlb_page - Flush a page's entry from the local CPU's TLBs | 54 | * local_flush_tlb_page - Flush a page's entry from the local CPU's TLBs |
43 | * @mm: The MM to flush for | 55 | * @mm: The MM to flush for |
44 | * @addr: The address of the target page in RAM (not its page struct) | 56 | * @addr: The address of the target page in RAM (not its page struct) |
45 | */ | 57 | */ |
46 | extern void local_flush_tlb_page(struct mm_struct *mm, unsigned long addr); | 58 | static inline |
59 | void local_flush_tlb_page(struct mm_struct *mm, unsigned long addr) | ||
60 | { | ||
61 | unsigned long pteu, flags, cnx; | ||
62 | |||
63 | addr &= PAGE_MASK; | ||
47 | 64 | ||
65 | local_irq_save(flags); | ||
66 | |||
67 | cnx = 1; | ||
68 | #ifdef CONFIG_MN10300_TLB_USE_PIDR | ||
69 | cnx = mm->context.tlbpid[smp_processor_id()]; | ||
70 | #endif | ||
71 | if (cnx) { | ||
72 | pteu = addr; | ||
73 | #ifdef CONFIG_MN10300_TLB_USE_PIDR | ||
74 | pteu |= cnx & xPTEU_PID; | ||
75 | #endif | ||
76 | IPTEU = pteu; | ||
77 | DPTEU = pteu; | ||
78 | if (IPTEL & xPTEL_V) | ||
79 | IPTEL = 0; | ||
80 | if (DPTEL & xPTEL_V) | ||
81 | DPTEL = 0; | ||
82 | } | ||
83 | local_irq_restore(flags); | ||
84 | } | ||
48 | 85 | ||
49 | /* | 86 | /* |
50 | * TLB flushing: | 87 | * TLB flushing: |
diff --git a/arch/mn10300/mm/mmu-context.c b/arch/mn10300/mm/mmu-context.c index 3d83966e30e1..a4f7d3dcc6e6 100644 --- a/arch/mn10300/mm/mmu-context.c +++ b/arch/mn10300/mm/mmu-context.c | |||
@@ -13,40 +13,15 @@ | |||
13 | #include <asm/mmu_context.h> | 13 | #include <asm/mmu_context.h> |
14 | #include <asm/tlbflush.h> | 14 | #include <asm/tlbflush.h> |
15 | 15 | ||
16 | #ifdef CONFIG_MN10300_TLB_USE_PIDR | ||
16 | /* | 17 | /* |
17 | * list of the MMU contexts last allocated on each CPU | 18 | * list of the MMU contexts last allocated on each CPU |
18 | */ | 19 | */ |
19 | unsigned long mmu_context_cache[NR_CPUS] = { | 20 | unsigned long mmu_context_cache[NR_CPUS] = { |
20 | [0 ... NR_CPUS - 1] = MMU_CONTEXT_FIRST_VERSION * 2 - 1, | 21 | [0 ... NR_CPUS - 1] = |
22 | MMU_CONTEXT_FIRST_VERSION * 2 - (1 - MMU_CONTEXT_TLBPID_LOCK_NR), | ||
21 | }; | 23 | }; |
22 | 24 | #endif /* CONFIG_MN10300_TLB_USE_PIDR */ | |
23 | /* | ||
24 | * flush the specified TLB entry | ||
25 | */ | ||
26 | void local_flush_tlb_page(struct mm_struct *mm, unsigned long addr) | ||
27 | { | ||
28 | unsigned long pteu, cnx, flags; | ||
29 | |||
30 | addr &= PAGE_MASK; | ||
31 | |||
32 | /* make sure the context doesn't migrate and defend against | ||
33 | * interference from vmalloc'd regions */ | ||
34 | local_irq_save(flags); | ||
35 | |||
36 | cnx = mm_context(mm); | ||
37 | |||
38 | if (cnx != MMU_NO_CONTEXT) { | ||
39 | pteu = addr | (cnx & 0x000000ffUL); | ||
40 | IPTEU = pteu; | ||
41 | DPTEU = pteu; | ||
42 | if (IPTEL & xPTEL_V) | ||
43 | IPTEL = 0; | ||
44 | if (DPTEL & xPTEL_V) | ||
45 | DPTEL = 0; | ||
46 | } | ||
47 | |||
48 | local_irq_restore(flags); | ||
49 | } | ||
50 | 25 | ||
51 | /* | 26 | /* |
52 | * preemptively set a TLB entry | 27 | * preemptively set a TLB entry |
@@ -63,10 +38,16 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, pte_t *pte | |||
63 | * interference from vmalloc'd regions */ | 38 | * interference from vmalloc'd regions */ |
64 | local_irq_save(flags); | 39 | local_irq_save(flags); |
65 | 40 | ||
41 | cnx = ~MMU_NO_CONTEXT; | ||
42 | #ifdef CONFIG_MN10300_TLB_USE_PIDR | ||
66 | cnx = mm_context(vma->vm_mm); | 43 | cnx = mm_context(vma->vm_mm); |
44 | #endif | ||
67 | 45 | ||
68 | if (cnx != MMU_NO_CONTEXT) { | 46 | if (cnx != MMU_NO_CONTEXT) { |
69 | pteu = addr | (cnx & 0x000000ffUL); | 47 | pteu = addr; |
48 | #ifdef CONFIG_MN10300_TLB_USE_PIDR | ||
49 | pteu |= cnx & MMU_CONTEXT_TLBPID_MASK; | ||
50 | #endif | ||
70 | if (!(pte_val(pte) & _PAGE_NX)) { | 51 | if (!(pte_val(pte) & _PAGE_NX)) { |
71 | IPTEU = pteu; | 52 | IPTEU = pteu; |
72 | if (IPTEL & xPTEL_V) | 53 | if (IPTEL & xPTEL_V) |