diff options
author | Russell King <rmk@dyn-67.arm.linux.org.uk> | 2007-05-05 15:03:35 -0400 |
---|---|---|
committer | Russell King <rmk+kernel@arm.linux.org.uk> | 2007-05-05 15:03:35 -0400 |
commit | 9ef7963503abd3287943125681c2dc17879e8d4e (patch) | |
tree | 21fc04bda77ea16d3f54c52d265f924ad954f8f9 | |
parent | 0058ca32c3004547ede575668a2be31862b92000 (diff) |
[ARM] mm 8: define mem_types table L1 bit 4 to be for ARMv6
Change the memory types table to define the L1 descriptor bit 4 to
be in terms of the ARMv6 definition - execute never.
Signed-off-by: Russell King <rmk+kernel@arm.linux.org.uk>
-rw-r--r-- | arch/arm/mm/mmu.c | 44 |
1 files changed, 20 insertions, 24 deletions
diff --git a/arch/arm/mm/mmu.c b/arch/arm/mm/mmu.c index 360405515bbd..44f385a3eb3f 100644 --- a/arch/arm/mm/mmu.c +++ b/arch/arm/mm/mmu.c | |||
@@ -181,16 +181,16 @@ static struct mem_type mem_types[] = { | |||
181 | .prot_pte = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY | | 181 | .prot_pte = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY | |
182 | L_PTE_WRITE, | 182 | L_PTE_WRITE, |
183 | .prot_l1 = PMD_TYPE_TABLE, | 183 | .prot_l1 = PMD_TYPE_TABLE, |
184 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4 | PMD_SECT_UNCACHED | | 184 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN | PMD_SECT_UNCACHED | |
185 | PMD_SECT_AP_WRITE, | 185 | PMD_SECT_AP_WRITE, |
186 | .domain = DOMAIN_IO, | 186 | .domain = DOMAIN_IO, |
187 | }, | 187 | }, |
188 | [MT_CACHECLEAN] = { | 188 | [MT_CACHECLEAN] = { |
189 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4, | 189 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN, |
190 | .domain = DOMAIN_KERNEL, | 190 | .domain = DOMAIN_KERNEL, |
191 | }, | 191 | }, |
192 | [MT_MINICLEAN] = { | 192 | [MT_MINICLEAN] = { |
193 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4 | PMD_SECT_MINICACHE, | 193 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN | PMD_SECT_MINICACHE, |
194 | .domain = DOMAIN_KERNEL, | 194 | .domain = DOMAIN_KERNEL, |
195 | }, | 195 | }, |
196 | [MT_LOW_VECTORS] = { | 196 | [MT_LOW_VECTORS] = { |
@@ -206,25 +206,25 @@ static struct mem_type mem_types[] = { | |||
206 | .domain = DOMAIN_USER, | 206 | .domain = DOMAIN_USER, |
207 | }, | 207 | }, |
208 | [MT_MEMORY] = { | 208 | [MT_MEMORY] = { |
209 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4 | PMD_SECT_AP_WRITE, | 209 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_AP_WRITE, |
210 | .domain = DOMAIN_KERNEL, | 210 | .domain = DOMAIN_KERNEL, |
211 | }, | 211 | }, |
212 | [MT_ROM] = { | 212 | [MT_ROM] = { |
213 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4, | 213 | .prot_sect = PMD_TYPE_SECT, |
214 | .domain = DOMAIN_KERNEL, | 214 | .domain = DOMAIN_KERNEL, |
215 | }, | 215 | }, |
216 | [MT_IXP2000_DEVICE] = { /* IXP2400 requires XCB=101 for on-chip I/O */ | 216 | [MT_IXP2000_DEVICE] = { /* IXP2400 requires XCB=101 for on-chip I/O */ |
217 | .prot_pte = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY | | 217 | .prot_pte = L_PTE_PRESENT | L_PTE_YOUNG | L_PTE_DIRTY | |
218 | L_PTE_WRITE, | 218 | L_PTE_WRITE, |
219 | .prot_l1 = PMD_TYPE_TABLE, | 219 | .prot_l1 = PMD_TYPE_TABLE, |
220 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4 | PMD_SECT_UNCACHED | | 220 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN | PMD_SECT_UNCACHED | |
221 | PMD_SECT_AP_WRITE | PMD_SECT_BUFFERABLE | | 221 | PMD_SECT_AP_WRITE | PMD_SECT_BUFFERABLE | |
222 | PMD_SECT_TEX(1), | 222 | PMD_SECT_TEX(1), |
223 | .domain = DOMAIN_IO, | 223 | .domain = DOMAIN_IO, |
224 | }, | 224 | }, |
225 | [MT_NONSHARED_DEVICE] = { | 225 | [MT_NONSHARED_DEVICE] = { |
226 | .prot_l1 = PMD_TYPE_TABLE, | 226 | .prot_l1 = PMD_TYPE_TABLE, |
227 | .prot_sect = PMD_TYPE_SECT | PMD_BIT4 | PMD_SECT_NONSHARED_DEV | | 227 | .prot_sect = PMD_TYPE_SECT | PMD_SECT_XN | PMD_SECT_NONSHARED_DEV | |
228 | PMD_SECT_AP_WRITE, | 228 | PMD_SECT_AP_WRITE, |
229 | .domain = DOMAIN_IO, | 229 | .domain = DOMAIN_IO, |
230 | } | 230 | } |
@@ -260,20 +260,23 @@ static void __init build_mem_type_table(void) | |||
260 | } | 260 | } |
261 | 261 | ||
262 | /* | 262 | /* |
263 | * Xscale must not have PMD bit 4 set for section mappings. | 263 | * ARMv5 and lower, bit 4 must be set for page tables. |
264 | * (was: cache "update-able on write" bit on ARM610) | ||
265 | * However, Xscale cores require this bit to be cleared. | ||
264 | */ | 266 | */ |
265 | if (cpu_is_xscale()) | 267 | if (cpu_is_xscale()) { |
266 | for (i = 0; i < ARRAY_SIZE(mem_types); i++) | 268 | for (i = 0; i < ARRAY_SIZE(mem_types); i++) { |
267 | mem_types[i].prot_sect &= ~PMD_BIT4; | 269 | mem_types[i].prot_sect &= ~PMD_BIT4; |
268 | 270 | mem_types[i].prot_l1 &= ~PMD_BIT4; | |
269 | /* | 271 | } |
270 | * ARMv5 and lower, excluding Xscale, bit 4 must be set for | 272 | } else if (cpu_arch < CPU_ARCH_ARMv6) { |
271 | * page tables. | 273 | for (i = 0; i < ARRAY_SIZE(mem_types); i++) { |
272 | */ | ||
273 | if (cpu_arch < CPU_ARCH_ARMv6 && !cpu_is_xscale()) | ||
274 | for (i = 0; i < ARRAY_SIZE(mem_types); i++) | ||
275 | if (mem_types[i].prot_l1) | 274 | if (mem_types[i].prot_l1) |
276 | mem_types[i].prot_l1 |= PMD_BIT4; | 275 | mem_types[i].prot_l1 |= PMD_BIT4; |
276 | if (mem_types[i].prot_sect) | ||
277 | mem_types[i].prot_sect |= PMD_BIT4; | ||
278 | } | ||
279 | } | ||
277 | 280 | ||
278 | cp = &cache_policies[cachepolicy]; | 281 | cp = &cache_policies[cachepolicy]; |
279 | kern_pgprot = user_pgprot = cp->pte; | 282 | kern_pgprot = user_pgprot = cp->pte; |
@@ -294,13 +297,6 @@ static void __init build_mem_type_table(void) | |||
294 | */ | 297 | */ |
295 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { | 298 | if (cpu_arch >= CPU_ARCH_ARMv6 && (cr & CR_XP)) { |
296 | /* | 299 | /* |
297 | * bit 4 becomes XN which we must clear for the | ||
298 | * kernel memory mapping. | ||
299 | */ | ||
300 | mem_types[MT_MEMORY].prot_sect &= ~PMD_SECT_XN; | ||
301 | mem_types[MT_ROM].prot_sect &= ~PMD_SECT_XN; | ||
302 | |||
303 | /* | ||
304 | * Mark cache clean areas and XIP ROM read only | 300 | * Mark cache clean areas and XIP ROM read only |
305 | * from SVC mode and no access from userspace. | 301 | * from SVC mode and no access from userspace. |
306 | */ | 302 | */ |