diff options
Diffstat (limited to 'arch')
-rw-r--r-- | arch/x86/Kconfig.debug | 2 | ||||
-rw-r--r-- | arch/x86/mm/init_32.c | 4 | ||||
-rw-r--r-- | arch/x86/mm/init_64.c | 16 | ||||
-rw-r--r-- | arch/x86/mm/pageattr-test.c | 15 | ||||
-rw-r--r-- | arch/x86/mm/pageattr.c | 74 |
5 files changed, 47 insertions, 64 deletions
diff --git a/arch/x86/Kconfig.debug b/arch/x86/Kconfig.debug index 2d0bd33b73aa..2e1e3af28c3a 100644 --- a/arch/x86/Kconfig.debug +++ b/arch/x86/Kconfig.debug | |||
@@ -40,7 +40,7 @@ comment "Page alloc debug is incompatible with Software Suspend on i386" | |||
40 | 40 | ||
41 | config DEBUG_PAGEALLOC | 41 | config DEBUG_PAGEALLOC |
42 | bool "Debug page memory allocations" | 42 | bool "Debug page memory allocations" |
43 | depends on DEBUG_KERNEL | 43 | depends on DEBUG_KERNEL && X86_32 |
44 | help | 44 | help |
45 | Unmap pages from the kernel linear mapping after free_pages(). | 45 | Unmap pages from the kernel linear mapping after free_pages(). |
46 | This results in a large slowdown, but helps to find certain types | 46 | This results in a large slowdown, but helps to find certain types |
diff --git a/arch/x86/mm/init_32.c b/arch/x86/mm/init_32.c index 8d7f723cfc28..8ed5c189d7aa 100644 --- a/arch/x86/mm/init_32.c +++ b/arch/x86/mm/init_32.c | |||
@@ -781,8 +781,6 @@ void mark_rodata_ro(void) | |||
781 | 781 | ||
782 | void free_init_pages(char *what, unsigned long begin, unsigned long end) | 782 | void free_init_pages(char *what, unsigned long begin, unsigned long end) |
783 | { | 783 | { |
784 | unsigned long addr; | ||
785 | |||
786 | #ifdef CONFIG_DEBUG_PAGEALLOC | 784 | #ifdef CONFIG_DEBUG_PAGEALLOC |
787 | /* | 785 | /* |
788 | * If debugging page accesses then do not free this memory but | 786 | * If debugging page accesses then do not free this memory but |
@@ -793,6 +791,8 @@ void free_init_pages(char *what, unsigned long begin, unsigned long end) | |||
793 | begin, PAGE_ALIGN(end)); | 791 | begin, PAGE_ALIGN(end)); |
794 | set_memory_np(begin, (end - begin) >> PAGE_SHIFT); | 792 | set_memory_np(begin, (end - begin) >> PAGE_SHIFT); |
795 | #else | 793 | #else |
794 | unsigned long addr; | ||
795 | |||
796 | /* | 796 | /* |
797 | * We just marked the kernel text read only above, now that | 797 | * We just marked the kernel text read only above, now that |
798 | * we are going to free part of that, we need to make that | 798 | * we are going to free part of that, we need to make that |
diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c index e0c1e98ad1bf..8a7b725ce3c7 100644 --- a/arch/x86/mm/init_64.c +++ b/arch/x86/mm/init_64.c | |||
@@ -569,22 +569,6 @@ void free_init_pages(char *what, unsigned long begin, unsigned long end) | |||
569 | free_page(addr); | 569 | free_page(addr); |
570 | totalram_pages++; | 570 | totalram_pages++; |
571 | } | 571 | } |
572 | #ifdef CONFIG_DEBUG_RODATA | ||
573 | /* | ||
574 | * This will make the __init pages not present and | ||
575 | * not executable, so that any attempt to use a | ||
576 | * __init function from now on will fault immediately | ||
577 | * rather than supriously later when memory gets reused. | ||
578 | * | ||
579 | * We only do this for DEBUG_RODATA to not break up the | ||
580 | * 2Mb kernel mapping just for this debug feature. | ||
581 | */ | ||
582 | if (begin >= __START_KERNEL_map) { | ||
583 | set_memory_rw(begin, (end - begin)/PAGE_SIZE); | ||
584 | set_memory_np(begin, (end - begin)/PAGE_SIZE); | ||
585 | set_memory_nx(begin, (end - begin)/PAGE_SIZE); | ||
586 | } | ||
587 | #endif | ||
588 | #endif | 572 | #endif |
589 | } | 573 | } |
590 | 574 | ||
diff --git a/arch/x86/mm/pageattr-test.c b/arch/x86/mm/pageattr-test.c index 554820265b95..06353d43f72e 100644 --- a/arch/x86/mm/pageattr-test.c +++ b/arch/x86/mm/pageattr-test.c | |||
@@ -15,8 +15,7 @@ | |||
15 | #include <asm/kdebug.h> | 15 | #include <asm/kdebug.h> |
16 | 16 | ||
17 | enum { | 17 | enum { |
18 | NTEST = 400, | 18 | NTEST = 4000, |
19 | LOWEST_LEVEL = PG_LEVEL_4K, | ||
20 | #ifdef CONFIG_X86_64 | 19 | #ifdef CONFIG_X86_64 |
21 | LPS = (1 << PMD_SHIFT), | 20 | LPS = (1 << PMD_SHIFT), |
22 | #elif defined(CONFIG_X86_PAE) | 21 | #elif defined(CONFIG_X86_PAE) |
@@ -59,10 +58,10 @@ static __init int print_split(struct split_state *s) | |||
59 | continue; | 58 | continue; |
60 | } | 59 | } |
61 | 60 | ||
62 | if (level == 2 && sizeof(long) == 8) { | 61 | if (level == PG_LEVEL_1G && sizeof(long) == 8) { |
63 | s->gpg++; | 62 | s->gpg++; |
64 | i += GPS/PAGE_SIZE; | 63 | i += GPS/PAGE_SIZE; |
65 | } else if (level != LOWEST_LEVEL) { | 64 | } else if (level == PG_LEVEL_2M) { |
66 | if (!(pte_val(*pte) & _PAGE_PSE)) { | 65 | if (!(pte_val(*pte) & _PAGE_PSE)) { |
67 | printk(KERN_ERR | 66 | printk(KERN_ERR |
68 | "%lx level %d but not PSE %Lx\n", | 67 | "%lx level %d but not PSE %Lx\n", |
@@ -162,7 +161,7 @@ static __init int exercise_pageattr(void) | |||
162 | continue; | 161 | continue; |
163 | } | 162 | } |
164 | 163 | ||
165 | err = __change_page_attr_clear(addr[i], len[i], | 164 | err = change_page_attr_clear(addr[i], len[i], |
166 | __pgprot(_PAGE_GLOBAL)); | 165 | __pgprot(_PAGE_GLOBAL)); |
167 | if (err < 0) { | 166 | if (err < 0) { |
168 | printk(KERN_ERR "CPA %d failed %d\n", i, err); | 167 | printk(KERN_ERR "CPA %d failed %d\n", i, err); |
@@ -175,7 +174,7 @@ static __init int exercise_pageattr(void) | |||
175 | pte ? (u64)pte_val(*pte) : 0ULL); | 174 | pte ? (u64)pte_val(*pte) : 0ULL); |
176 | failed++; | 175 | failed++; |
177 | } | 176 | } |
178 | if (level != LOWEST_LEVEL) { | 177 | if (level != PG_LEVEL_4K) { |
179 | printk(KERN_ERR "CPA %lx: unexpected level %d\n", | 178 | printk(KERN_ERR "CPA %lx: unexpected level %d\n", |
180 | addr[i], level); | 179 | addr[i], level); |
181 | failed++; | 180 | failed++; |
@@ -183,7 +182,6 @@ static __init int exercise_pageattr(void) | |||
183 | 182 | ||
184 | } | 183 | } |
185 | vfree(bm); | 184 | vfree(bm); |
186 | cpa_flush_all(); | ||
187 | 185 | ||
188 | failed += print_split(&sb); | 186 | failed += print_split(&sb); |
189 | 187 | ||
@@ -197,7 +195,7 @@ static __init int exercise_pageattr(void) | |||
197 | failed++; | 195 | failed++; |
198 | continue; | 196 | continue; |
199 | } | 197 | } |
200 | err = __change_page_attr_set(addr[i], len[i], | 198 | err = change_page_attr_set(addr[i], len[i], |
201 | __pgprot(_PAGE_GLOBAL)); | 199 | __pgprot(_PAGE_GLOBAL)); |
202 | if (err < 0) { | 200 | if (err < 0) { |
203 | printk(KERN_ERR "CPA reverting failed: %d\n", err); | 201 | printk(KERN_ERR "CPA reverting failed: %d\n", err); |
@@ -211,7 +209,6 @@ static __init int exercise_pageattr(void) | |||
211 | } | 209 | } |
212 | 210 | ||
213 | } | 211 | } |
214 | cpa_flush_all(); | ||
215 | 212 | ||
216 | failed += print_split(&sc); | 213 | failed += print_split(&sc); |
217 | 214 | ||
diff --git a/arch/x86/mm/pageattr.c b/arch/x86/mm/pageattr.c index 97ec9e7d29d9..532a40bc0e7e 100644 --- a/arch/x86/mm/pageattr.c +++ b/arch/x86/mm/pageattr.c | |||
@@ -197,10 +197,11 @@ static int split_large_page(pte_t *kpte, unsigned long address) | |||
197 | unsigned long addr; | 197 | unsigned long addr; |
198 | pte_t *pbase, *tmp; | 198 | pte_t *pbase, *tmp; |
199 | struct page *base; | 199 | struct page *base; |
200 | int i, level; | 200 | unsigned int i, level; |
201 | 201 | ||
202 | #ifdef CONFIG_DEBUG_PAGEALLOC | 202 | #ifdef CONFIG_DEBUG_PAGEALLOC |
203 | gfp_flags = GFP_ATOMIC; | 203 | gfp_flags = __GFP_HIGH | __GFP_NOFAIL | __GFP_NOWARN; |
204 | gfp_flags = GFP_ATOMIC | __GFP_NOWARN; | ||
204 | #endif | 205 | #endif |
205 | base = alloc_pages(gfp_flags, 0); | 206 | base = alloc_pages(gfp_flags, 0); |
206 | if (!base) | 207 | if (!base) |
@@ -224,6 +225,7 @@ static int split_large_page(pte_t *kpte, unsigned long address) | |||
224 | paravirt_alloc_pt(&init_mm, page_to_pfn(base)); | 225 | paravirt_alloc_pt(&init_mm, page_to_pfn(base)); |
225 | #endif | 226 | #endif |
226 | 227 | ||
228 | pgprot_val(ref_prot) &= ~_PAGE_NX; | ||
227 | for (i = 0; i < PTRS_PER_PTE; i++, addr += PAGE_SIZE) | 229 | for (i = 0; i < PTRS_PER_PTE; i++, addr += PAGE_SIZE) |
228 | set_pte(&pbase[i], pfn_pte(addr >> PAGE_SHIFT, ref_prot)); | 230 | set_pte(&pbase[i], pfn_pte(addr >> PAGE_SHIFT, ref_prot)); |
229 | 231 | ||
@@ -248,7 +250,8 @@ out_unlock: | |||
248 | } | 250 | } |
249 | 251 | ||
250 | static int | 252 | static int |
251 | __change_page_attr(unsigned long address, unsigned long pfn, pgprot_t prot) | 253 | __change_page_attr(unsigned long address, unsigned long pfn, |
254 | pgprot_t mask_set, pgprot_t mask_clr) | ||
252 | { | 255 | { |
253 | struct page *kpte_page; | 256 | struct page *kpte_page; |
254 | int level, err = 0; | 257 | int level, err = 0; |
@@ -267,15 +270,20 @@ repeat: | |||
267 | BUG_ON(PageLRU(kpte_page)); | 270 | BUG_ON(PageLRU(kpte_page)); |
268 | BUG_ON(PageCompound(kpte_page)); | 271 | BUG_ON(PageCompound(kpte_page)); |
269 | 272 | ||
270 | prot = static_protections(prot, address); | ||
271 | |||
272 | if (level == PG_LEVEL_4K) { | 273 | if (level == PG_LEVEL_4K) { |
273 | WARN_ON_ONCE(pgprot_val(prot) & _PAGE_PSE); | 274 | pgprot_t new_prot = pte_pgprot(*kpte); |
274 | set_pte_atomic(kpte, pfn_pte(pfn, canon_pgprot(prot))); | 275 | pte_t new_pte, old_pte = *kpte; |
275 | } else { | 276 | |
276 | /* Clear the PSE bit for the 4k level pages ! */ | 277 | pgprot_val(new_prot) &= ~pgprot_val(mask_clr); |
277 | pgprot_val(prot) = pgprot_val(prot) & ~_PAGE_PSE; | 278 | pgprot_val(new_prot) |= pgprot_val(mask_set); |
279 | |||
280 | new_prot = static_protections(new_prot, address); | ||
281 | |||
282 | new_pte = pfn_pte(pfn, canon_pgprot(new_prot)); | ||
283 | BUG_ON(pte_pfn(new_pte) != pte_pfn(old_pte)); | ||
278 | 284 | ||
285 | set_pte_atomic(kpte, new_pte); | ||
286 | } else { | ||
279 | err = split_large_page(kpte, address); | 287 | err = split_large_page(kpte, address); |
280 | if (!err) | 288 | if (!err) |
281 | goto repeat; | 289 | goto repeat; |
@@ -297,22 +305,26 @@ repeat: | |||
297 | * Modules and drivers should use the set_memory_* APIs instead. | 305 | * Modules and drivers should use the set_memory_* APIs instead. |
298 | */ | 306 | */ |
299 | 307 | ||
300 | static int change_page_attr_addr(unsigned long address, pgprot_t prot) | 308 | static int |
309 | change_page_attr_addr(unsigned long address, pgprot_t mask_set, | ||
310 | pgprot_t mask_clr) | ||
301 | { | 311 | { |
302 | int err = 0, kernel_map = 0; | 312 | int err = 0, kernel_map = 0; |
303 | unsigned long pfn = __pa(address) >> PAGE_SHIFT; | 313 | unsigned long pfn; |
304 | 314 | ||
305 | #ifdef CONFIG_X86_64 | 315 | #ifdef CONFIG_X86_64 |
306 | if (address >= __START_KERNEL_map && | 316 | if (address >= __START_KERNEL_map && |
307 | address < __START_KERNEL_map + KERNEL_TEXT_SIZE) { | 317 | address < __START_KERNEL_map + KERNEL_TEXT_SIZE) { |
308 | 318 | ||
309 | address = (unsigned long)__va(__pa(address)); | 319 | address = (unsigned long)__va(__pa((void *)address)); |
310 | kernel_map = 1; | 320 | kernel_map = 1; |
311 | } | 321 | } |
312 | #endif | 322 | #endif |
313 | 323 | ||
314 | if (!kernel_map || pte_present(pfn_pte(0, prot))) { | 324 | pfn = __pa(address) >> PAGE_SHIFT; |
315 | err = __change_page_attr(address, pfn, prot); | 325 | |
326 | if (!kernel_map || 1) { | ||
327 | err = __change_page_attr(address, pfn, mask_set, mask_clr); | ||
316 | if (err) | 328 | if (err) |
317 | return err; | 329 | return err; |
318 | } | 330 | } |
@@ -324,12 +336,15 @@ static int change_page_attr_addr(unsigned long address, pgprot_t prot) | |||
324 | */ | 336 | */ |
325 | if (__pa(address) < KERNEL_TEXT_SIZE) { | 337 | if (__pa(address) < KERNEL_TEXT_SIZE) { |
326 | unsigned long addr2; | 338 | unsigned long addr2; |
327 | pgprot_t prot2; | ||
328 | 339 | ||
329 | addr2 = __START_KERNEL_map + __pa(address); | 340 | addr2 = __pa(address) + __START_KERNEL_map - phys_base; |
330 | /* Make sure the kernel mappings stay executable */ | 341 | /* Make sure the kernel mappings stay executable */ |
331 | prot2 = pte_pgprot(pte_mkexec(pfn_pte(0, prot))); | 342 | pgprot_val(mask_clr) |= _PAGE_NX; |
332 | err = __change_page_attr(addr2, pfn, prot2); | 343 | /* |
344 | * Our high aliases are imprecise, so do not propagate | ||
345 | * failures back to users: | ||
346 | */ | ||
347 | __change_page_attr(addr2, pfn, mask_set, mask_clr); | ||
333 | } | 348 | } |
334 | #endif | 349 | #endif |
335 | 350 | ||
@@ -339,26 +354,13 @@ static int change_page_attr_addr(unsigned long address, pgprot_t prot) | |||
339 | static int __change_page_attr_set_clr(unsigned long addr, int numpages, | 354 | static int __change_page_attr_set_clr(unsigned long addr, int numpages, |
340 | pgprot_t mask_set, pgprot_t mask_clr) | 355 | pgprot_t mask_set, pgprot_t mask_clr) |
341 | { | 356 | { |
342 | pgprot_t new_prot; | 357 | unsigned int i; |
343 | int level; | 358 | int ret; |
344 | pte_t *pte; | ||
345 | int i, ret; | ||
346 | |||
347 | for (i = 0; i < numpages ; i++) { | ||
348 | |||
349 | pte = lookup_address(addr, &level); | ||
350 | if (!pte) | ||
351 | return -EINVAL; | ||
352 | |||
353 | new_prot = pte_pgprot(*pte); | ||
354 | |||
355 | pgprot_val(new_prot) &= ~pgprot_val(mask_clr); | ||
356 | pgprot_val(new_prot) |= pgprot_val(mask_set); | ||
357 | 359 | ||
358 | ret = change_page_attr_addr(addr, new_prot); | 360 | for (i = 0; i < numpages ; i++, addr += PAGE_SIZE) { |
361 | ret = change_page_attr_addr(addr, mask_set, mask_clr); | ||
359 | if (ret) | 362 | if (ret) |
360 | return ret; | 363 | return ret; |
361 | addr += PAGE_SIZE; | ||
362 | } | 364 | } |
363 | 365 | ||
364 | return 0; | 366 | return 0; |