diff options
author | Thomas Gleixner <tglx@linutronix.de> | 2008-02-18 14:54:14 -0500 |
---|---|---|
committer | Thomas Gleixner <tglx@linutronix.de> | 2008-02-18 14:54:14 -0500 |
commit | c31c7d4844ea4817692ae16bf70f9c96c05a50eb (patch) | |
tree | b06e86bda1545f3f84fee09e4600bb3aa25c5c33 /arch | |
parent | 1309d4e68497184d2fd87e892ddf14076c2bda98 (diff) |
x86: CPA, fix alias checks
c_p_a() did not discover all aliases correctly. (such as when called
on vmalloc()-ed areas or ioremap()-ed areas)
Push the alias checks to the lower, physical level and consistently
discover all aliases that might exist: the low direct mappings and
the high linear kernel-text mappings (on 64-bit).
Thanks to Andi Kleen for pointing out that this was buggy.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r-- | arch/x86/mm/pageattr.c | 192 |
1 files changed, 92 insertions, 100 deletions
diff --git a/arch/x86/mm/pageattr.c b/arch/x86/mm/pageattr.c index 4119379f80ff..7b4a17236972 100644 --- a/arch/x86/mm/pageattr.c +++ b/arch/x86/mm/pageattr.c | |||
@@ -16,6 +16,7 @@ | |||
16 | #include <asm/sections.h> | 16 | #include <asm/sections.h> |
17 | #include <asm/uaccess.h> | 17 | #include <asm/uaccess.h> |
18 | #include <asm/pgalloc.h> | 18 | #include <asm/pgalloc.h> |
19 | #include <asm/proto.h> | ||
19 | 20 | ||
20 | /* | 21 | /* |
21 | * The current flushing context - we pass it instead of 5 arguments: | 22 | * The current flushing context - we pass it instead of 5 arguments: |
@@ -26,8 +27,23 @@ struct cpa_data { | |||
26 | pgprot_t mask_clr; | 27 | pgprot_t mask_clr; |
27 | int numpages; | 28 | int numpages; |
28 | int flushtlb; | 29 | int flushtlb; |
30 | unsigned long pfn; | ||
29 | }; | 31 | }; |
30 | 32 | ||
33 | #ifdef CONFIG_X86_64 | ||
34 | |||
35 | static inline unsigned long highmap_start_pfn(void) | ||
36 | { | ||
37 | return __pa(_text) >> PAGE_SHIFT; | ||
38 | } | ||
39 | |||
40 | static inline unsigned long highmap_end_pfn(void) | ||
41 | { | ||
42 | return __pa(round_up((unsigned long)_end, PMD_SIZE)) >> PAGE_SHIFT; | ||
43 | } | ||
44 | |||
45 | #endif | ||
46 | |||
31 | static inline int | 47 | static inline int |
32 | within(unsigned long addr, unsigned long start, unsigned long end) | 48 | within(unsigned long addr, unsigned long start, unsigned long end) |
33 | { | 49 | { |
@@ -123,29 +139,14 @@ static void cpa_flush_range(unsigned long start, int numpages, int cache) | |||
123 | } | 139 | } |
124 | } | 140 | } |
125 | 141 | ||
126 | #define HIGH_MAP_START __START_KERNEL_map | ||
127 | #define HIGH_MAP_END (__START_KERNEL_map + KERNEL_TEXT_SIZE) | ||
128 | |||
129 | |||
130 | /* | ||
131 | * Converts a virtual address to a X86-64 highmap address | ||
132 | */ | ||
133 | static unsigned long virt_to_highmap(void *address) | ||
134 | { | ||
135 | #ifdef CONFIG_X86_64 | ||
136 | return __pa((unsigned long)address) + HIGH_MAP_START - phys_base; | ||
137 | #else | ||
138 | return (unsigned long)address; | ||
139 | #endif | ||
140 | } | ||
141 | |||
142 | /* | 142 | /* |
143 | * Certain areas of memory on x86 require very specific protection flags, | 143 | * Certain areas of memory on x86 require very specific protection flags, |
144 | * for example the BIOS area or kernel text. Callers don't always get this | 144 | * for example the BIOS area or kernel text. Callers don't always get this |
145 | * right (again, ioremap() on BIOS memory is not uncommon) so this function | 145 | * right (again, ioremap() on BIOS memory is not uncommon) so this function |
146 | * checks and fixes these known static required protection bits. | 146 | * checks and fixes these known static required protection bits. |
147 | */ | 147 | */ |
148 | static inline pgprot_t static_protections(pgprot_t prot, unsigned long address) | 148 | static inline pgprot_t static_protections(pgprot_t prot, unsigned long address, |
149 | unsigned long pfn) | ||
149 | { | 150 | { |
150 | pgprot_t forbidden = __pgprot(0); | 151 | pgprot_t forbidden = __pgprot(0); |
151 | 152 | ||
@@ -153,30 +154,23 @@ static inline pgprot_t static_protections(pgprot_t prot, unsigned long address) | |||
153 | * The BIOS area between 640k and 1Mb needs to be executable for | 154 | * The BIOS area between 640k and 1Mb needs to be executable for |
154 | * PCI BIOS based config access (CONFIG_PCI_GOBIOS) support. | 155 | * PCI BIOS based config access (CONFIG_PCI_GOBIOS) support. |
155 | */ | 156 | */ |
156 | if (within(__pa(address), BIOS_BEGIN, BIOS_END)) | 157 | if (within(pfn, BIOS_BEGIN >> PAGE_SHIFT, BIOS_END >> PAGE_SHIFT)) |
157 | pgprot_val(forbidden) |= _PAGE_NX; | 158 | pgprot_val(forbidden) |= _PAGE_NX; |
158 | 159 | ||
159 | /* | 160 | /* |
160 | * The kernel text needs to be executable for obvious reasons | 161 | * The kernel text needs to be executable for obvious reasons |
161 | * Does not cover __inittext since that is gone later on | 162 | * Does not cover __inittext since that is gone later on. On |
163 | * 64bit we do not enforce !NX on the low mapping | ||
162 | */ | 164 | */ |
163 | if (within(address, (unsigned long)_text, (unsigned long)_etext)) | 165 | if (within(address, (unsigned long)_text, (unsigned long)_etext)) |
164 | pgprot_val(forbidden) |= _PAGE_NX; | 166 | pgprot_val(forbidden) |= _PAGE_NX; |
165 | /* | ||
166 | * Do the same for the x86-64 high kernel mapping | ||
167 | */ | ||
168 | if (within(address, virt_to_highmap(_text), virt_to_highmap(_etext))) | ||
169 | pgprot_val(forbidden) |= _PAGE_NX; | ||
170 | 167 | ||
171 | /* The .rodata section needs to be read-only */ | ||
172 | if (within(address, (unsigned long)__start_rodata, | ||
173 | (unsigned long)__end_rodata)) | ||
174 | pgprot_val(forbidden) |= _PAGE_RW; | ||
175 | /* | 168 | /* |
176 | * Do the same for the x86-64 high kernel mapping | 169 | * The .rodata section needs to be read-only. Using the pfn |
170 | * catches all aliases. | ||
177 | */ | 171 | */ |
178 | if (within(address, virt_to_highmap(__start_rodata), | 172 | if (within(pfn, __pa((unsigned long)__start_rodata) >> PAGE_SHIFT, |
179 | virt_to_highmap(__end_rodata))) | 173 | __pa((unsigned long)__end_rodata) >> PAGE_SHIFT)) |
180 | pgprot_val(forbidden) |= _PAGE_RW; | 174 | pgprot_val(forbidden) |= _PAGE_RW; |
181 | 175 | ||
182 | prot = __pgprot(pgprot_val(prot) & ~pgprot_val(forbidden)); | 176 | prot = __pgprot(pgprot_val(prot) & ~pgprot_val(forbidden)); |
@@ -253,7 +247,7 @@ static int | |||
253 | try_preserve_large_page(pte_t *kpte, unsigned long address, | 247 | try_preserve_large_page(pte_t *kpte, unsigned long address, |
254 | struct cpa_data *cpa) | 248 | struct cpa_data *cpa) |
255 | { | 249 | { |
256 | unsigned long nextpage_addr, numpages, pmask, psize, flags, addr; | 250 | unsigned long nextpage_addr, numpages, pmask, psize, flags, addr, pfn; |
257 | pte_t new_pte, old_pte, *tmp; | 251 | pte_t new_pte, old_pte, *tmp; |
258 | pgprot_t old_prot, new_prot; | 252 | pgprot_t old_prot, new_prot; |
259 | int i, do_split = 1; | 253 | int i, do_split = 1; |
@@ -301,7 +295,15 @@ try_preserve_large_page(pte_t *kpte, unsigned long address, | |||
301 | 295 | ||
302 | pgprot_val(new_prot) &= ~pgprot_val(cpa->mask_clr); | 296 | pgprot_val(new_prot) &= ~pgprot_val(cpa->mask_clr); |
303 | pgprot_val(new_prot) |= pgprot_val(cpa->mask_set); | 297 | pgprot_val(new_prot) |= pgprot_val(cpa->mask_set); |
304 | new_prot = static_protections(new_prot, address); | 298 | |
299 | /* | ||
300 | * old_pte points to the large page base address. So we need | ||
301 | * to add the offset of the virtual address: | ||
302 | */ | ||
303 | pfn = pte_pfn(old_pte) + ((address & (psize - 1)) >> PAGE_SHIFT); | ||
304 | cpa->pfn = pfn; | ||
305 | |||
306 | new_prot = static_protections(new_prot, address, pfn); | ||
305 | 307 | ||
306 | /* | 308 | /* |
307 | * We need to check the full range, whether | 309 | * We need to check the full range, whether |
@@ -309,8 +311,9 @@ try_preserve_large_page(pte_t *kpte, unsigned long address, | |||
309 | * the pages in the range we try to preserve: | 311 | * the pages in the range we try to preserve: |
310 | */ | 312 | */ |
311 | addr = address + PAGE_SIZE; | 313 | addr = address + PAGE_SIZE; |
312 | for (i = 1; i < cpa->numpages; i++, addr += PAGE_SIZE) { | 314 | pfn++; |
313 | pgprot_t chk_prot = static_protections(new_prot, addr); | 315 | for (i = 1; i < cpa->numpages; i++, addr += PAGE_SIZE, pfn++) { |
316 | pgprot_t chk_prot = static_protections(new_prot, addr, pfn); | ||
314 | 317 | ||
315 | if (pgprot_val(chk_prot) != pgprot_val(new_prot)) | 318 | if (pgprot_val(chk_prot) != pgprot_val(new_prot)) |
316 | goto out_unlock; | 319 | goto out_unlock; |
@@ -505,46 +508,51 @@ out_unlock: | |||
505 | return 0; | 508 | return 0; |
506 | } | 509 | } |
507 | 510 | ||
508 | static int __change_page_attr(unsigned long address, struct cpa_data *cpa) | 511 | static int __change_page_attr(struct cpa_data *cpa, int primary) |
509 | { | 512 | { |
513 | unsigned long address = cpa->vaddr; | ||
510 | int do_split, err; | 514 | int do_split, err; |
511 | unsigned int level; | 515 | unsigned int level; |
512 | struct page *kpte_page; | 516 | struct page *kpte_page; |
513 | pte_t *kpte; | 517 | pte_t *kpte, old_pte; |
514 | 518 | ||
515 | repeat: | 519 | repeat: |
516 | kpte = lookup_address(address, &level); | 520 | kpte = lookup_address(address, &level); |
517 | if (!kpte) | 521 | if (!kpte) |
522 | return primary ? -EINVAL : 0; | ||
523 | |||
524 | old_pte = *kpte; | ||
525 | if (!pte_val(old_pte)) { | ||
526 | if (!primary) | ||
527 | return 0; | ||
528 | printk(KERN_WARNING "CPA: called for zero pte. " | ||
529 | "vaddr = %lx cpa->vaddr = %lx\n", address, | ||
530 | cpa->vaddr); | ||
531 | WARN_ON(1); | ||
518 | return -EINVAL; | 532 | return -EINVAL; |
533 | } | ||
519 | 534 | ||
520 | kpte_page = virt_to_page(kpte); | 535 | kpte_page = virt_to_page(kpte); |
521 | BUG_ON(PageLRU(kpte_page)); | 536 | BUG_ON(PageLRU(kpte_page)); |
522 | BUG_ON(PageCompound(kpte_page)); | 537 | BUG_ON(PageCompound(kpte_page)); |
523 | 538 | ||
524 | if (level == PG_LEVEL_4K) { | 539 | if (level == PG_LEVEL_4K) { |
525 | pte_t new_pte, old_pte = *kpte; | 540 | pte_t new_pte; |
526 | pgprot_t new_prot = pte_pgprot(old_pte); | 541 | pgprot_t new_prot = pte_pgprot(old_pte); |
527 | 542 | unsigned long pfn = pte_pfn(old_pte); | |
528 | if(!pte_val(old_pte)) { | ||
529 | printk(KERN_WARNING "CPA: called for zero pte. " | ||
530 | "vaddr = %lx cpa->vaddr = %lx\n", address, | ||
531 | cpa->vaddr); | ||
532 | WARN_ON(1); | ||
533 | return -EINVAL; | ||
534 | } | ||
535 | 543 | ||
536 | pgprot_val(new_prot) &= ~pgprot_val(cpa->mask_clr); | 544 | pgprot_val(new_prot) &= ~pgprot_val(cpa->mask_clr); |
537 | pgprot_val(new_prot) |= pgprot_val(cpa->mask_set); | 545 | pgprot_val(new_prot) |= pgprot_val(cpa->mask_set); |
538 | 546 | ||
539 | new_prot = static_protections(new_prot, address); | 547 | new_prot = static_protections(new_prot, address, pfn); |
540 | 548 | ||
541 | /* | 549 | /* |
542 | * We need to keep the pfn from the existing PTE, | 550 | * We need to keep the pfn from the existing PTE, |
543 | * after all we're only going to change it's attributes | 551 | * after all we're only going to change it's attributes |
544 | * not the memory it points to | 552 | * not the memory it points to |
545 | */ | 553 | */ |
546 | new_pte = pfn_pte(pte_pfn(old_pte), canon_pgprot(new_prot)); | 554 | new_pte = pfn_pte(pfn, canon_pgprot(new_prot)); |
547 | 555 | cpa->pfn = pfn; | |
548 | /* | 556 | /* |
549 | * Do we really change anything ? | 557 | * Do we really change anything ? |
550 | */ | 558 | */ |
@@ -581,67 +589,44 @@ repeat: | |||
581 | return err; | 589 | return err; |
582 | } | 590 | } |
583 | 591 | ||
584 | /** | 592 | static int __change_page_attr_set_clr(struct cpa_data *cpa, int checkalias); |
585 | * change_page_attr_addr - Change page table attributes in linear mapping | 593 | |
586 | * @address: Virtual address in linear mapping. | 594 | static int cpa_process_alias(struct cpa_data *cpa) |
587 | * @prot: New page table attribute (PAGE_*) | ||
588 | * | ||
589 | * Change page attributes of a page in the direct mapping. This is a variant | ||
590 | * of change_page_attr() that also works on memory holes that do not have | ||
591 | * mem_map entry (pfn_valid() is false). | ||
592 | * | ||
593 | * See change_page_attr() documentation for more details. | ||
594 | * | ||
595 | * Modules and drivers should use the set_memory_* APIs instead. | ||
596 | */ | ||
597 | static int change_page_attr_addr(struct cpa_data *cpa) | ||
598 | { | 595 | { |
599 | int err; | 596 | struct cpa_data alias_cpa; |
600 | unsigned long address = cpa->vaddr; | 597 | int ret; |
601 | 598 | ||
602 | #ifdef CONFIG_X86_64 | 599 | if (cpa->pfn > max_pfn_mapped) |
603 | unsigned long phys_addr = __pa(address); | 600 | return 0; |
604 | 601 | ||
605 | /* | 602 | alias_cpa = *cpa; |
606 | * If we are inside the high mapped kernel range, then we | 603 | alias_cpa.vaddr = (unsigned long) __va(cpa->pfn << PAGE_SHIFT); |
607 | * fixup the low mapping first. __va() returns the virtual | ||
608 | * address in the linear mapping: | ||
609 | */ | ||
610 | if (within(address, HIGH_MAP_START, HIGH_MAP_END)) | ||
611 | address = (unsigned long) __va(phys_addr); | ||
612 | #endif | ||
613 | 604 | ||
614 | err = __change_page_attr(address, cpa); | 605 | ret = __change_page_attr_set_clr(&alias_cpa, 0); |
615 | if (err) | ||
616 | return err; | ||
617 | 606 | ||
618 | #ifdef CONFIG_X86_64 | 607 | #ifdef CONFIG_X86_64 |
608 | if (ret) | ||
609 | return ret; | ||
619 | /* | 610 | /* |
620 | * If the physical address is inside the kernel map, we need | 611 | * If the physical address is inside the kernel map, we need |
621 | * to touch the high mapped kernel as well: | 612 | * to touch the high mapped kernel as well: |
622 | */ | 613 | */ |
623 | if (within(phys_addr, 0, KERNEL_TEXT_SIZE)) { | 614 | if (!within(cpa->pfn, highmap_start_pfn(), highmap_end_pfn())) |
624 | /* | 615 | return 0; |
625 | * Calc the high mapping address. See __phys_addr() | ||
626 | * for the non obvious details. | ||
627 | * | ||
628 | * Note that NX and other required permissions are | ||
629 | * checked in static_protections(). | ||
630 | */ | ||
631 | address = phys_addr + HIGH_MAP_START - phys_base; | ||
632 | 616 | ||
633 | /* | 617 | alias_cpa = *cpa; |
634 | * Our high aliases are imprecise, because we check | 618 | alias_cpa.vaddr = |
635 | * everything between 0 and KERNEL_TEXT_SIZE, so do | 619 | (cpa->pfn << PAGE_SHIFT) + __START_KERNEL_map - phys_base; |
636 | * not propagate lookup failures back to users: | 620 | |
637 | */ | 621 | /* |
638 | __change_page_attr(address, cpa); | 622 | * The high mapping range is imprecise, so ignore the return value. |
639 | } | 623 | */ |
624 | __change_page_attr_set_clr(&alias_cpa, 0); | ||
640 | #endif | 625 | #endif |
641 | return err; | 626 | return ret; |
642 | } | 627 | } |
643 | 628 | ||
644 | static int __change_page_attr_set_clr(struct cpa_data *cpa) | 629 | static int __change_page_attr_set_clr(struct cpa_data *cpa, int checkalias) |
645 | { | 630 | { |
646 | int ret, numpages = cpa->numpages; | 631 | int ret, numpages = cpa->numpages; |
647 | 632 | ||
@@ -651,10 +636,17 @@ static int __change_page_attr_set_clr(struct cpa_data *cpa) | |||
651 | * preservation check. | 636 | * preservation check. |
652 | */ | 637 | */ |
653 | cpa->numpages = numpages; | 638 | cpa->numpages = numpages; |
654 | ret = change_page_attr_addr(cpa); | 639 | |
640 | ret = __change_page_attr(cpa, checkalias); | ||
655 | if (ret) | 641 | if (ret) |
656 | return ret; | 642 | return ret; |
657 | 643 | ||
644 | if (checkalias) { | ||
645 | ret = cpa_process_alias(cpa); | ||
646 | if (ret) | ||
647 | return ret; | ||
648 | } | ||
649 | |||
658 | /* | 650 | /* |
659 | * Adjust the number of pages with the result of the | 651 | * Adjust the number of pages with the result of the |
660 | * CPA operation. Either a large page has been | 652 | * CPA operation. Either a large page has been |
@@ -703,7 +695,7 @@ static int change_page_attr_set_clr(unsigned long addr, int numpages, | |||
703 | cpa.mask_clr = mask_clr; | 695 | cpa.mask_clr = mask_clr; |
704 | cpa.flushtlb = 0; | 696 | cpa.flushtlb = 0; |
705 | 697 | ||
706 | ret = __change_page_attr_set_clr(&cpa); | 698 | ret = __change_page_attr_set_clr(&cpa, 1); |
707 | 699 | ||
708 | /* | 700 | /* |
709 | * Check whether we really changed something: | 701 | * Check whether we really changed something: |
@@ -841,7 +833,7 @@ static int __set_pages_p(struct page *page, int numpages) | |||
841 | .mask_set = __pgprot(_PAGE_PRESENT | _PAGE_RW), | 833 | .mask_set = __pgprot(_PAGE_PRESENT | _PAGE_RW), |
842 | .mask_clr = __pgprot(0)}; | 834 | .mask_clr = __pgprot(0)}; |
843 | 835 | ||
844 | return __change_page_attr_set_clr(&cpa); | 836 | return __change_page_attr_set_clr(&cpa, 1); |
845 | } | 837 | } |
846 | 838 | ||
847 | static int __set_pages_np(struct page *page, int numpages) | 839 | static int __set_pages_np(struct page *page, int numpages) |
@@ -851,7 +843,7 @@ static int __set_pages_np(struct page *page, int numpages) | |||
851 | .mask_set = __pgprot(0), | 843 | .mask_set = __pgprot(0), |
852 | .mask_clr = __pgprot(_PAGE_PRESENT | _PAGE_RW)}; | 844 | .mask_clr = __pgprot(_PAGE_PRESENT | _PAGE_RW)}; |
853 | 845 | ||
854 | return __change_page_attr_set_clr(&cpa); | 846 | return __change_page_attr_set_clr(&cpa, 1); |
855 | } | 847 | } |
856 | 848 | ||
857 | void kernel_map_pages(struct page *page, int numpages, int enable) | 849 | void kernel_map_pages(struct page *page, int numpages, int enable) |