aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorWilliam Lee Irwin III <wli@holomorphy.com>2005-05-24 22:31:09 -0400
committerLinus Torvalds <torvalds@ppc970.osdl.org>2005-05-24 23:08:13 -0400
commitcafdd8ba08935d9b161bb781851dc4c0e6f70427 (patch)
treeee6a5305902cfd6abf818e398c538694db24dcb7 /mm
parentc33880aaddbbab1ccf36f4457ed1090621f2e39a (diff)
[PATCH] try_to_unmap_cluster() passes out-of-bounds pte to pte_unmap()
try_to_unmap_cluster() does: for (pte = pte_offset_map(pmd, address); address < end; pte++, address += PAGE_SIZE) { ... } pte_unmap(pte); It may take a little staring to notice, but pte can actually fall off the end of the pte page in this iteration, which makes life difficult for kmap_atomic() and the users not expecting it to BUG(). Of course, we're somewhat lucky in that arithmetic elsewhere in the function guarantees that at least one iteration is made, lest this force larger rearrangements to be made. This issue and patch also apply to non-mm mainline and with trivial adjustments, at least two related kernels. Discovered during internal testing at Oracle. Signed-off-by: William Irwin <wli@holomorphy.com> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/rmap.c6
1 files changed, 3 insertions, 3 deletions
diff --git a/mm/rmap.c b/mm/rmap.c
index a6203b4e1278..9827409eb7c7 100644
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -626,7 +626,7 @@ static void try_to_unmap_cluster(unsigned long cursor,
626 pgd_t *pgd; 626 pgd_t *pgd;
627 pud_t *pud; 627 pud_t *pud;
628 pmd_t *pmd; 628 pmd_t *pmd;
629 pte_t *pte; 629 pte_t *pte, *original_pte;
630 pte_t pteval; 630 pte_t pteval;
631 struct page *page; 631 struct page *page;
632 unsigned long address; 632 unsigned long address;
@@ -658,7 +658,7 @@ static void try_to_unmap_cluster(unsigned long cursor,
658 if (!pmd_present(*pmd)) 658 if (!pmd_present(*pmd))
659 goto out_unlock; 659 goto out_unlock;
660 660
661 for (pte = pte_offset_map(pmd, address); 661 for (original_pte = pte = pte_offset_map(pmd, address);
662 address < end; pte++, address += PAGE_SIZE) { 662 address < end; pte++, address += PAGE_SIZE) {
663 663
664 if (!pte_present(*pte)) 664 if (!pte_present(*pte))
@@ -694,7 +694,7 @@ static void try_to_unmap_cluster(unsigned long cursor,
694 (*mapcount)--; 694 (*mapcount)--;
695 } 695 }
696 696
697 pte_unmap(pte); 697 pte_unmap(original_pte);
698out_unlock: 698out_unlock:
699 spin_unlock(&mm->page_table_lock); 699 spin_unlock(&mm->page_table_lock);
700} 700}