aboutsummaryrefslogtreecommitdiffstats
path: root/mm/memory.c
diff options
context:
space:
mode:
authorRik van Riel <riel@redhat.com>2010-03-05 16:42:09 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2010-03-06 14:26:26 -0500
commitc44b674323f4a2480dbeb65d4b487fa5f06f49e0 (patch)
treeb753050e6752eb2fc961ad3ea5dfdf88ef88364d /mm/memory.c
parent033a64b56aed798991de18d226085dfb1ccd858d (diff)
rmap: move exclusively owned pages to own anon_vma in do_wp_page()
When the parent process breaks the COW on a page, both the original which is mapped at child and the new page which is mapped parent end up in that same anon_vma. Generally this won't be a problem, but for some workloads it could preserve the O(N) rmap scanning complexity. A simple fix is to ensure that, when a page which is mapped child gets reused in do_wp_page, because we already are the exclusive owner, the page gets moved to our own exclusive child's anon_vma. Signed-off-by: Rik van Riel <riel@redhat.com> Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com> Cc: Larry Woodman <lwoodman@redhat.com> Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com> Reviewed-by: Minchan Kim <minchan.kim@gmail.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Hugh Dickins <hugh.dickins@tiscali.co.uk> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/memory.c')
-rw-r--r--mm/memory.c7
1 files changed, 7 insertions, 0 deletions
diff --git a/mm/memory.c b/mm/memory.c
index dc785b438d70..d1153e37e9ba 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -2138,6 +2138,13 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma,
2138 page_cache_release(old_page); 2138 page_cache_release(old_page);
2139 } 2139 }
2140 reuse = reuse_swap_page(old_page); 2140 reuse = reuse_swap_page(old_page);
2141 if (reuse)
2142 /*
2143 * The page is all ours. Move it to our anon_vma so
2144 * the rmap code will not search our parent or siblings.
2145 * Protected against the rmap code by the page lock.
2146 */
2147 page_move_anon_rmap(old_page, vma, address);
2141 unlock_page(old_page); 2148 unlock_page(old_page);
2142 } else if (unlikely((vma->vm_flags & (VM_WRITE|VM_SHARED)) == 2149 } else if (unlikely((vma->vm_flags & (VM_WRITE|VM_SHARED)) ==
2143 (VM_WRITE|VM_SHARED))) { 2150 (VM_WRITE|VM_SHARED))) {