diff options
author | Kirill A. Shutemov <kirill.shutemov@linux.intel.com> | 2013-11-14 17:30:48 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2013-11-14 19:32:14 -0500 |
commit | e1f56c89b040134add93f686931cc266541d239a (patch) | |
tree | 36efea76d1803cca3e2a7716850d872db70afd8f /mm | |
parent | 57c1ffcefb5acb3c8b5f8436c325a6bdbd8e9c78 (diff) |
mm: convert mm->nr_ptes to atomic_long_t
With split page table lock for PMD level we can't hold mm->page_table_lock
while updating nr_ptes.
Let's convert it to atomic_long_t to avoid races.
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Tested-by: Alex Thorlton <athorlton@sgi.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Cc: "Eric W . Biederman" <ebiederm@xmission.com>
Cc: "Paul E . McKenney" <paulmck@linux.vnet.ibm.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: Dave Hansen <dave.hansen@intel.com>
Cc: Dave Jones <davej@redhat.com>
Cc: David Howells <dhowells@redhat.com>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Kees Cook <keescook@chromium.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Michael Kerrisk <mtk.manpages@gmail.com>
Cc: Oleg Nesterov <oleg@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Rik van Riel <riel@redhat.com>
Cc: Robin Holt <robinmholt@gmail.com>
Cc: Sedat Dilek <sedat.dilek@gmail.com>
Cc: Srikar Dronamraju <srikar@linux.vnet.ibm.com>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Hugh Dickins <hughd@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/huge_memory.c | 10 | ||||
-rw-r--r-- | mm/memory.c | 4 | ||||
-rw-r--r-- | mm/mmap.c | 3 | ||||
-rw-r--r-- | mm/oom_kill.c | 6 |
4 files changed, 12 insertions, 11 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 0556c6a44959..e5b2d316be2e 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c | |||
@@ -738,7 +738,7 @@ static int __do_huge_pmd_anonymous_page(struct mm_struct *mm, | |||
738 | pgtable_trans_huge_deposit(mm, pmd, pgtable); | 738 | pgtable_trans_huge_deposit(mm, pmd, pgtable); |
739 | set_pmd_at(mm, haddr, pmd, entry); | 739 | set_pmd_at(mm, haddr, pmd, entry); |
740 | add_mm_counter(mm, MM_ANONPAGES, HPAGE_PMD_NR); | 740 | add_mm_counter(mm, MM_ANONPAGES, HPAGE_PMD_NR); |
741 | mm->nr_ptes++; | 741 | atomic_long_inc(&mm->nr_ptes); |
742 | spin_unlock(&mm->page_table_lock); | 742 | spin_unlock(&mm->page_table_lock); |
743 | } | 743 | } |
744 | 744 | ||
@@ -771,7 +771,7 @@ static bool set_huge_zero_page(pgtable_t pgtable, struct mm_struct *mm, | |||
771 | entry = pmd_mkhuge(entry); | 771 | entry = pmd_mkhuge(entry); |
772 | pgtable_trans_huge_deposit(mm, pmd, pgtable); | 772 | pgtable_trans_huge_deposit(mm, pmd, pgtable); |
773 | set_pmd_at(mm, haddr, pmd, entry); | 773 | set_pmd_at(mm, haddr, pmd, entry); |
774 | mm->nr_ptes++; | 774 | atomic_long_inc(&mm->nr_ptes); |
775 | return true; | 775 | return true; |
776 | } | 776 | } |
777 | 777 | ||
@@ -896,7 +896,7 @@ int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm, | |||
896 | pmd = pmd_mkold(pmd_wrprotect(pmd)); | 896 | pmd = pmd_mkold(pmd_wrprotect(pmd)); |
897 | pgtable_trans_huge_deposit(dst_mm, dst_pmd, pgtable); | 897 | pgtable_trans_huge_deposit(dst_mm, dst_pmd, pgtable); |
898 | set_pmd_at(dst_mm, addr, dst_pmd, pmd); | 898 | set_pmd_at(dst_mm, addr, dst_pmd, pmd); |
899 | dst_mm->nr_ptes++; | 899 | atomic_long_inc(&dst_mm->nr_ptes); |
900 | 900 | ||
901 | ret = 0; | 901 | ret = 0; |
902 | out_unlock: | 902 | out_unlock: |
@@ -1392,7 +1392,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, | |||
1392 | tlb_remove_pmd_tlb_entry(tlb, pmd, addr); | 1392 | tlb_remove_pmd_tlb_entry(tlb, pmd, addr); |
1393 | pgtable = pgtable_trans_huge_withdraw(tlb->mm, pmd); | 1393 | pgtable = pgtable_trans_huge_withdraw(tlb->mm, pmd); |
1394 | if (is_huge_zero_pmd(orig_pmd)) { | 1394 | if (is_huge_zero_pmd(orig_pmd)) { |
1395 | tlb->mm->nr_ptes--; | 1395 | atomic_long_dec(&tlb->mm->nr_ptes); |
1396 | spin_unlock(&tlb->mm->page_table_lock); | 1396 | spin_unlock(&tlb->mm->page_table_lock); |
1397 | put_huge_zero_page(); | 1397 | put_huge_zero_page(); |
1398 | } else { | 1398 | } else { |
@@ -1401,7 +1401,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma, | |||
1401 | VM_BUG_ON(page_mapcount(page) < 0); | 1401 | VM_BUG_ON(page_mapcount(page) < 0); |
1402 | add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR); | 1402 | add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR); |
1403 | VM_BUG_ON(!PageHead(page)); | 1403 | VM_BUG_ON(!PageHead(page)); |
1404 | tlb->mm->nr_ptes--; | 1404 | atomic_long_dec(&tlb->mm->nr_ptes); |
1405 | spin_unlock(&tlb->mm->page_table_lock); | 1405 | spin_unlock(&tlb->mm->page_table_lock); |
1406 | tlb_remove_page(tlb, page); | 1406 | tlb_remove_page(tlb, page); |
1407 | } | 1407 | } |
diff --git a/mm/memory.c b/mm/memory.c index bf8665849a5f..0b5a93a49f27 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
@@ -382,7 +382,7 @@ static void free_pte_range(struct mmu_gather *tlb, pmd_t *pmd, | |||
382 | pgtable_t token = pmd_pgtable(*pmd); | 382 | pgtable_t token = pmd_pgtable(*pmd); |
383 | pmd_clear(pmd); | 383 | pmd_clear(pmd); |
384 | pte_free_tlb(tlb, token, addr); | 384 | pte_free_tlb(tlb, token, addr); |
385 | tlb->mm->nr_ptes--; | 385 | atomic_long_dec(&tlb->mm->nr_ptes); |
386 | } | 386 | } |
387 | 387 | ||
388 | static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, | 388 | static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, |
@@ -573,7 +573,7 @@ int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma, | |||
573 | spin_lock(&mm->page_table_lock); | 573 | spin_lock(&mm->page_table_lock); |
574 | wait_split_huge_page = 0; | 574 | wait_split_huge_page = 0; |
575 | if (likely(pmd_none(*pmd))) { /* Has another populated it ? */ | 575 | if (likely(pmd_none(*pmd))) { /* Has another populated it ? */ |
576 | mm->nr_ptes++; | 576 | atomic_long_inc(&mm->nr_ptes); |
577 | pmd_populate(mm, pmd, new); | 577 | pmd_populate(mm, pmd, new); |
578 | new = NULL; | 578 | new = NULL; |
579 | } else if (unlikely(pmd_trans_splitting(*pmd))) | 579 | } else if (unlikely(pmd_trans_splitting(*pmd))) |
@@ -2724,7 +2724,8 @@ void exit_mmap(struct mm_struct *mm) | |||
2724 | } | 2724 | } |
2725 | vm_unacct_memory(nr_accounted); | 2725 | vm_unacct_memory(nr_accounted); |
2726 | 2726 | ||
2727 | WARN_ON(mm->nr_ptes > (FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT); | 2727 | WARN_ON(atomic_long_read(&mm->nr_ptes) > |
2728 | (FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT); | ||
2728 | } | 2729 | } |
2729 | 2730 | ||
2730 | /* Insert vm structure into process list sorted by address | 2731 | /* Insert vm structure into process list sorted by address |
diff --git a/mm/oom_kill.c b/mm/oom_kill.c index 6738c47f1f72..1e4a600a6163 100644 --- a/mm/oom_kill.c +++ b/mm/oom_kill.c | |||
@@ -161,7 +161,7 @@ unsigned long oom_badness(struct task_struct *p, struct mem_cgroup *memcg, | |||
161 | * The baseline for the badness score is the proportion of RAM that each | 161 | * The baseline for the badness score is the proportion of RAM that each |
162 | * task's rss, pagetable and swap space use. | 162 | * task's rss, pagetable and swap space use. |
163 | */ | 163 | */ |
164 | points = get_mm_rss(p->mm) + p->mm->nr_ptes + | 164 | points = get_mm_rss(p->mm) + atomic_long_read(&p->mm->nr_ptes) + |
165 | get_mm_counter(p->mm, MM_SWAPENTS); | 165 | get_mm_counter(p->mm, MM_SWAPENTS); |
166 | task_unlock(p); | 166 | task_unlock(p); |
167 | 167 | ||
@@ -364,10 +364,10 @@ static void dump_tasks(const struct mem_cgroup *memcg, const nodemask_t *nodemas | |||
364 | continue; | 364 | continue; |
365 | } | 365 | } |
366 | 366 | ||
367 | pr_info("[%5d] %5d %5d %8lu %8lu %7lu %8lu %5hd %s\n", | 367 | pr_info("[%5d] %5d %5d %8lu %8lu %7ld %8lu %5hd %s\n", |
368 | task->pid, from_kuid(&init_user_ns, task_uid(task)), | 368 | task->pid, from_kuid(&init_user_ns, task_uid(task)), |
369 | task->tgid, task->mm->total_vm, get_mm_rss(task->mm), | 369 | task->tgid, task->mm->total_vm, get_mm_rss(task->mm), |
370 | task->mm->nr_ptes, | 370 | atomic_long_read(&task->mm->nr_ptes), |
371 | get_mm_counter(task->mm, MM_SWAPENTS), | 371 | get_mm_counter(task->mm, MM_SWAPENTS), |
372 | task->signal->oom_score_adj, task->comm); | 372 | task->signal->oom_score_adj, task->comm); |
373 | task_unlock(task); | 373 | task_unlock(task); |