aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorKirill A. Shutemov <kirill.shutemov@linux.intel.com>2013-11-14 17:30:48 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2013-11-14 19:32:14 -0500
commite1f56c89b040134add93f686931cc266541d239a (patch)
tree36efea76d1803cca3e2a7716850d872db70afd8f /mm
parent57c1ffcefb5acb3c8b5f8436c325a6bdbd8e9c78 (diff)
mm: convert mm->nr_ptes to atomic_long_t
With split page table lock for PMD level we can't hold mm->page_table_lock while updating nr_ptes. Let's convert it to atomic_long_t to avoid races. Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Tested-by: Alex Thorlton <athorlton@sgi.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com> Cc: "Eric W . Biederman" <ebiederm@xmission.com> Cc: "Paul E . McKenney" <paulmck@linux.vnet.ibm.com> Cc: Al Viro <viro@zeniv.linux.org.uk> Cc: Andi Kleen <ak@linux.intel.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Dave Hansen <dave.hansen@intel.com> Cc: Dave Jones <davej@redhat.com> Cc: David Howells <dhowells@redhat.com> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Kees Cook <keescook@chromium.org> Cc: Mel Gorman <mgorman@suse.de> Cc: Michael Kerrisk <mtk.manpages@gmail.com> Cc: Oleg Nesterov <oleg@redhat.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Rik van Riel <riel@redhat.com> Cc: Robin Holt <robinmholt@gmail.com> Cc: Sedat Dilek <sedat.dilek@gmail.com> Cc: Srikar Dronamraju <srikar@linux.vnet.ibm.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Hugh Dickins <hughd@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/huge_memory.c10
-rw-r--r--mm/memory.c4
-rw-r--r--mm/mmap.c3
-rw-r--r--mm/oom_kill.c6
4 files changed, 12 insertions, 11 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 0556c6a44959..e5b2d316be2e 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -738,7 +738,7 @@ static int __do_huge_pmd_anonymous_page(struct mm_struct *mm,
738 pgtable_trans_huge_deposit(mm, pmd, pgtable); 738 pgtable_trans_huge_deposit(mm, pmd, pgtable);
739 set_pmd_at(mm, haddr, pmd, entry); 739 set_pmd_at(mm, haddr, pmd, entry);
740 add_mm_counter(mm, MM_ANONPAGES, HPAGE_PMD_NR); 740 add_mm_counter(mm, MM_ANONPAGES, HPAGE_PMD_NR);
741 mm->nr_ptes++; 741 atomic_long_inc(&mm->nr_ptes);
742 spin_unlock(&mm->page_table_lock); 742 spin_unlock(&mm->page_table_lock);
743 } 743 }
744 744
@@ -771,7 +771,7 @@ static bool set_huge_zero_page(pgtable_t pgtable, struct mm_struct *mm,
771 entry = pmd_mkhuge(entry); 771 entry = pmd_mkhuge(entry);
772 pgtable_trans_huge_deposit(mm, pmd, pgtable); 772 pgtable_trans_huge_deposit(mm, pmd, pgtable);
773 set_pmd_at(mm, haddr, pmd, entry); 773 set_pmd_at(mm, haddr, pmd, entry);
774 mm->nr_ptes++; 774 atomic_long_inc(&mm->nr_ptes);
775 return true; 775 return true;
776} 776}
777 777
@@ -896,7 +896,7 @@ int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm,
896 pmd = pmd_mkold(pmd_wrprotect(pmd)); 896 pmd = pmd_mkold(pmd_wrprotect(pmd));
897 pgtable_trans_huge_deposit(dst_mm, dst_pmd, pgtable); 897 pgtable_trans_huge_deposit(dst_mm, dst_pmd, pgtable);
898 set_pmd_at(dst_mm, addr, dst_pmd, pmd); 898 set_pmd_at(dst_mm, addr, dst_pmd, pmd);
899 dst_mm->nr_ptes++; 899 atomic_long_inc(&dst_mm->nr_ptes);
900 900
901 ret = 0; 901 ret = 0;
902out_unlock: 902out_unlock:
@@ -1392,7 +1392,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
1392 tlb_remove_pmd_tlb_entry(tlb, pmd, addr); 1392 tlb_remove_pmd_tlb_entry(tlb, pmd, addr);
1393 pgtable = pgtable_trans_huge_withdraw(tlb->mm, pmd); 1393 pgtable = pgtable_trans_huge_withdraw(tlb->mm, pmd);
1394 if (is_huge_zero_pmd(orig_pmd)) { 1394 if (is_huge_zero_pmd(orig_pmd)) {
1395 tlb->mm->nr_ptes--; 1395 atomic_long_dec(&tlb->mm->nr_ptes);
1396 spin_unlock(&tlb->mm->page_table_lock); 1396 spin_unlock(&tlb->mm->page_table_lock);
1397 put_huge_zero_page(); 1397 put_huge_zero_page();
1398 } else { 1398 } else {
@@ -1401,7 +1401,7 @@ int zap_huge_pmd(struct mmu_gather *tlb, struct vm_area_struct *vma,
1401 VM_BUG_ON(page_mapcount(page) < 0); 1401 VM_BUG_ON(page_mapcount(page) < 0);
1402 add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR); 1402 add_mm_counter(tlb->mm, MM_ANONPAGES, -HPAGE_PMD_NR);
1403 VM_BUG_ON(!PageHead(page)); 1403 VM_BUG_ON(!PageHead(page));
1404 tlb->mm->nr_ptes--; 1404 atomic_long_dec(&tlb->mm->nr_ptes);
1405 spin_unlock(&tlb->mm->page_table_lock); 1405 spin_unlock(&tlb->mm->page_table_lock);
1406 tlb_remove_page(tlb, page); 1406 tlb_remove_page(tlb, page);
1407 } 1407 }
diff --git a/mm/memory.c b/mm/memory.c
index bf8665849a5f..0b5a93a49f27 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -382,7 +382,7 @@ static void free_pte_range(struct mmu_gather *tlb, pmd_t *pmd,
382 pgtable_t token = pmd_pgtable(*pmd); 382 pgtable_t token = pmd_pgtable(*pmd);
383 pmd_clear(pmd); 383 pmd_clear(pmd);
384 pte_free_tlb(tlb, token, addr); 384 pte_free_tlb(tlb, token, addr);
385 tlb->mm->nr_ptes--; 385 atomic_long_dec(&tlb->mm->nr_ptes);
386} 386}
387 387
388static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud, 388static inline void free_pmd_range(struct mmu_gather *tlb, pud_t *pud,
@@ -573,7 +573,7 @@ int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
573 spin_lock(&mm->page_table_lock); 573 spin_lock(&mm->page_table_lock);
574 wait_split_huge_page = 0; 574 wait_split_huge_page = 0;
575 if (likely(pmd_none(*pmd))) { /* Has another populated it ? */ 575 if (likely(pmd_none(*pmd))) { /* Has another populated it ? */
576 mm->nr_ptes++; 576 atomic_long_inc(&mm->nr_ptes);
577 pmd_populate(mm, pmd, new); 577 pmd_populate(mm, pmd, new);
578 new = NULL; 578 new = NULL;
579 } else if (unlikely(pmd_trans_splitting(*pmd))) 579 } else if (unlikely(pmd_trans_splitting(*pmd)))
diff --git a/mm/mmap.c b/mm/mmap.c
index 5a6baddde15d..834b2d785f1e 100644
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -2724,7 +2724,8 @@ void exit_mmap(struct mm_struct *mm)
2724 } 2724 }
2725 vm_unacct_memory(nr_accounted); 2725 vm_unacct_memory(nr_accounted);
2726 2726
2727 WARN_ON(mm->nr_ptes > (FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT); 2727 WARN_ON(atomic_long_read(&mm->nr_ptes) >
2728 (FIRST_USER_ADDRESS+PMD_SIZE-1)>>PMD_SHIFT);
2728} 2729}
2729 2730
2730/* Insert vm structure into process list sorted by address 2731/* Insert vm structure into process list sorted by address
diff --git a/mm/oom_kill.c b/mm/oom_kill.c
index 6738c47f1f72..1e4a600a6163 100644
--- a/mm/oom_kill.c
+++ b/mm/oom_kill.c
@@ -161,7 +161,7 @@ unsigned long oom_badness(struct task_struct *p, struct mem_cgroup *memcg,
161 * The baseline for the badness score is the proportion of RAM that each 161 * The baseline for the badness score is the proportion of RAM that each
162 * task's rss, pagetable and swap space use. 162 * task's rss, pagetable and swap space use.
163 */ 163 */
164 points = get_mm_rss(p->mm) + p->mm->nr_ptes + 164 points = get_mm_rss(p->mm) + atomic_long_read(&p->mm->nr_ptes) +
165 get_mm_counter(p->mm, MM_SWAPENTS); 165 get_mm_counter(p->mm, MM_SWAPENTS);
166 task_unlock(p); 166 task_unlock(p);
167 167
@@ -364,10 +364,10 @@ static void dump_tasks(const struct mem_cgroup *memcg, const nodemask_t *nodemas
364 continue; 364 continue;
365 } 365 }
366 366
367 pr_info("[%5d] %5d %5d %8lu %8lu %7lu %8lu %5hd %s\n", 367 pr_info("[%5d] %5d %5d %8lu %8lu %7ld %8lu %5hd %s\n",
368 task->pid, from_kuid(&init_user_ns, task_uid(task)), 368 task->pid, from_kuid(&init_user_ns, task_uid(task)),
369 task->tgid, task->mm->total_vm, get_mm_rss(task->mm), 369 task->tgid, task->mm->total_vm, get_mm_rss(task->mm),
370 task->mm->nr_ptes, 370 atomic_long_read(&task->mm->nr_ptes),
371 get_mm_counter(task->mm, MM_SWAPENTS), 371 get_mm_counter(task->mm, MM_SWAPENTS),
372 task->signal->oom_score_adj, task->comm); 372 task->signal->oom_score_adj, task->comm);
373 task_unlock(task); 373 task_unlock(task);