summaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorPeter Xu <peterx@redhat.com>2018-12-21 17:30:50 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2018-12-21 17:51:18 -0500
commit2e83ee1d8694a61d0d95a5b694f2e61e8dde8627 (patch)
tree3b0948cfda175594379af6ec874c98f3b9ec338d /mm
parent2830bf6f05fb3e05bc4743274b806c821807a684 (diff)
mm: thp: fix flags for pmd migration when split
When splitting a huge migrating PMD, we'll transfer all the existing PMD bits and apply them again onto the small PTEs. However we are fetching the bits unconditionally via pmd_soft_dirty(), pmd_write() or pmd_yound() while actually they don't make sense at all when it's a migration entry. Fix them up. Since at it, drop the ifdef together as not needed. Note that if my understanding is correct about the problem then if without the patch there is chance to lose some of the dirty bits in the migrating pmd pages (on x86_64 we're fetching bit 11 which is part of swap offset instead of bit 2) and it could potentially corrupt the memory of an userspace program which depends on the dirty bit. Link: http://lkml.kernel.org/r/20181213051510.20306-1-peterx@redhat.com Signed-off-by: Peter Xu <peterx@redhat.com> Reviewed-by: Konstantin Khlebnikov <khlebnikov@yandex-team.ru> Reviewed-by: William Kucharski <william.kucharski@oracle.com> Acked-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com> Cc: Andrea Arcangeli <aarcange@redhat.com> Cc: Matthew Wilcox <willy@infradead.org> Cc: Michal Hocko <mhocko@suse.com> Cc: Dave Jiang <dave.jiang@intel.com> Cc: "Aneesh Kumar K.V" <aneesh.kumar@linux.vnet.ibm.com> Cc: Souptick Joarder <jrdr.linux@gmail.com> Cc: Konstantin Khlebnikov <khlebnikov@yandex-team.ru> Cc: Zi Yan <zi.yan@cs.rutgers.edu> Cc: <stable@vger.kernel.org> [4.14+] Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/huge_memory.c20
1 files changed, 11 insertions, 9 deletions
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 5da55b38b1b7..e84a10b0d310 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -2144,23 +2144,25 @@ static void __split_huge_pmd_locked(struct vm_area_struct *vma, pmd_t *pmd,
2144 */ 2144 */
2145 old_pmd = pmdp_invalidate(vma, haddr, pmd); 2145 old_pmd = pmdp_invalidate(vma, haddr, pmd);
2146 2146
2147#ifdef CONFIG_ARCH_ENABLE_THP_MIGRATION
2148 pmd_migration = is_pmd_migration_entry(old_pmd); 2147 pmd_migration = is_pmd_migration_entry(old_pmd);
2149 if (pmd_migration) { 2148 if (unlikely(pmd_migration)) {
2150 swp_entry_t entry; 2149 swp_entry_t entry;
2151 2150
2152 entry = pmd_to_swp_entry(old_pmd); 2151 entry = pmd_to_swp_entry(old_pmd);
2153 page = pfn_to_page(swp_offset(entry)); 2152 page = pfn_to_page(swp_offset(entry));
2154 } else 2153 write = is_write_migration_entry(entry);
2155#endif 2154 young = false;
2155 soft_dirty = pmd_swp_soft_dirty(old_pmd);
2156 } else {
2156 page = pmd_page(old_pmd); 2157 page = pmd_page(old_pmd);
2158 if (pmd_dirty(old_pmd))
2159 SetPageDirty(page);
2160 write = pmd_write(old_pmd);
2161 young = pmd_young(old_pmd);
2162 soft_dirty = pmd_soft_dirty(old_pmd);
2163 }
2157 VM_BUG_ON_PAGE(!page_count(page), page); 2164 VM_BUG_ON_PAGE(!page_count(page), page);
2158 page_ref_add(page, HPAGE_PMD_NR - 1); 2165 page_ref_add(page, HPAGE_PMD_NR - 1);
2159 if (pmd_dirty(old_pmd))
2160 SetPageDirty(page);
2161 write = pmd_write(old_pmd);
2162 young = pmd_young(old_pmd);
2163 soft_dirty = pmd_soft_dirty(old_pmd);
2164 2166
2165 /* 2167 /*
2166 * Withdraw the table only after we mark the pmd entry invalid. 2168 * Withdraw the table only after we mark the pmd entry invalid.