diff options
Diffstat (limited to 'mm/memory.c')
-rw-r--r-- | mm/memory.c | 150 |
1 files changed, 90 insertions, 60 deletions
diff --git a/mm/memory.c b/mm/memory.c index a40e4b1cee4f..24ba688876d6 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
@@ -1219,6 +1219,30 @@ int remap_pfn_range(struct vm_area_struct *vma, unsigned long addr, | |||
1219 | EXPORT_SYMBOL(remap_pfn_range); | 1219 | EXPORT_SYMBOL(remap_pfn_range); |
1220 | 1220 | ||
1221 | /* | 1221 | /* |
1222 | * handle_pte_fault chooses page fault handler according to an entry | ||
1223 | * which was read non-atomically. Before making any commitment, on | ||
1224 | * those architectures or configurations (e.g. i386 with PAE) which | ||
1225 | * might give a mix of unmatched parts, do_swap_page and do_file_page | ||
1226 | * must check under lock before unmapping the pte and proceeding | ||
1227 | * (but do_wp_page is only called after already making such a check; | ||
1228 | * and do_anonymous_page and do_no_page can safely check later on). | ||
1229 | */ | ||
1230 | static inline int pte_unmap_same(struct mm_struct *mm, | ||
1231 | pte_t *page_table, pte_t orig_pte) | ||
1232 | { | ||
1233 | int same = 1; | ||
1234 | #if defined(CONFIG_SMP) || defined(CONFIG_PREEMPT) | ||
1235 | if (sizeof(pte_t) > sizeof(unsigned long)) { | ||
1236 | spin_lock(&mm->page_table_lock); | ||
1237 | same = pte_same(*page_table, orig_pte); | ||
1238 | spin_unlock(&mm->page_table_lock); | ||
1239 | } | ||
1240 | #endif | ||
1241 | pte_unmap(page_table); | ||
1242 | return same; | ||
1243 | } | ||
1244 | |||
1245 | /* | ||
1222 | * Do pte_mkwrite, but only if the vma says VM_WRITE. We do this when | 1246 | * Do pte_mkwrite, but only if the vma says VM_WRITE. We do this when |
1223 | * servicing faults for write access. In the normal case, do always want | 1247 | * servicing faults for write access. In the normal case, do always want |
1224 | * pte_mkwrite. But get_user_pages can cause write faults for mappings | 1248 | * pte_mkwrite. But get_user_pages can cause write faults for mappings |
@@ -1245,12 +1269,13 @@ static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) | |||
1245 | * change only once the write actually happens. This avoids a few races, | 1269 | * change only once the write actually happens. This avoids a few races, |
1246 | * and potentially makes it more efficient. | 1270 | * and potentially makes it more efficient. |
1247 | * | 1271 | * |
1248 | * We hold the mm semaphore and the page_table_lock on entry and exit | 1272 | * We enter with non-exclusive mmap_sem (to exclude vma changes, |
1249 | * with the page_table_lock released. | 1273 | * but allow concurrent faults), with pte both mapped and locked. |
1274 | * We return with mmap_sem still held, but pte unmapped and unlocked. | ||
1250 | */ | 1275 | */ |
1251 | static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | 1276 | static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, |
1252 | unsigned long address, pte_t *page_table, pmd_t *pmd, | 1277 | unsigned long address, pte_t *page_table, pmd_t *pmd, |
1253 | pte_t orig_pte) | 1278 | spinlock_t *ptl, pte_t orig_pte) |
1254 | { | 1279 | { |
1255 | struct page *old_page, *new_page; | 1280 | struct page *old_page, *new_page; |
1256 | unsigned long pfn = pte_pfn(orig_pte); | 1281 | unsigned long pfn = pte_pfn(orig_pte); |
@@ -1288,8 +1313,7 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1288 | * Ok, we need to copy. Oh, well.. | 1313 | * Ok, we need to copy. Oh, well.. |
1289 | */ | 1314 | */ |
1290 | page_cache_get(old_page); | 1315 | page_cache_get(old_page); |
1291 | pte_unmap(page_table); | 1316 | pte_unmap_unlock(page_table, ptl); |
1292 | spin_unlock(&mm->page_table_lock); | ||
1293 | 1317 | ||
1294 | if (unlikely(anon_vma_prepare(vma))) | 1318 | if (unlikely(anon_vma_prepare(vma))) |
1295 | goto oom; | 1319 | goto oom; |
@@ -1307,8 +1331,7 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1307 | /* | 1331 | /* |
1308 | * Re-check the pte - we dropped the lock | 1332 | * Re-check the pte - we dropped the lock |
1309 | */ | 1333 | */ |
1310 | spin_lock(&mm->page_table_lock); | 1334 | page_table = pte_offset_map_lock(mm, pmd, address, &ptl); |
1311 | page_table = pte_offset_map(pmd, address); | ||
1312 | if (likely(pte_same(*page_table, orig_pte))) { | 1335 | if (likely(pte_same(*page_table, orig_pte))) { |
1313 | page_remove_rmap(old_page); | 1336 | page_remove_rmap(old_page); |
1314 | if (!PageAnon(old_page)) { | 1337 | if (!PageAnon(old_page)) { |
@@ -1321,7 +1344,6 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1321 | ptep_establish(vma, address, page_table, entry); | 1344 | ptep_establish(vma, address, page_table, entry); |
1322 | update_mmu_cache(vma, address, entry); | 1345 | update_mmu_cache(vma, address, entry); |
1323 | lazy_mmu_prot_update(entry); | 1346 | lazy_mmu_prot_update(entry); |
1324 | |||
1325 | lru_cache_add_active(new_page); | 1347 | lru_cache_add_active(new_page); |
1326 | page_add_anon_rmap(new_page, vma, address); | 1348 | page_add_anon_rmap(new_page, vma, address); |
1327 | 1349 | ||
@@ -1332,8 +1354,7 @@ static int do_wp_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1332 | page_cache_release(new_page); | 1354 | page_cache_release(new_page); |
1333 | page_cache_release(old_page); | 1355 | page_cache_release(old_page); |
1334 | unlock: | 1356 | unlock: |
1335 | pte_unmap(page_table); | 1357 | pte_unmap_unlock(page_table, ptl); |
1336 | spin_unlock(&mm->page_table_lock); | ||
1337 | return ret; | 1358 | return ret; |
1338 | oom: | 1359 | oom: |
1339 | page_cache_release(old_page); | 1360 | page_cache_release(old_page); |
@@ -1660,20 +1681,22 @@ void swapin_readahead(swp_entry_t entry, unsigned long addr,struct vm_area_struc | |||
1660 | } | 1681 | } |
1661 | 1682 | ||
1662 | /* | 1683 | /* |
1663 | * We hold the mm semaphore and the page_table_lock on entry and | 1684 | * We enter with non-exclusive mmap_sem (to exclude vma changes, |
1664 | * should release the pagetable lock on exit.. | 1685 | * but allow concurrent faults), and pte mapped but not yet locked. |
1686 | * We return with mmap_sem still held, but pte unmapped and unlocked. | ||
1665 | */ | 1687 | */ |
1666 | static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, | 1688 | static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, |
1667 | unsigned long address, pte_t *page_table, pmd_t *pmd, | 1689 | unsigned long address, pte_t *page_table, pmd_t *pmd, |
1668 | int write_access, pte_t orig_pte) | 1690 | int write_access, pte_t orig_pte) |
1669 | { | 1691 | { |
1692 | spinlock_t *ptl; | ||
1670 | struct page *page; | 1693 | struct page *page; |
1671 | swp_entry_t entry; | 1694 | swp_entry_t entry; |
1672 | pte_t pte; | 1695 | pte_t pte; |
1673 | int ret = VM_FAULT_MINOR; | 1696 | int ret = VM_FAULT_MINOR; |
1674 | 1697 | ||
1675 | pte_unmap(page_table); | 1698 | if (!pte_unmap_same(mm, page_table, orig_pte)) |
1676 | spin_unlock(&mm->page_table_lock); | 1699 | goto out; |
1677 | 1700 | ||
1678 | entry = pte_to_swp_entry(orig_pte); | 1701 | entry = pte_to_swp_entry(orig_pte); |
1679 | page = lookup_swap_cache(entry); | 1702 | page = lookup_swap_cache(entry); |
@@ -1682,11 +1705,10 @@ static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1682 | page = read_swap_cache_async(entry, vma, address); | 1705 | page = read_swap_cache_async(entry, vma, address); |
1683 | if (!page) { | 1706 | if (!page) { |
1684 | /* | 1707 | /* |
1685 | * Back out if somebody else faulted in this pte while | 1708 | * Back out if somebody else faulted in this pte |
1686 | * we released the page table lock. | 1709 | * while we released the pte lock. |
1687 | */ | 1710 | */ |
1688 | spin_lock(&mm->page_table_lock); | 1711 | page_table = pte_offset_map_lock(mm, pmd, address, &ptl); |
1689 | page_table = pte_offset_map(pmd, address); | ||
1690 | if (likely(pte_same(*page_table, orig_pte))) | 1712 | if (likely(pte_same(*page_table, orig_pte))) |
1691 | ret = VM_FAULT_OOM; | 1713 | ret = VM_FAULT_OOM; |
1692 | goto unlock; | 1714 | goto unlock; |
@@ -1702,11 +1724,9 @@ static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1702 | lock_page(page); | 1724 | lock_page(page); |
1703 | 1725 | ||
1704 | /* | 1726 | /* |
1705 | * Back out if somebody else faulted in this pte while we | 1727 | * Back out if somebody else already faulted in this pte. |
1706 | * released the page table lock. | ||
1707 | */ | 1728 | */ |
1708 | spin_lock(&mm->page_table_lock); | 1729 | page_table = pte_offset_map_lock(mm, pmd, address, &ptl); |
1709 | page_table = pte_offset_map(pmd, address); | ||
1710 | if (unlikely(!pte_same(*page_table, orig_pte))) | 1730 | if (unlikely(!pte_same(*page_table, orig_pte))) |
1711 | goto out_nomap; | 1731 | goto out_nomap; |
1712 | 1732 | ||
@@ -1735,7 +1755,7 @@ static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1735 | 1755 | ||
1736 | if (write_access) { | 1756 | if (write_access) { |
1737 | if (do_wp_page(mm, vma, address, | 1757 | if (do_wp_page(mm, vma, address, |
1738 | page_table, pmd, pte) == VM_FAULT_OOM) | 1758 | page_table, pmd, ptl, pte) == VM_FAULT_OOM) |
1739 | ret = VM_FAULT_OOM; | 1759 | ret = VM_FAULT_OOM; |
1740 | goto out; | 1760 | goto out; |
1741 | } | 1761 | } |
@@ -1744,37 +1764,32 @@ static int do_swap_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1744 | update_mmu_cache(vma, address, pte); | 1764 | update_mmu_cache(vma, address, pte); |
1745 | lazy_mmu_prot_update(pte); | 1765 | lazy_mmu_prot_update(pte); |
1746 | unlock: | 1766 | unlock: |
1747 | pte_unmap(page_table); | 1767 | pte_unmap_unlock(page_table, ptl); |
1748 | spin_unlock(&mm->page_table_lock); | ||
1749 | out: | 1768 | out: |
1750 | return ret; | 1769 | return ret; |
1751 | out_nomap: | 1770 | out_nomap: |
1752 | pte_unmap(page_table); | 1771 | pte_unmap_unlock(page_table, ptl); |
1753 | spin_unlock(&mm->page_table_lock); | ||
1754 | unlock_page(page); | 1772 | unlock_page(page); |
1755 | page_cache_release(page); | 1773 | page_cache_release(page); |
1756 | return ret; | 1774 | return ret; |
1757 | } | 1775 | } |
1758 | 1776 | ||
1759 | /* | 1777 | /* |
1760 | * We are called with the MM semaphore and page_table_lock | 1778 | * We enter with non-exclusive mmap_sem (to exclude vma changes, |
1761 | * spinlock held to protect against concurrent faults in | 1779 | * but allow concurrent faults), and pte mapped but not yet locked. |
1762 | * multithreaded programs. | 1780 | * We return with mmap_sem still held, but pte unmapped and unlocked. |
1763 | */ | 1781 | */ |
1764 | static int do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, | 1782 | static int do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, |
1765 | unsigned long address, pte_t *page_table, pmd_t *pmd, | 1783 | unsigned long address, pte_t *page_table, pmd_t *pmd, |
1766 | int write_access) | 1784 | int write_access) |
1767 | { | 1785 | { |
1768 | struct page *page = ZERO_PAGE(addr); | 1786 | struct page *page; |
1787 | spinlock_t *ptl; | ||
1769 | pte_t entry; | 1788 | pte_t entry; |
1770 | 1789 | ||
1771 | /* Mapping of ZERO_PAGE - vm_page_prot is readonly */ | ||
1772 | entry = mk_pte(page, vma->vm_page_prot); | ||
1773 | |||
1774 | if (write_access) { | 1790 | if (write_access) { |
1775 | /* Allocate our own private page. */ | 1791 | /* Allocate our own private page. */ |
1776 | pte_unmap(page_table); | 1792 | pte_unmap(page_table); |
1777 | spin_unlock(&mm->page_table_lock); | ||
1778 | 1793 | ||
1779 | if (unlikely(anon_vma_prepare(vma))) | 1794 | if (unlikely(anon_vma_prepare(vma))) |
1780 | goto oom; | 1795 | goto oom; |
@@ -1782,23 +1797,28 @@ static int do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1782 | if (!page) | 1797 | if (!page) |
1783 | goto oom; | 1798 | goto oom; |
1784 | 1799 | ||
1785 | spin_lock(&mm->page_table_lock); | ||
1786 | page_table = pte_offset_map(pmd, address); | ||
1787 | |||
1788 | if (!pte_none(*page_table)) { | ||
1789 | page_cache_release(page); | ||
1790 | goto unlock; | ||
1791 | } | ||
1792 | inc_mm_counter(mm, anon_rss); | ||
1793 | entry = mk_pte(page, vma->vm_page_prot); | 1800 | entry = mk_pte(page, vma->vm_page_prot); |
1794 | entry = maybe_mkwrite(pte_mkdirty(entry), vma); | 1801 | entry = maybe_mkwrite(pte_mkdirty(entry), vma); |
1802 | |||
1803 | page_table = pte_offset_map_lock(mm, pmd, address, &ptl); | ||
1804 | if (!pte_none(*page_table)) | ||
1805 | goto release; | ||
1806 | inc_mm_counter(mm, anon_rss); | ||
1795 | lru_cache_add_active(page); | 1807 | lru_cache_add_active(page); |
1796 | SetPageReferenced(page); | 1808 | SetPageReferenced(page); |
1797 | page_add_anon_rmap(page, vma, address); | 1809 | page_add_anon_rmap(page, vma, address); |
1798 | } else { | 1810 | } else { |
1811 | /* Map the ZERO_PAGE - vm_page_prot is readonly */ | ||
1812 | page = ZERO_PAGE(address); | ||
1813 | page_cache_get(page); | ||
1814 | entry = mk_pte(page, vma->vm_page_prot); | ||
1815 | |||
1816 | ptl = &mm->page_table_lock; | ||
1817 | spin_lock(ptl); | ||
1818 | if (!pte_none(*page_table)) | ||
1819 | goto release; | ||
1799 | inc_mm_counter(mm, file_rss); | 1820 | inc_mm_counter(mm, file_rss); |
1800 | page_add_file_rmap(page); | 1821 | page_add_file_rmap(page); |
1801 | page_cache_get(page); | ||
1802 | } | 1822 | } |
1803 | 1823 | ||
1804 | set_pte_at(mm, address, page_table, entry); | 1824 | set_pte_at(mm, address, page_table, entry); |
@@ -1807,9 +1827,11 @@ static int do_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1807 | update_mmu_cache(vma, address, entry); | 1827 | update_mmu_cache(vma, address, entry); |
1808 | lazy_mmu_prot_update(entry); | 1828 | lazy_mmu_prot_update(entry); |
1809 | unlock: | 1829 | unlock: |
1810 | pte_unmap(page_table); | 1830 | pte_unmap_unlock(page_table, ptl); |
1811 | spin_unlock(&mm->page_table_lock); | ||
1812 | return VM_FAULT_MINOR; | 1831 | return VM_FAULT_MINOR; |
1832 | release: | ||
1833 | page_cache_release(page); | ||
1834 | goto unlock; | ||
1813 | oom: | 1835 | oom: |
1814 | return VM_FAULT_OOM; | 1836 | return VM_FAULT_OOM; |
1815 | } | 1837 | } |
@@ -1823,13 +1845,15 @@ oom: | |||
1823 | * As this is called only for pages that do not currently exist, we | 1845 | * As this is called only for pages that do not currently exist, we |
1824 | * do not need to flush old virtual caches or the TLB. | 1846 | * do not need to flush old virtual caches or the TLB. |
1825 | * | 1847 | * |
1826 | * This is called with the MM semaphore held and the page table | 1848 | * We enter with non-exclusive mmap_sem (to exclude vma changes, |
1827 | * spinlock held. Exit with the spinlock released. | 1849 | * but allow concurrent faults), and pte mapped but not yet locked. |
1850 | * We return with mmap_sem still held, but pte unmapped and unlocked. | ||
1828 | */ | 1851 | */ |
1829 | static int do_no_page(struct mm_struct *mm, struct vm_area_struct *vma, | 1852 | static int do_no_page(struct mm_struct *mm, struct vm_area_struct *vma, |
1830 | unsigned long address, pte_t *page_table, pmd_t *pmd, | 1853 | unsigned long address, pte_t *page_table, pmd_t *pmd, |
1831 | int write_access) | 1854 | int write_access) |
1832 | { | 1855 | { |
1856 | spinlock_t *ptl; | ||
1833 | struct page *new_page; | 1857 | struct page *new_page; |
1834 | struct address_space *mapping = NULL; | 1858 | struct address_space *mapping = NULL; |
1835 | pte_t entry; | 1859 | pte_t entry; |
@@ -1838,7 +1862,6 @@ static int do_no_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1838 | int anon = 0; | 1862 | int anon = 0; |
1839 | 1863 | ||
1840 | pte_unmap(page_table); | 1864 | pte_unmap(page_table); |
1841 | spin_unlock(&mm->page_table_lock); | ||
1842 | 1865 | ||
1843 | if (vma->vm_file) { | 1866 | if (vma->vm_file) { |
1844 | mapping = vma->vm_file->f_mapping; | 1867 | mapping = vma->vm_file->f_mapping; |
@@ -1878,21 +1901,20 @@ retry: | |||
1878 | anon = 1; | 1901 | anon = 1; |
1879 | } | 1902 | } |
1880 | 1903 | ||
1881 | spin_lock(&mm->page_table_lock); | 1904 | page_table = pte_offset_map_lock(mm, pmd, address, &ptl); |
1882 | /* | 1905 | /* |
1883 | * For a file-backed vma, someone could have truncated or otherwise | 1906 | * For a file-backed vma, someone could have truncated or otherwise |
1884 | * invalidated this page. If unmap_mapping_range got called, | 1907 | * invalidated this page. If unmap_mapping_range got called, |
1885 | * retry getting the page. | 1908 | * retry getting the page. |
1886 | */ | 1909 | */ |
1887 | if (mapping && unlikely(sequence != mapping->truncate_count)) { | 1910 | if (mapping && unlikely(sequence != mapping->truncate_count)) { |
1888 | spin_unlock(&mm->page_table_lock); | 1911 | pte_unmap_unlock(page_table, ptl); |
1889 | page_cache_release(new_page); | 1912 | page_cache_release(new_page); |
1890 | cond_resched(); | 1913 | cond_resched(); |
1891 | sequence = mapping->truncate_count; | 1914 | sequence = mapping->truncate_count; |
1892 | smp_rmb(); | 1915 | smp_rmb(); |
1893 | goto retry; | 1916 | goto retry; |
1894 | } | 1917 | } |
1895 | page_table = pte_offset_map(pmd, address); | ||
1896 | 1918 | ||
1897 | /* | 1919 | /* |
1898 | * This silly early PAGE_DIRTY setting removes a race | 1920 | * This silly early PAGE_DIRTY setting removes a race |
@@ -1929,8 +1951,7 @@ retry: | |||
1929 | update_mmu_cache(vma, address, entry); | 1951 | update_mmu_cache(vma, address, entry); |
1930 | lazy_mmu_prot_update(entry); | 1952 | lazy_mmu_prot_update(entry); |
1931 | unlock: | 1953 | unlock: |
1932 | pte_unmap(page_table); | 1954 | pte_unmap_unlock(page_table, ptl); |
1933 | spin_unlock(&mm->page_table_lock); | ||
1934 | return ret; | 1955 | return ret; |
1935 | oom: | 1956 | oom: |
1936 | page_cache_release(new_page); | 1957 | page_cache_release(new_page); |
@@ -1941,6 +1962,10 @@ oom: | |||
1941 | * Fault of a previously existing named mapping. Repopulate the pte | 1962 | * Fault of a previously existing named mapping. Repopulate the pte |
1942 | * from the encoded file_pte if possible. This enables swappable | 1963 | * from the encoded file_pte if possible. This enables swappable |
1943 | * nonlinear vmas. | 1964 | * nonlinear vmas. |
1965 | * | ||
1966 | * We enter with non-exclusive mmap_sem (to exclude vma changes, | ||
1967 | * but allow concurrent faults), and pte mapped but not yet locked. | ||
1968 | * We return with mmap_sem still held, but pte unmapped and unlocked. | ||
1944 | */ | 1969 | */ |
1945 | static int do_file_page(struct mm_struct *mm, struct vm_area_struct *vma, | 1970 | static int do_file_page(struct mm_struct *mm, struct vm_area_struct *vma, |
1946 | unsigned long address, pte_t *page_table, pmd_t *pmd, | 1971 | unsigned long address, pte_t *page_table, pmd_t *pmd, |
@@ -1949,8 +1974,8 @@ static int do_file_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
1949 | pgoff_t pgoff; | 1974 | pgoff_t pgoff; |
1950 | int err; | 1975 | int err; |
1951 | 1976 | ||
1952 | pte_unmap(page_table); | 1977 | if (!pte_unmap_same(mm, page_table, orig_pte)) |
1953 | spin_unlock(&mm->page_table_lock); | 1978 | return VM_FAULT_MINOR; |
1954 | 1979 | ||
1955 | if (unlikely(!(vma->vm_flags & VM_NONLINEAR))) { | 1980 | if (unlikely(!(vma->vm_flags & VM_NONLINEAR))) { |
1956 | /* | 1981 | /* |
@@ -1989,8 +2014,8 @@ static inline int handle_pte_fault(struct mm_struct *mm, | |||
1989 | pte_t *pte, pmd_t *pmd, int write_access) | 2014 | pte_t *pte, pmd_t *pmd, int write_access) |
1990 | { | 2015 | { |
1991 | pte_t entry; | 2016 | pte_t entry; |
2017 | spinlock_t *ptl; | ||
1992 | 2018 | ||
1993 | spin_lock(&mm->page_table_lock); | ||
1994 | entry = *pte; | 2019 | entry = *pte; |
1995 | if (!pte_present(entry)) { | 2020 | if (!pte_present(entry)) { |
1996 | if (pte_none(entry)) { | 2021 | if (pte_none(entry)) { |
@@ -2007,17 +2032,22 @@ static inline int handle_pte_fault(struct mm_struct *mm, | |||
2007 | pte, pmd, write_access, entry); | 2032 | pte, pmd, write_access, entry); |
2008 | } | 2033 | } |
2009 | 2034 | ||
2035 | ptl = &mm->page_table_lock; | ||
2036 | spin_lock(ptl); | ||
2037 | if (unlikely(!pte_same(*pte, entry))) | ||
2038 | goto unlock; | ||
2010 | if (write_access) { | 2039 | if (write_access) { |
2011 | if (!pte_write(entry)) | 2040 | if (!pte_write(entry)) |
2012 | return do_wp_page(mm, vma, address, pte, pmd, entry); | 2041 | return do_wp_page(mm, vma, address, |
2042 | pte, pmd, ptl, entry); | ||
2013 | entry = pte_mkdirty(entry); | 2043 | entry = pte_mkdirty(entry); |
2014 | } | 2044 | } |
2015 | entry = pte_mkyoung(entry); | 2045 | entry = pte_mkyoung(entry); |
2016 | ptep_set_access_flags(vma, address, pte, entry, write_access); | 2046 | ptep_set_access_flags(vma, address, pte, entry, write_access); |
2017 | update_mmu_cache(vma, address, entry); | 2047 | update_mmu_cache(vma, address, entry); |
2018 | lazy_mmu_prot_update(entry); | 2048 | lazy_mmu_prot_update(entry); |
2019 | pte_unmap(pte); | 2049 | unlock: |
2020 | spin_unlock(&mm->page_table_lock); | 2050 | pte_unmap_unlock(pte, ptl); |
2021 | return VM_FAULT_MINOR; | 2051 | return VM_FAULT_MINOR; |
2022 | } | 2052 | } |
2023 | 2053 | ||