diff options
Diffstat (limited to 'mm/rmap.c')
-rw-r--r-- | mm/rmap.c | 22 |
1 files changed, 10 insertions, 12 deletions
@@ -225,7 +225,9 @@ vma_address(struct page *page, struct vm_area_struct *vma) | |||
225 | 225 | ||
226 | /* | 226 | /* |
227 | * At what user virtual address is page expected in vma? checking that the | 227 | * At what user virtual address is page expected in vma? checking that the |
228 | * page matches the vma: currently only used by unuse_process, on anon pages. | 228 | * page matches the vma: currently only used on anon pages, by unuse_vma; |
229 | * and by extraordinary checks on anon pages in VM_UNPAGED vmas, taking | ||
230 | * care that an mmap of /dev/mem might window free and foreign pages. | ||
229 | */ | 231 | */ |
230 | unsigned long page_address_in_vma(struct page *page, struct vm_area_struct *vma) | 232 | unsigned long page_address_in_vma(struct page *page, struct vm_area_struct *vma) |
231 | { | 233 | { |
@@ -234,7 +236,8 @@ unsigned long page_address_in_vma(struct page *page, struct vm_area_struct *vma) | |||
234 | (void *)page->mapping - PAGE_MAPPING_ANON) | 236 | (void *)page->mapping - PAGE_MAPPING_ANON) |
235 | return -EFAULT; | 237 | return -EFAULT; |
236 | } else if (page->mapping && !(vma->vm_flags & VM_NONLINEAR)) { | 238 | } else if (page->mapping && !(vma->vm_flags & VM_NONLINEAR)) { |
237 | if (vma->vm_file->f_mapping != page->mapping) | 239 | if (!vma->vm_file || |
240 | vma->vm_file->f_mapping != page->mapping) | ||
238 | return -EFAULT; | 241 | return -EFAULT; |
239 | } else | 242 | } else |
240 | return -EFAULT; | 243 | return -EFAULT; |
@@ -529,10 +532,8 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma) | |||
529 | * If the page is mlock()d, we cannot swap it out. | 532 | * If the page is mlock()d, we cannot swap it out. |
530 | * If it's recently referenced (perhaps page_referenced | 533 | * If it's recently referenced (perhaps page_referenced |
531 | * skipped over this mm) then we should reactivate it. | 534 | * skipped over this mm) then we should reactivate it. |
532 | * | ||
533 | * Pages belonging to VM_RESERVED regions should not happen here. | ||
534 | */ | 535 | */ |
535 | if ((vma->vm_flags & (VM_LOCKED|VM_RESERVED)) || | 536 | if ((vma->vm_flags & VM_LOCKED) || |
536 | ptep_clear_flush_young(vma, address, pte)) { | 537 | ptep_clear_flush_young(vma, address, pte)) { |
537 | ret = SWAP_FAIL; | 538 | ret = SWAP_FAIL; |
538 | goto out_unmap; | 539 | goto out_unmap; |
@@ -727,7 +728,7 @@ static int try_to_unmap_file(struct page *page) | |||
727 | 728 | ||
728 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, | 729 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, |
729 | shared.vm_set.list) { | 730 | shared.vm_set.list) { |
730 | if (vma->vm_flags & (VM_LOCKED|VM_RESERVED)) | 731 | if (vma->vm_flags & VM_LOCKED) |
731 | continue; | 732 | continue; |
732 | cursor = (unsigned long) vma->vm_private_data; | 733 | cursor = (unsigned long) vma->vm_private_data; |
733 | if (cursor > max_nl_cursor) | 734 | if (cursor > max_nl_cursor) |
@@ -761,7 +762,7 @@ static int try_to_unmap_file(struct page *page) | |||
761 | do { | 762 | do { |
762 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, | 763 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, |
763 | shared.vm_set.list) { | 764 | shared.vm_set.list) { |
764 | if (vma->vm_flags & (VM_LOCKED|VM_RESERVED)) | 765 | if (vma->vm_flags & VM_LOCKED) |
765 | continue; | 766 | continue; |
766 | cursor = (unsigned long) vma->vm_private_data; | 767 | cursor = (unsigned long) vma->vm_private_data; |
767 | while ( cursor < max_nl_cursor && | 768 | while ( cursor < max_nl_cursor && |
@@ -783,11 +784,8 @@ static int try_to_unmap_file(struct page *page) | |||
783 | * in locked vmas). Reset cursor on all unreserved nonlinear | 784 | * in locked vmas). Reset cursor on all unreserved nonlinear |
784 | * vmas, now forgetting on which ones it had fallen behind. | 785 | * vmas, now forgetting on which ones it had fallen behind. |
785 | */ | 786 | */ |
786 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, | 787 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, shared.vm_set.list) |
787 | shared.vm_set.list) { | 788 | vma->vm_private_data = NULL; |
788 | if (!(vma->vm_flags & VM_RESERVED)) | ||
789 | vma->vm_private_data = NULL; | ||
790 | } | ||
791 | out: | 789 | out: |
792 | spin_unlock(&mapping->i_mmap_lock); | 790 | spin_unlock(&mapping->i_mmap_lock); |
793 | return ret; | 791 | return ret; |