aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/mmap.c30
-rw-r--r--mm/swap_state.c1
2 files changed, 25 insertions, 6 deletions
diff --git a/mm/mmap.c b/mm/mmap.c
index 64c9d736155c..c1f2ea4a0b99 100644
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -221,7 +221,7 @@ static void __remove_shared_vm_struct(struct vm_area_struct *vma,
221 if (vma->vm_flags & VM_DENYWRITE) 221 if (vma->vm_flags & VM_DENYWRITE)
222 atomic_inc(&file_inode(file)->i_writecount); 222 atomic_inc(&file_inode(file)->i_writecount);
223 if (vma->vm_flags & VM_SHARED) 223 if (vma->vm_flags & VM_SHARED)
224 mapping->i_mmap_writable--; 224 mapping_unmap_writable(mapping);
225 225
226 flush_dcache_mmap_lock(mapping); 226 flush_dcache_mmap_lock(mapping);
227 if (unlikely(vma->vm_flags & VM_NONLINEAR)) 227 if (unlikely(vma->vm_flags & VM_NONLINEAR))
@@ -622,7 +622,7 @@ static void __vma_link_file(struct vm_area_struct *vma)
622 if (vma->vm_flags & VM_DENYWRITE) 622 if (vma->vm_flags & VM_DENYWRITE)
623 atomic_dec(&file_inode(file)->i_writecount); 623 atomic_dec(&file_inode(file)->i_writecount);
624 if (vma->vm_flags & VM_SHARED) 624 if (vma->vm_flags & VM_SHARED)
625 mapping->i_mmap_writable++; 625 atomic_inc(&mapping->i_mmap_writable);
626 626
627 flush_dcache_mmap_lock(mapping); 627 flush_dcache_mmap_lock(mapping);
628 if (unlikely(vma->vm_flags & VM_NONLINEAR)) 628 if (unlikely(vma->vm_flags & VM_NONLINEAR))
@@ -1577,6 +1577,17 @@ munmap_back:
1577 if (error) 1577 if (error)
1578 goto free_vma; 1578 goto free_vma;
1579 } 1579 }
1580 if (vm_flags & VM_SHARED) {
1581 error = mapping_map_writable(file->f_mapping);
1582 if (error)
1583 goto allow_write_and_free_vma;
1584 }
1585
1586 /* ->mmap() can change vma->vm_file, but must guarantee that
1587 * vma_link() below can deny write-access if VM_DENYWRITE is set
1588 * and map writably if VM_SHARED is set. This usually means the
1589 * new file must not have been exposed to user-space, yet.
1590 */
1580 vma->vm_file = get_file(file); 1591 vma->vm_file = get_file(file);
1581 error = file->f_op->mmap(file, vma); 1592 error = file->f_op->mmap(file, vma);
1582 if (error) 1593 if (error)
@@ -1616,8 +1627,12 @@ munmap_back:
1616 1627
1617 vma_link(mm, vma, prev, rb_link, rb_parent); 1628 vma_link(mm, vma, prev, rb_link, rb_parent);
1618 /* Once vma denies write, undo our temporary denial count */ 1629 /* Once vma denies write, undo our temporary denial count */
1619 if (vm_flags & VM_DENYWRITE) 1630 if (file) {
1620 allow_write_access(file); 1631 if (vm_flags & VM_SHARED)
1632 mapping_unmap_writable(file->f_mapping);
1633 if (vm_flags & VM_DENYWRITE)
1634 allow_write_access(file);
1635 }
1621 file = vma->vm_file; 1636 file = vma->vm_file;
1622out: 1637out:
1623 perf_event_mmap(vma); 1638 perf_event_mmap(vma);
@@ -1646,14 +1661,17 @@ out:
1646 return addr; 1661 return addr;
1647 1662
1648unmap_and_free_vma: 1663unmap_and_free_vma:
1649 if (vm_flags & VM_DENYWRITE)
1650 allow_write_access(file);
1651 vma->vm_file = NULL; 1664 vma->vm_file = NULL;
1652 fput(file); 1665 fput(file);
1653 1666
1654 /* Undo any partial mapping done by a device driver. */ 1667 /* Undo any partial mapping done by a device driver. */
1655 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end); 1668 unmap_region(mm, vma, prev, vma->vm_start, vma->vm_end);
1656 charged = 0; 1669 charged = 0;
1670 if (vm_flags & VM_SHARED)
1671 mapping_unmap_writable(file->f_mapping);
1672allow_write_and_free_vma:
1673 if (vm_flags & VM_DENYWRITE)
1674 allow_write_access(file);
1657free_vma: 1675free_vma:
1658 kmem_cache_free(vm_area_cachep, vma); 1676 kmem_cache_free(vm_area_cachep, vma);
1659unacct_error: 1677unacct_error:
diff --git a/mm/swap_state.c b/mm/swap_state.c
index e160151da6b8..3e0ec83d000c 100644
--- a/mm/swap_state.c
+++ b/mm/swap_state.c
@@ -39,6 +39,7 @@ static struct backing_dev_info swap_backing_dev_info = {
39struct address_space swapper_spaces[MAX_SWAPFILES] = { 39struct address_space swapper_spaces[MAX_SWAPFILES] = {
40 [0 ... MAX_SWAPFILES - 1] = { 40 [0 ... MAX_SWAPFILES - 1] = {
41 .page_tree = RADIX_TREE_INIT(GFP_ATOMIC|__GFP_NOWARN), 41 .page_tree = RADIX_TREE_INIT(GFP_ATOMIC|__GFP_NOWARN),
42 .i_mmap_writable = ATOMIC_INIT(0),
42 .a_ops = &swap_aops, 43 .a_ops = &swap_aops,
43 .backing_dev_info = &swap_backing_dev_info, 44 .backing_dev_info = &swap_backing_dev_info,
44 } 45 }