diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2011-05-24 20:12:06 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2011-05-25 11:39:18 -0400 |
commit | 3d48ae45e72390ddf8cc5256ac32ed6f7a19cbea (patch) | |
tree | 1f46db3a8424090dd8e0b58991fa5acc1a73e680 | |
parent | 97a894136f29802da19a15541de3c019e1ca147e (diff) |
mm: Convert i_mmap_lock to a mutex
Straightforward conversion of i_mmap_lock to a mutex.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Acked-by: Hugh Dickins <hughd@google.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: David Miller <davem@davemloft.net>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Russell King <rmk@arm.linux.org.uk>
Cc: Paul Mundt <lethal@linux-sh.org>
Cc: Jeff Dike <jdike@addtoit.com>
Cc: Richard Weinberger <richard@nod.at>
Cc: Tony Luck <tony.luck@intel.com>
Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Mel Gorman <mel@csn.ul.ie>
Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Cc: Nick Piggin <npiggin@kernel.dk>
Cc: Namhyung Kim <namhyung@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | Documentation/lockstat.txt | 2 | ||||
-rw-r--r-- | Documentation/vm/locking | 2 | ||||
-rw-r--r-- | arch/x86/mm/hugetlbpage.c | 4 | ||||
-rw-r--r-- | fs/hugetlbfs/inode.c | 4 | ||||
-rw-r--r-- | fs/inode.c | 2 | ||||
-rw-r--r-- | include/linux/fs.h | 2 | ||||
-rw-r--r-- | include/linux/mmu_notifier.h | 2 | ||||
-rw-r--r-- | kernel/fork.c | 4 | ||||
-rw-r--r-- | mm/filemap.c | 10 | ||||
-rw-r--r-- | mm/filemap_xip.c | 4 | ||||
-rw-r--r-- | mm/fremap.c | 4 | ||||
-rw-r--r-- | mm/hugetlb.c | 14 | ||||
-rw-r--r-- | mm/memory-failure.c | 4 | ||||
-rw-r--r-- | mm/memory.c | 4 | ||||
-rw-r--r-- | mm/mmap.c | 22 | ||||
-rw-r--r-- | mm/mremap.c | 4 | ||||
-rw-r--r-- | mm/rmap.c | 28 |
17 files changed, 58 insertions, 58 deletions
diff --git a/Documentation/lockstat.txt b/Documentation/lockstat.txt index 65f4c795015d..9c0a80d17a23 100644 --- a/Documentation/lockstat.txt +++ b/Documentation/lockstat.txt | |||
@@ -136,7 +136,7 @@ View the top contending locks: | |||
136 | dcache_lock: 1037 1161 0.38 45.32 774.51 6611 243371 0.15 306.48 77387.24 | 136 | dcache_lock: 1037 1161 0.38 45.32 774.51 6611 243371 0.15 306.48 77387.24 |
137 | &inode->i_mutex: 161 286 18446744073709 62882.54 1244614.55 3653 20598 18446744073709 62318.60 1693822.74 | 137 | &inode->i_mutex: 161 286 18446744073709 62882.54 1244614.55 3653 20598 18446744073709 62318.60 1693822.74 |
138 | &zone->lru_lock: 94 94 0.53 7.33 92.10 4366 32690 0.29 59.81 16350.06 | 138 | &zone->lru_lock: 94 94 0.53 7.33 92.10 4366 32690 0.29 59.81 16350.06 |
139 | &inode->i_data.i_mmap_lock: 79 79 0.40 3.77 53.03 11779 87755 0.28 116.93 29898.44 | 139 | &inode->i_data.i_mmap_mutex: 79 79 0.40 3.77 53.03 11779 87755 0.28 116.93 29898.44 |
140 | &q->__queue_lock: 48 50 0.52 31.62 86.31 774 13131 0.17 113.08 12277.52 | 140 | &q->__queue_lock: 48 50 0.52 31.62 86.31 774 13131 0.17 113.08 12277.52 |
141 | &rq->rq_lock_key: 43 47 0.74 68.50 170.63 3706 33929 0.22 107.99 17460.62 | 141 | &rq->rq_lock_key: 43 47 0.74 68.50 170.63 3706 33929 0.22 107.99 17460.62 |
142 | &rq->rq_lock_key#2: 39 46 0.75 6.68 49.03 2979 32292 0.17 125.17 17137.63 | 142 | &rq->rq_lock_key#2: 39 46 0.75 6.68 49.03 2979 32292 0.17 125.17 17137.63 |
diff --git a/Documentation/vm/locking b/Documentation/vm/locking index 25fadb448760..f61228bd6395 100644 --- a/Documentation/vm/locking +++ b/Documentation/vm/locking | |||
@@ -66,7 +66,7 @@ in some cases it is not really needed. Eg, vm_start is modified by | |||
66 | expand_stack(), it is hard to come up with a destructive scenario without | 66 | expand_stack(), it is hard to come up with a destructive scenario without |
67 | having the vmlist protection in this case. | 67 | having the vmlist protection in this case. |
68 | 68 | ||
69 | The page_table_lock nests with the inode i_mmap_lock and the kmem cache | 69 | The page_table_lock nests with the inode i_mmap_mutex and the kmem cache |
70 | c_spinlock spinlocks. This is okay, since the kmem code asks for pages after | 70 | c_spinlock spinlocks. This is okay, since the kmem code asks for pages after |
71 | dropping c_spinlock. The page_table_lock also nests with pagecache_lock and | 71 | dropping c_spinlock. The page_table_lock also nests with pagecache_lock and |
72 | pagemap_lru_lock spinlocks, and no code asks for memory with these locks | 72 | pagemap_lru_lock spinlocks, and no code asks for memory with these locks |
diff --git a/arch/x86/mm/hugetlbpage.c b/arch/x86/mm/hugetlbpage.c index d4203988504a..f581a18c0d4d 100644 --- a/arch/x86/mm/hugetlbpage.c +++ b/arch/x86/mm/hugetlbpage.c | |||
@@ -72,7 +72,7 @@ static void huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) | |||
72 | if (!vma_shareable(vma, addr)) | 72 | if (!vma_shareable(vma, addr)) |
73 | return; | 73 | return; |
74 | 74 | ||
75 | spin_lock(&mapping->i_mmap_lock); | 75 | mutex_lock(&mapping->i_mmap_mutex); |
76 | vma_prio_tree_foreach(svma, &iter, &mapping->i_mmap, idx, idx) { | 76 | vma_prio_tree_foreach(svma, &iter, &mapping->i_mmap, idx, idx) { |
77 | if (svma == vma) | 77 | if (svma == vma) |
78 | continue; | 78 | continue; |
@@ -97,7 +97,7 @@ static void huge_pmd_share(struct mm_struct *mm, unsigned long addr, pud_t *pud) | |||
97 | put_page(virt_to_page(spte)); | 97 | put_page(virt_to_page(spte)); |
98 | spin_unlock(&mm->page_table_lock); | 98 | spin_unlock(&mm->page_table_lock); |
99 | out: | 99 | out: |
100 | spin_unlock(&mapping->i_mmap_lock); | 100 | mutex_unlock(&mapping->i_mmap_mutex); |
101 | } | 101 | } |
102 | 102 | ||
103 | /* | 103 | /* |
diff --git a/fs/hugetlbfs/inode.c b/fs/hugetlbfs/inode.c index b9eeb1cd03ff..e7a035781b7d 100644 --- a/fs/hugetlbfs/inode.c +++ b/fs/hugetlbfs/inode.c | |||
@@ -412,10 +412,10 @@ static int hugetlb_vmtruncate(struct inode *inode, loff_t offset) | |||
412 | pgoff = offset >> PAGE_SHIFT; | 412 | pgoff = offset >> PAGE_SHIFT; |
413 | 413 | ||
414 | i_size_write(inode, offset); | 414 | i_size_write(inode, offset); |
415 | spin_lock(&mapping->i_mmap_lock); | 415 | mutex_lock(&mapping->i_mmap_mutex); |
416 | if (!prio_tree_empty(&mapping->i_mmap)) | 416 | if (!prio_tree_empty(&mapping->i_mmap)) |
417 | hugetlb_vmtruncate_list(&mapping->i_mmap, pgoff); | 417 | hugetlb_vmtruncate_list(&mapping->i_mmap, pgoff); |
418 | spin_unlock(&mapping->i_mmap_lock); | 418 | mutex_unlock(&mapping->i_mmap_mutex); |
419 | truncate_hugepages(inode, offset); | 419 | truncate_hugepages(inode, offset); |
420 | return 0; | 420 | return 0; |
421 | } | 421 | } |
diff --git a/fs/inode.c b/fs/inode.c index 7a7284c71abd..88aa3bcc7681 100644 --- a/fs/inode.c +++ b/fs/inode.c | |||
@@ -326,7 +326,7 @@ void address_space_init_once(struct address_space *mapping) | |||
326 | memset(mapping, 0, sizeof(*mapping)); | 326 | memset(mapping, 0, sizeof(*mapping)); |
327 | INIT_RADIX_TREE(&mapping->page_tree, GFP_ATOMIC); | 327 | INIT_RADIX_TREE(&mapping->page_tree, GFP_ATOMIC); |
328 | spin_lock_init(&mapping->tree_lock); | 328 | spin_lock_init(&mapping->tree_lock); |
329 | spin_lock_init(&mapping->i_mmap_lock); | 329 | mutex_init(&mapping->i_mmap_mutex); |
330 | INIT_LIST_HEAD(&mapping->private_list); | 330 | INIT_LIST_HEAD(&mapping->private_list); |
331 | spin_lock_init(&mapping->private_lock); | 331 | spin_lock_init(&mapping->private_lock); |
332 | INIT_RAW_PRIO_TREE_ROOT(&mapping->i_mmap); | 332 | INIT_RAW_PRIO_TREE_ROOT(&mapping->i_mmap); |
diff --git a/include/linux/fs.h b/include/linux/fs.h index 5d2c86bdf5ba..5bb9e826019b 100644 --- a/include/linux/fs.h +++ b/include/linux/fs.h | |||
@@ -634,7 +634,7 @@ struct address_space { | |||
634 | unsigned int i_mmap_writable;/* count VM_SHARED mappings */ | 634 | unsigned int i_mmap_writable;/* count VM_SHARED mappings */ |
635 | struct prio_tree_root i_mmap; /* tree of private and shared mappings */ | 635 | struct prio_tree_root i_mmap; /* tree of private and shared mappings */ |
636 | struct list_head i_mmap_nonlinear;/*list VM_NONLINEAR mappings */ | 636 | struct list_head i_mmap_nonlinear;/*list VM_NONLINEAR mappings */ |
637 | spinlock_t i_mmap_lock; /* protect tree, count, list */ | 637 | struct mutex i_mmap_mutex; /* protect tree, count, list */ |
638 | unsigned long nrpages; /* number of total pages */ | 638 | unsigned long nrpages; /* number of total pages */ |
639 | pgoff_t writeback_index;/* writeback starts here */ | 639 | pgoff_t writeback_index;/* writeback starts here */ |
640 | const struct address_space_operations *a_ops; /* methods */ | 640 | const struct address_space_operations *a_ops; /* methods */ |
diff --git a/include/linux/mmu_notifier.h b/include/linux/mmu_notifier.h index cc2e7dfea9d7..a877dfc243eb 100644 --- a/include/linux/mmu_notifier.h +++ b/include/linux/mmu_notifier.h | |||
@@ -150,7 +150,7 @@ struct mmu_notifier_ops { | |||
150 | * Therefore notifier chains can only be traversed when either | 150 | * Therefore notifier chains can only be traversed when either |
151 | * | 151 | * |
152 | * 1. mmap_sem is held. | 152 | * 1. mmap_sem is held. |
153 | * 2. One of the reverse map locks is held (i_mmap_lock or anon_vma->lock). | 153 | * 2. One of the reverse map locks is held (i_mmap_mutex or anon_vma->lock). |
154 | * 3. No other concurrent thread can access the list (release) | 154 | * 3. No other concurrent thread can access the list (release) |
155 | */ | 155 | */ |
156 | struct mmu_notifier { | 156 | struct mmu_notifier { |
diff --git a/kernel/fork.c b/kernel/fork.c index 4eef925477fc..927692734bcf 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
@@ -383,14 +383,14 @@ static int dup_mmap(struct mm_struct *mm, struct mm_struct *oldmm) | |||
383 | get_file(file); | 383 | get_file(file); |
384 | if (tmp->vm_flags & VM_DENYWRITE) | 384 | if (tmp->vm_flags & VM_DENYWRITE) |
385 | atomic_dec(&inode->i_writecount); | 385 | atomic_dec(&inode->i_writecount); |
386 | spin_lock(&mapping->i_mmap_lock); | 386 | mutex_lock(&mapping->i_mmap_mutex); |
387 | if (tmp->vm_flags & VM_SHARED) | 387 | if (tmp->vm_flags & VM_SHARED) |
388 | mapping->i_mmap_writable++; | 388 | mapping->i_mmap_writable++; |
389 | flush_dcache_mmap_lock(mapping); | 389 | flush_dcache_mmap_lock(mapping); |
390 | /* insert tmp into the share list, just after mpnt */ | 390 | /* insert tmp into the share list, just after mpnt */ |
391 | vma_prio_tree_add(tmp, mpnt); | 391 | vma_prio_tree_add(tmp, mpnt); |
392 | flush_dcache_mmap_unlock(mapping); | 392 | flush_dcache_mmap_unlock(mapping); |
393 | spin_unlock(&mapping->i_mmap_lock); | 393 | mutex_unlock(&mapping->i_mmap_mutex); |
394 | } | 394 | } |
395 | 395 | ||
396 | /* | 396 | /* |
diff --git a/mm/filemap.c b/mm/filemap.c index 8144f87dcbb4..88354ae0b1fd 100644 --- a/mm/filemap.c +++ b/mm/filemap.c | |||
@@ -58,16 +58,16 @@ | |||
58 | /* | 58 | /* |
59 | * Lock ordering: | 59 | * Lock ordering: |
60 | * | 60 | * |
61 | * ->i_mmap_lock (truncate_pagecache) | 61 | * ->i_mmap_mutex (truncate_pagecache) |
62 | * ->private_lock (__free_pte->__set_page_dirty_buffers) | 62 | * ->private_lock (__free_pte->__set_page_dirty_buffers) |
63 | * ->swap_lock (exclusive_swap_page, others) | 63 | * ->swap_lock (exclusive_swap_page, others) |
64 | * ->mapping->tree_lock | 64 | * ->mapping->tree_lock |
65 | * | 65 | * |
66 | * ->i_mutex | 66 | * ->i_mutex |
67 | * ->i_mmap_lock (truncate->unmap_mapping_range) | 67 | * ->i_mmap_mutex (truncate->unmap_mapping_range) |
68 | * | 68 | * |
69 | * ->mmap_sem | 69 | * ->mmap_sem |
70 | * ->i_mmap_lock | 70 | * ->i_mmap_mutex |
71 | * ->page_table_lock or pte_lock (various, mainly in memory.c) | 71 | * ->page_table_lock or pte_lock (various, mainly in memory.c) |
72 | * ->mapping->tree_lock (arch-dependent flush_dcache_mmap_lock) | 72 | * ->mapping->tree_lock (arch-dependent flush_dcache_mmap_lock) |
73 | * | 73 | * |
@@ -84,7 +84,7 @@ | |||
84 | * sb_lock (fs/fs-writeback.c) | 84 | * sb_lock (fs/fs-writeback.c) |
85 | * ->mapping->tree_lock (__sync_single_inode) | 85 | * ->mapping->tree_lock (__sync_single_inode) |
86 | * | 86 | * |
87 | * ->i_mmap_lock | 87 | * ->i_mmap_mutex |
88 | * ->anon_vma.lock (vma_adjust) | 88 | * ->anon_vma.lock (vma_adjust) |
89 | * | 89 | * |
90 | * ->anon_vma.lock | 90 | * ->anon_vma.lock |
@@ -106,7 +106,7 @@ | |||
106 | * | 106 | * |
107 | * (code doesn't rely on that order, so you could switch it around) | 107 | * (code doesn't rely on that order, so you could switch it around) |
108 | * ->tasklist_lock (memory_failure, collect_procs_ao) | 108 | * ->tasklist_lock (memory_failure, collect_procs_ao) |
109 | * ->i_mmap_lock | 109 | * ->i_mmap_mutex |
110 | */ | 110 | */ |
111 | 111 | ||
112 | /* | 112 | /* |
diff --git a/mm/filemap_xip.c b/mm/filemap_xip.c index 83364df74a33..93356cd12828 100644 --- a/mm/filemap_xip.c +++ b/mm/filemap_xip.c | |||
@@ -183,7 +183,7 @@ __xip_unmap (struct address_space * mapping, | |||
183 | return; | 183 | return; |
184 | 184 | ||
185 | retry: | 185 | retry: |
186 | spin_lock(&mapping->i_mmap_lock); | 186 | mutex_lock(&mapping->i_mmap_mutex); |
187 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { | 187 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { |
188 | mm = vma->vm_mm; | 188 | mm = vma->vm_mm; |
189 | address = vma->vm_start + | 189 | address = vma->vm_start + |
@@ -201,7 +201,7 @@ retry: | |||
201 | page_cache_release(page); | 201 | page_cache_release(page); |
202 | } | 202 | } |
203 | } | 203 | } |
204 | spin_unlock(&mapping->i_mmap_lock); | 204 | mutex_unlock(&mapping->i_mmap_mutex); |
205 | 205 | ||
206 | if (locked) { | 206 | if (locked) { |
207 | mutex_unlock(&xip_sparse_mutex); | 207 | mutex_unlock(&xip_sparse_mutex); |
diff --git a/mm/fremap.c b/mm/fremap.c index ec520c7b28df..7f4123056e06 100644 --- a/mm/fremap.c +++ b/mm/fremap.c | |||
@@ -211,13 +211,13 @@ SYSCALL_DEFINE5(remap_file_pages, unsigned long, start, unsigned long, size, | |||
211 | } | 211 | } |
212 | goto out; | 212 | goto out; |
213 | } | 213 | } |
214 | spin_lock(&mapping->i_mmap_lock); | 214 | mutex_lock(&mapping->i_mmap_mutex); |
215 | flush_dcache_mmap_lock(mapping); | 215 | flush_dcache_mmap_lock(mapping); |
216 | vma->vm_flags |= VM_NONLINEAR; | 216 | vma->vm_flags |= VM_NONLINEAR; |
217 | vma_prio_tree_remove(vma, &mapping->i_mmap); | 217 | vma_prio_tree_remove(vma, &mapping->i_mmap); |
218 | vma_nonlinear_insert(vma, &mapping->i_mmap_nonlinear); | 218 | vma_nonlinear_insert(vma, &mapping->i_mmap_nonlinear); |
219 | flush_dcache_mmap_unlock(mapping); | 219 | flush_dcache_mmap_unlock(mapping); |
220 | spin_unlock(&mapping->i_mmap_lock); | 220 | mutex_unlock(&mapping->i_mmap_mutex); |
221 | } | 221 | } |
222 | 222 | ||
223 | if (vma->vm_flags & VM_LOCKED) { | 223 | if (vma->vm_flags & VM_LOCKED) { |
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index bbb4a5bbb958..5fd68b95c671 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -2205,7 +2205,7 @@ void __unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, | |||
2205 | unsigned long sz = huge_page_size(h); | 2205 | unsigned long sz = huge_page_size(h); |
2206 | 2206 | ||
2207 | /* | 2207 | /* |
2208 | * A page gathering list, protected by per file i_mmap_lock. The | 2208 | * A page gathering list, protected by per file i_mmap_mutex. The |
2209 | * lock is used to avoid list corruption from multiple unmapping | 2209 | * lock is used to avoid list corruption from multiple unmapping |
2210 | * of the same page since we are using page->lru. | 2210 | * of the same page since we are using page->lru. |
2211 | */ | 2211 | */ |
@@ -2274,9 +2274,9 @@ void __unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, | |||
2274 | void unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, | 2274 | void unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start, |
2275 | unsigned long end, struct page *ref_page) | 2275 | unsigned long end, struct page *ref_page) |
2276 | { | 2276 | { |
2277 | spin_lock(&vma->vm_file->f_mapping->i_mmap_lock); | 2277 | mutex_lock(&vma->vm_file->f_mapping->i_mmap_mutex); |
2278 | __unmap_hugepage_range(vma, start, end, ref_page); | 2278 | __unmap_hugepage_range(vma, start, end, ref_page); |
2279 | spin_unlock(&vma->vm_file->f_mapping->i_mmap_lock); | 2279 | mutex_unlock(&vma->vm_file->f_mapping->i_mmap_mutex); |
2280 | } | 2280 | } |
2281 | 2281 | ||
2282 | /* | 2282 | /* |
@@ -2308,7 +2308,7 @@ static int unmap_ref_private(struct mm_struct *mm, struct vm_area_struct *vma, | |||
2308 | * this mapping should be shared between all the VMAs, | 2308 | * this mapping should be shared between all the VMAs, |
2309 | * __unmap_hugepage_range() is called as the lock is already held | 2309 | * __unmap_hugepage_range() is called as the lock is already held |
2310 | */ | 2310 | */ |
2311 | spin_lock(&mapping->i_mmap_lock); | 2311 | mutex_lock(&mapping->i_mmap_mutex); |
2312 | vma_prio_tree_foreach(iter_vma, &iter, &mapping->i_mmap, pgoff, pgoff) { | 2312 | vma_prio_tree_foreach(iter_vma, &iter, &mapping->i_mmap, pgoff, pgoff) { |
2313 | /* Do not unmap the current VMA */ | 2313 | /* Do not unmap the current VMA */ |
2314 | if (iter_vma == vma) | 2314 | if (iter_vma == vma) |
@@ -2326,7 +2326,7 @@ static int unmap_ref_private(struct mm_struct *mm, struct vm_area_struct *vma, | |||
2326 | address, address + huge_page_size(h), | 2326 | address, address + huge_page_size(h), |
2327 | page); | 2327 | page); |
2328 | } | 2328 | } |
2329 | spin_unlock(&mapping->i_mmap_lock); | 2329 | mutex_unlock(&mapping->i_mmap_mutex); |
2330 | 2330 | ||
2331 | return 1; | 2331 | return 1; |
2332 | } | 2332 | } |
@@ -2810,7 +2810,7 @@ void hugetlb_change_protection(struct vm_area_struct *vma, | |||
2810 | BUG_ON(address >= end); | 2810 | BUG_ON(address >= end); |
2811 | flush_cache_range(vma, address, end); | 2811 | flush_cache_range(vma, address, end); |
2812 | 2812 | ||
2813 | spin_lock(&vma->vm_file->f_mapping->i_mmap_lock); | 2813 | mutex_lock(&vma->vm_file->f_mapping->i_mmap_mutex); |
2814 | spin_lock(&mm->page_table_lock); | 2814 | spin_lock(&mm->page_table_lock); |
2815 | for (; address < end; address += huge_page_size(h)) { | 2815 | for (; address < end; address += huge_page_size(h)) { |
2816 | ptep = huge_pte_offset(mm, address); | 2816 | ptep = huge_pte_offset(mm, address); |
@@ -2825,7 +2825,7 @@ void hugetlb_change_protection(struct vm_area_struct *vma, | |||
2825 | } | 2825 | } |
2826 | } | 2826 | } |
2827 | spin_unlock(&mm->page_table_lock); | 2827 | spin_unlock(&mm->page_table_lock); |
2828 | spin_unlock(&vma->vm_file->f_mapping->i_mmap_lock); | 2828 | mutex_unlock(&vma->vm_file->f_mapping->i_mmap_mutex); |
2829 | 2829 | ||
2830 | flush_tlb_range(vma, start, end); | 2830 | flush_tlb_range(vma, start, end); |
2831 | } | 2831 | } |
diff --git a/mm/memory-failure.c b/mm/memory-failure.c index 2b9a5eef39e0..12178ec32ab5 100644 --- a/mm/memory-failure.c +++ b/mm/memory-failure.c | |||
@@ -429,7 +429,7 @@ static void collect_procs_file(struct page *page, struct list_head *to_kill, | |||
429 | */ | 429 | */ |
430 | 430 | ||
431 | read_lock(&tasklist_lock); | 431 | read_lock(&tasklist_lock); |
432 | spin_lock(&mapping->i_mmap_lock); | 432 | mutex_lock(&mapping->i_mmap_mutex); |
433 | for_each_process(tsk) { | 433 | for_each_process(tsk) { |
434 | pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); | 434 | pgoff_t pgoff = page->index << (PAGE_CACHE_SHIFT - PAGE_SHIFT); |
435 | 435 | ||
@@ -449,7 +449,7 @@ static void collect_procs_file(struct page *page, struct list_head *to_kill, | |||
449 | add_to_kill(tsk, page, vma, to_kill, tkc); | 449 | add_to_kill(tsk, page, vma, to_kill, tkc); |
450 | } | 450 | } |
451 | } | 451 | } |
452 | spin_unlock(&mapping->i_mmap_lock); | 452 | mutex_unlock(&mapping->i_mmap_mutex); |
453 | read_unlock(&tasklist_lock); | 453 | read_unlock(&tasklist_lock); |
454 | } | 454 | } |
455 | 455 | ||
diff --git a/mm/memory.c b/mm/memory.c index 18655878b9f8..7bbe4d3df756 100644 --- a/mm/memory.c +++ b/mm/memory.c | |||
@@ -2667,12 +2667,12 @@ void unmap_mapping_range(struct address_space *mapping, | |||
2667 | details.last_index = ULONG_MAX; | 2667 | details.last_index = ULONG_MAX; |
2668 | 2668 | ||
2669 | 2669 | ||
2670 | spin_lock(&mapping->i_mmap_lock); | 2670 | mutex_lock(&mapping->i_mmap_mutex); |
2671 | if (unlikely(!prio_tree_empty(&mapping->i_mmap))) | 2671 | if (unlikely(!prio_tree_empty(&mapping->i_mmap))) |
2672 | unmap_mapping_range_tree(&mapping->i_mmap, &details); | 2672 | unmap_mapping_range_tree(&mapping->i_mmap, &details); |
2673 | if (unlikely(!list_empty(&mapping->i_mmap_nonlinear))) | 2673 | if (unlikely(!list_empty(&mapping->i_mmap_nonlinear))) |
2674 | unmap_mapping_range_list(&mapping->i_mmap_nonlinear, &details); | 2674 | unmap_mapping_range_list(&mapping->i_mmap_nonlinear, &details); |
2675 | spin_unlock(&mapping->i_mmap_lock); | 2675 | mutex_unlock(&mapping->i_mmap_mutex); |
2676 | } | 2676 | } |
2677 | EXPORT_SYMBOL(unmap_mapping_range); | 2677 | EXPORT_SYMBOL(unmap_mapping_range); |
2678 | 2678 | ||
@@ -194,7 +194,7 @@ error: | |||
194 | } | 194 | } |
195 | 195 | ||
196 | /* | 196 | /* |
197 | * Requires inode->i_mapping->i_mmap_lock | 197 | * Requires inode->i_mapping->i_mmap_mutex |
198 | */ | 198 | */ |
199 | static void __remove_shared_vm_struct(struct vm_area_struct *vma, | 199 | static void __remove_shared_vm_struct(struct vm_area_struct *vma, |
200 | struct file *file, struct address_space *mapping) | 200 | struct file *file, struct address_space *mapping) |
@@ -222,9 +222,9 @@ void unlink_file_vma(struct vm_area_struct *vma) | |||
222 | 222 | ||
223 | if (file) { | 223 | if (file) { |
224 | struct address_space *mapping = file->f_mapping; | 224 | struct address_space *mapping = file->f_mapping; |
225 | spin_lock(&mapping->i_mmap_lock); | 225 | mutex_lock(&mapping->i_mmap_mutex); |
226 | __remove_shared_vm_struct(vma, file, mapping); | 226 | __remove_shared_vm_struct(vma, file, mapping); |
227 | spin_unlock(&mapping->i_mmap_lock); | 227 | mutex_unlock(&mapping->i_mmap_mutex); |
228 | } | 228 | } |
229 | } | 229 | } |
230 | 230 | ||
@@ -446,13 +446,13 @@ static void vma_link(struct mm_struct *mm, struct vm_area_struct *vma, | |||
446 | mapping = vma->vm_file->f_mapping; | 446 | mapping = vma->vm_file->f_mapping; |
447 | 447 | ||
448 | if (mapping) | 448 | if (mapping) |
449 | spin_lock(&mapping->i_mmap_lock); | 449 | mutex_lock(&mapping->i_mmap_mutex); |
450 | 450 | ||
451 | __vma_link(mm, vma, prev, rb_link, rb_parent); | 451 | __vma_link(mm, vma, prev, rb_link, rb_parent); |
452 | __vma_link_file(vma); | 452 | __vma_link_file(vma); |
453 | 453 | ||
454 | if (mapping) | 454 | if (mapping) |
455 | spin_unlock(&mapping->i_mmap_lock); | 455 | mutex_unlock(&mapping->i_mmap_mutex); |
456 | 456 | ||
457 | mm->map_count++; | 457 | mm->map_count++; |
458 | validate_mm(mm); | 458 | validate_mm(mm); |
@@ -555,7 +555,7 @@ again: remove_next = 1 + (end > next->vm_end); | |||
555 | mapping = file->f_mapping; | 555 | mapping = file->f_mapping; |
556 | if (!(vma->vm_flags & VM_NONLINEAR)) | 556 | if (!(vma->vm_flags & VM_NONLINEAR)) |
557 | root = &mapping->i_mmap; | 557 | root = &mapping->i_mmap; |
558 | spin_lock(&mapping->i_mmap_lock); | 558 | mutex_lock(&mapping->i_mmap_mutex); |
559 | if (insert) { | 559 | if (insert) { |
560 | /* | 560 | /* |
561 | * Put into prio_tree now, so instantiated pages | 561 | * Put into prio_tree now, so instantiated pages |
@@ -622,7 +622,7 @@ again: remove_next = 1 + (end > next->vm_end); | |||
622 | if (anon_vma) | 622 | if (anon_vma) |
623 | anon_vma_unlock(anon_vma); | 623 | anon_vma_unlock(anon_vma); |
624 | if (mapping) | 624 | if (mapping) |
625 | spin_unlock(&mapping->i_mmap_lock); | 625 | mutex_unlock(&mapping->i_mmap_mutex); |
626 | 626 | ||
627 | if (remove_next) { | 627 | if (remove_next) { |
628 | if (file) { | 628 | if (file) { |
@@ -2290,7 +2290,7 @@ void exit_mmap(struct mm_struct *mm) | |||
2290 | 2290 | ||
2291 | /* Insert vm structure into process list sorted by address | 2291 | /* Insert vm structure into process list sorted by address |
2292 | * and into the inode's i_mmap tree. If vm_file is non-NULL | 2292 | * and into the inode's i_mmap tree. If vm_file is non-NULL |
2293 | * then i_mmap_lock is taken here. | 2293 | * then i_mmap_mutex is taken here. |
2294 | */ | 2294 | */ |
2295 | int insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) | 2295 | int insert_vm_struct(struct mm_struct * mm, struct vm_area_struct * vma) |
2296 | { | 2296 | { |
@@ -2532,7 +2532,7 @@ static void vm_lock_mapping(struct mm_struct *mm, struct address_space *mapping) | |||
2532 | */ | 2532 | */ |
2533 | if (test_and_set_bit(AS_MM_ALL_LOCKS, &mapping->flags)) | 2533 | if (test_and_set_bit(AS_MM_ALL_LOCKS, &mapping->flags)) |
2534 | BUG(); | 2534 | BUG(); |
2535 | spin_lock_nest_lock(&mapping->i_mmap_lock, &mm->mmap_sem); | 2535 | mutex_lock_nest_lock(&mapping->i_mmap_mutex, &mm->mmap_sem); |
2536 | } | 2536 | } |
2537 | } | 2537 | } |
2538 | 2538 | ||
@@ -2559,7 +2559,7 @@ static void vm_lock_mapping(struct mm_struct *mm, struct address_space *mapping) | |||
2559 | * vma in this mm is backed by the same anon_vma or address_space. | 2559 | * vma in this mm is backed by the same anon_vma or address_space. |
2560 | * | 2560 | * |
2561 | * We can take all the locks in random order because the VM code | 2561 | * We can take all the locks in random order because the VM code |
2562 | * taking i_mmap_lock or anon_vma->lock outside the mmap_sem never | 2562 | * taking i_mmap_mutex or anon_vma->lock outside the mmap_sem never |
2563 | * takes more than one of them in a row. Secondly we're protected | 2563 | * takes more than one of them in a row. Secondly we're protected |
2564 | * against a concurrent mm_take_all_locks() by the mm_all_locks_mutex. | 2564 | * against a concurrent mm_take_all_locks() by the mm_all_locks_mutex. |
2565 | * | 2565 | * |
@@ -2631,7 +2631,7 @@ static void vm_unlock_mapping(struct address_space *mapping) | |||
2631 | * AS_MM_ALL_LOCKS can't change to 0 from under us | 2631 | * AS_MM_ALL_LOCKS can't change to 0 from under us |
2632 | * because we hold the mm_all_locks_mutex. | 2632 | * because we hold the mm_all_locks_mutex. |
2633 | */ | 2633 | */ |
2634 | spin_unlock(&mapping->i_mmap_lock); | 2634 | mutex_unlock(&mapping->i_mmap_mutex); |
2635 | if (!test_and_clear_bit(AS_MM_ALL_LOCKS, | 2635 | if (!test_and_clear_bit(AS_MM_ALL_LOCKS, |
2636 | &mapping->flags)) | 2636 | &mapping->flags)) |
2637 | BUG(); | 2637 | BUG(); |
diff --git a/mm/mremap.c b/mm/mremap.c index 909e1e1e99b1..506fa44403df 100644 --- a/mm/mremap.c +++ b/mm/mremap.c | |||
@@ -93,7 +93,7 @@ static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, | |||
93 | * and we propagate stale pages into the dst afterward. | 93 | * and we propagate stale pages into the dst afterward. |
94 | */ | 94 | */ |
95 | mapping = vma->vm_file->f_mapping; | 95 | mapping = vma->vm_file->f_mapping; |
96 | spin_lock(&mapping->i_mmap_lock); | 96 | mutex_lock(&mapping->i_mmap_mutex); |
97 | } | 97 | } |
98 | 98 | ||
99 | /* | 99 | /* |
@@ -122,7 +122,7 @@ static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, | |||
122 | pte_unmap(new_pte - 1); | 122 | pte_unmap(new_pte - 1); |
123 | pte_unmap_unlock(old_pte - 1, old_ptl); | 123 | pte_unmap_unlock(old_pte - 1, old_ptl); |
124 | if (mapping) | 124 | if (mapping) |
125 | spin_unlock(&mapping->i_mmap_lock); | 125 | mutex_unlock(&mapping->i_mmap_mutex); |
126 | mmu_notifier_invalidate_range_end(vma->vm_mm, old_start, old_end); | 126 | mmu_notifier_invalidate_range_end(vma->vm_mm, old_start, old_end); |
127 | } | 127 | } |
128 | 128 | ||
@@ -24,7 +24,7 @@ | |||
24 | * inode->i_alloc_sem (vmtruncate_range) | 24 | * inode->i_alloc_sem (vmtruncate_range) |
25 | * mm->mmap_sem | 25 | * mm->mmap_sem |
26 | * page->flags PG_locked (lock_page) | 26 | * page->flags PG_locked (lock_page) |
27 | * mapping->i_mmap_lock | 27 | * mapping->i_mmap_mutex |
28 | * anon_vma->lock | 28 | * anon_vma->lock |
29 | * mm->page_table_lock or pte_lock | 29 | * mm->page_table_lock or pte_lock |
30 | * zone->lru_lock (in mark_page_accessed, isolate_lru_page) | 30 | * zone->lru_lock (in mark_page_accessed, isolate_lru_page) |
@@ -646,14 +646,14 @@ static int page_referenced_file(struct page *page, | |||
646 | * The page lock not only makes sure that page->mapping cannot | 646 | * The page lock not only makes sure that page->mapping cannot |
647 | * suddenly be NULLified by truncation, it makes sure that the | 647 | * suddenly be NULLified by truncation, it makes sure that the |
648 | * structure at mapping cannot be freed and reused yet, | 648 | * structure at mapping cannot be freed and reused yet, |
649 | * so we can safely take mapping->i_mmap_lock. | 649 | * so we can safely take mapping->i_mmap_mutex. |
650 | */ | 650 | */ |
651 | BUG_ON(!PageLocked(page)); | 651 | BUG_ON(!PageLocked(page)); |
652 | 652 | ||
653 | spin_lock(&mapping->i_mmap_lock); | 653 | mutex_lock(&mapping->i_mmap_mutex); |
654 | 654 | ||
655 | /* | 655 | /* |
656 | * i_mmap_lock does not stabilize mapcount at all, but mapcount | 656 | * i_mmap_mutex does not stabilize mapcount at all, but mapcount |
657 | * is more likely to be accurate if we note it after spinning. | 657 | * is more likely to be accurate if we note it after spinning. |
658 | */ | 658 | */ |
659 | mapcount = page_mapcount(page); | 659 | mapcount = page_mapcount(page); |
@@ -675,7 +675,7 @@ static int page_referenced_file(struct page *page, | |||
675 | break; | 675 | break; |
676 | } | 676 | } |
677 | 677 | ||
678 | spin_unlock(&mapping->i_mmap_lock); | 678 | mutex_unlock(&mapping->i_mmap_mutex); |
679 | return referenced; | 679 | return referenced; |
680 | } | 680 | } |
681 | 681 | ||
@@ -762,7 +762,7 @@ static int page_mkclean_file(struct address_space *mapping, struct page *page) | |||
762 | 762 | ||
763 | BUG_ON(PageAnon(page)); | 763 | BUG_ON(PageAnon(page)); |
764 | 764 | ||
765 | spin_lock(&mapping->i_mmap_lock); | 765 | mutex_lock(&mapping->i_mmap_mutex); |
766 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { | 766 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { |
767 | if (vma->vm_flags & VM_SHARED) { | 767 | if (vma->vm_flags & VM_SHARED) { |
768 | unsigned long address = vma_address(page, vma); | 768 | unsigned long address = vma_address(page, vma); |
@@ -771,7 +771,7 @@ static int page_mkclean_file(struct address_space *mapping, struct page *page) | |||
771 | ret += page_mkclean_one(page, vma, address); | 771 | ret += page_mkclean_one(page, vma, address); |
772 | } | 772 | } |
773 | } | 773 | } |
774 | spin_unlock(&mapping->i_mmap_lock); | 774 | mutex_unlock(&mapping->i_mmap_mutex); |
775 | return ret; | 775 | return ret; |
776 | } | 776 | } |
777 | 777 | ||
@@ -1119,7 +1119,7 @@ out_mlock: | |||
1119 | /* | 1119 | /* |
1120 | * We need mmap_sem locking, Otherwise VM_LOCKED check makes | 1120 | * We need mmap_sem locking, Otherwise VM_LOCKED check makes |
1121 | * unstable result and race. Plus, We can't wait here because | 1121 | * unstable result and race. Plus, We can't wait here because |
1122 | * we now hold anon_vma->lock or mapping->i_mmap_lock. | 1122 | * we now hold anon_vma->lock or mapping->i_mmap_mutex. |
1123 | * if trylock failed, the page remain in evictable lru and later | 1123 | * if trylock failed, the page remain in evictable lru and later |
1124 | * vmscan could retry to move the page to unevictable lru if the | 1124 | * vmscan could retry to move the page to unevictable lru if the |
1125 | * page is actually mlocked. | 1125 | * page is actually mlocked. |
@@ -1345,7 +1345,7 @@ static int try_to_unmap_file(struct page *page, enum ttu_flags flags) | |||
1345 | unsigned long max_nl_size = 0; | 1345 | unsigned long max_nl_size = 0; |
1346 | unsigned int mapcount; | 1346 | unsigned int mapcount; |
1347 | 1347 | ||
1348 | spin_lock(&mapping->i_mmap_lock); | 1348 | mutex_lock(&mapping->i_mmap_mutex); |
1349 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { | 1349 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { |
1350 | unsigned long address = vma_address(page, vma); | 1350 | unsigned long address = vma_address(page, vma); |
1351 | if (address == -EFAULT) | 1351 | if (address == -EFAULT) |
@@ -1391,7 +1391,7 @@ static int try_to_unmap_file(struct page *page, enum ttu_flags flags) | |||
1391 | mapcount = page_mapcount(page); | 1391 | mapcount = page_mapcount(page); |
1392 | if (!mapcount) | 1392 | if (!mapcount) |
1393 | goto out; | 1393 | goto out; |
1394 | cond_resched_lock(&mapping->i_mmap_lock); | 1394 | cond_resched(); |
1395 | 1395 | ||
1396 | max_nl_size = (max_nl_size + CLUSTER_SIZE - 1) & CLUSTER_MASK; | 1396 | max_nl_size = (max_nl_size + CLUSTER_SIZE - 1) & CLUSTER_MASK; |
1397 | if (max_nl_cursor == 0) | 1397 | if (max_nl_cursor == 0) |
@@ -1413,7 +1413,7 @@ static int try_to_unmap_file(struct page *page, enum ttu_flags flags) | |||
1413 | } | 1413 | } |
1414 | vma->vm_private_data = (void *) max_nl_cursor; | 1414 | vma->vm_private_data = (void *) max_nl_cursor; |
1415 | } | 1415 | } |
1416 | cond_resched_lock(&mapping->i_mmap_lock); | 1416 | cond_resched(); |
1417 | max_nl_cursor += CLUSTER_SIZE; | 1417 | max_nl_cursor += CLUSTER_SIZE; |
1418 | } while (max_nl_cursor <= max_nl_size); | 1418 | } while (max_nl_cursor <= max_nl_size); |
1419 | 1419 | ||
@@ -1425,7 +1425,7 @@ static int try_to_unmap_file(struct page *page, enum ttu_flags flags) | |||
1425 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, shared.vm_set.list) | 1425 | list_for_each_entry(vma, &mapping->i_mmap_nonlinear, shared.vm_set.list) |
1426 | vma->vm_private_data = NULL; | 1426 | vma->vm_private_data = NULL; |
1427 | out: | 1427 | out: |
1428 | spin_unlock(&mapping->i_mmap_lock); | 1428 | mutex_unlock(&mapping->i_mmap_mutex); |
1429 | return ret; | 1429 | return ret; |
1430 | } | 1430 | } |
1431 | 1431 | ||
@@ -1544,7 +1544,7 @@ static int rmap_walk_file(struct page *page, int (*rmap_one)(struct page *, | |||
1544 | 1544 | ||
1545 | if (!mapping) | 1545 | if (!mapping) |
1546 | return ret; | 1546 | return ret; |
1547 | spin_lock(&mapping->i_mmap_lock); | 1547 | mutex_lock(&mapping->i_mmap_mutex); |
1548 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { | 1548 | vma_prio_tree_foreach(vma, &iter, &mapping->i_mmap, pgoff, pgoff) { |
1549 | unsigned long address = vma_address(page, vma); | 1549 | unsigned long address = vma_address(page, vma); |
1550 | if (address == -EFAULT) | 1550 | if (address == -EFAULT) |
@@ -1558,7 +1558,7 @@ static int rmap_walk_file(struct page *page, int (*rmap_one)(struct page *, | |||
1558 | * never contain migration ptes. Decide what to do about this | 1558 | * never contain migration ptes. Decide what to do about this |
1559 | * limitation to linear when we need rmap_walk() on nonlinear. | 1559 | * limitation to linear when we need rmap_walk() on nonlinear. |
1560 | */ | 1560 | */ |
1561 | spin_unlock(&mapping->i_mmap_lock); | 1561 | mutex_unlock(&mapping->i_mmap_mutex); |
1562 | return ret; | 1562 | return ret; |
1563 | } | 1563 | } |
1564 | 1564 | ||