aboutsummaryrefslogtreecommitdiffstats
path: root/mm/mlock.c
diff options
context:
space:
mode:
authorRik van Riel <riel@redhat.com>2008-10-18 23:26:50 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2008-10-20 11:52:31 -0400
commitba470de43188cdbff795b5da43a1474523c6c2fb (patch)
tree0477460fa8c3e61edd9f1534cd2193656e586f8b /mm/mlock.c
parent8edb08caf68184fb170f4f69c7445929e199eaea (diff)
mmap: handle mlocked pages during map, remap, unmap
Originally by Nick Piggin <npiggin@suse.de> Remove mlocked pages from the LRU using "unevictable infrastructure" during mmap(), munmap(), mremap() and truncate(). Try to move back to normal LRU lists on munmap() when last mlocked mapping removed. Remove PageMlocked() status when page truncated from file. [akpm@linux-foundation.org: cleanup] [kamezawa.hiroyu@jp.fujitsu.com: fix double unlock_page()] [kosaki.motohiro@jp.fujitsu.com: split LRU: munlock rework] [lee.schermerhorn@hp.com: mlock: fix __mlock_vma_pages_range comment block] [akpm@linux-foundation.org: remove bogus kerneldoc token] Signed-off-by: Nick Piggin <npiggin@suse.de> Signed-off-by: Lee Schermerhorn <lee.schermerhorn@hp.com> Signed-off-by: Rik van Riel <riel@redhat.com> Signed-off-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com> Signed-off-by: KAMEZAWA Hiroyuki <kamewzawa.hiroyu@jp.fujitsu.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/mlock.c')
-rw-r--r--mm/mlock.c221
1 files changed, 90 insertions, 131 deletions
diff --git a/mm/mlock.c b/mm/mlock.c
index c83896a72504..8b478350a2a1 100644
--- a/mm/mlock.c
+++ b/mm/mlock.c
@@ -112,26 +112,49 @@ static void munlock_vma_page(struct page *page)
112 } 112 }
113} 113}
114 114
115/* 115/**
116 * mlock a range of pages in the vma. 116 * __mlock_vma_pages_range() - mlock/munlock a range of pages in the vma.
117 * @vma: target vma
118 * @start: start address
119 * @end: end address
120 * @mlock: 0 indicate munlock, otherwise mlock.
121 *
122 * If @mlock == 0, unlock an mlocked range;
123 * else mlock the range of pages. This takes care of making the pages present ,
124 * too.
117 * 125 *
118 * This takes care of making the pages present too. 126 * return 0 on success, negative error code on error.
119 * 127 *
120 * vma->vm_mm->mmap_sem must be held for write. 128 * vma->vm_mm->mmap_sem must be held for at least read.
121 */ 129 */
122static int __mlock_vma_pages_range(struct vm_area_struct *vma, 130static long __mlock_vma_pages_range(struct vm_area_struct *vma,
123 unsigned long start, unsigned long end) 131 unsigned long start, unsigned long end,
132 int mlock)
124{ 133{
125 struct mm_struct *mm = vma->vm_mm; 134 struct mm_struct *mm = vma->vm_mm;
126 unsigned long addr = start; 135 unsigned long addr = start;
127 struct page *pages[16]; /* 16 gives a reasonable batch */ 136 struct page *pages[16]; /* 16 gives a reasonable batch */
128 int write = !!(vma->vm_flags & VM_WRITE);
129 int nr_pages = (end - start) / PAGE_SIZE; 137 int nr_pages = (end - start) / PAGE_SIZE;
130 int ret; 138 int ret;
139 int gup_flags = 0;
131 140
132 VM_BUG_ON(start & ~PAGE_MASK || end & ~PAGE_MASK); 141 VM_BUG_ON(start & ~PAGE_MASK);
133 VM_BUG_ON(start < vma->vm_start || end > vma->vm_end); 142 VM_BUG_ON(end & ~PAGE_MASK);
134 VM_BUG_ON(!rwsem_is_locked(&vma->vm_mm->mmap_sem)); 143 VM_BUG_ON(start < vma->vm_start);
144 VM_BUG_ON(end > vma->vm_end);
145 VM_BUG_ON((!rwsem_is_locked(&mm->mmap_sem)) &&
146 (atomic_read(&mm->mm_users) != 0));
147
148 /*
149 * mlock: don't page populate if page has PROT_NONE permission.
150 * munlock: the pages always do munlock althrough
151 * its has PROT_NONE permission.
152 */
153 if (!mlock)
154 gup_flags |= GUP_FLAGS_IGNORE_VMA_PERMISSIONS;
155
156 if (vma->vm_flags & VM_WRITE)
157 gup_flags |= GUP_FLAGS_WRITE;
135 158
136 lru_add_drain_all(); /* push cached pages to LRU */ 159 lru_add_drain_all(); /* push cached pages to LRU */
137 160
@@ -146,9 +169,9 @@ static int __mlock_vma_pages_range(struct vm_area_struct *vma,
146 * disable migration of this page. However, page may 169 * disable migration of this page. However, page may
147 * still be truncated out from under us. 170 * still be truncated out from under us.
148 */ 171 */
149 ret = get_user_pages(current, mm, addr, 172 ret = __get_user_pages(current, mm, addr,
150 min_t(int, nr_pages, ARRAY_SIZE(pages)), 173 min_t(int, nr_pages, ARRAY_SIZE(pages)),
151 write, 0, pages, NULL); 174 gup_flags, pages, NULL);
152 /* 175 /*
153 * This can happen for, e.g., VM_NONLINEAR regions before 176 * This can happen for, e.g., VM_NONLINEAR regions before
154 * a page has been allocated and mapped at a given offset, 177 * a page has been allocated and mapped at a given offset,
@@ -178,8 +201,12 @@ static int __mlock_vma_pages_range(struct vm_area_struct *vma,
178 * by the elevated reference, we need only check for 201 * by the elevated reference, we need only check for
179 * page truncation (file-cache only). 202 * page truncation (file-cache only).
180 */ 203 */
181 if (page->mapping) 204 if (page->mapping) {
182 mlock_vma_page(page); 205 if (mlock)
206 mlock_vma_page(page);
207 else
208 munlock_vma_page(page);
209 }
183 unlock_page(page); 210 unlock_page(page);
184 put_page(page); /* ref from get_user_pages() */ 211 put_page(page); /* ref from get_user_pages() */
185 212
@@ -197,125 +224,38 @@ static int __mlock_vma_pages_range(struct vm_area_struct *vma,
197 return 0; /* count entire vma as locked_vm */ 224 return 0; /* count entire vma as locked_vm */
198} 225}
199 226
200/*
201 * private structure for munlock page table walk
202 */
203struct munlock_page_walk {
204 struct vm_area_struct *vma;
205 pmd_t *pmd; /* for migration_entry_wait() */
206};
207
208/*
209 * munlock normal pages for present ptes
210 */
211static int __munlock_pte_handler(pte_t *ptep, unsigned long addr,
212 unsigned long end, struct mm_walk *walk)
213{
214 struct munlock_page_walk *mpw = walk->private;
215 swp_entry_t entry;
216 struct page *page;
217 pte_t pte;
218
219retry:
220 pte = *ptep;
221 /*
222 * If it's a swap pte, we might be racing with page migration.
223 */
224 if (unlikely(!pte_present(pte))) {
225 if (!is_swap_pte(pte))
226 goto out;
227 entry = pte_to_swp_entry(pte);
228 if (is_migration_entry(entry)) {
229 migration_entry_wait(mpw->vma->vm_mm, mpw->pmd, addr);
230 goto retry;
231 }
232 goto out;
233 }
234
235 page = vm_normal_page(mpw->vma, addr, pte);
236 if (!page)
237 goto out;
238
239 lock_page(page);
240 if (!page->mapping) {
241 unlock_page(page);
242 goto retry;
243 }
244 munlock_vma_page(page);
245 unlock_page(page);
246
247out:
248 return 0;
249}
250
251/*
252 * Save pmd for pte handler for waiting on migration entries
253 */
254static int __munlock_pmd_handler(pmd_t *pmd, unsigned long addr,
255 unsigned long end, struct mm_walk *walk)
256{
257 struct munlock_page_walk *mpw = walk->private;
258
259 mpw->pmd = pmd;
260 return 0;
261}
262
263
264/*
265 * munlock a range of pages in the vma using standard page table walk.
266 *
267 * vma->vm_mm->mmap_sem must be held for write.
268 */
269static void __munlock_vma_pages_range(struct vm_area_struct *vma,
270 unsigned long start, unsigned long end)
271{
272 struct mm_struct *mm = vma->vm_mm;
273 struct munlock_page_walk mpw = {
274 .vma = vma,
275 };
276 struct mm_walk munlock_page_walk = {
277 .pmd_entry = __munlock_pmd_handler,
278 .pte_entry = __munlock_pte_handler,
279 .private = &mpw,
280 .mm = mm,
281 };
282
283 VM_BUG_ON(start & ~PAGE_MASK || end & ~PAGE_MASK);
284 VM_BUG_ON(!rwsem_is_locked(&vma->vm_mm->mmap_sem));
285 VM_BUG_ON(start < vma->vm_start);
286 VM_BUG_ON(end > vma->vm_end);
287
288 lru_add_drain_all(); /* push cached pages to LRU */
289 walk_page_range(start, end, &munlock_page_walk);
290 lru_add_drain_all(); /* to update stats */
291}
292
293#else /* CONFIG_UNEVICTABLE_LRU */ 227#else /* CONFIG_UNEVICTABLE_LRU */
294 228
295/* 229/*
296 * Just make pages present if VM_LOCKED. No-op if unlocking. 230 * Just make pages present if VM_LOCKED. No-op if unlocking.
297 */ 231 */
298static int __mlock_vma_pages_range(struct vm_area_struct *vma, 232static long __mlock_vma_pages_range(struct vm_area_struct *vma,
299 unsigned long start, unsigned long end) 233 unsigned long start, unsigned long end,
234 int mlock)
300{ 235{
301 if (vma->vm_flags & VM_LOCKED) 236 if (mlock && (vma->vm_flags & VM_LOCKED))
302 make_pages_present(start, end); 237 make_pages_present(start, end);
303 return 0; 238 return 0;
304} 239}
305
306/*
307 * munlock a range of pages in the vma -- no-op.
308 */
309static void __munlock_vma_pages_range(struct vm_area_struct *vma,
310 unsigned long start, unsigned long end)
311{
312}
313#endif /* CONFIG_UNEVICTABLE_LRU */ 240#endif /* CONFIG_UNEVICTABLE_LRU */
314 241
315/* 242/**
316 * mlock all pages in this vma range. For mmap()/mremap()/... 243 * mlock_vma_pages_range() - mlock pages in specified vma range.
244 * @vma - the vma containing the specfied address range
245 * @start - starting address in @vma to mlock
246 * @end - end address [+1] in @vma to mlock
247 *
248 * For mmap()/mremap()/expansion of mlocked vma.
249 *
250 * return 0 on success for "normal" vmas.
251 *
252 * return number of pages [> 0] to be removed from locked_vm on success
253 * of "special" vmas.
254 *
255 * return negative error if vma spanning @start-@range disappears while
256 * mmap semaphore is dropped. Unlikely?
317 */ 257 */
318int mlock_vma_pages_range(struct vm_area_struct *vma, 258long mlock_vma_pages_range(struct vm_area_struct *vma,
319 unsigned long start, unsigned long end) 259 unsigned long start, unsigned long end)
320{ 260{
321 struct mm_struct *mm = vma->vm_mm; 261 struct mm_struct *mm = vma->vm_mm;
@@ -331,8 +271,10 @@ int mlock_vma_pages_range(struct vm_area_struct *vma,
331 if (!((vma->vm_flags & (VM_DONTEXPAND | VM_RESERVED)) || 271 if (!((vma->vm_flags & (VM_DONTEXPAND | VM_RESERVED)) ||
332 is_vm_hugetlb_page(vma) || 272 is_vm_hugetlb_page(vma) ||
333 vma == get_gate_vma(current))) { 273 vma == get_gate_vma(current))) {
274 long error;
334 downgrade_write(&mm->mmap_sem); 275 downgrade_write(&mm->mmap_sem);
335 nr_pages = __mlock_vma_pages_range(vma, start, end); 276
277 error = __mlock_vma_pages_range(vma, start, end, 1);
336 278
337 up_read(&mm->mmap_sem); 279 up_read(&mm->mmap_sem);
338 /* vma can change or disappear */ 280 /* vma can change or disappear */
@@ -340,8 +282,9 @@ int mlock_vma_pages_range(struct vm_area_struct *vma,
340 vma = find_vma(mm, start); 282 vma = find_vma(mm, start);
341 /* non-NULL vma must contain @start, but need to check @end */ 283 /* non-NULL vma must contain @start, but need to check @end */
342 if (!vma || end > vma->vm_end) 284 if (!vma || end > vma->vm_end)
343 return -EAGAIN; 285 return -ENOMEM;
344 return nr_pages; 286
287 return 0; /* hide other errors from mmap(), et al */
345 } 288 }
346 289
347 /* 290 /*
@@ -356,17 +299,33 @@ int mlock_vma_pages_range(struct vm_area_struct *vma,
356 299
357no_mlock: 300no_mlock:
358 vma->vm_flags &= ~VM_LOCKED; /* and don't come back! */ 301 vma->vm_flags &= ~VM_LOCKED; /* and don't come back! */
359 return nr_pages; /* pages NOT mlocked */ 302 return nr_pages; /* error or pages NOT mlocked */
360} 303}
361 304
362 305
363/* 306/*
364 * munlock all pages in vma. For munmap() and exit(). 307 * munlock_vma_pages_range() - munlock all pages in the vma range.'
308 * @vma - vma containing range to be munlock()ed.
309 * @start - start address in @vma of the range
310 * @end - end of range in @vma.
311 *
312 * For mremap(), munmap() and exit().
313 *
314 * Called with @vma VM_LOCKED.
315 *
316 * Returns with VM_LOCKED cleared. Callers must be prepared to
317 * deal with this.
318 *
319 * We don't save and restore VM_LOCKED here because pages are
320 * still on lru. In unmap path, pages might be scanned by reclaim
321 * and re-mlocked by try_to_{munlock|unmap} before we unmap and
322 * free them. This will result in freeing mlocked pages.
365 */ 323 */
366void munlock_vma_pages_all(struct vm_area_struct *vma) 324void munlock_vma_pages_range(struct vm_area_struct *vma,
325 unsigned long start, unsigned long end)
367{ 326{
368 vma->vm_flags &= ~VM_LOCKED; 327 vma->vm_flags &= ~VM_LOCKED;
369 __munlock_vma_pages_range(vma, vma->vm_start, vma->vm_end); 328 __mlock_vma_pages_range(vma, start, end, 0);
370} 329}
371 330
372/* 331/*
@@ -443,7 +402,7 @@ success:
443 */ 402 */
444 downgrade_write(&mm->mmap_sem); 403 downgrade_write(&mm->mmap_sem);
445 404
446 ret = __mlock_vma_pages_range(vma, start, end); 405 ret = __mlock_vma_pages_range(vma, start, end, 1);
447 if (ret > 0) { 406 if (ret > 0) {
448 mm->locked_vm -= ret; 407 mm->locked_vm -= ret;
449 ret = 0; 408 ret = 0;
@@ -460,7 +419,7 @@ success:
460 *prev = find_vma(mm, start); 419 *prev = find_vma(mm, start);
461 /* non-NULL *prev must contain @start, but need to check @end */ 420 /* non-NULL *prev must contain @start, but need to check @end */
462 if (!(*prev) || end > (*prev)->vm_end) 421 if (!(*prev) || end > (*prev)->vm_end)
463 ret = -EAGAIN; 422 ret = -ENOMEM;
464 } else { 423 } else {
465 /* 424 /*
466 * TODO: for unlocking, pages will already be resident, so 425 * TODO: for unlocking, pages will already be resident, so
@@ -469,7 +428,7 @@ success:
469 * while. Should we downgrade the semaphore for both lock 428 * while. Should we downgrade the semaphore for both lock
470 * AND unlock ? 429 * AND unlock ?
471 */ 430 */
472 __munlock_vma_pages_range(vma, start, end); 431 __mlock_vma_pages_range(vma, start, end, 0);
473 } 432 }
474 433
475out: 434out: