aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorMichel Lespinasse <walken@google.com>2013-02-22 19:32:44 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2013-02-23 20:50:11 -0500
commitcea10a19b7972a1954c4a2d05a7de8db48b444fb (patch)
tree694b3c906259cfbfc7b7cb1b0eb507ecf0d1d63c /mm
parentc22c0d6344c362b1dde5d8e160d3d07536aca120 (diff)
mm: directly use __mlock_vma_pages_range() in find_extend_vma()
In find_extend_vma(), we don't need mlock_vma_pages_range() to verify the vma type - we know we're working with a stack. So, we can call directly into __mlock_vma_pages_range(), and remove the last make_pages_present() call site. Note that we don't use mm_populate() here, so we can't release the mmap_sem while allocating new stack pages. This is deemed acceptable, because the stack vmas grow by a bounded number of pages at a time, and these are anon pages so we don't have to read from disk to populate them. Signed-off-by: Michel Lespinasse <walken@google.com> Acked-by: Rik van Riel <riel@redhat.com> Tested-by: Andy Lutomirski <luto@amacapital.net> Cc: Greg Ungerer <gregungerer@westnet.com.au> Cc: David Howells <dhowells@redhat.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r--mm/internal.h4
-rw-r--r--mm/memory.c24
-rw-r--r--mm/mlock.c57
-rw-r--r--mm/mmap.c10
4 files changed, 9 insertions, 86 deletions
diff --git a/mm/internal.h b/mm/internal.h
index 9ba21100ebf3..1c0c4cc0fcf7 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -162,8 +162,8 @@ void __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma,
162 struct vm_area_struct *prev, struct rb_node *rb_parent); 162 struct vm_area_struct *prev, struct rb_node *rb_parent);
163 163
164#ifdef CONFIG_MMU 164#ifdef CONFIG_MMU
165extern long mlock_vma_pages_range(struct vm_area_struct *vma, 165extern long __mlock_vma_pages_range(struct vm_area_struct *vma,
166 unsigned long start, unsigned long end); 166 unsigned long start, unsigned long end, int *nonblocking);
167extern void munlock_vma_pages_range(struct vm_area_struct *vma, 167extern void munlock_vma_pages_range(struct vm_area_struct *vma,
168 unsigned long start, unsigned long end); 168 unsigned long start, unsigned long end);
169static inline void munlock_vma_pages_all(struct vm_area_struct *vma) 169static inline void munlock_vma_pages_all(struct vm_area_struct *vma)
diff --git a/mm/memory.c b/mm/memory.c
index 0abd07097ec6..7837ceacf090 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3824,30 +3824,6 @@ int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address)
3824} 3824}
3825#endif /* __PAGETABLE_PMD_FOLDED */ 3825#endif /* __PAGETABLE_PMD_FOLDED */
3826 3826
3827int make_pages_present(unsigned long addr, unsigned long end)
3828{
3829 int ret, len, write;
3830 struct vm_area_struct * vma;
3831
3832 vma = find_vma(current->mm, addr);
3833 if (!vma)
3834 return -ENOMEM;
3835 /*
3836 * We want to touch writable mappings with a write fault in order
3837 * to break COW, except for shared mappings because these don't COW
3838 * and we would not want to dirty them for nothing.
3839 */
3840 write = (vma->vm_flags & (VM_WRITE | VM_SHARED)) == VM_WRITE;
3841 BUG_ON(addr >= end);
3842 BUG_ON(end > vma->vm_end);
3843 len = DIV_ROUND_UP(end, PAGE_SIZE) - addr/PAGE_SIZE;
3844 ret = get_user_pages(current, current->mm, addr,
3845 len, write, 0, NULL, NULL);
3846 if (ret < 0)
3847 return ret;
3848 return ret == len ? 0 : -EFAULT;
3849}
3850
3851#if !defined(__HAVE_ARCH_GATE_AREA) 3827#if !defined(__HAVE_ARCH_GATE_AREA)
3852 3828
3853#if defined(AT_SYSINFO_EHDR) 3829#if defined(AT_SYSINFO_EHDR)
diff --git a/mm/mlock.c b/mm/mlock.c
index a296a49865df..569400a5d079 100644
--- a/mm/mlock.c
+++ b/mm/mlock.c
@@ -155,9 +155,8 @@ void munlock_vma_page(struct page *page)
155 * 155 *
156 * vma->vm_mm->mmap_sem must be held for at least read. 156 * vma->vm_mm->mmap_sem must be held for at least read.
157 */ 157 */
158static long __mlock_vma_pages_range(struct vm_area_struct *vma, 158long __mlock_vma_pages_range(struct vm_area_struct *vma,
159 unsigned long start, unsigned long end, 159 unsigned long start, unsigned long end, int *nonblocking)
160 int *nonblocking)
161{ 160{
162 struct mm_struct *mm = vma->vm_mm; 161 struct mm_struct *mm = vma->vm_mm;
163 unsigned long addr = start; 162 unsigned long addr = start;
@@ -202,56 +201,6 @@ static int __mlock_posix_error_return(long retval)
202 return retval; 201 return retval;
203} 202}
204 203
205/**
206 * mlock_vma_pages_range() - mlock pages in specified vma range.
207 * @vma - the vma containing the specfied address range
208 * @start - starting address in @vma to mlock
209 * @end - end address [+1] in @vma to mlock
210 *
211 * For mmap()/mremap()/expansion of mlocked vma.
212 *
213 * return 0 on success for "normal" vmas.
214 *
215 * return number of pages [> 0] to be removed from locked_vm on success
216 * of "special" vmas.
217 */
218long mlock_vma_pages_range(struct vm_area_struct *vma,
219 unsigned long start, unsigned long end)
220{
221 int nr_pages = (end - start) / PAGE_SIZE;
222 BUG_ON(!(vma->vm_flags & VM_LOCKED));
223
224 /*
225 * filter unlockable vmas
226 */
227 if (vma->vm_flags & (VM_IO | VM_PFNMAP))
228 goto no_mlock;
229
230 if (!((vma->vm_flags & VM_DONTEXPAND) ||
231 is_vm_hugetlb_page(vma) ||
232 vma == get_gate_vma(current->mm))) {
233
234 __mlock_vma_pages_range(vma, start, end, NULL);
235
236 /* Hide errors from mmap() and other callers */
237 return 0;
238 }
239
240 /*
241 * User mapped kernel pages or huge pages:
242 * make these pages present to populate the ptes, but
243 * fall thru' to reset VM_LOCKED--no need to unlock, and
244 * return nr_pages so these don't get counted against task's
245 * locked limit. huge pages are already counted against
246 * locked vm limit.
247 */
248 make_pages_present(start, end);
249
250no_mlock:
251 vma->vm_flags &= ~VM_LOCKED; /* and don't come back! */
252 return nr_pages; /* error or pages NOT mlocked */
253}
254
255/* 204/*
256 * munlock_vma_pages_range() - munlock all pages in the vma range.' 205 * munlock_vma_pages_range() - munlock all pages in the vma range.'
257 * @vma - vma containing range to be munlock()ed. 206 * @vma - vma containing range to be munlock()ed.
@@ -303,7 +252,7 @@ void munlock_vma_pages_range(struct vm_area_struct *vma,
303 * 252 *
304 * Filters out "special" vmas -- VM_LOCKED never gets set for these, and 253 * Filters out "special" vmas -- VM_LOCKED never gets set for these, and
305 * munlock is a no-op. However, for some special vmas, we go ahead and 254 * munlock is a no-op. However, for some special vmas, we go ahead and
306 * populate the ptes via make_pages_present(). 255 * populate the ptes.
307 * 256 *
308 * For vmas that pass the filters, merge/split as appropriate. 257 * For vmas that pass the filters, merge/split as appropriate.
309 */ 258 */
diff --git a/mm/mmap.c b/mm/mmap.c
index d6bc39e939ab..8826c77513a9 100644
--- a/mm/mmap.c
+++ b/mm/mmap.c
@@ -2204,9 +2204,8 @@ find_extend_vma(struct mm_struct *mm, unsigned long addr)
2204 return vma; 2204 return vma;
2205 if (!prev || expand_stack(prev, addr)) 2205 if (!prev || expand_stack(prev, addr))
2206 return NULL; 2206 return NULL;
2207 if (prev->vm_flags & VM_LOCKED) { 2207 if (prev->vm_flags & VM_LOCKED)
2208 mlock_vma_pages_range(prev, addr, prev->vm_end); 2208 __mlock_vma_pages_range(prev, addr, prev->vm_end, NULL);
2209 }
2210 return prev; 2209 return prev;
2211} 2210}
2212#else 2211#else
@@ -2232,9 +2231,8 @@ find_extend_vma(struct mm_struct * mm, unsigned long addr)
2232 start = vma->vm_start; 2231 start = vma->vm_start;
2233 if (expand_stack(vma, addr)) 2232 if (expand_stack(vma, addr))
2234 return NULL; 2233 return NULL;
2235 if (vma->vm_flags & VM_LOCKED) { 2234 if (vma->vm_flags & VM_LOCKED)
2236 mlock_vma_pages_range(vma, addr, start); 2235 __mlock_vma_pages_range(vma, addr, start, NULL);
2237 }
2238 return vma; 2236 return vma;
2239} 2237}
2240#endif 2238#endif