diff options
author | Hugh Dickins <hugh@veritas.com> | 2005-10-29 21:16:00 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2005-10-30 00:40:37 -0400 |
commit | 7be7a546994f1222b2312fd348da14e16b6b7b42 (patch) | |
tree | e1f8dae8783274372a0f136be6eb64102877e9f6 /mm/mremap.c | |
parent | 65500d234e74fc4e8f18e1a429bc24e51e75de4a (diff) |
[PATCH] mm: move_page_tables by extents
Speeding up mremap's moving of ptes has never been a priority, but the locking
will get more complicated shortly, and is already too baroque.
Scrap the current one-by-one moving, do an extent at a time: curtailed by end
of src and dst pmds (have to use PMD_SIZE: the way pmd_addr_end gets elided
doesn't match this usage), and by latency considerations.
One nice property of the old method is lost: it never allocated a page table
unless absolutely necessary, so you could free empty page tables by mremapping
to and fro. Whereas this way, it allocates a dst table wherever there was a
src table. I keep diving in to reinstate the old behaviour, then come out
preferring not to clutter how it now is.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'mm/mremap.c')
-rw-r--r-- | mm/mremap.c | 168 |
1 files changed, 72 insertions, 96 deletions
diff --git a/mm/mremap.c b/mm/mremap.c index 55df8f53e84d..f4e562098500 100644 --- a/mm/mremap.c +++ b/mm/mremap.c | |||
@@ -22,40 +22,15 @@ | |||
22 | #include <asm/cacheflush.h> | 22 | #include <asm/cacheflush.h> |
23 | #include <asm/tlbflush.h> | 23 | #include <asm/tlbflush.h> |
24 | 24 | ||
25 | static pte_t *get_one_pte_map_nested(struct mm_struct *mm, unsigned long addr) | 25 | static pmd_t *get_old_pmd(struct mm_struct *mm, unsigned long addr) |
26 | { | ||
27 | pgd_t *pgd; | ||
28 | pud_t *pud; | ||
29 | pmd_t *pmd; | ||
30 | pte_t *pte = NULL; | ||
31 | |||
32 | pgd = pgd_offset(mm, addr); | ||
33 | if (pgd_none_or_clear_bad(pgd)) | ||
34 | goto end; | ||
35 | |||
36 | pud = pud_offset(pgd, addr); | ||
37 | if (pud_none_or_clear_bad(pud)) | ||
38 | goto end; | ||
39 | |||
40 | pmd = pmd_offset(pud, addr); | ||
41 | if (pmd_none_or_clear_bad(pmd)) | ||
42 | goto end; | ||
43 | |||
44 | pte = pte_offset_map_nested(pmd, addr); | ||
45 | if (pte_none(*pte)) { | ||
46 | pte_unmap_nested(pte); | ||
47 | pte = NULL; | ||
48 | } | ||
49 | end: | ||
50 | return pte; | ||
51 | } | ||
52 | |||
53 | static pte_t *get_one_pte_map(struct mm_struct *mm, unsigned long addr) | ||
54 | { | 26 | { |
55 | pgd_t *pgd; | 27 | pgd_t *pgd; |
56 | pud_t *pud; | 28 | pud_t *pud; |
57 | pmd_t *pmd; | 29 | pmd_t *pmd; |
58 | 30 | ||
31 | /* | ||
32 | * We don't need page_table_lock: we have mmap_sem exclusively. | ||
33 | */ | ||
59 | pgd = pgd_offset(mm, addr); | 34 | pgd = pgd_offset(mm, addr); |
60 | if (pgd_none_or_clear_bad(pgd)) | 35 | if (pgd_none_or_clear_bad(pgd)) |
61 | return NULL; | 36 | return NULL; |
@@ -68,35 +43,48 @@ static pte_t *get_one_pte_map(struct mm_struct *mm, unsigned long addr) | |||
68 | if (pmd_none_or_clear_bad(pmd)) | 43 | if (pmd_none_or_clear_bad(pmd)) |
69 | return NULL; | 44 | return NULL; |
70 | 45 | ||
71 | return pte_offset_map(pmd, addr); | 46 | return pmd; |
72 | } | 47 | } |
73 | 48 | ||
74 | static inline pte_t *alloc_one_pte_map(struct mm_struct *mm, unsigned long addr) | 49 | static pmd_t *alloc_new_pmd(struct mm_struct *mm, unsigned long addr) |
75 | { | 50 | { |
76 | pgd_t *pgd; | 51 | pgd_t *pgd; |
77 | pud_t *pud; | 52 | pud_t *pud; |
78 | pmd_t *pmd; | 53 | pmd_t *pmd = NULL; |
79 | pte_t *pte = NULL; | 54 | pte_t *pte; |
80 | 55 | ||
56 | /* | ||
57 | * We do need page_table_lock: because allocators expect that. | ||
58 | */ | ||
59 | spin_lock(&mm->page_table_lock); | ||
81 | pgd = pgd_offset(mm, addr); | 60 | pgd = pgd_offset(mm, addr); |
82 | |||
83 | pud = pud_alloc(mm, pgd, addr); | 61 | pud = pud_alloc(mm, pgd, addr); |
84 | if (!pud) | 62 | if (!pud) |
85 | return NULL; | 63 | goto out; |
64 | |||
86 | pmd = pmd_alloc(mm, pud, addr); | 65 | pmd = pmd_alloc(mm, pud, addr); |
87 | if (pmd) | 66 | if (!pmd) |
88 | pte = pte_alloc_map(mm, pmd, addr); | 67 | goto out; |
89 | return pte; | 68 | |
69 | pte = pte_alloc_map(mm, pmd, addr); | ||
70 | if (!pte) { | ||
71 | pmd = NULL; | ||
72 | goto out; | ||
73 | } | ||
74 | pte_unmap(pte); | ||
75 | out: | ||
76 | spin_unlock(&mm->page_table_lock); | ||
77 | return pmd; | ||
90 | } | 78 | } |
91 | 79 | ||
92 | static int | 80 | static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, |
93 | move_one_page(struct vm_area_struct *vma, unsigned long old_addr, | 81 | unsigned long old_addr, unsigned long old_end, |
94 | struct vm_area_struct *new_vma, unsigned long new_addr) | 82 | struct vm_area_struct *new_vma, pmd_t *new_pmd, |
83 | unsigned long new_addr) | ||
95 | { | 84 | { |
96 | struct address_space *mapping = NULL; | 85 | struct address_space *mapping = NULL; |
97 | struct mm_struct *mm = vma->vm_mm; | 86 | struct mm_struct *mm = vma->vm_mm; |
98 | int error = 0; | 87 | pte_t *old_pte, *new_pte, pte; |
99 | pte_t *src, *dst; | ||
100 | 88 | ||
101 | if (vma->vm_file) { | 89 | if (vma->vm_file) { |
102 | /* | 90 | /* |
@@ -111,74 +99,62 @@ move_one_page(struct vm_area_struct *vma, unsigned long old_addr, | |||
111 | new_vma->vm_truncate_count != vma->vm_truncate_count) | 99 | new_vma->vm_truncate_count != vma->vm_truncate_count) |
112 | new_vma->vm_truncate_count = 0; | 100 | new_vma->vm_truncate_count = 0; |
113 | } | 101 | } |
114 | spin_lock(&mm->page_table_lock); | ||
115 | 102 | ||
116 | src = get_one_pte_map_nested(mm, old_addr); | 103 | spin_lock(&mm->page_table_lock); |
117 | if (src) { | 104 | old_pte = pte_offset_map(old_pmd, old_addr); |
118 | /* | 105 | new_pte = pte_offset_map_nested(new_pmd, new_addr); |
119 | * Look to see whether alloc_one_pte_map needs to perform a | 106 | |
120 | * memory allocation. If it does then we need to drop the | 107 | for (; old_addr < old_end; old_pte++, old_addr += PAGE_SIZE, |
121 | * atomic kmap | 108 | new_pte++, new_addr += PAGE_SIZE) { |
122 | */ | 109 | if (pte_none(*old_pte)) |
123 | dst = get_one_pte_map(mm, new_addr); | 110 | continue; |
124 | if (unlikely(!dst)) { | 111 | pte = ptep_clear_flush(vma, old_addr, old_pte); |
125 | pte_unmap_nested(src); | 112 | /* ZERO_PAGE can be dependant on virtual addr */ |
126 | if (mapping) | 113 | pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr); |
127 | spin_unlock(&mapping->i_mmap_lock); | 114 | set_pte_at(mm, new_addr, new_pte, pte); |
128 | dst = alloc_one_pte_map(mm, new_addr); | ||
129 | if (mapping && !spin_trylock(&mapping->i_mmap_lock)) { | ||
130 | spin_unlock(&mm->page_table_lock); | ||
131 | spin_lock(&mapping->i_mmap_lock); | ||
132 | spin_lock(&mm->page_table_lock); | ||
133 | } | ||
134 | src = get_one_pte_map_nested(mm, old_addr); | ||
135 | } | ||
136 | /* | ||
137 | * Since alloc_one_pte_map can drop and re-acquire | ||
138 | * page_table_lock, we should re-check the src entry... | ||
139 | */ | ||
140 | if (src) { | ||
141 | if (dst) { | ||
142 | pte_t pte; | ||
143 | pte = ptep_clear_flush(vma, old_addr, src); | ||
144 | |||
145 | /* ZERO_PAGE can be dependant on virtual addr */ | ||
146 | pte = move_pte(pte, new_vma->vm_page_prot, | ||
147 | old_addr, new_addr); | ||
148 | set_pte_at(mm, new_addr, dst, pte); | ||
149 | } else | ||
150 | error = -ENOMEM; | ||
151 | pte_unmap_nested(src); | ||
152 | } | ||
153 | if (dst) | ||
154 | pte_unmap(dst); | ||
155 | } | 115 | } |
116 | |||
117 | pte_unmap_nested(new_pte - 1); | ||
118 | pte_unmap(old_pte - 1); | ||
156 | spin_unlock(&mm->page_table_lock); | 119 | spin_unlock(&mm->page_table_lock); |
157 | if (mapping) | 120 | if (mapping) |
158 | spin_unlock(&mapping->i_mmap_lock); | 121 | spin_unlock(&mapping->i_mmap_lock); |
159 | return error; | ||
160 | } | 122 | } |
161 | 123 | ||
124 | #define LATENCY_LIMIT (64 * PAGE_SIZE) | ||
125 | |||
162 | static unsigned long move_page_tables(struct vm_area_struct *vma, | 126 | static unsigned long move_page_tables(struct vm_area_struct *vma, |
163 | unsigned long old_addr, struct vm_area_struct *new_vma, | 127 | unsigned long old_addr, struct vm_area_struct *new_vma, |
164 | unsigned long new_addr, unsigned long len) | 128 | unsigned long new_addr, unsigned long len) |
165 | { | 129 | { |
166 | unsigned long offset; | 130 | unsigned long extent, next, old_end; |
131 | pmd_t *old_pmd, *new_pmd; | ||
167 | 132 | ||
168 | flush_cache_range(vma, old_addr, old_addr + len); | 133 | old_end = old_addr + len; |
134 | flush_cache_range(vma, old_addr, old_end); | ||
169 | 135 | ||
170 | /* | 136 | for (; old_addr < old_end; old_addr += extent, new_addr += extent) { |
171 | * This is not the clever way to do this, but we're taking the | ||
172 | * easy way out on the assumption that most remappings will be | ||
173 | * only a few pages.. This also makes error recovery easier. | ||
174 | */ | ||
175 | for (offset = 0; offset < len; offset += PAGE_SIZE) { | ||
176 | if (move_one_page(vma, old_addr + offset, | ||
177 | new_vma, new_addr + offset) < 0) | ||
178 | break; | ||
179 | cond_resched(); | 137 | cond_resched(); |
138 | next = (old_addr + PMD_SIZE) & PMD_MASK; | ||
139 | if (next - 1 > old_end) | ||
140 | next = old_end; | ||
141 | extent = next - old_addr; | ||
142 | old_pmd = get_old_pmd(vma->vm_mm, old_addr); | ||
143 | if (!old_pmd) | ||
144 | continue; | ||
145 | new_pmd = alloc_new_pmd(vma->vm_mm, new_addr); | ||
146 | if (!new_pmd) | ||
147 | break; | ||
148 | next = (new_addr + PMD_SIZE) & PMD_MASK; | ||
149 | if (extent > next - new_addr) | ||
150 | extent = next - new_addr; | ||
151 | if (extent > LATENCY_LIMIT) | ||
152 | extent = LATENCY_LIMIT; | ||
153 | move_ptes(vma, old_pmd, old_addr, old_addr + extent, | ||
154 | new_vma, new_pmd, new_addr); | ||
180 | } | 155 | } |
181 | return offset; | 156 | |
157 | return len + old_addr - old_end; /* how much done */ | ||
182 | } | 158 | } |
183 | 159 | ||
184 | static unsigned long move_vma(struct vm_area_struct *vma, | 160 | static unsigned long move_vma(struct vm_area_struct *vma, |