diff options
| author | Hugh Dickins <hugh@veritas.com> | 2005-10-29 21:16:00 -0400 |
|---|---|---|
| committer | Linus Torvalds <torvalds@g5.osdl.org> | 2005-10-30 00:40:37 -0400 |
| commit | 7be7a546994f1222b2312fd348da14e16b6b7b42 (patch) | |
| tree | e1f8dae8783274372a0f136be6eb64102877e9f6 | |
| parent | 65500d234e74fc4e8f18e1a429bc24e51e75de4a (diff) | |
[PATCH] mm: move_page_tables by extents
Speeding up mremap's moving of ptes has never been a priority, but the locking
will get more complicated shortly, and is already too baroque.
Scrap the current one-by-one moving, do an extent at a time: curtailed by end
of src and dst pmds (have to use PMD_SIZE: the way pmd_addr_end gets elided
doesn't match this usage), and by latency considerations.
One nice property of the old method is lost: it never allocated a page table
unless absolutely necessary, so you could free empty page tables by mremapping
to and fro. Whereas this way, it allocates a dst table wherever there was a
src table. I keep diving in to reinstate the old behaviour, then come out
preferring not to clutter how it now is.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
| -rw-r--r-- | mm/mremap.c | 168 |
1 files changed, 72 insertions, 96 deletions
diff --git a/mm/mremap.c b/mm/mremap.c index 55df8f53e8..f4e5620985 100644 --- a/mm/mremap.c +++ b/mm/mremap.c | |||
| @@ -22,40 +22,15 @@ | |||
| 22 | #include <asm/cacheflush.h> | 22 | #include <asm/cacheflush.h> |
| 23 | #include <asm/tlbflush.h> | 23 | #include <asm/tlbflush.h> |
| 24 | 24 | ||
| 25 | static pte_t *get_one_pte_map_nested(struct mm_struct *mm, unsigned long addr) | 25 | static pmd_t *get_old_pmd(struct mm_struct *mm, unsigned long addr) |
| 26 | { | ||
| 27 | pgd_t *pgd; | ||
| 28 | pud_t *pud; | ||
| 29 | pmd_t *pmd; | ||
| 30 | pte_t *pte = NULL; | ||
| 31 | |||
| 32 | pgd = pgd_offset(mm, addr); | ||
| 33 | if (pgd_none_or_clear_bad(pgd)) | ||
| 34 | goto end; | ||
| 35 | |||
| 36 | pud = pud_offset(pgd, addr); | ||
| 37 | if (pud_none_or_clear_bad(pud)) | ||
| 38 | goto end; | ||
| 39 | |||
| 40 | pmd = pmd_offset(pud, addr); | ||
| 41 | if (pmd_none_or_clear_bad(pmd)) | ||
| 42 | goto end; | ||
| 43 | |||
| 44 | pte = pte_offset_map_nested(pmd, addr); | ||
| 45 | if (pte_none(*pte)) { | ||
| 46 | pte_unmap_nested(pte); | ||
| 47 | pte = NULL; | ||
| 48 | } | ||
| 49 | end: | ||
| 50 | return pte; | ||
| 51 | } | ||
| 52 | |||
| 53 | static pte_t *get_one_pte_map(struct mm_struct *mm, unsigned long addr) | ||
| 54 | { | 26 | { |
| 55 | pgd_t *pgd; | 27 | pgd_t *pgd; |
| 56 | pud_t *pud; | 28 | pud_t *pud; |
| 57 | pmd_t *pmd; | 29 | pmd_t *pmd; |
| 58 | 30 | ||
| 31 | /* | ||
| 32 | * We don't need page_table_lock: we have mmap_sem exclusively. | ||
| 33 | */ | ||
| 59 | pgd = pgd_offset(mm, addr); | 34 | pgd = pgd_offset(mm, addr); |
| 60 | if (pgd_none_or_clear_bad(pgd)) | 35 | if (pgd_none_or_clear_bad(pgd)) |
| 61 | return NULL; | 36 | return NULL; |
| @@ -68,35 +43,48 @@ static pte_t *get_one_pte_map(struct mm_struct *mm, unsigned long addr) | |||
| 68 | if (pmd_none_or_clear_bad(pmd)) | 43 | if (pmd_none_or_clear_bad(pmd)) |
| 69 | return NULL; | 44 | return NULL; |
| 70 | 45 | ||
| 71 | return pte_offset_map(pmd, addr); | 46 | return pmd; |
| 72 | } | 47 | } |
| 73 | 48 | ||
| 74 | static inline pte_t *alloc_one_pte_map(struct mm_struct *mm, unsigned long addr) | 49 | static pmd_t *alloc_new_pmd(struct mm_struct *mm, unsigned long addr) |
| 75 | { | 50 | { |
| 76 | pgd_t *pgd; | 51 | pgd_t *pgd; |
| 77 | pud_t *pud; | 52 | pud_t *pud; |
| 78 | pmd_t *pmd; | 53 | pmd_t *pmd = NULL; |
| 79 | pte_t *pte = NULL; | 54 | pte_t *pte; |
| 80 | 55 | ||
| 56 | /* | ||
| 57 | * We do need page_table_lock: because allocators expect that. | ||
| 58 | */ | ||
| 59 | spin_lock(&mm->page_table_lock); | ||
| 81 | pgd = pgd_offset(mm, addr); | 60 | pgd = pgd_offset(mm, addr); |
| 82 | |||
| 83 | pud = pud_alloc(mm, pgd, addr); | 61 | pud = pud_alloc(mm, pgd, addr); |
| 84 | if (!pud) | 62 | if (!pud) |
| 85 | return NULL; | 63 | goto out; |
| 64 | |||
| 86 | pmd = pmd_alloc(mm, pud, addr); | 65 | pmd = pmd_alloc(mm, pud, addr); |
| 87 | if (pmd) | 66 | if (!pmd) |
| 88 | pte = pte_alloc_map(mm, pmd, addr); | 67 | goto out; |
| 89 | return pte; | 68 | |
| 69 | pte = pte_alloc_map(mm, pmd, addr); | ||
| 70 | if (!pte) { | ||
| 71 | pmd = NULL; | ||
| 72 | goto out; | ||
| 73 | } | ||
| 74 | pte_unmap(pte); | ||
| 75 | out: | ||
| 76 | spin_unlock(&mm->page_table_lock); | ||
| 77 | return pmd; | ||
| 90 | } | 78 | } |
| 91 | 79 | ||
| 92 | static int | 80 | static void move_ptes(struct vm_area_struct *vma, pmd_t *old_pmd, |
| 93 | move_one_page(struct vm_area_struct *vma, unsigned long old_addr, | 81 | unsigned long old_addr, unsigned long old_end, |
| 94 | struct vm_area_struct *new_vma, unsigned long new_addr) | 82 | struct vm_area_struct *new_vma, pmd_t *new_pmd, |
| 83 | unsigned long new_addr) | ||
| 95 | { | 84 | { |
| 96 | struct address_space *mapping = NULL; | 85 | struct address_space *mapping = NULL; |
| 97 | struct mm_struct *mm = vma->vm_mm; | 86 | struct mm_struct *mm = vma->vm_mm; |
| 98 | int error = 0; | 87 | pte_t *old_pte, *new_pte, pte; |
| 99 | pte_t *src, *dst; | ||
| 100 | 88 | ||
| 101 | if (vma->vm_file) { | 89 | if (vma->vm_file) { |
| 102 | /* | 90 | /* |
| @@ -111,74 +99,62 @@ move_one_page(struct vm_area_struct *vma, unsigned long old_addr, | |||
| 111 | new_vma->vm_truncate_count != vma->vm_truncate_count) | 99 | new_vma->vm_truncate_count != vma->vm_truncate_count) |
| 112 | new_vma->vm_truncate_count = 0; | 100 | new_vma->vm_truncate_count = 0; |
| 113 | } | 101 | } |
| 114 | spin_lock(&mm->page_table_lock); | ||
| 115 | 102 | ||
| 116 | src = get_one_pte_map_nested(mm, old_addr); | 103 | spin_lock(&mm->page_table_lock); |
| 117 | if (src) { | 104 | old_pte = pte_offset_map(old_pmd, old_addr); |
| 118 | /* | 105 | new_pte = pte_offset_map_nested(new_pmd, new_addr); |
| 119 | * Look to see whether alloc_one_pte_map needs to perform a | 106 | |
| 120 | * memory allocation. If it does then we need to drop the | 107 | for (; old_addr < old_end; old_pte++, old_addr += PAGE_SIZE, |
| 121 | * atomic kmap | 108 | new_pte++, new_addr += PAGE_SIZE) { |
| 122 | */ | 109 | if (pte_none(*old_pte)) |
| 123 | dst = get_one_pte_map(mm, new_addr); | 110 | continue; |
| 124 | if (unlikely(!dst)) { | 111 | pte = ptep_clear_flush(vma, old_addr, old_pte); |
| 125 | pte_unmap_nested(src); | 112 | /* ZERO_PAGE can be dependant on virtual addr */ |
| 126 | if (mapping) | 113 | pte = move_pte(pte, new_vma->vm_page_prot, old_addr, new_addr); |
| 127 | spin_unlock(&mapping->i_mmap_lock); | 114 | set_pte_at(mm, new_addr, new_pte, pte); |
| 128 | dst = alloc_one_pte_map(mm, new_addr); | ||
| 129 | if (mapping && !spin_trylock(&mapping->i_mmap_lock)) { | ||
| 130 | spin_unlock(&mm->page_table_lock); | ||
| 131 | spin_lock(&mapping->i_mmap_lock); | ||
| 132 | spin_lock(&mm->page_table_lock); | ||
| 133 | } | ||
| 134 | src = get_one_pte_map_nested(mm, old_addr); | ||
| 135 | } | ||
| 136 | /* | ||
| 137 | * Since alloc_one_pte_map can drop and re-acquire | ||
| 138 | * page_table_lock, we should re-check the src entry... | ||
| 139 | */ | ||
| 140 | if (src) { | ||
| 141 | if (dst) { | ||
| 142 | pte_t pte; | ||
| 143 | pte = ptep_clear_flush(vma, old_addr, src); | ||
| 144 | |||
| 145 | /* ZERO_PAGE can be dependant on virtual addr */ | ||
| 146 | pte = move_pte(pte, new_vma->vm_page_prot, | ||
| 147 | old_addr, new_addr); | ||
| 148 | set_pte_at(mm, new_addr, dst, pte); | ||
| 149 | } else | ||
| 150 | error = -ENOMEM; | ||
| 151 | pte_unmap_nested(src); | ||
| 152 | } | ||
| 153 | if (dst) | ||
| 154 | pte_unmap(dst); | ||
| 155 | } | 115 | } |
| 116 | |||
| 117 | pte_unmap_nested(new_pte - 1); | ||
| 118 | pte_unmap(old_pte - 1); | ||
| 156 | spin_unlock(&mm->page_table_lock); | 119 | spin_unlock(&mm->page_table_lock); |
| 157 | if (mapping) | 120 | if (mapping) |
| 158 | spin_unlock(&mapping->i_mmap_lock); | 121 | spin_unlock(&mapping->i_mmap_lock); |
| 159 | return error; | ||
| 160 | } | 122 | } |
| 161 | 123 | ||
| 124 | #define LATENCY_LIMIT (64 * PAGE_SIZE) | ||
| 125 | |||
| 162 | static unsigned long move_page_tables(struct vm_area_struct *vma, | 126 | static unsigned long move_page_tables(struct vm_area_struct *vma, |
| 163 | unsigned long old_addr, struct vm_area_struct *new_vma, | 127 | unsigned long old_addr, struct vm_area_struct *new_vma, |
| 164 | unsigned long new_addr, unsigned long len) | 128 | unsigned long new_addr, unsigned long len) |
| 165 | { | 129 | { |
| 166 | unsigned long offset; | 130 | unsigned long extent, next, old_end; |
| 131 | pmd_t *old_pmd, *new_pmd; | ||
| 167 | 132 | ||
| 168 | flush_cache_range(vma, old_addr, old_addr + len); | 133 | old_end = old_addr + len; |
| 134 | flush_cache_range(vma, old_addr, old_end); | ||
| 169 | 135 | ||
| 170 | /* | 136 | for (; old_addr < old_end; old_addr += extent, new_addr += extent) { |
| 171 | * This is not the clever way to do this, but we're taking the | ||
| 172 | * easy way out on the assumption that most remappings will be | ||
| 173 | * only a few pages.. This also makes error recovery easier. | ||
| 174 | */ | ||
| 175 | for (offset = 0; offset < len; offset += PAGE_SIZE) { | ||
| 176 | if (move_one_page(vma, old_addr + offset, | ||
| 177 | new_vma, new_addr + offset) < 0) | ||
| 178 | break; | ||
| 179 | cond_resched(); | 137 | cond_resched(); |
| 138 | next = (old_addr + PMD_SIZE) & PMD_MASK; | ||
| 139 | if (next - 1 > old_end) | ||
| 140 | next = old_end; | ||
| 141 | extent = next - old_addr; | ||
| 142 | old_pmd = get_old_pmd(vma->vm_mm, old_addr); | ||
| 143 | if (!old_pmd) | ||
| 144 | continue; | ||
| 145 | new_pmd = alloc_new_pmd(vma->vm_mm, new_addr); | ||
| 146 | if (!new_pmd) | ||
| 147 | break; | ||
| 148 | next = (new_addr + PMD_SIZE) & PMD_MASK; | ||
| 149 | if (extent > next - new_addr) | ||
| 150 | extent = next - new_addr; | ||
| 151 | if (extent > LATENCY_LIMIT) | ||
| 152 | extent = LATENCY_LIMIT; | ||
| 153 | move_ptes(vma, old_pmd, old_addr, old_addr + extent, | ||
| 154 | new_vma, new_pmd, new_addr); | ||
| 180 | } | 155 | } |
| 181 | return offset; | 156 | |
| 157 | return len + old_addr - old_end; /* how much done */ | ||
| 182 | } | 158 | } |
| 183 | 159 | ||
| 184 | static unsigned long move_vma(struct vm_area_struct *vma, | 160 | static unsigned long move_vma(struct vm_area_struct *vma, |
