diff options
author | Dave Hansen <dave@linux.vnet.ibm.com> | 2012-05-29 18:06:46 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-05-29 19:22:24 -0400 |
commit | c50ac050811d6485616a193eb0f37bfbd191cc89 (patch) | |
tree | e237c67e83182bad5744e17c867bb6bceb207544 /mm | |
parent | 5c2b8a162b5f8616f709bf20d5ec88f709485522 (diff) |
hugetlb: fix resv_map leak in error path
When called for anonymous (non-shared) mappings, hugetlb_reserve_pages()
does a resv_map_alloc(). It depends on code in hugetlbfs's
vm_ops->close() to release that allocation.
However, in the mmap() failure path, we do a plain unmap_region() without
the remove_vma() which actually calls vm_ops->close().
This is a decent fix. This leak could get reintroduced if new code (say,
after hugetlb_reserve_pages() in hugetlbfs_file_mmap()) decides to return
an error. But, I think it would have to unroll the reservation anyway.
Christoph's test case:
http://marc.info/?l=linux-mm&m=133728900729735
This patch applies to 3.4 and later. A version for earlier kernels is at
https://lkml.org/lkml/2012/5/22/418.
Signed-off-by: Dave Hansen <dave@linux.vnet.ibm.com>
Acked-by: Mel Gorman <mel@csn.ul.ie>
Acked-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Reported-by: Christoph Lameter <cl@linux.com>
Tested-by: Christoph Lameter <cl@linux.com>
Cc: Andrea Arcangeli <aarcange@redhat.com>
Cc: <stable@vger.kernel.org> [2.6.32+]
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/hugetlb.c | 28 |
1 files changed, 22 insertions, 6 deletions
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 41a647dfb738..285a81e87ec8 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -2157,6 +2157,15 @@ static void hugetlb_vm_op_open(struct vm_area_struct *vma) | |||
2157 | kref_get(&reservations->refs); | 2157 | kref_get(&reservations->refs); |
2158 | } | 2158 | } |
2159 | 2159 | ||
2160 | static void resv_map_put(struct vm_area_struct *vma) | ||
2161 | { | ||
2162 | struct resv_map *reservations = vma_resv_map(vma); | ||
2163 | |||
2164 | if (!reservations) | ||
2165 | return; | ||
2166 | kref_put(&reservations->refs, resv_map_release); | ||
2167 | } | ||
2168 | |||
2160 | static void hugetlb_vm_op_close(struct vm_area_struct *vma) | 2169 | static void hugetlb_vm_op_close(struct vm_area_struct *vma) |
2161 | { | 2170 | { |
2162 | struct hstate *h = hstate_vma(vma); | 2171 | struct hstate *h = hstate_vma(vma); |
@@ -2173,7 +2182,7 @@ static void hugetlb_vm_op_close(struct vm_area_struct *vma) | |||
2173 | reserve = (end - start) - | 2182 | reserve = (end - start) - |
2174 | region_count(&reservations->regions, start, end); | 2183 | region_count(&reservations->regions, start, end); |
2175 | 2184 | ||
2176 | kref_put(&reservations->refs, resv_map_release); | 2185 | resv_map_put(vma); |
2177 | 2186 | ||
2178 | if (reserve) { | 2187 | if (reserve) { |
2179 | hugetlb_acct_memory(h, -reserve); | 2188 | hugetlb_acct_memory(h, -reserve); |
@@ -2991,12 +3000,16 @@ int hugetlb_reserve_pages(struct inode *inode, | |||
2991 | set_vma_resv_flags(vma, HPAGE_RESV_OWNER); | 3000 | set_vma_resv_flags(vma, HPAGE_RESV_OWNER); |
2992 | } | 3001 | } |
2993 | 3002 | ||
2994 | if (chg < 0) | 3003 | if (chg < 0) { |
2995 | return chg; | 3004 | ret = chg; |
3005 | goto out_err; | ||
3006 | } | ||
2996 | 3007 | ||
2997 | /* There must be enough pages in the subpool for the mapping */ | 3008 | /* There must be enough pages in the subpool for the mapping */ |
2998 | if (hugepage_subpool_get_pages(spool, chg)) | 3009 | if (hugepage_subpool_get_pages(spool, chg)) { |
2999 | return -ENOSPC; | 3010 | ret = -ENOSPC; |
3011 | goto out_err; | ||
3012 | } | ||
3000 | 3013 | ||
3001 | /* | 3014 | /* |
3002 | * Check enough hugepages are available for the reservation. | 3015 | * Check enough hugepages are available for the reservation. |
@@ -3005,7 +3018,7 @@ int hugetlb_reserve_pages(struct inode *inode, | |||
3005 | ret = hugetlb_acct_memory(h, chg); | 3018 | ret = hugetlb_acct_memory(h, chg); |
3006 | if (ret < 0) { | 3019 | if (ret < 0) { |
3007 | hugepage_subpool_put_pages(spool, chg); | 3020 | hugepage_subpool_put_pages(spool, chg); |
3008 | return ret; | 3021 | goto out_err; |
3009 | } | 3022 | } |
3010 | 3023 | ||
3011 | /* | 3024 | /* |
@@ -3022,6 +3035,9 @@ int hugetlb_reserve_pages(struct inode *inode, | |||
3022 | if (!vma || vma->vm_flags & VM_MAYSHARE) | 3035 | if (!vma || vma->vm_flags & VM_MAYSHARE) |
3023 | region_add(&inode->i_mapping->private_list, from, to); | 3036 | region_add(&inode->i_mapping->private_list, from, to); |
3024 | return 0; | 3037 | return 0; |
3038 | out_err: | ||
3039 | resv_map_put(vma); | ||
3040 | return ret; | ||
3025 | } | 3041 | } |
3026 | 3042 | ||
3027 | void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed) | 3043 | void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed) |