diff options
author | Adam Litke <agl@us.ibm.com> | 2007-11-14 19:59:39 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-11-14 21:45:40 -0500 |
commit | 2fc39cec6a9b5b41727d3386b780b69422a15152 (patch) | |
tree | 4b01dcef394a109ae3ff83189f7386ff3cdb3f61 | |
parent | c79fb75e5a514a5a35f22c229042aa29f4237e3a (diff) |
hugetlb: debit quota in alloc_huge_page
Now that quota is credited by free_huge_page(), calls to hugetlb_get_quota()
seem out of place. The alloc/free API is unbalanced because we handle the
hugetlb_put_quota() but expect the caller to open-code hugetlb_get_quota().
Move the get inside alloc_huge_page to clean up this disparity.
This patch has been kept apart from the previous patch because of the somewhat
dodgy ERR_PTR() use herein. Moving the quota logic means that
alloc_huge_page() has two failure modes. Quota failure must result in a
SIGBUS while a standard allocation failure is OOM. Unfortunately, ERR_PTR()
doesn't like the small positive errnos we have in VM_FAULT_* so they must be
negated before they are used.
Does anyone take issue with the way I am using PTR_ERR. If so, what are your
thoughts on how to clean this up (without needing an if,else if,else block at
each alloc_huge_page() callsite)?
Signed-off-by: Adam Litke <agl@us.ibm.com>
Cc: Ken Chen <kenchen@google.com>
Cc: Andy Whitcroft <apw@shadowen.org>
Cc: Dave Hansen <haveblue@us.ibm.com>
Cc: David Gibson <hermes@gibson.dropbear.id.au>
Cc: William Lee Irwin III <wli@holomorphy.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | mm/hugetlb.c | 24 |
1 files changed, 12 insertions, 12 deletions
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 3992bd5120e7..bc12b0adfa87 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -388,6 +388,10 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, | |||
388 | unsigned long addr) | 388 | unsigned long addr) |
389 | { | 389 | { |
390 | struct page *page; | 390 | struct page *page; |
391 | struct address_space *mapping = vma->vm_file->f_mapping; | ||
392 | |||
393 | if (hugetlb_get_quota(mapping)) | ||
394 | return ERR_PTR(-VM_FAULT_SIGBUS); | ||
391 | 395 | ||
392 | if (vma->vm_flags & VM_MAYSHARE) | 396 | if (vma->vm_flags & VM_MAYSHARE) |
393 | page = alloc_huge_page_shared(vma, addr); | 397 | page = alloc_huge_page_shared(vma, addr); |
@@ -395,9 +399,10 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, | |||
395 | page = alloc_huge_page_private(vma, addr); | 399 | page = alloc_huge_page_private(vma, addr); |
396 | if (page) { | 400 | if (page) { |
397 | set_page_refcounted(page); | 401 | set_page_refcounted(page); |
398 | set_page_private(page, (unsigned long) vma->vm_file->f_mapping); | 402 | set_page_private(page, (unsigned long) mapping); |
399 | } | 403 | return page; |
400 | return page; | 404 | } else |
405 | return ERR_PTR(-VM_FAULT_OOM); | ||
401 | } | 406 | } |
402 | 407 | ||
403 | static int __init hugetlb_init(void) | 408 | static int __init hugetlb_init(void) |
@@ -737,15 +742,13 @@ static int hugetlb_cow(struct mm_struct *mm, struct vm_area_struct *vma, | |||
737 | set_huge_ptep_writable(vma, address, ptep); | 742 | set_huge_ptep_writable(vma, address, ptep); |
738 | return 0; | 743 | return 0; |
739 | } | 744 | } |
740 | if (hugetlb_get_quota(vma->vm_file->f_mapping)) | ||
741 | return VM_FAULT_SIGBUS; | ||
742 | 745 | ||
743 | page_cache_get(old_page); | 746 | page_cache_get(old_page); |
744 | new_page = alloc_huge_page(vma, address); | 747 | new_page = alloc_huge_page(vma, address); |
745 | 748 | ||
746 | if (!new_page) { | 749 | if (IS_ERR(new_page)) { |
747 | page_cache_release(old_page); | 750 | page_cache_release(old_page); |
748 | return VM_FAULT_OOM; | 751 | return -PTR_ERR(new_page); |
749 | } | 752 | } |
750 | 753 | ||
751 | spin_unlock(&mm->page_table_lock); | 754 | spin_unlock(&mm->page_table_lock); |
@@ -789,12 +792,9 @@ retry: | |||
789 | size = i_size_read(mapping->host) >> HPAGE_SHIFT; | 792 | size = i_size_read(mapping->host) >> HPAGE_SHIFT; |
790 | if (idx >= size) | 793 | if (idx >= size) |
791 | goto out; | 794 | goto out; |
792 | if (hugetlb_get_quota(mapping)) | ||
793 | goto out; | ||
794 | page = alloc_huge_page(vma, address); | 795 | page = alloc_huge_page(vma, address); |
795 | if (!page) { | 796 | if (IS_ERR(page)) { |
796 | hugetlb_put_quota(mapping); | 797 | ret = -PTR_ERR(page); |
797 | ret = VM_FAULT_OOM; | ||
798 | goto out; | 798 | goto out; |
799 | } | 799 | } |
800 | clear_huge_page(page, address); | 800 | clear_huge_page(page, address); |