diff options
Diffstat (limited to 'mm/hugetlb.c')
-rw-r--r-- | mm/hugetlb.c | 10 |
1 files changed, 5 insertions, 5 deletions
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 06de5aa4d644..8ee3bd8ec5b5 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -146,7 +146,7 @@ static long region_chg(struct list_head *head, long f, long t) | |||
146 | if (rg->from > t) | 146 | if (rg->from > t) |
147 | return chg; | 147 | return chg; |
148 | 148 | ||
149 | /* We overlap with this area, if it extends futher than | 149 | /* We overlap with this area, if it extends further than |
150 | * us then we must extend ourselves. Account for its | 150 | * us then we must extend ourselves. Account for its |
151 | * existing reservation. */ | 151 | * existing reservation. */ |
152 | if (rg->to > t) { | 152 | if (rg->to > t) { |
@@ -842,7 +842,7 @@ struct page *alloc_huge_page_node(struct hstate *h, int nid) | |||
842 | } | 842 | } |
843 | 843 | ||
844 | /* | 844 | /* |
845 | * Increase the hugetlb pool such that it can accomodate a reservation | 845 | * Increase the hugetlb pool such that it can accommodate a reservation |
846 | * of size 'delta'. | 846 | * of size 'delta'. |
847 | */ | 847 | */ |
848 | static int gather_surplus_pages(struct hstate *h, int delta) | 848 | static int gather_surplus_pages(struct hstate *h, int delta) |
@@ -890,7 +890,7 @@ retry: | |||
890 | 890 | ||
891 | /* | 891 | /* |
892 | * The surplus_list now contains _at_least_ the number of extra pages | 892 | * The surplus_list now contains _at_least_ the number of extra pages |
893 | * needed to accomodate the reservation. Add the appropriate number | 893 | * needed to accommodate the reservation. Add the appropriate number |
894 | * of pages to the hugetlb pool and free the extras back to the buddy | 894 | * of pages to the hugetlb pool and free the extras back to the buddy |
895 | * allocator. Commit the entire reservation here to prevent another | 895 | * allocator. Commit the entire reservation here to prevent another |
896 | * process from stealing the pages as they are added to the pool but | 896 | * process from stealing the pages as they are added to the pool but |
@@ -2043,7 +2043,7 @@ static void hugetlb_vm_op_open(struct vm_area_struct *vma) | |||
2043 | * This new VMA should share its siblings reservation map if present. | 2043 | * This new VMA should share its siblings reservation map if present. |
2044 | * The VMA will only ever have a valid reservation map pointer where | 2044 | * The VMA will only ever have a valid reservation map pointer where |
2045 | * it is being copied for another still existing VMA. As that VMA | 2045 | * it is being copied for another still existing VMA. As that VMA |
2046 | * has a reference to the reservation map it cannot dissappear until | 2046 | * has a reference to the reservation map it cannot disappear until |
2047 | * after this open call completes. It is therefore safe to take a | 2047 | * after this open call completes. It is therefore safe to take a |
2048 | * new reference here without additional locking. | 2048 | * new reference here without additional locking. |
2049 | */ | 2049 | */ |
@@ -2490,7 +2490,7 @@ static int hugetlb_no_page(struct mm_struct *mm, struct vm_area_struct *vma, | |||
2490 | /* | 2490 | /* |
2491 | * Currently, we are forced to kill the process in the event the | 2491 | * Currently, we are forced to kill the process in the event the |
2492 | * original mapper has unmapped pages from the child due to a failed | 2492 | * original mapper has unmapped pages from the child due to a failed |
2493 | * COW. Warn that such a situation has occured as it may not be obvious | 2493 | * COW. Warn that such a situation has occurred as it may not be obvious |
2494 | */ | 2494 | */ |
2495 | if (is_vma_resv_set(vma, HPAGE_RESV_UNMAPPED)) { | 2495 | if (is_vma_resv_set(vma, HPAGE_RESV_UNMAPPED)) { |
2496 | printk(KERN_WARNING | 2496 | printk(KERN_WARNING |