diff options
author | Naoya Horiguchi <n-horiguchi@ah.jp.nec.com> | 2010-09-07 21:19:36 -0400 |
---|---|---|
committer | Andi Kleen <ak@linux.intel.com> | 2010-10-08 03:32:45 -0400 |
commit | 6de2b1aab94355482bd2accdc115666509667458 (patch) | |
tree | 630a88dd5afe2c49a428179770140b22bbc9d7da | |
parent | 290408d4a25002f099efeee7b6a5778d431154d6 (diff) |
HWPOISON, hugetlb: add free check to dequeue_hwpoison_huge_page()
This check is necessary to avoid race between dequeue and allocation,
which can cause a free hugepage to be dequeued twice and get kernel unstable.
Signed-off-by: Naoya Horiguchi <n-horiguchi@ah.jp.nec.com>
Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
Acked-by: Mel Gorman <mel@csn.ul.ie>
Reviewed-by: Christoph Lameter <cl@linux.com>
Signed-off-by: Andi Kleen <ak@linux.intel.com>
-rw-r--r-- | include/linux/hugetlb.h | 4 | ||||
-rw-r--r-- | mm/hugetlb.c | 29 | ||||
-rw-r--r-- | mm/memory-failure.c | 6 |
3 files changed, 31 insertions, 8 deletions
diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h index 9e51f77d44ca..796f30e00806 100644 --- a/include/linux/hugetlb.h +++ b/include/linux/hugetlb.h | |||
@@ -43,7 +43,7 @@ int hugetlb_reserve_pages(struct inode *inode, long from, long to, | |||
43 | struct vm_area_struct *vma, | 43 | struct vm_area_struct *vma, |
44 | int acctflags); | 44 | int acctflags); |
45 | void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed); | 45 | void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed); |
46 | void __isolate_hwpoisoned_huge_page(struct page *page); | 46 | int dequeue_hwpoisoned_huge_page(struct page *page); |
47 | void copy_huge_page(struct page *dst, struct page *src); | 47 | void copy_huge_page(struct page *dst, struct page *src); |
48 | 48 | ||
49 | extern unsigned long hugepages_treat_as_movable; | 49 | extern unsigned long hugepages_treat_as_movable; |
@@ -102,7 +102,7 @@ static inline void hugetlb_report_meminfo(struct seq_file *m) | |||
102 | #define hugetlb_free_pgd_range(tlb, addr, end, floor, ceiling) ({BUG(); 0; }) | 102 | #define hugetlb_free_pgd_range(tlb, addr, end, floor, ceiling) ({BUG(); 0; }) |
103 | #define hugetlb_fault(mm, vma, addr, flags) ({ BUG(); 0; }) | 103 | #define hugetlb_fault(mm, vma, addr, flags) ({ BUG(); 0; }) |
104 | #define huge_pte_offset(mm, address) 0 | 104 | #define huge_pte_offset(mm, address) 0 |
105 | #define __isolate_hwpoisoned_huge_page(page) 0 | 105 | #define dequeue_hwpoisoned_huge_page(page) 0 |
106 | static inline void copy_huge_page(struct page *dst, struct page *src) | 106 | static inline void copy_huge_page(struct page *dst, struct page *src) |
107 | { | 107 | { |
108 | } | 108 | } |
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 0fa9de8361bd..deb7bebefe68 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -2955,18 +2955,39 @@ void hugetlb_unreserve_pages(struct inode *inode, long offset, long freed) | |||
2955 | hugetlb_acct_memory(h, -(chg - freed)); | 2955 | hugetlb_acct_memory(h, -(chg - freed)); |
2956 | } | 2956 | } |
2957 | 2957 | ||
2958 | /* Should be called in hugetlb_lock */ | ||
2959 | static int is_hugepage_on_freelist(struct page *hpage) | ||
2960 | { | ||
2961 | struct page *page; | ||
2962 | struct page *tmp; | ||
2963 | struct hstate *h = page_hstate(hpage); | ||
2964 | int nid = page_to_nid(hpage); | ||
2965 | |||
2966 | list_for_each_entry_safe(page, tmp, &h->hugepage_freelists[nid], lru) | ||
2967 | if (page == hpage) | ||
2968 | return 1; | ||
2969 | return 0; | ||
2970 | } | ||
2971 | |||
2972 | #ifdef CONFIG_MEMORY_FAILURE | ||
2958 | /* | 2973 | /* |
2959 | * This function is called from memory failure code. | 2974 | * This function is called from memory failure code. |
2960 | * Assume the caller holds page lock of the head page. | 2975 | * Assume the caller holds page lock of the head page. |
2961 | */ | 2976 | */ |
2962 | void __isolate_hwpoisoned_huge_page(struct page *hpage) | 2977 | int dequeue_hwpoisoned_huge_page(struct page *hpage) |
2963 | { | 2978 | { |
2964 | struct hstate *h = page_hstate(hpage); | 2979 | struct hstate *h = page_hstate(hpage); |
2965 | int nid = page_to_nid(hpage); | 2980 | int nid = page_to_nid(hpage); |
2981 | int ret = -EBUSY; | ||
2966 | 2982 | ||
2967 | spin_lock(&hugetlb_lock); | 2983 | spin_lock(&hugetlb_lock); |
2968 | list_del(&hpage->lru); | 2984 | if (is_hugepage_on_freelist(hpage)) { |
2969 | h->free_huge_pages--; | 2985 | list_del(&hpage->lru); |
2970 | h->free_huge_pages_node[nid]--; | 2986 | h->free_huge_pages--; |
2987 | h->free_huge_pages_node[nid]--; | ||
2988 | ret = 0; | ||
2989 | } | ||
2971 | spin_unlock(&hugetlb_lock); | 2990 | spin_unlock(&hugetlb_lock); |
2991 | return ret; | ||
2972 | } | 2992 | } |
2993 | #endif | ||
diff --git a/mm/memory-failure.c b/mm/memory-failure.c index 757f6b0accfe..5c7158a11592 100644 --- a/mm/memory-failure.c +++ b/mm/memory-failure.c | |||
@@ -698,6 +698,7 @@ static int me_swapcache_clean(struct page *p, unsigned long pfn) | |||
698 | */ | 698 | */ |
699 | static int me_huge_page(struct page *p, unsigned long pfn) | 699 | static int me_huge_page(struct page *p, unsigned long pfn) |
700 | { | 700 | { |
701 | int res = 0; | ||
701 | struct page *hpage = compound_head(p); | 702 | struct page *hpage = compound_head(p); |
702 | /* | 703 | /* |
703 | * We can safely recover from error on free or reserved (i.e. | 704 | * We can safely recover from error on free or reserved (i.e. |
@@ -710,8 +711,9 @@ static int me_huge_page(struct page *p, unsigned long pfn) | |||
710 | * so there is no race between isolation and mapping/unmapping. | 711 | * so there is no race between isolation and mapping/unmapping. |
711 | */ | 712 | */ |
712 | if (!(page_mapping(hpage) || PageAnon(hpage))) { | 713 | if (!(page_mapping(hpage) || PageAnon(hpage))) { |
713 | __isolate_hwpoisoned_huge_page(hpage); | 714 | res = dequeue_hwpoisoned_huge_page(hpage); |
714 | return RECOVERED; | 715 | if (!res) |
716 | return RECOVERED; | ||
715 | } | 717 | } |
716 | return DELAYED; | 718 | return DELAYED; |
717 | } | 719 | } |