diff options
Diffstat (limited to 'mm/hugetlb.c')
-rw-r--r-- | mm/hugetlb.c | 42 |
1 files changed, 5 insertions, 37 deletions
diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 05407831016b..271e4432734c 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c | |||
@@ -3896,20 +3896,6 @@ follow_huge_pud(struct mm_struct *mm, unsigned long address, | |||
3896 | 3896 | ||
3897 | #ifdef CONFIG_MEMORY_FAILURE | 3897 | #ifdef CONFIG_MEMORY_FAILURE |
3898 | 3898 | ||
3899 | /* Should be called in hugetlb_lock */ | ||
3900 | static int is_hugepage_on_freelist(struct page *hpage) | ||
3901 | { | ||
3902 | struct page *page; | ||
3903 | struct page *tmp; | ||
3904 | struct hstate *h = page_hstate(hpage); | ||
3905 | int nid = page_to_nid(hpage); | ||
3906 | |||
3907 | list_for_each_entry_safe(page, tmp, &h->hugepage_freelists[nid], lru) | ||
3908 | if (page == hpage) | ||
3909 | return 1; | ||
3910 | return 0; | ||
3911 | } | ||
3912 | |||
3913 | /* | 3899 | /* |
3914 | * This function is called from memory failure code. | 3900 | * This function is called from memory failure code. |
3915 | * Assume the caller holds page lock of the head page. | 3901 | * Assume the caller holds page lock of the head page. |
@@ -3921,7 +3907,11 @@ int dequeue_hwpoisoned_huge_page(struct page *hpage) | |||
3921 | int ret = -EBUSY; | 3907 | int ret = -EBUSY; |
3922 | 3908 | ||
3923 | spin_lock(&hugetlb_lock); | 3909 | spin_lock(&hugetlb_lock); |
3924 | if (is_hugepage_on_freelist(hpage)) { | 3910 | /* |
3911 | * Just checking !page_huge_active is not enough, because that could be | ||
3912 | * an isolated/hwpoisoned hugepage (which have >0 refcount). | ||
3913 | */ | ||
3914 | if (!page_huge_active(hpage) && !page_count(hpage)) { | ||
3925 | /* | 3915 | /* |
3926 | * Hwpoisoned hugepage isn't linked to activelist or freelist, | 3916 | * Hwpoisoned hugepage isn't linked to activelist or freelist, |
3927 | * but dangling hpage->lru can trigger list-debug warnings | 3917 | * but dangling hpage->lru can trigger list-debug warnings |
@@ -3965,25 +3955,3 @@ void putback_active_hugepage(struct page *page) | |||
3965 | spin_unlock(&hugetlb_lock); | 3955 | spin_unlock(&hugetlb_lock); |
3966 | put_page(page); | 3956 | put_page(page); |
3967 | } | 3957 | } |
3968 | |||
3969 | bool is_hugepage_active(struct page *page) | ||
3970 | { | ||
3971 | VM_BUG_ON_PAGE(!PageHuge(page), page); | ||
3972 | /* | ||
3973 | * This function can be called for a tail page because the caller, | ||
3974 | * scan_movable_pages, scans through a given pfn-range which typically | ||
3975 | * covers one memory block. In systems using gigantic hugepage (1GB | ||
3976 | * for x86_64,) a hugepage is larger than a memory block, and we don't | ||
3977 | * support migrating such large hugepages for now, so return false | ||
3978 | * when called for tail pages. | ||
3979 | */ | ||
3980 | if (PageTail(page)) | ||
3981 | return false; | ||
3982 | /* | ||
3983 | * Refcount of a hwpoisoned hugepages is 1, but they are not active, | ||
3984 | * so we should return false for them. | ||
3985 | */ | ||
3986 | if (unlikely(PageHWPoison(page))) | ||
3987 | return false; | ||
3988 | return page_count(page) > 0; | ||
3989 | } | ||