diff options
author | Hugh Dickins <hugh@veritas.com> | 2009-01-06 17:40:06 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2009-01-06 18:59:07 -0500 |
commit | 8cc3b39221b0ecbd83a338948a8396df097fc656 (patch) | |
tree | 8a8617590f6bc9842c28a0fc4e5e5fbc93e95780 /mm | |
parent | 79f4b7bf393e67bbffec807cc68caaefc72b82ee (diff) |
badpage: keep any bad page out of circulation
Until now the bad_page() checkers have special-cased PageReserved, keeping
those pages out of circulation thereafter. Now extend the special case to
all: we want to keep ANY page with bad state out of circulation - the
"free" page may well be in use by something.
Leave the bad state of those pages untouched, for examination by
debuggers; except for PageBuddy - leaving that set would risk bringing the
page back.
Signed-off-by: Hugh Dickins <hugh@veritas.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Cc: Christoph Lameter <cl@linux-foundation.org>
Cc: Mel Gorman <mel@csn.ul.ie>
Cc: Rik van Riel <riel@redhat.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/page_alloc.c | 52 |
1 files changed, 24 insertions, 28 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index b90a74d28485..bd330252fc77 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
@@ -231,9 +231,9 @@ static void bad_page(struct page *page) | |||
231 | printk(KERN_EMERG "Trying to fix it up, but a reboot is needed\n" | 231 | printk(KERN_EMERG "Trying to fix it up, but a reboot is needed\n" |
232 | KERN_EMERG "Backtrace:\n"); | 232 | KERN_EMERG "Backtrace:\n"); |
233 | dump_stack(); | 233 | dump_stack(); |
234 | set_page_count(page, 0); | 234 | |
235 | reset_page_mapcount(page); | 235 | /* Leave bad fields for debug, except PageBuddy could make trouble */ |
236 | page->mapping = NULL; | 236 | __ClearPageBuddy(page); |
237 | add_taint(TAINT_BAD_PAGE); | 237 | add_taint(TAINT_BAD_PAGE); |
238 | } | 238 | } |
239 | 239 | ||
@@ -290,25 +290,31 @@ void prep_compound_gigantic_page(struct page *page, unsigned long order) | |||
290 | } | 290 | } |
291 | #endif | 291 | #endif |
292 | 292 | ||
293 | static void destroy_compound_page(struct page *page, unsigned long order) | 293 | static int destroy_compound_page(struct page *page, unsigned long order) |
294 | { | 294 | { |
295 | int i; | 295 | int i; |
296 | int nr_pages = 1 << order; | 296 | int nr_pages = 1 << order; |
297 | int bad = 0; | ||
297 | 298 | ||
298 | if (unlikely(compound_order(page) != order)) | 299 | if (unlikely(compound_order(page) != order) || |
300 | unlikely(!PageHead(page))) { | ||
299 | bad_page(page); | 301 | bad_page(page); |
302 | bad++; | ||
303 | } | ||
300 | 304 | ||
301 | if (unlikely(!PageHead(page))) | ||
302 | bad_page(page); | ||
303 | __ClearPageHead(page); | 305 | __ClearPageHead(page); |
306 | |||
304 | for (i = 1; i < nr_pages; i++) { | 307 | for (i = 1; i < nr_pages; i++) { |
305 | struct page *p = page + i; | 308 | struct page *p = page + i; |
306 | 309 | ||
307 | if (unlikely(!PageTail(p) | | 310 | if (unlikely(!PageTail(p) | (p->first_page != page))) { |
308 | (p->first_page != page))) | ||
309 | bad_page(page); | 311 | bad_page(page); |
312 | bad++; | ||
313 | } | ||
310 | __ClearPageTail(p); | 314 | __ClearPageTail(p); |
311 | } | 315 | } |
316 | |||
317 | return bad; | ||
312 | } | 318 | } |
313 | 319 | ||
314 | static inline void prep_zero_page(struct page *page, int order, gfp_t gfp_flags) | 320 | static inline void prep_zero_page(struct page *page, int order, gfp_t gfp_flags) |
@@ -428,7 +434,8 @@ static inline void __free_one_page(struct page *page, | |||
428 | int migratetype = get_pageblock_migratetype(page); | 434 | int migratetype = get_pageblock_migratetype(page); |
429 | 435 | ||
430 | if (unlikely(PageCompound(page))) | 436 | if (unlikely(PageCompound(page))) |
431 | destroy_compound_page(page, order); | 437 | if (unlikely(destroy_compound_page(page, order))) |
438 | return; | ||
432 | 439 | ||
433 | page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); | 440 | page_idx = page_to_pfn(page) & ((1 << MAX_ORDER) - 1); |
434 | 441 | ||
@@ -465,15 +472,10 @@ static inline int free_pages_check(struct page *page) | |||
465 | if (unlikely(page_mapcount(page) | | 472 | if (unlikely(page_mapcount(page) | |
466 | (page->mapping != NULL) | | 473 | (page->mapping != NULL) | |
467 | (page_count(page) != 0) | | 474 | (page_count(page) != 0) | |
468 | (page->flags & PAGE_FLAGS_CHECK_AT_FREE))) | 475 | (page->flags & PAGE_FLAGS_CHECK_AT_FREE))) { |
469 | bad_page(page); | 476 | bad_page(page); |
470 | /* | ||
471 | * For now, we report if PG_reserved was found set, but do not | ||
472 | * clear it, and do not free the page. But we shall soon need | ||
473 | * to do more, for when the ZERO_PAGE count wraps negative. | ||
474 | */ | ||
475 | if (PageReserved(page)) | ||
476 | return 1; | 477 | return 1; |
478 | } | ||
477 | if (page->flags & PAGE_FLAGS_CHECK_AT_PREP) | 479 | if (page->flags & PAGE_FLAGS_CHECK_AT_PREP) |
478 | page->flags &= ~PAGE_FLAGS_CHECK_AT_PREP; | 480 | page->flags &= ~PAGE_FLAGS_CHECK_AT_PREP; |
479 | return 0; | 481 | return 0; |
@@ -521,11 +523,11 @@ static void __free_pages_ok(struct page *page, unsigned int order) | |||
521 | { | 523 | { |
522 | unsigned long flags; | 524 | unsigned long flags; |
523 | int i; | 525 | int i; |
524 | int reserved = 0; | 526 | int bad = 0; |
525 | 527 | ||
526 | for (i = 0 ; i < (1 << order) ; ++i) | 528 | for (i = 0 ; i < (1 << order) ; ++i) |
527 | reserved += free_pages_check(page + i); | 529 | bad += free_pages_check(page + i); |
528 | if (reserved) | 530 | if (bad) |
529 | return; | 531 | return; |
530 | 532 | ||
531 | if (!PageHighMem(page)) { | 533 | if (!PageHighMem(page)) { |
@@ -610,17 +612,11 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags) | |||
610 | if (unlikely(page_mapcount(page) | | 612 | if (unlikely(page_mapcount(page) | |
611 | (page->mapping != NULL) | | 613 | (page->mapping != NULL) | |
612 | (page_count(page) != 0) | | 614 | (page_count(page) != 0) | |
613 | (page->flags & PAGE_FLAGS_CHECK_AT_PREP))) | 615 | (page->flags & PAGE_FLAGS_CHECK_AT_PREP))) { |
614 | bad_page(page); | 616 | bad_page(page); |
615 | |||
616 | /* | ||
617 | * For now, we report if PG_reserved was found set, but do not | ||
618 | * clear it, and do not allocate the page: as a safety net. | ||
619 | */ | ||
620 | if (PageReserved(page)) | ||
621 | return 1; | 617 | return 1; |
618 | } | ||
622 | 619 | ||
623 | page->flags &= ~PAGE_FLAGS_CHECK_AT_PREP; | ||
624 | set_page_private(page, 0); | 620 | set_page_private(page, 0); |
625 | set_page_refcounted(page); | 621 | set_page_refcounted(page); |
626 | 622 | ||