diff options
Diffstat (limited to 'mm/page_alloc.c')
-rw-r--r-- | mm/page_alloc.c | 138 |
1 files changed, 110 insertions, 28 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index b864584c92b4..f7873a47fa8e 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
@@ -687,16 +687,48 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags) | |||
687 | return 0; | 687 | return 0; |
688 | } | 688 | } |
689 | 689 | ||
690 | /* | ||
691 | * Go through the free lists for the given migratetype and remove | ||
692 | * the smallest available page from the freelists | ||
693 | */ | ||
694 | static struct page *__rmqueue_smallest(struct zone *zone, unsigned int order, | ||
695 | int migratetype) | ||
696 | { | ||
697 | unsigned int current_order; | ||
698 | struct free_area * area; | ||
699 | struct page *page; | ||
700 | |||
701 | /* Find a page of the appropriate size in the preferred list */ | ||
702 | for (current_order = order; current_order < MAX_ORDER; ++current_order) { | ||
703 | area = &(zone->free_area[current_order]); | ||
704 | if (list_empty(&area->free_list[migratetype])) | ||
705 | continue; | ||
706 | |||
707 | page = list_entry(area->free_list[migratetype].next, | ||
708 | struct page, lru); | ||
709 | list_del(&page->lru); | ||
710 | rmv_page_order(page); | ||
711 | area->nr_free--; | ||
712 | __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order)); | ||
713 | expand(zone, page, order, current_order, area, migratetype); | ||
714 | return page; | ||
715 | } | ||
716 | |||
717 | return NULL; | ||
718 | } | ||
719 | |||
720 | |||
690 | #ifdef CONFIG_PAGE_GROUP_BY_MOBILITY | 721 | #ifdef CONFIG_PAGE_GROUP_BY_MOBILITY |
691 | /* | 722 | /* |
692 | * This array describes the order lists are fallen back to when | 723 | * This array describes the order lists are fallen back to when |
693 | * the free lists for the desirable migrate type are depleted | 724 | * the free lists for the desirable migrate type are depleted |
694 | */ | 725 | */ |
695 | static int fallbacks[MIGRATE_TYPES][MIGRATE_TYPES-1] = { | 726 | static int fallbacks[MIGRATE_TYPES][MIGRATE_TYPES-1] = { |
696 | [MIGRATE_UNMOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC }, | 727 | [MIGRATE_UNMOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE }, |
697 | [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC }, | 728 | [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE }, |
698 | [MIGRATE_MOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_HIGHATOMIC }, | 729 | [MIGRATE_MOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE }, |
699 | [MIGRATE_HIGHATOMIC] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_MOVABLE}, | 730 | [MIGRATE_HIGHATOMIC] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_RESERVE }, |
731 | [MIGRATE_RESERVE] = { MIGRATE_RESERVE, MIGRATE_RESERVE, MIGRATE_RESERVE, MIGRATE_RESERVE }, /* Never used */ | ||
700 | }; | 732 | }; |
701 | 733 | ||
702 | /* | 734 | /* |
@@ -799,6 +831,9 @@ retry: | |||
799 | for (i = 0; i < MIGRATE_TYPES - 1; i++) { | 831 | for (i = 0; i < MIGRATE_TYPES - 1; i++) { |
800 | migratetype = fallbacks[start_migratetype][i]; | 832 | migratetype = fallbacks[start_migratetype][i]; |
801 | 833 | ||
834 | /* MIGRATE_RESERVE handled later if necessary */ | ||
835 | if (migratetype == MIGRATE_RESERVE) | ||
836 | continue; | ||
802 | /* | 837 | /* |
803 | * Make it hard to fallback to blocks used for | 838 | * Make it hard to fallback to blocks used for |
804 | * high-order atomic allocations | 839 | * high-order atomic allocations |
@@ -861,7 +896,8 @@ retry: | |||
861 | goto retry; | 896 | goto retry; |
862 | } | 897 | } |
863 | 898 | ||
864 | return NULL; | 899 | /* Use MIGRATE_RESERVE rather than fail an allocation */ |
900 | return __rmqueue_smallest(zone, order, MIGRATE_RESERVE); | ||
865 | } | 901 | } |
866 | #else | 902 | #else |
867 | static struct page *__rmqueue_fallback(struct zone *zone, int order, | 903 | static struct page *__rmqueue_fallback(struct zone *zone, int order, |
@@ -871,36 +907,19 @@ static struct page *__rmqueue_fallback(struct zone *zone, int order, | |||
871 | } | 907 | } |
872 | #endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */ | 908 | #endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */ |
873 | 909 | ||
874 | /* | 910 | /* |
875 | * Do the hard work of removing an element from the buddy allocator. | 911 | * Do the hard work of removing an element from the buddy allocator. |
876 | * Call me with the zone->lock already held. | 912 | * Call me with the zone->lock already held. |
877 | */ | 913 | */ |
878 | static struct page *__rmqueue(struct zone *zone, unsigned int order, | 914 | static struct page *__rmqueue(struct zone *zone, unsigned int order, |
879 | int migratetype) | 915 | int migratetype) |
880 | { | 916 | { |
881 | struct free_area * area; | ||
882 | unsigned int current_order; | ||
883 | struct page *page; | 917 | struct page *page; |
884 | 918 | ||
885 | /* Find a page of the appropriate size in the preferred list */ | 919 | page = __rmqueue_smallest(zone, order, migratetype); |
886 | for (current_order = order; current_order < MAX_ORDER; ++current_order) { | ||
887 | area = &(zone->free_area[current_order]); | ||
888 | if (list_empty(&area->free_list[migratetype])) | ||
889 | continue; | ||
890 | |||
891 | page = list_entry(area->free_list[migratetype].next, | ||
892 | struct page, lru); | ||
893 | list_del(&page->lru); | ||
894 | rmv_page_order(page); | ||
895 | area->nr_free--; | ||
896 | __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order)); | ||
897 | expand(zone, page, order, current_order, area, migratetype); | ||
898 | goto got_page; | ||
899 | } | ||
900 | |||
901 | page = __rmqueue_fallback(zone, order, migratetype); | ||
902 | 920 | ||
903 | got_page: | 921 | if (unlikely(!page)) |
922 | page = __rmqueue_fallback(zone, order, migratetype); | ||
904 | 923 | ||
905 | return page; | 924 | return page; |
906 | } | 925 | } |
@@ -2506,6 +2525,65 @@ static inline unsigned long wait_table_bits(unsigned long size) | |||
2506 | 2525 | ||
2507 | #define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1)) | 2526 | #define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1)) |
2508 | 2527 | ||
2528 | #ifdef CONFIG_PAGE_GROUP_BY_MOBILITY | ||
2529 | /* | ||
2530 | * Mark a number of MAX_ORDER_NR_PAGES blocks as MIGRATE_RESERVE. The number | ||
2531 | * of blocks reserved is based on zone->pages_min. The memory within the | ||
2532 | * reserve will tend to store contiguous free pages. Setting min_free_kbytes | ||
2533 | * higher will lead to a bigger reserve which will get freed as contiguous | ||
2534 | * blocks as reclaim kicks in | ||
2535 | */ | ||
2536 | static void setup_zone_migrate_reserve(struct zone *zone) | ||
2537 | { | ||
2538 | unsigned long start_pfn, pfn, end_pfn; | ||
2539 | struct page *page; | ||
2540 | unsigned long reserve, block_migratetype; | ||
2541 | |||
2542 | /* Get the start pfn, end pfn and the number of blocks to reserve */ | ||
2543 | start_pfn = zone->zone_start_pfn; | ||
2544 | end_pfn = start_pfn + zone->spanned_pages; | ||
2545 | reserve = roundup(zone->pages_min, MAX_ORDER_NR_PAGES) >> (MAX_ORDER-1); | ||
2546 | |||
2547 | for (pfn = start_pfn; pfn < end_pfn; pfn += MAX_ORDER_NR_PAGES) { | ||
2548 | if (!pfn_valid(pfn)) | ||
2549 | continue; | ||
2550 | page = pfn_to_page(pfn); | ||
2551 | |||
2552 | /* Blocks with reserved pages will never free, skip them. */ | ||
2553 | if (PageReserved(page)) | ||
2554 | continue; | ||
2555 | |||
2556 | block_migratetype = get_pageblock_migratetype(page); | ||
2557 | |||
2558 | /* If this block is reserved, account for it */ | ||
2559 | if (reserve > 0 && block_migratetype == MIGRATE_RESERVE) { | ||
2560 | reserve--; | ||
2561 | continue; | ||
2562 | } | ||
2563 | |||
2564 | /* Suitable for reserving if this block is movable */ | ||
2565 | if (reserve > 0 && block_migratetype == MIGRATE_MOVABLE) { | ||
2566 | set_pageblock_migratetype(page, MIGRATE_RESERVE); | ||
2567 | move_freepages_block(zone, page, MIGRATE_RESERVE); | ||
2568 | reserve--; | ||
2569 | continue; | ||
2570 | } | ||
2571 | |||
2572 | /* | ||
2573 | * If the reserve is met and this is a previous reserved block, | ||
2574 | * take it back | ||
2575 | */ | ||
2576 | if (block_migratetype == MIGRATE_RESERVE) { | ||
2577 | set_pageblock_migratetype(page, MIGRATE_MOVABLE); | ||
2578 | move_freepages_block(zone, page, MIGRATE_MOVABLE); | ||
2579 | } | ||
2580 | } | ||
2581 | } | ||
2582 | #else | ||
2583 | static inline void setup_zone_migrate_reserve(struct zone *zone) | ||
2584 | { | ||
2585 | } | ||
2586 | #endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */ | ||
2509 | /* | 2587 | /* |
2510 | * Initially all pages are reserved - free ones are freed | 2588 | * Initially all pages are reserved - free ones are freed |
2511 | * up by free_all_bootmem() once the early boot process is | 2589 | * up by free_all_bootmem() once the early boot process is |
@@ -2541,9 +2619,12 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone, | |||
2541 | * movable at startup. This will force kernel allocations | 2619 | * movable at startup. This will force kernel allocations |
2542 | * to reserve their blocks rather than leaking throughout | 2620 | * to reserve their blocks rather than leaking throughout |
2543 | * the address space during boot when many long-lived | 2621 | * the address space during boot when many long-lived |
2544 | * kernel allocations are made | 2622 | * kernel allocations are made. Later some blocks near |
2623 | * the start are marked MIGRATE_RESERVE by | ||
2624 | * setup_zone_migrate_reserve() | ||
2545 | */ | 2625 | */ |
2546 | set_pageblock_migratetype(page, MIGRATE_MOVABLE); | 2626 | if ((pfn & (MAX_ORDER_NR_PAGES-1))) |
2627 | set_pageblock_migratetype(page, MIGRATE_MOVABLE); | ||
2547 | 2628 | ||
2548 | INIT_LIST_HEAD(&page->lru); | 2629 | INIT_LIST_HEAD(&page->lru); |
2549 | #ifdef WANT_PAGE_VIRTUAL | 2630 | #ifdef WANT_PAGE_VIRTUAL |
@@ -4078,6 +4159,7 @@ void setup_per_zone_pages_min(void) | |||
4078 | 4159 | ||
4079 | zone->pages_low = zone->pages_min + (tmp >> 2); | 4160 | zone->pages_low = zone->pages_min + (tmp >> 2); |
4080 | zone->pages_high = zone->pages_min + (tmp >> 1); | 4161 | zone->pages_high = zone->pages_min + (tmp >> 1); |
4162 | setup_zone_migrate_reserve(zone); | ||
4081 | spin_unlock_irqrestore(&zone->lru_lock, flags); | 4163 | spin_unlock_irqrestore(&zone->lru_lock, flags); |
4082 | } | 4164 | } |
4083 | 4165 | ||