aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/page_alloc.c138
1 files changed, 110 insertions, 28 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index b864584c92b4..f7873a47fa8e 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -687,16 +687,48 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags)
687 return 0; 687 return 0;
688} 688}
689 689
690/*
691 * Go through the free lists for the given migratetype and remove
692 * the smallest available page from the freelists
693 */
694static struct page *__rmqueue_smallest(struct zone *zone, unsigned int order,
695 int migratetype)
696{
697 unsigned int current_order;
698 struct free_area * area;
699 struct page *page;
700
701 /* Find a page of the appropriate size in the preferred list */
702 for (current_order = order; current_order < MAX_ORDER; ++current_order) {
703 area = &(zone->free_area[current_order]);
704 if (list_empty(&area->free_list[migratetype]))
705 continue;
706
707 page = list_entry(area->free_list[migratetype].next,
708 struct page, lru);
709 list_del(&page->lru);
710 rmv_page_order(page);
711 area->nr_free--;
712 __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order));
713 expand(zone, page, order, current_order, area, migratetype);
714 return page;
715 }
716
717 return NULL;
718}
719
720
690#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY 721#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY
691/* 722/*
692 * This array describes the order lists are fallen back to when 723 * This array describes the order lists are fallen back to when
693 * the free lists for the desirable migrate type are depleted 724 * the free lists for the desirable migrate type are depleted
694 */ 725 */
695static int fallbacks[MIGRATE_TYPES][MIGRATE_TYPES-1] = { 726static int fallbacks[MIGRATE_TYPES][MIGRATE_TYPES-1] = {
696 [MIGRATE_UNMOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC }, 727 [MIGRATE_UNMOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
697 [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC }, 728 [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
698 [MIGRATE_MOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_HIGHATOMIC }, 729 [MIGRATE_MOVABLE] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
699 [MIGRATE_HIGHATOMIC] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_MOVABLE}, 730 [MIGRATE_HIGHATOMIC] = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_MOVABLE, MIGRATE_RESERVE },
731 [MIGRATE_RESERVE] = { MIGRATE_RESERVE, MIGRATE_RESERVE, MIGRATE_RESERVE, MIGRATE_RESERVE }, /* Never used */
700}; 732};
701 733
702/* 734/*
@@ -799,6 +831,9 @@ retry:
799 for (i = 0; i < MIGRATE_TYPES - 1; i++) { 831 for (i = 0; i < MIGRATE_TYPES - 1; i++) {
800 migratetype = fallbacks[start_migratetype][i]; 832 migratetype = fallbacks[start_migratetype][i];
801 833
834 /* MIGRATE_RESERVE handled later if necessary */
835 if (migratetype == MIGRATE_RESERVE)
836 continue;
802 /* 837 /*
803 * Make it hard to fallback to blocks used for 838 * Make it hard to fallback to blocks used for
804 * high-order atomic allocations 839 * high-order atomic allocations
@@ -861,7 +896,8 @@ retry:
861 goto retry; 896 goto retry;
862 } 897 }
863 898
864 return NULL; 899 /* Use MIGRATE_RESERVE rather than fail an allocation */
900 return __rmqueue_smallest(zone, order, MIGRATE_RESERVE);
865} 901}
866#else 902#else
867static struct page *__rmqueue_fallback(struct zone *zone, int order, 903static struct page *__rmqueue_fallback(struct zone *zone, int order,
@@ -871,36 +907,19 @@ static struct page *__rmqueue_fallback(struct zone *zone, int order,
871} 907}
872#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */ 908#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */
873 909
874/* 910/*
875 * Do the hard work of removing an element from the buddy allocator. 911 * Do the hard work of removing an element from the buddy allocator.
876 * Call me with the zone->lock already held. 912 * Call me with the zone->lock already held.
877 */ 913 */
878static struct page *__rmqueue(struct zone *zone, unsigned int order, 914static struct page *__rmqueue(struct zone *zone, unsigned int order,
879 int migratetype) 915 int migratetype)
880{ 916{
881 struct free_area * area;
882 unsigned int current_order;
883 struct page *page; 917 struct page *page;
884 918
885 /* Find a page of the appropriate size in the preferred list */ 919 page = __rmqueue_smallest(zone, order, migratetype);
886 for (current_order = order; current_order < MAX_ORDER; ++current_order) {
887 area = &(zone->free_area[current_order]);
888 if (list_empty(&area->free_list[migratetype]))
889 continue;
890
891 page = list_entry(area->free_list[migratetype].next,
892 struct page, lru);
893 list_del(&page->lru);
894 rmv_page_order(page);
895 area->nr_free--;
896 __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order));
897 expand(zone, page, order, current_order, area, migratetype);
898 goto got_page;
899 }
900
901 page = __rmqueue_fallback(zone, order, migratetype);
902 920
903got_page: 921 if (unlikely(!page))
922 page = __rmqueue_fallback(zone, order, migratetype);
904 923
905 return page; 924 return page;
906} 925}
@@ -2506,6 +2525,65 @@ static inline unsigned long wait_table_bits(unsigned long size)
2506 2525
2507#define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1)) 2526#define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1))
2508 2527
2528#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY
2529/*
2530 * Mark a number of MAX_ORDER_NR_PAGES blocks as MIGRATE_RESERVE. The number
2531 * of blocks reserved is based on zone->pages_min. The memory within the
2532 * reserve will tend to store contiguous free pages. Setting min_free_kbytes
2533 * higher will lead to a bigger reserve which will get freed as contiguous
2534 * blocks as reclaim kicks in
2535 */
2536static void setup_zone_migrate_reserve(struct zone *zone)
2537{
2538 unsigned long start_pfn, pfn, end_pfn;
2539 struct page *page;
2540 unsigned long reserve, block_migratetype;
2541
2542 /* Get the start pfn, end pfn and the number of blocks to reserve */
2543 start_pfn = zone->zone_start_pfn;
2544 end_pfn = start_pfn + zone->spanned_pages;
2545 reserve = roundup(zone->pages_min, MAX_ORDER_NR_PAGES) >> (MAX_ORDER-1);
2546
2547 for (pfn = start_pfn; pfn < end_pfn; pfn += MAX_ORDER_NR_PAGES) {
2548 if (!pfn_valid(pfn))
2549 continue;
2550 page = pfn_to_page(pfn);
2551
2552 /* Blocks with reserved pages will never free, skip them. */
2553 if (PageReserved(page))
2554 continue;
2555
2556 block_migratetype = get_pageblock_migratetype(page);
2557
2558 /* If this block is reserved, account for it */
2559 if (reserve > 0 && block_migratetype == MIGRATE_RESERVE) {
2560 reserve--;
2561 continue;
2562 }
2563
2564 /* Suitable for reserving if this block is movable */
2565 if (reserve > 0 && block_migratetype == MIGRATE_MOVABLE) {
2566 set_pageblock_migratetype(page, MIGRATE_RESERVE);
2567 move_freepages_block(zone, page, MIGRATE_RESERVE);
2568 reserve--;
2569 continue;
2570 }
2571
2572 /*
2573 * If the reserve is met and this is a previous reserved block,
2574 * take it back
2575 */
2576 if (block_migratetype == MIGRATE_RESERVE) {
2577 set_pageblock_migratetype(page, MIGRATE_MOVABLE);
2578 move_freepages_block(zone, page, MIGRATE_MOVABLE);
2579 }
2580 }
2581}
2582#else
2583static inline void setup_zone_migrate_reserve(struct zone *zone)
2584{
2585}
2586#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */
2509/* 2587/*
2510 * Initially all pages are reserved - free ones are freed 2588 * Initially all pages are reserved - free ones are freed
2511 * up by free_all_bootmem() once the early boot process is 2589 * up by free_all_bootmem() once the early boot process is
@@ -2541,9 +2619,12 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone,
2541 * movable at startup. This will force kernel allocations 2619 * movable at startup. This will force kernel allocations
2542 * to reserve their blocks rather than leaking throughout 2620 * to reserve their blocks rather than leaking throughout
2543 * the address space during boot when many long-lived 2621 * the address space during boot when many long-lived
2544 * kernel allocations are made 2622 * kernel allocations are made. Later some blocks near
2623 * the start are marked MIGRATE_RESERVE by
2624 * setup_zone_migrate_reserve()
2545 */ 2625 */
2546 set_pageblock_migratetype(page, MIGRATE_MOVABLE); 2626 if ((pfn & (MAX_ORDER_NR_PAGES-1)))
2627 set_pageblock_migratetype(page, MIGRATE_MOVABLE);
2547 2628
2548 INIT_LIST_HEAD(&page->lru); 2629 INIT_LIST_HEAD(&page->lru);
2549#ifdef WANT_PAGE_VIRTUAL 2630#ifdef WANT_PAGE_VIRTUAL
@@ -4078,6 +4159,7 @@ void setup_per_zone_pages_min(void)
4078 4159
4079 zone->pages_low = zone->pages_min + (tmp >> 2); 4160 zone->pages_low = zone->pages_min + (tmp >> 2);
4080 zone->pages_high = zone->pages_min + (tmp >> 1); 4161 zone->pages_high = zone->pages_min + (tmp >> 1);
4162 setup_zone_migrate_reserve(zone);
4081 spin_unlock_irqrestore(&zone->lru_lock, flags); 4163 spin_unlock_irqrestore(&zone->lru_lock, flags);
4082 } 4164 }
4083 4165