aboutsummaryrefslogtreecommitdiffstats
path: root/mm/page_alloc.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/page_alloc.c')
-rw-r--r--mm/page_alloc.c63
1 files changed, 33 insertions, 30 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index dde04ff4be31..234bd4895d14 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -56,6 +56,7 @@ long nr_swap_pages;
56int percpu_pagelist_fraction; 56int percpu_pagelist_fraction;
57 57
58static void fastcall free_hot_cold_page(struct page *page, int cold); 58static void fastcall free_hot_cold_page(struct page *page, int cold);
59static void __free_pages_ok(struct page *page, unsigned int order);
59 60
60/* 61/*
61 * results with 256, 32 in the lowmem_reserve sysctl: 62 * results with 256, 32 in the lowmem_reserve sysctl:
@@ -169,20 +170,23 @@ static void bad_page(struct page *page)
169 * All pages have PG_compound set. All pages have their ->private pointing at 170 * All pages have PG_compound set. All pages have their ->private pointing at
170 * the head page (even the head page has this). 171 * the head page (even the head page has this).
171 * 172 *
172 * The first tail page's ->mapping, if non-zero, holds the address of the 173 * The first tail page's ->lru.next holds the address of the compound page's
173 * compound page's put_page() function. 174 * put_page() function. Its ->lru.prev holds the order of allocation.
174 * 175 * This usage means that zero-order pages may not be compound.
175 * The order of the allocation is stored in the first tail page's ->index
176 * This is only for debug at present. This usage means that zero-order pages
177 * may not be compound.
178 */ 176 */
177
178static void free_compound_page(struct page *page)
179{
180 __free_pages_ok(page, (unsigned long)page[1].lru.prev);
181}
182
179static void prep_compound_page(struct page *page, unsigned long order) 183static void prep_compound_page(struct page *page, unsigned long order)
180{ 184{
181 int i; 185 int i;
182 int nr_pages = 1 << order; 186 int nr_pages = 1 << order;
183 187
184 page[1].mapping = NULL; 188 page[1].lru.next = (void *)free_compound_page; /* set dtor */
185 page[1].index = order; 189 page[1].lru.prev = (void *)order;
186 for (i = 0; i < nr_pages; i++) { 190 for (i = 0; i < nr_pages; i++) {
187 struct page *p = page + i; 191 struct page *p = page + i;
188 192
@@ -196,7 +200,7 @@ static void destroy_compound_page(struct page *page, unsigned long order)
196 int i; 200 int i;
197 int nr_pages = 1 << order; 201 int nr_pages = 1 << order;
198 202
199 if (unlikely(page[1].index != order)) 203 if (unlikely((unsigned long)page[1].lru.prev != order))
200 bad_page(page); 204 bad_page(page);
201 205
202 for (i = 0; i < nr_pages; i++) { 206 for (i = 0; i < nr_pages; i++) {
@@ -586,21 +590,20 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
586} 590}
587 591
588#ifdef CONFIG_NUMA 592#ifdef CONFIG_NUMA
589/* Called from the slab reaper to drain remote pagesets */ 593/*
590void drain_remote_pages(void) 594 * Called from the slab reaper to drain pagesets on a particular node that
595 * belong to the currently executing processor.
596 */
597void drain_node_pages(int nodeid)
591{ 598{
592 struct zone *zone; 599 int i, z;
593 int i;
594 unsigned long flags; 600 unsigned long flags;
595 601
596 local_irq_save(flags); 602 local_irq_save(flags);
597 for_each_zone(zone) { 603 for (z = 0; z < MAX_NR_ZONES; z++) {
604 struct zone *zone = NODE_DATA(nodeid)->node_zones + z;
598 struct per_cpu_pageset *pset; 605 struct per_cpu_pageset *pset;
599 606
600 /* Do not drain local pagesets */
601 if (zone->zone_pgdat->node_id == numa_node_id())
602 continue;
603
604 pset = zone_pcp(zone, smp_processor_id()); 607 pset = zone_pcp(zone, smp_processor_id());
605 for (i = 0; i < ARRAY_SIZE(pset->pcp); i++) { 608 for (i = 0; i < ARRAY_SIZE(pset->pcp); i++) {
606 struct per_cpu_pages *pcp; 609 struct per_cpu_pages *pcp;
@@ -1011,7 +1014,7 @@ rebalance:
1011 if (page) 1014 if (page)
1012 goto got_pg; 1015 goto got_pg;
1013 1016
1014 out_of_memory(gfp_mask, order); 1017 out_of_memory(zonelist, gfp_mask, order);
1015 goto restart; 1018 goto restart;
1016 } 1019 }
1017 1020
@@ -1537,29 +1540,29 @@ static int __initdata node_load[MAX_NUMNODES];
1537 */ 1540 */
1538static int __init find_next_best_node(int node, nodemask_t *used_node_mask) 1541static int __init find_next_best_node(int node, nodemask_t *used_node_mask)
1539{ 1542{
1540 int i, n, val; 1543 int n, val;
1541 int min_val = INT_MAX; 1544 int min_val = INT_MAX;
1542 int best_node = -1; 1545 int best_node = -1;
1543 1546
1544 for_each_online_node(i) { 1547 /* Use the local node if we haven't already */
1545 cpumask_t tmp; 1548 if (!node_isset(node, *used_node_mask)) {
1549 node_set(node, *used_node_mask);
1550 return node;
1551 }
1546 1552
1547 /* Start from local node */ 1553 for_each_online_node(n) {
1548 n = (node+i) % num_online_nodes(); 1554 cpumask_t tmp;
1549 1555
1550 /* Don't want a node to appear more than once */ 1556 /* Don't want a node to appear more than once */
1551 if (node_isset(n, *used_node_mask)) 1557 if (node_isset(n, *used_node_mask))
1552 continue; 1558 continue;
1553 1559
1554 /* Use the local node if we haven't already */
1555 if (!node_isset(node, *used_node_mask)) {
1556 best_node = node;
1557 break;
1558 }
1559
1560 /* Use the distance array to find the distance */ 1560 /* Use the distance array to find the distance */
1561 val = node_distance(node, n); 1561 val = node_distance(node, n);
1562 1562
1563 /* Penalize nodes under us ("prefer the next node") */
1564 val += (n < node);
1565
1563 /* Give preference to headless and unused nodes */ 1566 /* Give preference to headless and unused nodes */
1564 tmp = node_to_cpumask(n); 1567 tmp = node_to_cpumask(n);
1565 if (!cpus_empty(tmp)) 1568 if (!cpus_empty(tmp))