aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
Diffstat (limited to 'mm')
-rw-r--r--mm/vmscan.c108
1 files changed, 62 insertions, 46 deletions
diff --git a/mm/vmscan.c b/mm/vmscan.c
index f7c4f37c3b18..5feef4d4650e 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1431,13 +1431,14 @@ int try_to_free_pages(struct zone **zones, gfp_t gfp_mask)
1431 int ret = 0; 1431 int ret = 0;
1432 int total_scanned = 0, total_reclaimed = 0; 1432 int total_scanned = 0, total_reclaimed = 0;
1433 struct reclaim_state *reclaim_state = current->reclaim_state; 1433 struct reclaim_state *reclaim_state = current->reclaim_state;
1434 struct scan_control sc;
1435 unsigned long lru_pages = 0; 1434 unsigned long lru_pages = 0;
1436 int i; 1435 int i;
1437 1436 struct scan_control sc = {
1438 sc.gfp_mask = gfp_mask; 1437 .gfp_mask = gfp_mask,
1439 sc.may_writepage = !laptop_mode; 1438 .may_writepage = !laptop_mode,
1440 sc.may_swap = 1; 1439 .swap_cluster_max = SWAP_CLUSTER_MAX,
1440 .may_swap = 1,
1441 };
1441 1442
1442 inc_page_state(allocstall); 1443 inc_page_state(allocstall);
1443 1444
@@ -1455,7 +1456,6 @@ int try_to_free_pages(struct zone **zones, gfp_t gfp_mask)
1455 sc.nr_mapped = read_page_state(nr_mapped); 1456 sc.nr_mapped = read_page_state(nr_mapped);
1456 sc.nr_scanned = 0; 1457 sc.nr_scanned = 0;
1457 sc.nr_reclaimed = 0; 1458 sc.nr_reclaimed = 0;
1458 sc.swap_cluster_max = SWAP_CLUSTER_MAX;
1459 if (!priority) 1459 if (!priority)
1460 disable_swap_token(); 1460 disable_swap_token();
1461 shrink_caches(priority, zones, &sc); 1461 shrink_caches(priority, zones, &sc);
@@ -1478,7 +1478,8 @@ int try_to_free_pages(struct zone **zones, gfp_t gfp_mask)
1478 * that's undesirable in laptop mode, where we *want* lumpy 1478 * that's undesirable in laptop mode, where we *want* lumpy
1479 * writeout. So in laptop mode, write out the whole world. 1479 * writeout. So in laptop mode, write out the whole world.
1480 */ 1480 */
1481 if (total_scanned > sc.swap_cluster_max + sc.swap_cluster_max/2) { 1481 if (total_scanned > sc.swap_cluster_max +
1482 sc.swap_cluster_max / 2) {
1482 wakeup_pdflush(laptop_mode ? 0 : total_scanned); 1483 wakeup_pdflush(laptop_mode ? 0 : total_scanned);
1483 sc.may_writepage = 1; 1484 sc.may_writepage = 1;
1484 } 1485 }
@@ -1532,14 +1533,16 @@ static int balance_pgdat(pg_data_t *pgdat, int nr_pages, int order)
1532 int i; 1533 int i;
1533 int total_scanned, total_reclaimed; 1534 int total_scanned, total_reclaimed;
1534 struct reclaim_state *reclaim_state = current->reclaim_state; 1535 struct reclaim_state *reclaim_state = current->reclaim_state;
1535 struct scan_control sc; 1536 struct scan_control sc = {
1537 .gfp_mask = GFP_KERNEL,
1538 .may_swap = 1,
1539 .swap_cluster_max = nr_pages ? nr_pages : SWAP_CLUSTER_MAX,
1540 };
1536 1541
1537loop_again: 1542loop_again:
1538 total_scanned = 0; 1543 total_scanned = 0;
1539 total_reclaimed = 0; 1544 total_reclaimed = 0;
1540 sc.gfp_mask = GFP_KERNEL; 1545 sc.may_writepage = !laptop_mode,
1541 sc.may_writepage = !laptop_mode;
1542 sc.may_swap = 1;
1543 sc.nr_mapped = read_page_state(nr_mapped); 1546 sc.nr_mapped = read_page_state(nr_mapped);
1544 1547
1545 inc_page_state(pageoutrun); 1548 inc_page_state(pageoutrun);
@@ -1621,7 +1624,6 @@ scan:
1621 zone->prev_priority = priority; 1624 zone->prev_priority = priority;
1622 sc.nr_scanned = 0; 1625 sc.nr_scanned = 0;
1623 sc.nr_reclaimed = 0; 1626 sc.nr_reclaimed = 0;
1624 sc.swap_cluster_max = nr_pages? nr_pages : SWAP_CLUSTER_MAX;
1625 shrink_zone(priority, zone, &sc); 1627 shrink_zone(priority, zone, &sc);
1626 reclaim_state->reclaimed_slab = 0; 1628 reclaim_state->reclaimed_slab = 0;
1627 nr_slab = shrink_slab(sc.nr_scanned, GFP_KERNEL, 1629 nr_slab = shrink_slab(sc.nr_scanned, GFP_KERNEL,
@@ -1869,46 +1871,21 @@ int zone_reclaim_interval __read_mostly = 30*HZ;
1869/* 1871/*
1870 * Try to free up some pages from this zone through reclaim. 1872 * Try to free up some pages from this zone through reclaim.
1871 */ 1873 */
1872int zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order) 1874static int __zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order)
1873{ 1875{
1874 int nr_pages; 1876 const int nr_pages = 1 << order;
1875 struct task_struct *p = current; 1877 struct task_struct *p = current;
1876 struct reclaim_state reclaim_state; 1878 struct reclaim_state reclaim_state;
1877 struct scan_control sc;
1878 cpumask_t mask;
1879 int node_id;
1880 int priority; 1879 int priority;
1881 1880 struct scan_control sc = {
1882 if (time_before(jiffies, 1881 .may_writepage = !!(zone_reclaim_mode & RECLAIM_WRITE),
1883 zone->last_unsuccessful_zone_reclaim + zone_reclaim_interval)) 1882 .may_swap = !!(zone_reclaim_mode & RECLAIM_SWAP),
1884 return 0; 1883 .nr_mapped = read_page_state(nr_mapped),
1885 1884 .swap_cluster_max = max(nr_pages, SWAP_CLUSTER_MAX),
1886 if (!(gfp_mask & __GFP_WAIT) || 1885 .gfp_mask = gfp_mask,
1887 zone->all_unreclaimable || 1886 };
1888 atomic_read(&zone->reclaim_in_progress) > 0 ||
1889 (p->flags & PF_MEMALLOC))
1890 return 0;
1891
1892 node_id = zone->zone_pgdat->node_id;
1893 mask = node_to_cpumask(node_id);
1894 if (!cpus_empty(mask) && node_id != numa_node_id())
1895 return 0;
1896
1897 sc.may_writepage = !!(zone_reclaim_mode & RECLAIM_WRITE);
1898 sc.may_swap = !!(zone_reclaim_mode & RECLAIM_SWAP);
1899 sc.nr_scanned = 0;
1900 sc.nr_reclaimed = 0;
1901 sc.nr_mapped = read_page_state(nr_mapped);
1902 sc.gfp_mask = gfp_mask;
1903 1887
1904 disable_swap_token(); 1888 disable_swap_token();
1905
1906 nr_pages = 1 << order;
1907 if (nr_pages > SWAP_CLUSTER_MAX)
1908 sc.swap_cluster_max = nr_pages;
1909 else
1910 sc.swap_cluster_max = SWAP_CLUSTER_MAX;
1911
1912 cond_resched(); 1889 cond_resched();
1913 /* 1890 /*
1914 * We need to be able to allocate from the reserves for RECLAIM_SWAP 1891 * We need to be able to allocate from the reserves for RECLAIM_SWAP
@@ -1949,5 +1926,44 @@ int zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order)
1949 1926
1950 return sc.nr_reclaimed >= nr_pages; 1927 return sc.nr_reclaimed >= nr_pages;
1951} 1928}
1929
1930int zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order)
1931{
1932 cpumask_t mask;
1933 int node_id;
1934
1935 /*
1936 * Do not reclaim if there was a recent unsuccessful attempt at zone
1937 * reclaim. In that case we let allocations go off node for the
1938 * zone_reclaim_interval. Otherwise we would scan for each off-node
1939 * page allocation.
1940 */
1941 if (time_before(jiffies,
1942 zone->last_unsuccessful_zone_reclaim + zone_reclaim_interval))
1943 return 0;
1944
1945 /*
1946 * Avoid concurrent zone reclaims, do not reclaim in a zone that does
1947 * not have reclaimable pages and if we should not delay the allocation
1948 * then do not scan.
1949 */
1950 if (!(gfp_mask & __GFP_WAIT) ||
1951 zone->all_unreclaimable ||
1952 atomic_read(&zone->reclaim_in_progress) > 0 ||
1953 (current->flags & PF_MEMALLOC))
1954 return 0;
1955
1956 /*
1957 * Only run zone reclaim on the local zone or on zones that do not
1958 * have associated processors. This will favor the local processor
1959 * over remote processors and spread off node memory allocations
1960 * as wide as possible.
1961 */
1962 node_id = zone->zone_pgdat->node_id;
1963 mask = node_to_cpumask(node_id);
1964 if (!cpus_empty(mask) && node_id != numa_node_id())
1965 return 0;
1966 return __zone_reclaim(zone, gfp_mask, order);
1967}
1952#endif 1968#endif
1953 1969