summaryrefslogtreecommitdiffstats
path: root/mm/page_alloc.c
diff options
context:
space:
mode:
authorMel Gorman <mgorman@techsingularity.net>2016-07-28 18:47:31 -0400
committerLinus Torvalds <torvalds@linux-foundation.org>2016-07-28 19:07:41 -0400
commit5a1c84b404a7176b8b36e2a0041b6f0adb3151a3 (patch)
treeff98e242c5d4d3a24ca49f6ddc707028aeb938f9 /mm/page_alloc.c
parentbb4cc2bea6df7854d629bff114ca03237cc718d6 (diff)
mm: remove reclaim and compaction retry approximations
If per-zone LRU accounting is available then there is no point approximating whether reclaim and compaction should retry based on pgdat statistics. This is effectively a revert of "mm, vmstat: remove zone and node double accounting by approximating retries" with the difference that inactive/active stats are still available. This preserves the history of why the approximation was retried and why it had to be reverted to handle OOM kills on 32-bit systems. Link: http://lkml.kernel.org/r/1469110261-7365-4-git-send-email-mgorman@techsingularity.net Signed-off-by: Mel Gorman <mgorman@techsingularity.net> Acked-by: Johannes Weiner <hannes@cmpxchg.org> Acked-by: Minchan Kim <minchan@kernel.org> Cc: Michal Hocko <mhocko@suse.cz> Cc: Vlastimil Babka <vbabka@suse.cz> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/page_alloc.c')
-rw-r--r--mm/page_alloc.c49
1 files changed, 10 insertions, 39 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 759cfa8cbbeb..dfdb608f7b3d 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -3402,7 +3402,6 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
3402{ 3402{
3403 struct zone *zone; 3403 struct zone *zone;
3404 struct zoneref *z; 3404 struct zoneref *z;
3405 pg_data_t *current_pgdat = NULL;
3406 3405
3407 /* 3406 /*
3408 * Make sure we converge to OOM if we cannot make any progress 3407 * Make sure we converge to OOM if we cannot make any progress
@@ -3412,15 +3411,6 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
3412 return false; 3411 return false;
3413 3412
3414 /* 3413 /*
3415 * Blindly retry lowmem allocation requests that are often ignored by
3416 * the OOM killer up to MAX_RECLAIM_RETRIES as we not have a reliable
3417 * and fast means of calculating reclaimable, dirty and writeback pages
3418 * in eligible zones.
3419 */
3420 if (ac->high_zoneidx < ZONE_NORMAL)
3421 goto out;
3422
3423 /*
3424 * Keep reclaiming pages while there is a chance this will lead 3414 * Keep reclaiming pages while there is a chance this will lead
3425 * somewhere. If none of the target zones can satisfy our allocation 3415 * somewhere. If none of the target zones can satisfy our allocation
3426 * request even if all reclaimable pages are considered then we are 3416 * request even if all reclaimable pages are considered then we are
@@ -3430,38 +3420,18 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
3430 ac->nodemask) { 3420 ac->nodemask) {
3431 unsigned long available; 3421 unsigned long available;
3432 unsigned long reclaimable; 3422 unsigned long reclaimable;
3433 int zid;
3434 3423
3435 if (current_pgdat == zone->zone_pgdat) 3424 available = reclaimable = zone_reclaimable_pages(zone);
3436 continue;
3437
3438 current_pgdat = zone->zone_pgdat;
3439 available = reclaimable = pgdat_reclaimable_pages(current_pgdat);
3440 available -= DIV_ROUND_UP(no_progress_loops * available, 3425 available -= DIV_ROUND_UP(no_progress_loops * available,
3441 MAX_RECLAIM_RETRIES); 3426 MAX_RECLAIM_RETRIES);
3442 3427 available += zone_page_state_snapshot(zone, NR_FREE_PAGES);
3443 /* Account for all free pages on eligible zones */
3444 for (zid = 0; zid <= zone_idx(zone); zid++) {
3445 struct zone *acct_zone = &current_pgdat->node_zones[zid];
3446
3447 available += zone_page_state_snapshot(acct_zone, NR_FREE_PAGES);
3448 }
3449 3428
3450 /* 3429 /*
3451 * Would the allocation succeed if we reclaimed the whole 3430 * Would the allocation succeed if we reclaimed the whole
3452 * available? This is approximate because there is no 3431 * available?
3453 * accurate count of reclaimable pages per zone.
3454 */ 3432 */
3455 for (zid = 0; zid <= zone_idx(zone); zid++) { 3433 if (__zone_watermark_ok(zone, order, min_wmark_pages(zone),
3456 struct zone *check_zone = &current_pgdat->node_zones[zid]; 3434 ac_classzone_idx(ac), alloc_flags, available)) {
3457 unsigned long estimate;
3458
3459 estimate = min(check_zone->managed_pages, available);
3460 if (!__zone_watermark_ok(check_zone, order,
3461 min_wmark_pages(check_zone), ac_classzone_idx(ac),
3462 alloc_flags, estimate))
3463 continue;
3464
3465 /* 3435 /*
3466 * If we didn't make any progress and have a lot of 3436 * If we didn't make any progress and have a lot of
3467 * dirty + writeback pages then we should wait for 3437 * dirty + writeback pages then we should wait for
@@ -3471,16 +3441,15 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order,
3471 if (!did_some_progress) { 3441 if (!did_some_progress) {
3472 unsigned long write_pending; 3442 unsigned long write_pending;
3473 3443
3474 write_pending = 3444 write_pending = zone_page_state_snapshot(zone,
3475 node_page_state(current_pgdat, NR_WRITEBACK) + 3445 NR_ZONE_WRITE_PENDING);
3476 node_page_state(current_pgdat, NR_FILE_DIRTY);
3477 3446
3478 if (2 * write_pending > reclaimable) { 3447 if (2 * write_pending > reclaimable) {
3479 congestion_wait(BLK_RW_ASYNC, HZ/10); 3448 congestion_wait(BLK_RW_ASYNC, HZ/10);
3480 return true; 3449 return true;
3481 } 3450 }
3482 } 3451 }
3483out: 3452
3484 /* 3453 /*
3485 * Memory allocation/reclaim might be called from a WQ 3454 * Memory allocation/reclaim might be called from a WQ
3486 * context and the current implementation of the WQ 3455 * context and the current implementation of the WQ
@@ -4361,6 +4330,7 @@ void show_free_areas(unsigned int filter)
4361 " active_file:%lukB" 4330 " active_file:%lukB"
4362 " inactive_file:%lukB" 4331 " inactive_file:%lukB"
4363 " unevictable:%lukB" 4332 " unevictable:%lukB"
4333 " writepending:%lukB"
4364 " present:%lukB" 4334 " present:%lukB"
4365 " managed:%lukB" 4335 " managed:%lukB"
4366 " mlocked:%lukB" 4336 " mlocked:%lukB"
@@ -4383,6 +4353,7 @@ void show_free_areas(unsigned int filter)
4383 K(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)), 4353 K(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)),
4384 K(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)), 4354 K(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)),
4385 K(zone_page_state(zone, NR_ZONE_UNEVICTABLE)), 4355 K(zone_page_state(zone, NR_ZONE_UNEVICTABLE)),
4356 K(zone_page_state(zone, NR_ZONE_WRITE_PENDING)),
4386 K(zone->present_pages), 4357 K(zone->present_pages),
4387 K(zone->managed_pages), 4358 K(zone->managed_pages),
4388 K(zone_page_state(zone, NR_MLOCK)), 4359 K(zone_page_state(zone, NR_MLOCK)),