diff options
author | Mel Gorman <mgorman@techsingularity.net> | 2016-07-28 18:47:31 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2016-07-28 19:07:41 -0400 |
commit | 5a1c84b404a7176b8b36e2a0041b6f0adb3151a3 (patch) | |
tree | ff98e242c5d4d3a24ca49f6ddc707028aeb938f9 /mm/page_alloc.c | |
parent | bb4cc2bea6df7854d629bff114ca03237cc718d6 (diff) |
mm: remove reclaim and compaction retry approximations
If per-zone LRU accounting is available then there is no point
approximating whether reclaim and compaction should retry based on pgdat
statistics. This is effectively a revert of "mm, vmstat: remove zone
and node double accounting by approximating retries" with the difference
that inactive/active stats are still available. This preserves the
history of why the approximation was retried and why it had to be
reverted to handle OOM kills on 32-bit systems.
Link: http://lkml.kernel.org/r/1469110261-7365-4-git-send-email-mgorman@techsingularity.net
Signed-off-by: Mel Gorman <mgorman@techsingularity.net>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Minchan Kim <minchan@kernel.org>
Cc: Michal Hocko <mhocko@suse.cz>
Cc: Vlastimil Babka <vbabka@suse.cz>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'mm/page_alloc.c')
-rw-r--r-- | mm/page_alloc.c | 49 |
1 files changed, 10 insertions, 39 deletions
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 759cfa8cbbeb..dfdb608f7b3d 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
@@ -3402,7 +3402,6 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order, | |||
3402 | { | 3402 | { |
3403 | struct zone *zone; | 3403 | struct zone *zone; |
3404 | struct zoneref *z; | 3404 | struct zoneref *z; |
3405 | pg_data_t *current_pgdat = NULL; | ||
3406 | 3405 | ||
3407 | /* | 3406 | /* |
3408 | * Make sure we converge to OOM if we cannot make any progress | 3407 | * Make sure we converge to OOM if we cannot make any progress |
@@ -3412,15 +3411,6 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order, | |||
3412 | return false; | 3411 | return false; |
3413 | 3412 | ||
3414 | /* | 3413 | /* |
3415 | * Blindly retry lowmem allocation requests that are often ignored by | ||
3416 | * the OOM killer up to MAX_RECLAIM_RETRIES as we not have a reliable | ||
3417 | * and fast means of calculating reclaimable, dirty and writeback pages | ||
3418 | * in eligible zones. | ||
3419 | */ | ||
3420 | if (ac->high_zoneidx < ZONE_NORMAL) | ||
3421 | goto out; | ||
3422 | |||
3423 | /* | ||
3424 | * Keep reclaiming pages while there is a chance this will lead | 3414 | * Keep reclaiming pages while there is a chance this will lead |
3425 | * somewhere. If none of the target zones can satisfy our allocation | 3415 | * somewhere. If none of the target zones can satisfy our allocation |
3426 | * request even if all reclaimable pages are considered then we are | 3416 | * request even if all reclaimable pages are considered then we are |
@@ -3430,38 +3420,18 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order, | |||
3430 | ac->nodemask) { | 3420 | ac->nodemask) { |
3431 | unsigned long available; | 3421 | unsigned long available; |
3432 | unsigned long reclaimable; | 3422 | unsigned long reclaimable; |
3433 | int zid; | ||
3434 | 3423 | ||
3435 | if (current_pgdat == zone->zone_pgdat) | 3424 | available = reclaimable = zone_reclaimable_pages(zone); |
3436 | continue; | ||
3437 | |||
3438 | current_pgdat = zone->zone_pgdat; | ||
3439 | available = reclaimable = pgdat_reclaimable_pages(current_pgdat); | ||
3440 | available -= DIV_ROUND_UP(no_progress_loops * available, | 3425 | available -= DIV_ROUND_UP(no_progress_loops * available, |
3441 | MAX_RECLAIM_RETRIES); | 3426 | MAX_RECLAIM_RETRIES); |
3442 | 3427 | available += zone_page_state_snapshot(zone, NR_FREE_PAGES); | |
3443 | /* Account for all free pages on eligible zones */ | ||
3444 | for (zid = 0; zid <= zone_idx(zone); zid++) { | ||
3445 | struct zone *acct_zone = ¤t_pgdat->node_zones[zid]; | ||
3446 | |||
3447 | available += zone_page_state_snapshot(acct_zone, NR_FREE_PAGES); | ||
3448 | } | ||
3449 | 3428 | ||
3450 | /* | 3429 | /* |
3451 | * Would the allocation succeed if we reclaimed the whole | 3430 | * Would the allocation succeed if we reclaimed the whole |
3452 | * available? This is approximate because there is no | 3431 | * available? |
3453 | * accurate count of reclaimable pages per zone. | ||
3454 | */ | 3432 | */ |
3455 | for (zid = 0; zid <= zone_idx(zone); zid++) { | 3433 | if (__zone_watermark_ok(zone, order, min_wmark_pages(zone), |
3456 | struct zone *check_zone = ¤t_pgdat->node_zones[zid]; | 3434 | ac_classzone_idx(ac), alloc_flags, available)) { |
3457 | unsigned long estimate; | ||
3458 | |||
3459 | estimate = min(check_zone->managed_pages, available); | ||
3460 | if (!__zone_watermark_ok(check_zone, order, | ||
3461 | min_wmark_pages(check_zone), ac_classzone_idx(ac), | ||
3462 | alloc_flags, estimate)) | ||
3463 | continue; | ||
3464 | |||
3465 | /* | 3435 | /* |
3466 | * If we didn't make any progress and have a lot of | 3436 | * If we didn't make any progress and have a lot of |
3467 | * dirty + writeback pages then we should wait for | 3437 | * dirty + writeback pages then we should wait for |
@@ -3471,16 +3441,15 @@ should_reclaim_retry(gfp_t gfp_mask, unsigned order, | |||
3471 | if (!did_some_progress) { | 3441 | if (!did_some_progress) { |
3472 | unsigned long write_pending; | 3442 | unsigned long write_pending; |
3473 | 3443 | ||
3474 | write_pending = | 3444 | write_pending = zone_page_state_snapshot(zone, |
3475 | node_page_state(current_pgdat, NR_WRITEBACK) + | 3445 | NR_ZONE_WRITE_PENDING); |
3476 | node_page_state(current_pgdat, NR_FILE_DIRTY); | ||
3477 | 3446 | ||
3478 | if (2 * write_pending > reclaimable) { | 3447 | if (2 * write_pending > reclaimable) { |
3479 | congestion_wait(BLK_RW_ASYNC, HZ/10); | 3448 | congestion_wait(BLK_RW_ASYNC, HZ/10); |
3480 | return true; | 3449 | return true; |
3481 | } | 3450 | } |
3482 | } | 3451 | } |
3483 | out: | 3452 | |
3484 | /* | 3453 | /* |
3485 | * Memory allocation/reclaim might be called from a WQ | 3454 | * Memory allocation/reclaim might be called from a WQ |
3486 | * context and the current implementation of the WQ | 3455 | * context and the current implementation of the WQ |
@@ -4361,6 +4330,7 @@ void show_free_areas(unsigned int filter) | |||
4361 | " active_file:%lukB" | 4330 | " active_file:%lukB" |
4362 | " inactive_file:%lukB" | 4331 | " inactive_file:%lukB" |
4363 | " unevictable:%lukB" | 4332 | " unevictable:%lukB" |
4333 | " writepending:%lukB" | ||
4364 | " present:%lukB" | 4334 | " present:%lukB" |
4365 | " managed:%lukB" | 4335 | " managed:%lukB" |
4366 | " mlocked:%lukB" | 4336 | " mlocked:%lukB" |
@@ -4383,6 +4353,7 @@ void show_free_areas(unsigned int filter) | |||
4383 | K(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)), | 4353 | K(zone_page_state(zone, NR_ZONE_ACTIVE_FILE)), |
4384 | K(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)), | 4354 | K(zone_page_state(zone, NR_ZONE_INACTIVE_FILE)), |
4385 | K(zone_page_state(zone, NR_ZONE_UNEVICTABLE)), | 4355 | K(zone_page_state(zone, NR_ZONE_UNEVICTABLE)), |
4356 | K(zone_page_state(zone, NR_ZONE_WRITE_PENDING)), | ||
4386 | K(zone->present_pages), | 4357 | K(zone->present_pages), |
4387 | K(zone->managed_pages), | 4358 | K(zone->managed_pages), |
4388 | K(zone_page_state(zone, NR_MLOCK)), | 4359 | K(zone_page_state(zone, NR_MLOCK)), |