aboutsummaryrefslogtreecommitdiffstats
path: root/mm/page-writeback.c
diff options
context:
space:
mode:
authorWu Fengguang <fengguang.wu@intel.com>2011-06-20 00:18:42 -0400
committerWu Fengguang <fengguang.wu@intel.com>2011-07-10 01:09:02 -0400
commitffd1f609ab10532e8137b4b981fdf903ef4d0b32 (patch)
treeb691e22952a8d12782cc73ec21e7aa450859cf4f /mm/page-writeback.c
parentc42843f2f0bbc9d716a32caf667d18fc2bf3bc4c (diff)
writeback: introduce max-pause and pass-good dirty limits
The max-pause limit helps to keep the sleep time inside balance_dirty_pages() within MAX_PAUSE=200ms. The 200ms max sleep means per task rate limit of 8pages/200ms=160KB/s when dirty exceeded, which normally is enough to stop dirtiers from continue pushing the dirty pages high, unless there are a sufficient large number of slow dirtiers (eg. 500 tasks doing 160KB/s will still sum up to 80MB/s, exceeding the write bandwidth of a slow disk and hence accumulating more and more dirty pages). The pass-good limit helps to let go of the good bdi's in the presence of a blocked bdi (ie. NFS server not responding) or slow USB disk which for some reason build up a large number of initial dirty pages that refuse to go away anytime soon. For example, given two bdi's A and B and the initial state bdi_thresh_A = dirty_thresh / 2 bdi_thresh_B = dirty_thresh / 2 bdi_dirty_A = dirty_thresh / 2 bdi_dirty_B = dirty_thresh / 2 Then A get blocked, after a dozen seconds bdi_thresh_A = 0 bdi_thresh_B = dirty_thresh bdi_dirty_A = dirty_thresh / 2 bdi_dirty_B = dirty_thresh / 2 The (bdi_dirty_B < bdi_thresh_B) test is now useless and the dirty pages will be effectively throttled by condition (nr_dirty < dirty_thresh). This has two problems: (1) we lose the protections for light dirtiers (2) balance_dirty_pages() effectively becomes IO-less because the (bdi_nr_reclaimable > bdi_thresh) test won't be true. This is good for IO, but balance_dirty_pages() loses an important way to break out of the loop which leads to more spread out throttle delays. DIRTY_PASSGOOD_AREA can eliminate the above issues. The only problem is, DIRTY_PASSGOOD_AREA needs to be defined as 2 to fully cover the above example while this patch uses the more conservative value 8 so as not to surprise people with too many dirty pages than expected. The max-pause limit won't noticeably impact the speed dirty pages are knocked down when there is a sudden drop of global/bdi dirty thresholds. Because the heavy dirties will be throttled below 160KB/s which is slow enough. It does help to avoid long dirty throttle delays and especially will make light dirtiers more responsive. Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
Diffstat (limited to 'mm/page-writeback.c')
-rw-r--r--mm/page-writeback.c33
1 files changed, 33 insertions, 0 deletions
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index da959952b9f5..798842a22474 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -37,6 +37,11 @@
37#include <trace/events/writeback.h> 37#include <trace/events/writeback.h>
38 38
39/* 39/*
40 * Sleep at most 200ms at a time in balance_dirty_pages().
41 */
42#define MAX_PAUSE max(HZ/5, 1)
43
44/*
40 * Estimate write bandwidth at 200ms intervals. 45 * Estimate write bandwidth at 200ms intervals.
41 */ 46 */
42#define BANDWIDTH_INTERVAL max(HZ/5, 1) 47#define BANDWIDTH_INTERVAL max(HZ/5, 1)
@@ -399,6 +404,11 @@ unsigned long determine_dirtyable_memory(void)
399 return x + 1; /* Ensure that we never return 0 */ 404 return x + 1; /* Ensure that we never return 0 */
400} 405}
401 406
407static unsigned long hard_dirty_limit(unsigned long thresh)
408{
409 return max(thresh, global_dirty_limit);
410}
411
402/* 412/*
403 * global_dirty_limits - background-writeback and dirty-throttling thresholds 413 * global_dirty_limits - background-writeback and dirty-throttling thresholds
404 * 414 *
@@ -723,6 +733,29 @@ static void balance_dirty_pages(struct address_space *mapping,
723 io_schedule_timeout(pause); 733 io_schedule_timeout(pause);
724 trace_balance_dirty_wait(bdi); 734 trace_balance_dirty_wait(bdi);
725 735
736 dirty_thresh = hard_dirty_limit(dirty_thresh);
737 /*
738 * max-pause area. If dirty exceeded but still within this
739 * area, no need to sleep for more than 200ms: (a) 8 pages per
740 * 200ms is typically more than enough to curb heavy dirtiers;
741 * (b) the pause time limit makes the dirtiers more responsive.
742 */
743 if (nr_dirty < dirty_thresh +
744 dirty_thresh / DIRTY_MAXPAUSE_AREA &&
745 time_after(jiffies, start_time + MAX_PAUSE))
746 break;
747 /*
748 * pass-good area. When some bdi gets blocked (eg. NFS server
749 * not responding), or write bandwidth dropped dramatically due
750 * to concurrent reads, or dirty threshold suddenly dropped and
751 * the dirty pages cannot be brought down anytime soon (eg. on
752 * slow USB stick), at least let go of the good bdi's.
753 */
754 if (nr_dirty < dirty_thresh +
755 dirty_thresh / DIRTY_PASSGOOD_AREA &&
756 bdi_dirty < bdi_thresh)
757 break;
758
726 /* 759 /*
727 * Increase the delay for each loop, up to our previous 760 * Increase the delay for each loop, up to our previous
728 * default of taking a 100ms nap. 761 * default of taking a 100ms nap.