summaryrefslogtreecommitdiffstats
path: root/mm/page-writeback.c
diff options
context:
space:
mode:
Diffstat (limited to 'mm/page-writeback.c')
-rw-r--r--mm/page-writeback.c36
1 files changed, 18 insertions, 18 deletions
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 0b9c5cbe8eba..c518c845f202 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -1972,31 +1972,31 @@ bool wb_over_bg_thresh(struct bdi_writeback *wb)
1972int dirty_writeback_centisecs_handler(struct ctl_table *table, int write, 1972int dirty_writeback_centisecs_handler(struct ctl_table *table, int write,
1973 void __user *buffer, size_t *length, loff_t *ppos) 1973 void __user *buffer, size_t *length, loff_t *ppos)
1974{ 1974{
1975 proc_dointvec(table, write, buffer, length, ppos); 1975 unsigned int old_interval = dirty_writeback_interval;
1976 return 0; 1976 int ret;
1977
1978 ret = proc_dointvec(table, write, buffer, length, ppos);
1979
1980 /*
1981 * Writing 0 to dirty_writeback_interval will disable periodic writeback
1982 * and a different non-zero value will wakeup the writeback threads.
1983 * wb_wakeup_delayed() would be more appropriate, but it's a pain to
1984 * iterate over all bdis and wbs.
1985 * The reason we do this is to make the change take effect immediately.
1986 */
1987 if (!ret && write && dirty_writeback_interval &&
1988 dirty_writeback_interval != old_interval)
1989 wakeup_flusher_threads(WB_REASON_PERIODIC);
1990
1991 return ret;
1977} 1992}
1978 1993
1979#ifdef CONFIG_BLOCK 1994#ifdef CONFIG_BLOCK
1980void laptop_mode_timer_fn(unsigned long data) 1995void laptop_mode_timer_fn(unsigned long data)
1981{ 1996{
1982 struct request_queue *q = (struct request_queue *)data; 1997 struct request_queue *q = (struct request_queue *)data;
1983 int nr_pages = global_node_page_state(NR_FILE_DIRTY) +
1984 global_node_page_state(NR_UNSTABLE_NFS);
1985 struct bdi_writeback *wb;
1986 1998
1987 /* 1999 wakeup_flusher_threads_bdi(q->backing_dev_info, WB_REASON_LAPTOP_TIMER);
1988 * We want to write everything out, not just down to the dirty
1989 * threshold
1990 */
1991 if (!bdi_has_dirty_io(q->backing_dev_info))
1992 return;
1993
1994 rcu_read_lock();
1995 list_for_each_entry_rcu(wb, &q->backing_dev_info->wb_list, bdi_node)
1996 if (wb_has_dirty_io(wb))
1997 wb_start_writeback(wb, nr_pages, true,
1998 WB_REASON_LAPTOP_TIMER);
1999 rcu_read_unlock();
2000} 2000}
2001 2001
2002/* 2002/*