aboutsummaryrefslogtreecommitdiffstats
path: root/mm
diff options
context:
space:
mode:
authorJens Axboe <jens.axboe@oracle.com>2009-09-23 13:37:09 -0400
committerJens Axboe <jens.axboe@oracle.com>2009-09-25 12:08:25 -0400
commit5b0830cb9085f4b69f9d57d7f3aaff322ffbec26 (patch)
tree10040eb359269d4cd05487790b758144a69e8e39 /mm
parent71fd05a887e0f3f6bfff76ff81b33776177d0606 (diff)
writeback: get rid to incorrect references to pdflush in comments
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
Diffstat (limited to 'mm')
-rw-r--r--mm/page-writeback.c8
-rw-r--r--mm/shmem.c5
-rw-r--r--mm/vmscan.c8
3 files changed, 11 insertions, 10 deletions
diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 3c78fc316202..8bef063125b1 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -58,7 +58,7 @@ static inline long sync_writeback_pages(unsigned long dirtied)
58/* The following parameters are exported via /proc/sys/vm */ 58/* The following parameters are exported via /proc/sys/vm */
59 59
60/* 60/*
61 * Start background writeback (via pdflush) at this percentage 61 * Start background writeback (via writeback threads) at this percentage
62 */ 62 */
63int dirty_background_ratio = 10; 63int dirty_background_ratio = 10;
64 64
@@ -477,8 +477,8 @@ get_dirty_limits(unsigned long *pbackground, unsigned long *pdirty,
477 * balance_dirty_pages() must be called by processes which are generating dirty 477 * balance_dirty_pages() must be called by processes which are generating dirty
478 * data. It looks at the number of dirty pages in the machine and will force 478 * data. It looks at the number of dirty pages in the machine and will force
479 * the caller to perform writeback if the system is over `vm_dirty_ratio'. 479 * the caller to perform writeback if the system is over `vm_dirty_ratio'.
480 * If we're over `background_thresh' then pdflush is woken to perform some 480 * If we're over `background_thresh' then the writeback threads are woken to
481 * writeout. 481 * perform some writeout.
482 */ 482 */
483static void balance_dirty_pages(struct address_space *mapping, 483static void balance_dirty_pages(struct address_space *mapping,
484 unsigned long write_chunk) 484 unsigned long write_chunk)
@@ -582,7 +582,7 @@ static void balance_dirty_pages(struct address_space *mapping,
582 bdi->dirty_exceeded = 0; 582 bdi->dirty_exceeded = 0;
583 583
584 if (writeback_in_progress(bdi)) 584 if (writeback_in_progress(bdi))
585 return; /* pdflush is already working this queue */ 585 return;
586 586
587 /* 587 /*
588 * In laptop mode, we wait until hitting the higher threshold before 588 * In laptop mode, we wait until hitting the higher threshold before
diff --git a/mm/shmem.c b/mm/shmem.c
index b206a7a32e2a..aa9481166aae 100644
--- a/mm/shmem.c
+++ b/mm/shmem.c
@@ -1046,8 +1046,9 @@ static int shmem_writepage(struct page *page, struct writeback_control *wbc)
1046 * sync from ever calling shmem_writepage; but a stacking filesystem 1046 * sync from ever calling shmem_writepage; but a stacking filesystem
1047 * may use the ->writepage of its underlying filesystem, in which case 1047 * may use the ->writepage of its underlying filesystem, in which case
1048 * tmpfs should write out to swap only in response to memory pressure, 1048 * tmpfs should write out to swap only in response to memory pressure,
1049 * and not for pdflush or sync. However, in those cases, we do still 1049 * and not for the writeback threads or sync. However, in those cases,
1050 * want to check if there's a redundant swappage to be discarded. 1050 * we do still want to check if there's a redundant swappage to be
1051 * discarded.
1051 */ 1052 */
1052 if (wbc->for_reclaim) 1053 if (wbc->for_reclaim)
1053 swap = get_swap_page(); 1054 swap = get_swap_page();
diff --git a/mm/vmscan.c b/mm/vmscan.c
index 613e89f471d9..359c3c57ef85 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1709,10 +1709,10 @@ static void shrink_zones(int priority, struct zonelist *zonelist,
1709 * 1709 *
1710 * If the caller is !__GFP_FS then the probability of a failure is reasonably 1710 * If the caller is !__GFP_FS then the probability of a failure is reasonably
1711 * high - the zone may be full of dirty or under-writeback pages, which this 1711 * high - the zone may be full of dirty or under-writeback pages, which this
1712 * caller can't do much about. We kick pdflush and take explicit naps in the 1712 * caller can't do much about. We kick the writeback threads and take explicit
1713 * hope that some of these pages can be written. But if the allocating task 1713 * naps in the hope that some of these pages can be written. But if the
1714 * holds filesystem locks which prevent writeout this might not work, and the 1714 * allocating task holds filesystem locks which prevent writeout this might not
1715 * allocation attempt will fail. 1715 * work, and the allocation attempt will fail.
1716 * 1716 *
1717 * returns: 0, if no pages reclaimed 1717 * returns: 0, if no pages reclaimed
1718 * else, the number of pages reclaimed 1718 * else, the number of pages reclaimed