diff options
-rw-r--r-- | fs/fs-writeback.c | 61 |
1 files changed, 38 insertions, 23 deletions
diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c index c9d105ff7970..1f22fb5217c0 100644 --- a/fs/fs-writeback.c +++ b/fs/fs-writeback.c | |||
@@ -119,7 +119,7 @@ void __mark_inode_dirty(struct inode *inode, int flags) | |||
119 | goto out; | 119 | goto out; |
120 | 120 | ||
121 | /* | 121 | /* |
122 | * If the inode was already on s_dirty or s_io, don't | 122 | * If the inode was already on s_dirty/s_io/s_more_io, don't |
123 | * reposition it (that would break s_dirty time-ordering). | 123 | * reposition it (that would break s_dirty time-ordering). |
124 | */ | 124 | */ |
125 | if (!was_dirty) { | 125 | if (!was_dirty) { |
@@ -173,6 +173,33 @@ static void requeue_io(struct inode *inode) | |||
173 | } | 173 | } |
174 | 174 | ||
175 | /* | 175 | /* |
176 | * Move expired dirty inodes from @delaying_queue to @dispatch_queue. | ||
177 | */ | ||
178 | static void move_expired_inodes(struct list_head *delaying_queue, | ||
179 | struct list_head *dispatch_queue, | ||
180 | unsigned long *older_than_this) | ||
181 | { | ||
182 | while (!list_empty(delaying_queue)) { | ||
183 | struct inode *inode = list_entry(delaying_queue->prev, | ||
184 | struct inode, i_list); | ||
185 | if (older_than_this && | ||
186 | time_after(inode->dirtied_when, *older_than_this)) | ||
187 | break; | ||
188 | list_move(&inode->i_list, dispatch_queue); | ||
189 | } | ||
190 | } | ||
191 | |||
192 | /* | ||
193 | * Queue all expired dirty inodes for io, eldest first. | ||
194 | */ | ||
195 | static void queue_io(struct super_block *sb, | ||
196 | unsigned long *older_than_this) | ||
197 | { | ||
198 | list_splice_init(&sb->s_more_io, sb->s_io.prev); | ||
199 | move_expired_inodes(&sb->s_dirty, &sb->s_io, older_than_this); | ||
200 | } | ||
201 | |||
202 | /* | ||
176 | * Write a single inode's dirty pages and inode data out to disk. | 203 | * Write a single inode's dirty pages and inode data out to disk. |
177 | * If `wait' is set, wait on the writeout. | 204 | * If `wait' is set, wait on the writeout. |
178 | * | 205 | * |
@@ -222,7 +249,7 @@ __sync_single_inode(struct inode *inode, struct writeback_control *wbc) | |||
222 | /* | 249 | /* |
223 | * We didn't write back all the pages. nfs_writepages() | 250 | * We didn't write back all the pages. nfs_writepages() |
224 | * sometimes bales out without doing anything. Redirty | 251 | * sometimes bales out without doing anything. Redirty |
225 | * the inode. It is moved from s_io onto s_dirty. | 252 | * the inode; Move it from s_io onto s_more_io/s_dirty. |
226 | */ | 253 | */ |
227 | /* | 254 | /* |
228 | * akpm: if the caller was the kupdate function we put | 255 | * akpm: if the caller was the kupdate function we put |
@@ -235,10 +262,9 @@ __sync_single_inode(struct inode *inode, struct writeback_control *wbc) | |||
235 | */ | 262 | */ |
236 | if (wbc->for_kupdate) { | 263 | if (wbc->for_kupdate) { |
237 | /* | 264 | /* |
238 | * For the kupdate function we leave the inode | 265 | * For the kupdate function we move the inode |
239 | * at the head of sb_dirty so it will get more | 266 | * to s_more_io so it will get more writeout as |
240 | * writeout as soon as the queue becomes | 267 | * soon as the queue becomes uncongested. |
241 | * uncongested. | ||
242 | */ | 268 | */ |
243 | inode->i_state |= I_DIRTY_PAGES; | 269 | inode->i_state |= I_DIRTY_PAGES; |
244 | requeue_io(inode); | 270 | requeue_io(inode); |
@@ -296,10 +322,10 @@ __writeback_single_inode(struct inode *inode, struct writeback_control *wbc) | |||
296 | 322 | ||
297 | /* | 323 | /* |
298 | * We're skipping this inode because it's locked, and we're not | 324 | * We're skipping this inode because it's locked, and we're not |
299 | * doing writeback-for-data-integrity. Move it to the head of | 325 | * doing writeback-for-data-integrity. Move it to s_more_io so |
300 | * s_dirty so that writeback can proceed with the other inodes | 326 | * that writeback can proceed with the other inodes on s_io. |
301 | * on s_io. We'll have another go at writing back this inode | 327 | * We'll have another go at writing back this inode when we |
302 | * when the s_dirty iodes get moved back onto s_io. | 328 | * completed a full scan of s_io. |
303 | */ | 329 | */ |
304 | requeue_io(inode); | 330 | requeue_io(inode); |
305 | 331 | ||
@@ -366,7 +392,7 @@ sync_sb_inodes(struct super_block *sb, struct writeback_control *wbc) | |||
366 | const unsigned long start = jiffies; /* livelock avoidance */ | 392 | const unsigned long start = jiffies; /* livelock avoidance */ |
367 | 393 | ||
368 | if (!wbc->for_kupdate || list_empty(&sb->s_io)) | 394 | if (!wbc->for_kupdate || list_empty(&sb->s_io)) |
369 | list_splice_init(&sb->s_dirty, &sb->s_io); | 395 | queue_io(sb, wbc->older_than_this); |
370 | 396 | ||
371 | while (!list_empty(&sb->s_io)) { | 397 | while (!list_empty(&sb->s_io)) { |
372 | struct inode *inode = list_entry(sb->s_io.prev, | 398 | struct inode *inode = list_entry(sb->s_io.prev, |
@@ -411,13 +437,6 @@ sync_sb_inodes(struct super_block *sb, struct writeback_control *wbc) | |||
411 | if (time_after(inode->dirtied_when, start)) | 437 | if (time_after(inode->dirtied_when, start)) |
412 | break; | 438 | break; |
413 | 439 | ||
414 | /* Was this inode dirtied too recently? */ | ||
415 | if (wbc->older_than_this && time_after(inode->dirtied_when, | ||
416 | *wbc->older_than_this)) { | ||
417 | list_splice_init(&sb->s_io, sb->s_dirty.prev); | ||
418 | break; | ||
419 | } | ||
420 | |||
421 | /* Is another pdflush already flushing this queue? */ | 440 | /* Is another pdflush already flushing this queue? */ |
422 | if (current_is_pdflush() && !writeback_acquire(bdi)) | 441 | if (current_is_pdflush() && !writeback_acquire(bdi)) |
423 | break; | 442 | break; |
@@ -446,10 +465,6 @@ sync_sb_inodes(struct super_block *sb, struct writeback_control *wbc) | |||
446 | if (wbc->nr_to_write <= 0) | 465 | if (wbc->nr_to_write <= 0) |
447 | break; | 466 | break; |
448 | } | 467 | } |
449 | |||
450 | if (list_empty(&sb->s_io)) | ||
451 | list_splice_init(&sb->s_more_io, &sb->s_io); | ||
452 | |||
453 | return; /* Leave any unwritten inodes on s_io */ | 468 | return; /* Leave any unwritten inodes on s_io */ |
454 | } | 469 | } |
455 | 470 | ||
@@ -459,7 +474,7 @@ sync_sb_inodes(struct super_block *sb, struct writeback_control *wbc) | |||
459 | * Note: | 474 | * Note: |
460 | * We don't need to grab a reference to superblock here. If it has non-empty | 475 | * We don't need to grab a reference to superblock here. If it has non-empty |
461 | * ->s_dirty it's hadn't been killed yet and kill_super() won't proceed | 476 | * ->s_dirty it's hadn't been killed yet and kill_super() won't proceed |
462 | * past sync_inodes_sb() until both the ->s_dirty and ->s_io lists are | 477 | * past sync_inodes_sb() until the ->s_dirty/s_io/s_more_io lists are all |
463 | * empty. Since __sync_single_inode() regains inode_lock before it finally moves | 478 | * empty. Since __sync_single_inode() regains inode_lock before it finally moves |
464 | * inode from superblock lists we are OK. | 479 | * inode from superblock lists we are OK. |
465 | * | 480 | * |