diff options
-rw-r--r-- | fs/fs-writeback.c | 2 | ||||
-rw-r--r-- | include/linux/backing-dev.h | 2 | ||||
-rw-r--r-- | mm/backing-dev.c | 26 |
3 files changed, 15 insertions, 15 deletions
diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c index 261570deb22c..002be0ff2ab3 100644 --- a/fs/fs-writeback.c +++ b/fs/fs-writeback.c | |||
@@ -840,7 +840,7 @@ int bdi_writeback_thread(void *data) | |||
840 | 840 | ||
841 | /* | 841 | /* |
842 | * Longest period of inactivity that we tolerate. If we | 842 | * Longest period of inactivity that we tolerate. If we |
843 | * see dirty data again later, the task will get | 843 | * see dirty data again later, the thread will get |
844 | * recreated automatically. | 844 | * recreated automatically. |
845 | */ | 845 | */ |
846 | max_idle = max(5UL * 60 * HZ, wait_jiffies); | 846 | max_idle = max(5UL * 60 * HZ, wait_jiffies); |
diff --git a/include/linux/backing-dev.h b/include/linux/backing-dev.h index e536f3a74e60..f0936f5f85dd 100644 --- a/include/linux/backing-dev.h +++ b/include/linux/backing-dev.h | |||
@@ -50,7 +50,7 @@ struct bdi_writeback { | |||
50 | 50 | ||
51 | unsigned long last_old_flush; /* last old data flush */ | 51 | unsigned long last_old_flush; /* last old data flush */ |
52 | 52 | ||
53 | struct task_struct *task; /* writeback task */ | 53 | struct task_struct *task; /* writeback thread */ |
54 | struct list_head b_dirty; /* dirty inodes */ | 54 | struct list_head b_dirty; /* dirty inodes */ |
55 | struct list_head b_io; /* parked for writeback */ | 55 | struct list_head b_io; /* parked for writeback */ |
56 | struct list_head b_more_io; /* parked for more writeback */ | 56 | struct list_head b_more_io; /* parked for more writeback */ |
diff --git a/mm/backing-dev.c b/mm/backing-dev.c index ac78a3336181..4e9ed2a8521f 100644 --- a/mm/backing-dev.c +++ b/mm/backing-dev.c | |||
@@ -50,7 +50,7 @@ static struct timer_list sync_supers_timer; | |||
50 | static int bdi_sync_supers(void *); | 50 | static int bdi_sync_supers(void *); |
51 | static void sync_supers_timer_fn(unsigned long); | 51 | static void sync_supers_timer_fn(unsigned long); |
52 | 52 | ||
53 | static void bdi_add_default_flusher_task(struct backing_dev_info *bdi); | 53 | static void bdi_add_default_flusher_thread(struct backing_dev_info *bdi); |
54 | 54 | ||
55 | #ifdef CONFIG_DEBUG_FS | 55 | #ifdef CONFIG_DEBUG_FS |
56 | #include <linux/debugfs.h> | 56 | #include <linux/debugfs.h> |
@@ -279,10 +279,10 @@ static void bdi_flush_io(struct backing_dev_info *bdi) | |||
279 | } | 279 | } |
280 | 280 | ||
281 | /* | 281 | /* |
282 | * kupdated() used to do this. We cannot do it from the bdi_forker_task() | 282 | * kupdated() used to do this. We cannot do it from the bdi_forker_thread() |
283 | * or we risk deadlocking on ->s_umount. The longer term solution would be | 283 | * or we risk deadlocking on ->s_umount. The longer term solution would be |
284 | * to implement sync_supers_bdi() or similar and simply do it from the | 284 | * to implement sync_supers_bdi() or similar and simply do it from the |
285 | * bdi writeback tasks individually. | 285 | * bdi writeback thread individually. |
286 | */ | 286 | */ |
287 | static int bdi_sync_supers(void *unused) | 287 | static int bdi_sync_supers(void *unused) |
288 | { | 288 | { |
@@ -318,7 +318,7 @@ static void sync_supers_timer_fn(unsigned long unused) | |||
318 | bdi_arm_supers_timer(); | 318 | bdi_arm_supers_timer(); |
319 | } | 319 | } |
320 | 320 | ||
321 | static int bdi_forker_task(void *ptr) | 321 | static int bdi_forker_thread(void *ptr) |
322 | { | 322 | { |
323 | struct bdi_writeback *me = ptr; | 323 | struct bdi_writeback *me = ptr; |
324 | 324 | ||
@@ -354,7 +354,7 @@ static int bdi_forker_task(void *ptr) | |||
354 | !bdi_has_dirty_io(bdi)) | 354 | !bdi_has_dirty_io(bdi)) |
355 | continue; | 355 | continue; |
356 | 356 | ||
357 | bdi_add_default_flusher_task(bdi); | 357 | bdi_add_default_flusher_thread(bdi); |
358 | } | 358 | } |
359 | 359 | ||
360 | set_current_state(TASK_INTERRUPTIBLE); | 360 | set_current_state(TASK_INTERRUPTIBLE); |
@@ -376,7 +376,7 @@ static int bdi_forker_task(void *ptr) | |||
376 | 376 | ||
377 | /* | 377 | /* |
378 | * This is our real job - check for pending entries in | 378 | * This is our real job - check for pending entries in |
379 | * bdi_pending_list, and create the tasks that got added | 379 | * bdi_pending_list, and create the threads that got added |
380 | */ | 380 | */ |
381 | bdi = list_entry(bdi_pending_list.next, struct backing_dev_info, | 381 | bdi = list_entry(bdi_pending_list.next, struct backing_dev_info, |
382 | bdi_list); | 382 | bdi_list); |
@@ -387,7 +387,7 @@ static int bdi_forker_task(void *ptr) | |||
387 | wb->task = kthread_run(bdi_writeback_thread, wb, "flush-%s", | 387 | wb->task = kthread_run(bdi_writeback_thread, wb, "flush-%s", |
388 | dev_name(bdi->dev)); | 388 | dev_name(bdi->dev)); |
389 | /* | 389 | /* |
390 | * If task creation fails, then readd the bdi to | 390 | * If thread creation fails, then readd the bdi to |
391 | * the pending list and force writeout of the bdi | 391 | * the pending list and force writeout of the bdi |
392 | * from this forker thread. That will free some memory | 392 | * from this forker thread. That will free some memory |
393 | * and we can try again. | 393 | * and we can try again. |
@@ -430,10 +430,10 @@ static void bdi_add_to_pending(struct rcu_head *head) | |||
430 | } | 430 | } |
431 | 431 | ||
432 | /* | 432 | /* |
433 | * Add the default flusher task that gets created for any bdi | 433 | * Add the default flusher thread that gets created for any bdi |
434 | * that has dirty data pending writeout | 434 | * that has dirty data pending writeout |
435 | */ | 435 | */ |
436 | void static bdi_add_default_flusher_task(struct backing_dev_info *bdi) | 436 | static void bdi_add_default_flusher_thread(struct backing_dev_info *bdi) |
437 | { | 437 | { |
438 | if (!bdi_cap_writeback_dirty(bdi)) | 438 | if (!bdi_cap_writeback_dirty(bdi)) |
439 | return; | 439 | return; |
@@ -445,10 +445,10 @@ void static bdi_add_default_flusher_task(struct backing_dev_info *bdi) | |||
445 | } | 445 | } |
446 | 446 | ||
447 | /* | 447 | /* |
448 | * Check with the helper whether to proceed adding a task. Will only | 448 | * Check with the helper whether to proceed adding a thread. Will only |
449 | * abort if we two or more simultanous calls to | 449 | * abort if we two or more simultanous calls to |
450 | * bdi_add_default_flusher_task() occured, further additions will block | 450 | * bdi_add_default_flusher_thread() occured, further additions will |
451 | * waiting for previous additions to finish. | 451 | * block waiting for previous additions to finish. |
452 | */ | 452 | */ |
453 | if (!test_and_set_bit(BDI_pending, &bdi->state)) { | 453 | if (!test_and_set_bit(BDI_pending, &bdi->state)) { |
454 | list_del_rcu(&bdi->bdi_list); | 454 | list_del_rcu(&bdi->bdi_list); |
@@ -506,7 +506,7 @@ int bdi_register(struct backing_dev_info *bdi, struct device *parent, | |||
506 | if (bdi_cap_flush_forker(bdi)) { | 506 | if (bdi_cap_flush_forker(bdi)) { |
507 | struct bdi_writeback *wb = &bdi->wb; | 507 | struct bdi_writeback *wb = &bdi->wb; |
508 | 508 | ||
509 | wb->task = kthread_run(bdi_forker_task, wb, "bdi-%s", | 509 | wb->task = kthread_run(bdi_forker_thread, wb, "bdi-%s", |
510 | dev_name(dev)); | 510 | dev_name(dev)); |
511 | if (IS_ERR(wb->task)) { | 511 | if (IS_ERR(wb->task)) { |
512 | wb->task = NULL; | 512 | wb->task = NULL; |