diff options
-rw-r--r-- | include/linux/workqueue.h | 21 | ||||
-rw-r--r-- | kernel/workqueue.c | 355 |
2 files changed, 322 insertions, 54 deletions
diff --git a/include/linux/workqueue.h b/include/linux/workqueue.h index b90958a037dc..8762f62103d8 100644 --- a/include/linux/workqueue.h +++ b/include/linux/workqueue.h | |||
@@ -26,11 +26,13 @@ enum { | |||
26 | WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */ | 26 | WORK_STRUCT_PENDING_BIT = 0, /* work item is pending execution */ |
27 | #ifdef CONFIG_DEBUG_OBJECTS_WORK | 27 | #ifdef CONFIG_DEBUG_OBJECTS_WORK |
28 | WORK_STRUCT_STATIC_BIT = 1, /* static initializer (debugobjects) */ | 28 | WORK_STRUCT_STATIC_BIT = 1, /* static initializer (debugobjects) */ |
29 | WORK_STRUCT_FLAG_BITS = 2, | 29 | WORK_STRUCT_COLOR_SHIFT = 3, /* color for workqueue flushing */ |
30 | #else | 30 | #else |
31 | WORK_STRUCT_FLAG_BITS = 1, | 31 | WORK_STRUCT_COLOR_SHIFT = 2, /* color for workqueue flushing */ |
32 | #endif | 32 | #endif |
33 | 33 | ||
34 | WORK_STRUCT_COLOR_BITS = 4, | ||
35 | |||
34 | WORK_STRUCT_PENDING = 1 << WORK_STRUCT_PENDING_BIT, | 36 | WORK_STRUCT_PENDING = 1 << WORK_STRUCT_PENDING_BIT, |
35 | #ifdef CONFIG_DEBUG_OBJECTS_WORK | 37 | #ifdef CONFIG_DEBUG_OBJECTS_WORK |
36 | WORK_STRUCT_STATIC = 1 << WORK_STRUCT_STATIC_BIT, | 38 | WORK_STRUCT_STATIC = 1 << WORK_STRUCT_STATIC_BIT, |
@@ -38,6 +40,21 @@ enum { | |||
38 | WORK_STRUCT_STATIC = 0, | 40 | WORK_STRUCT_STATIC = 0, |
39 | #endif | 41 | #endif |
40 | 42 | ||
43 | /* | ||
44 | * The last color is no color used for works which don't | ||
45 | * participate in workqueue flushing. | ||
46 | */ | ||
47 | WORK_NR_COLORS = (1 << WORK_STRUCT_COLOR_BITS) - 1, | ||
48 | WORK_NO_COLOR = WORK_NR_COLORS, | ||
49 | |||
50 | /* | ||
51 | * Reserve 6 bits off of cwq pointer w/ debugobjects turned | ||
52 | * off. This makes cwqs aligned to 64 bytes which isn't too | ||
53 | * excessive while allowing 15 workqueue flush colors. | ||
54 | */ | ||
55 | WORK_STRUCT_FLAG_BITS = WORK_STRUCT_COLOR_SHIFT + | ||
56 | WORK_STRUCT_COLOR_BITS, | ||
57 | |||
41 | WORK_STRUCT_FLAG_MASK = (1UL << WORK_STRUCT_FLAG_BITS) - 1, | 58 | WORK_STRUCT_FLAG_MASK = (1UL << WORK_STRUCT_FLAG_BITS) - 1, |
42 | WORK_STRUCT_WQ_DATA_MASK = ~WORK_STRUCT_FLAG_MASK, | 59 | WORK_STRUCT_WQ_DATA_MASK = ~WORK_STRUCT_FLAG_MASK, |
43 | }; | 60 | }; |
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 74a38499b19a..56e47c59d73b 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
@@ -41,6 +41,8 @@ | |||
41 | * | 41 | * |
42 | * L: cwq->lock protected. Access with cwq->lock held. | 42 | * L: cwq->lock protected. Access with cwq->lock held. |
43 | * | 43 | * |
44 | * F: wq->flush_mutex protected. | ||
45 | * | ||
44 | * W: workqueue_lock protected. | 46 | * W: workqueue_lock protected. |
45 | */ | 47 | */ |
46 | 48 | ||
@@ -60,10 +62,23 @@ struct cpu_workqueue_struct { | |||
60 | unsigned int cpu; | 62 | unsigned int cpu; |
61 | 63 | ||
62 | struct workqueue_struct *wq; /* I: the owning workqueue */ | 64 | struct workqueue_struct *wq; /* I: the owning workqueue */ |
65 | int work_color; /* L: current color */ | ||
66 | int flush_color; /* L: flushing color */ | ||
67 | int nr_in_flight[WORK_NR_COLORS]; | ||
68 | /* L: nr of in_flight works */ | ||
63 | struct task_struct *thread; | 69 | struct task_struct *thread; |
64 | }; | 70 | }; |
65 | 71 | ||
66 | /* | 72 | /* |
73 | * Structure used to wait for workqueue flush. | ||
74 | */ | ||
75 | struct wq_flusher { | ||
76 | struct list_head list; /* F: list of flushers */ | ||
77 | int flush_color; /* F: flush color waiting for */ | ||
78 | struct completion done; /* flush completion */ | ||
79 | }; | ||
80 | |||
81 | /* | ||
67 | * The externally visible workqueue abstraction is an array of | 82 | * The externally visible workqueue abstraction is an array of |
68 | * per-CPU workqueues: | 83 | * per-CPU workqueues: |
69 | */ | 84 | */ |
@@ -71,6 +86,15 @@ struct workqueue_struct { | |||
71 | unsigned int flags; /* I: WQ_* flags */ | 86 | unsigned int flags; /* I: WQ_* flags */ |
72 | struct cpu_workqueue_struct *cpu_wq; /* I: cwq's */ | 87 | struct cpu_workqueue_struct *cpu_wq; /* I: cwq's */ |
73 | struct list_head list; /* W: list of all workqueues */ | 88 | struct list_head list; /* W: list of all workqueues */ |
89 | |||
90 | struct mutex flush_mutex; /* protects wq flushing */ | ||
91 | int work_color; /* F: current work color */ | ||
92 | int flush_color; /* F: current flush color */ | ||
93 | atomic_t nr_cwqs_to_flush; /* flush in progress */ | ||
94 | struct wq_flusher *first_flusher; /* F: first flusher */ | ||
95 | struct list_head flusher_queue; /* F: flush waiters */ | ||
96 | struct list_head flusher_overflow; /* F: flush overflow list */ | ||
97 | |||
74 | const char *name; /* I: workqueue name */ | 98 | const char *name; /* I: workqueue name */ |
75 | #ifdef CONFIG_LOCKDEP | 99 | #ifdef CONFIG_LOCKDEP |
76 | struct lockdep_map lockdep_map; | 100 | struct lockdep_map lockdep_map; |
@@ -207,6 +231,22 @@ static struct cpu_workqueue_struct *target_cwq(unsigned int cpu, | |||
207 | return get_cwq(cpu, wq); | 231 | return get_cwq(cpu, wq); |
208 | } | 232 | } |
209 | 233 | ||
234 | static unsigned int work_color_to_flags(int color) | ||
235 | { | ||
236 | return color << WORK_STRUCT_COLOR_SHIFT; | ||
237 | } | ||
238 | |||
239 | static int get_work_color(struct work_struct *work) | ||
240 | { | ||
241 | return (*work_data_bits(work) >> WORK_STRUCT_COLOR_SHIFT) & | ||
242 | ((1 << WORK_STRUCT_COLOR_BITS) - 1); | ||
243 | } | ||
244 | |||
245 | static int work_next_color(int color) | ||
246 | { | ||
247 | return (color + 1) % WORK_NR_COLORS; | ||
248 | } | ||
249 | |||
210 | /* | 250 | /* |
211 | * Set the workqueue on which a work item is to be run | 251 | * Set the workqueue on which a work item is to be run |
212 | * - Must *only* be called if the pending flag is set | 252 | * - Must *only* be called if the pending flag is set |
@@ -273,7 +313,9 @@ static void __queue_work(unsigned int cpu, struct workqueue_struct *wq, | |||
273 | debug_work_activate(work); | 313 | debug_work_activate(work); |
274 | spin_lock_irqsave(&cwq->lock, flags); | 314 | spin_lock_irqsave(&cwq->lock, flags); |
275 | BUG_ON(!list_empty(&work->entry)); | 315 | BUG_ON(!list_empty(&work->entry)); |
276 | insert_work(cwq, work, &cwq->worklist, 0); | 316 | cwq->nr_in_flight[cwq->work_color]++; |
317 | insert_work(cwq, work, &cwq->worklist, | ||
318 | work_color_to_flags(cwq->work_color)); | ||
277 | spin_unlock_irqrestore(&cwq->lock, flags); | 319 | spin_unlock_irqrestore(&cwq->lock, flags); |
278 | } | 320 | } |
279 | 321 | ||
@@ -387,6 +429,44 @@ int queue_delayed_work_on(int cpu, struct workqueue_struct *wq, | |||
387 | EXPORT_SYMBOL_GPL(queue_delayed_work_on); | 429 | EXPORT_SYMBOL_GPL(queue_delayed_work_on); |
388 | 430 | ||
389 | /** | 431 | /** |
432 | * cwq_dec_nr_in_flight - decrement cwq's nr_in_flight | ||
433 | * @cwq: cwq of interest | ||
434 | * @color: color of work which left the queue | ||
435 | * | ||
436 | * A work either has completed or is removed from pending queue, | ||
437 | * decrement nr_in_flight of its cwq and handle workqueue flushing. | ||
438 | * | ||
439 | * CONTEXT: | ||
440 | * spin_lock_irq(cwq->lock). | ||
441 | */ | ||
442 | static void cwq_dec_nr_in_flight(struct cpu_workqueue_struct *cwq, int color) | ||
443 | { | ||
444 | /* ignore uncolored works */ | ||
445 | if (color == WORK_NO_COLOR) | ||
446 | return; | ||
447 | |||
448 | cwq->nr_in_flight[color]--; | ||
449 | |||
450 | /* is flush in progress and are we at the flushing tip? */ | ||
451 | if (likely(cwq->flush_color != color)) | ||
452 | return; | ||
453 | |||
454 | /* are there still in-flight works? */ | ||
455 | if (cwq->nr_in_flight[color]) | ||
456 | return; | ||
457 | |||
458 | /* this cwq is done, clear flush_color */ | ||
459 | cwq->flush_color = -1; | ||
460 | |||
461 | /* | ||
462 | * If this was the last cwq, wake up the first flusher. It | ||
463 | * will handle the rest. | ||
464 | */ | ||
465 | if (atomic_dec_and_test(&cwq->wq->nr_cwqs_to_flush)) | ||
466 | complete(&cwq->wq->first_flusher->done); | ||
467 | } | ||
468 | |||
469 | /** | ||
390 | * process_one_work - process single work | 470 | * process_one_work - process single work |
391 | * @cwq: cwq to process work for | 471 | * @cwq: cwq to process work for |
392 | * @work: work to process | 472 | * @work: work to process |
@@ -404,6 +484,7 @@ static void process_one_work(struct cpu_workqueue_struct *cwq, | |||
404 | struct work_struct *work) | 484 | struct work_struct *work) |
405 | { | 485 | { |
406 | work_func_t f = work->func; | 486 | work_func_t f = work->func; |
487 | int work_color; | ||
407 | #ifdef CONFIG_LOCKDEP | 488 | #ifdef CONFIG_LOCKDEP |
408 | /* | 489 | /* |
409 | * It is permissible to free the struct work_struct from | 490 | * It is permissible to free the struct work_struct from |
@@ -417,6 +498,7 @@ static void process_one_work(struct cpu_workqueue_struct *cwq, | |||
417 | /* claim and process */ | 498 | /* claim and process */ |
418 | debug_work_deactivate(work); | 499 | debug_work_deactivate(work); |
419 | cwq->current_work = work; | 500 | cwq->current_work = work; |
501 | work_color = get_work_color(work); | ||
420 | list_del_init(&work->entry); | 502 | list_del_init(&work->entry); |
421 | 503 | ||
422 | spin_unlock_irq(&cwq->lock); | 504 | spin_unlock_irq(&cwq->lock); |
@@ -443,6 +525,7 @@ static void process_one_work(struct cpu_workqueue_struct *cwq, | |||
443 | 525 | ||
444 | /* we're done with it, release */ | 526 | /* we're done with it, release */ |
445 | cwq->current_work = NULL; | 527 | cwq->current_work = NULL; |
528 | cwq_dec_nr_in_flight(cwq, work_color); | ||
446 | } | 529 | } |
447 | 530 | ||
448 | static void run_workqueue(struct cpu_workqueue_struct *cwq) | 531 | static void run_workqueue(struct cpu_workqueue_struct *cwq) |
@@ -529,29 +612,78 @@ static void insert_wq_barrier(struct cpu_workqueue_struct *cwq, | |||
529 | init_completion(&barr->done); | 612 | init_completion(&barr->done); |
530 | 613 | ||
531 | debug_work_activate(&barr->work); | 614 | debug_work_activate(&barr->work); |
532 | insert_work(cwq, &barr->work, head, 0); | 615 | insert_work(cwq, &barr->work, head, work_color_to_flags(WORK_NO_COLOR)); |
533 | } | 616 | } |
534 | 617 | ||
535 | static int flush_cpu_workqueue(struct cpu_workqueue_struct *cwq) | 618 | /** |
619 | * flush_workqueue_prep_cwqs - prepare cwqs for workqueue flushing | ||
620 | * @wq: workqueue being flushed | ||
621 | * @flush_color: new flush color, < 0 for no-op | ||
622 | * @work_color: new work color, < 0 for no-op | ||
623 | * | ||
624 | * Prepare cwqs for workqueue flushing. | ||
625 | * | ||
626 | * If @flush_color is non-negative, flush_color on all cwqs should be | ||
627 | * -1. If no cwq has in-flight commands at the specified color, all | ||
628 | * cwq->flush_color's stay at -1 and %false is returned. If any cwq | ||
629 | * has in flight commands, its cwq->flush_color is set to | ||
630 | * @flush_color, @wq->nr_cwqs_to_flush is updated accordingly, cwq | ||
631 | * wakeup logic is armed and %true is returned. | ||
632 | * | ||
633 | * The caller should have initialized @wq->first_flusher prior to | ||
634 | * calling this function with non-negative @flush_color. If | ||
635 | * @flush_color is negative, no flush color update is done and %false | ||
636 | * is returned. | ||
637 | * | ||
638 | * If @work_color is non-negative, all cwqs should have the same | ||
639 | * work_color which is previous to @work_color and all will be | ||
640 | * advanced to @work_color. | ||
641 | * | ||
642 | * CONTEXT: | ||
643 | * mutex_lock(wq->flush_mutex). | ||
644 | * | ||
645 | * RETURNS: | ||
646 | * %true if @flush_color >= 0 and there's something to flush. %false | ||
647 | * otherwise. | ||
648 | */ | ||
649 | static bool flush_workqueue_prep_cwqs(struct workqueue_struct *wq, | ||
650 | int flush_color, int work_color) | ||
536 | { | 651 | { |
537 | int active = 0; | 652 | bool wait = false; |
538 | struct wq_barrier barr; | 653 | unsigned int cpu; |
539 | 654 | ||
540 | WARN_ON(cwq->thread == current); | 655 | if (flush_color >= 0) { |
541 | 656 | BUG_ON(atomic_read(&wq->nr_cwqs_to_flush)); | |
542 | spin_lock_irq(&cwq->lock); | 657 | atomic_set(&wq->nr_cwqs_to_flush, 1); |
543 | if (!list_empty(&cwq->worklist) || cwq->current_work != NULL) { | ||
544 | insert_wq_barrier(cwq, &barr, &cwq->worklist); | ||
545 | active = 1; | ||
546 | } | 658 | } |
547 | spin_unlock_irq(&cwq->lock); | ||
548 | 659 | ||
549 | if (active) { | 660 | for_each_possible_cpu(cpu) { |
550 | wait_for_completion(&barr.done); | 661 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); |
551 | destroy_work_on_stack(&barr.work); | 662 | |
663 | spin_lock_irq(&cwq->lock); | ||
664 | |||
665 | if (flush_color >= 0) { | ||
666 | BUG_ON(cwq->flush_color != -1); | ||
667 | |||
668 | if (cwq->nr_in_flight[flush_color]) { | ||
669 | cwq->flush_color = flush_color; | ||
670 | atomic_inc(&wq->nr_cwqs_to_flush); | ||
671 | wait = true; | ||
672 | } | ||
673 | } | ||
674 | |||
675 | if (work_color >= 0) { | ||
676 | BUG_ON(work_color != work_next_color(cwq->work_color)); | ||
677 | cwq->work_color = work_color; | ||
678 | } | ||
679 | |||
680 | spin_unlock_irq(&cwq->lock); | ||
552 | } | 681 | } |
553 | 682 | ||
554 | return active; | 683 | if (flush_color >= 0 && atomic_dec_and_test(&wq->nr_cwqs_to_flush)) |
684 | complete(&wq->first_flusher->done); | ||
685 | |||
686 | return wait; | ||
555 | } | 687 | } |
556 | 688 | ||
557 | /** | 689 | /** |
@@ -566,13 +698,143 @@ static int flush_cpu_workqueue(struct cpu_workqueue_struct *cwq) | |||
566 | */ | 698 | */ |
567 | void flush_workqueue(struct workqueue_struct *wq) | 699 | void flush_workqueue(struct workqueue_struct *wq) |
568 | { | 700 | { |
569 | int cpu; | 701 | struct wq_flusher this_flusher = { |
702 | .list = LIST_HEAD_INIT(this_flusher.list), | ||
703 | .flush_color = -1, | ||
704 | .done = COMPLETION_INITIALIZER_ONSTACK(this_flusher.done), | ||
705 | }; | ||
706 | int next_color; | ||
570 | 707 | ||
571 | might_sleep(); | ||
572 | lock_map_acquire(&wq->lockdep_map); | 708 | lock_map_acquire(&wq->lockdep_map); |
573 | lock_map_release(&wq->lockdep_map); | 709 | lock_map_release(&wq->lockdep_map); |
574 | for_each_possible_cpu(cpu) | 710 | |
575 | flush_cpu_workqueue(get_cwq(cpu, wq)); | 711 | mutex_lock(&wq->flush_mutex); |
712 | |||
713 | /* | ||
714 | * Start-to-wait phase | ||
715 | */ | ||
716 | next_color = work_next_color(wq->work_color); | ||
717 | |||
718 | if (next_color != wq->flush_color) { | ||
719 | /* | ||
720 | * Color space is not full. The current work_color | ||
721 | * becomes our flush_color and work_color is advanced | ||
722 | * by one. | ||
723 | */ | ||
724 | BUG_ON(!list_empty(&wq->flusher_overflow)); | ||
725 | this_flusher.flush_color = wq->work_color; | ||
726 | wq->work_color = next_color; | ||
727 | |||
728 | if (!wq->first_flusher) { | ||
729 | /* no flush in progress, become the first flusher */ | ||
730 | BUG_ON(wq->flush_color != this_flusher.flush_color); | ||
731 | |||
732 | wq->first_flusher = &this_flusher; | ||
733 | |||
734 | if (!flush_workqueue_prep_cwqs(wq, wq->flush_color, | ||
735 | wq->work_color)) { | ||
736 | /* nothing to flush, done */ | ||
737 | wq->flush_color = next_color; | ||
738 | wq->first_flusher = NULL; | ||
739 | goto out_unlock; | ||
740 | } | ||
741 | } else { | ||
742 | /* wait in queue */ | ||
743 | BUG_ON(wq->flush_color == this_flusher.flush_color); | ||
744 | list_add_tail(&this_flusher.list, &wq->flusher_queue); | ||
745 | flush_workqueue_prep_cwqs(wq, -1, wq->work_color); | ||
746 | } | ||
747 | } else { | ||
748 | /* | ||
749 | * Oops, color space is full, wait on overflow queue. | ||
750 | * The next flush completion will assign us | ||
751 | * flush_color and transfer to flusher_queue. | ||
752 | */ | ||
753 | list_add_tail(&this_flusher.list, &wq->flusher_overflow); | ||
754 | } | ||
755 | |||
756 | mutex_unlock(&wq->flush_mutex); | ||
757 | |||
758 | wait_for_completion(&this_flusher.done); | ||
759 | |||
760 | /* | ||
761 | * Wake-up-and-cascade phase | ||
762 | * | ||
763 | * First flushers are responsible for cascading flushes and | ||
764 | * handling overflow. Non-first flushers can simply return. | ||
765 | */ | ||
766 | if (wq->first_flusher != &this_flusher) | ||
767 | return; | ||
768 | |||
769 | mutex_lock(&wq->flush_mutex); | ||
770 | |||
771 | wq->first_flusher = NULL; | ||
772 | |||
773 | BUG_ON(!list_empty(&this_flusher.list)); | ||
774 | BUG_ON(wq->flush_color != this_flusher.flush_color); | ||
775 | |||
776 | while (true) { | ||
777 | struct wq_flusher *next, *tmp; | ||
778 | |||
779 | /* complete all the flushers sharing the current flush color */ | ||
780 | list_for_each_entry_safe(next, tmp, &wq->flusher_queue, list) { | ||
781 | if (next->flush_color != wq->flush_color) | ||
782 | break; | ||
783 | list_del_init(&next->list); | ||
784 | complete(&next->done); | ||
785 | } | ||
786 | |||
787 | BUG_ON(!list_empty(&wq->flusher_overflow) && | ||
788 | wq->flush_color != work_next_color(wq->work_color)); | ||
789 | |||
790 | /* this flush_color is finished, advance by one */ | ||
791 | wq->flush_color = work_next_color(wq->flush_color); | ||
792 | |||
793 | /* one color has been freed, handle overflow queue */ | ||
794 | if (!list_empty(&wq->flusher_overflow)) { | ||
795 | /* | ||
796 | * Assign the same color to all overflowed | ||
797 | * flushers, advance work_color and append to | ||
798 | * flusher_queue. This is the start-to-wait | ||
799 | * phase for these overflowed flushers. | ||
800 | */ | ||
801 | list_for_each_entry(tmp, &wq->flusher_overflow, list) | ||
802 | tmp->flush_color = wq->work_color; | ||
803 | |||
804 | wq->work_color = work_next_color(wq->work_color); | ||
805 | |||
806 | list_splice_tail_init(&wq->flusher_overflow, | ||
807 | &wq->flusher_queue); | ||
808 | flush_workqueue_prep_cwqs(wq, -1, wq->work_color); | ||
809 | } | ||
810 | |||
811 | if (list_empty(&wq->flusher_queue)) { | ||
812 | BUG_ON(wq->flush_color != wq->work_color); | ||
813 | break; | ||
814 | } | ||
815 | |||
816 | /* | ||
817 | * Need to flush more colors. Make the next flusher | ||
818 | * the new first flusher and arm cwqs. | ||
819 | */ | ||
820 | BUG_ON(wq->flush_color == wq->work_color); | ||
821 | BUG_ON(wq->flush_color != next->flush_color); | ||
822 | |||
823 | list_del_init(&next->list); | ||
824 | wq->first_flusher = next; | ||
825 | |||
826 | if (flush_workqueue_prep_cwqs(wq, wq->flush_color, -1)) | ||
827 | break; | ||
828 | |||
829 | /* | ||
830 | * Meh... this color is already done, clear first | ||
831 | * flusher and repeat cascading. | ||
832 | */ | ||
833 | wq->first_flusher = NULL; | ||
834 | } | ||
835 | |||
836 | out_unlock: | ||
837 | mutex_unlock(&wq->flush_mutex); | ||
576 | } | 838 | } |
577 | EXPORT_SYMBOL_GPL(flush_workqueue); | 839 | EXPORT_SYMBOL_GPL(flush_workqueue); |
578 | 840 | ||
@@ -659,6 +921,7 @@ static int try_to_grab_pending(struct work_struct *work) | |||
659 | if (cwq == get_wq_data(work)) { | 921 | if (cwq == get_wq_data(work)) { |
660 | debug_work_deactivate(work); | 922 | debug_work_deactivate(work); |
661 | list_del_init(&work->entry); | 923 | list_del_init(&work->entry); |
924 | cwq_dec_nr_in_flight(cwq, get_work_color(work)); | ||
662 | ret = 1; | 925 | ret = 1; |
663 | } | 926 | } |
664 | } | 927 | } |
@@ -1066,6 +1329,10 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
1066 | goto err; | 1329 | goto err; |
1067 | 1330 | ||
1068 | wq->flags = flags; | 1331 | wq->flags = flags; |
1332 | mutex_init(&wq->flush_mutex); | ||
1333 | atomic_set(&wq->nr_cwqs_to_flush, 0); | ||
1334 | INIT_LIST_HEAD(&wq->flusher_queue); | ||
1335 | INIT_LIST_HEAD(&wq->flusher_overflow); | ||
1069 | wq->name = name; | 1336 | wq->name = name; |
1070 | lockdep_init_map(&wq->lockdep_map, lock_name, key, 0); | 1337 | lockdep_init_map(&wq->lockdep_map, lock_name, key, 0); |
1071 | INIT_LIST_HEAD(&wq->list); | 1338 | INIT_LIST_HEAD(&wq->list); |
@@ -1083,6 +1350,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
1083 | BUG_ON((unsigned long)cwq & WORK_STRUCT_FLAG_MASK); | 1350 | BUG_ON((unsigned long)cwq & WORK_STRUCT_FLAG_MASK); |
1084 | cwq->wq = wq; | 1351 | cwq->wq = wq; |
1085 | cwq->cpu = cpu; | 1352 | cwq->cpu = cpu; |
1353 | cwq->flush_color = -1; | ||
1086 | spin_lock_init(&cwq->lock); | 1354 | spin_lock_init(&cwq->lock); |
1087 | INIT_LIST_HEAD(&cwq->worklist); | 1355 | INIT_LIST_HEAD(&cwq->worklist); |
1088 | init_waitqueue_head(&cwq->more_work); | 1356 | init_waitqueue_head(&cwq->more_work); |
@@ -1116,33 +1384,6 @@ err: | |||
1116 | } | 1384 | } |
1117 | EXPORT_SYMBOL_GPL(__create_workqueue_key); | 1385 | EXPORT_SYMBOL_GPL(__create_workqueue_key); |
1118 | 1386 | ||
1119 | static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq) | ||
1120 | { | ||
1121 | /* | ||
1122 | * Our caller is either destroy_workqueue() or CPU_POST_DEAD, | ||
1123 | * cpu_add_remove_lock protects cwq->thread. | ||
1124 | */ | ||
1125 | if (cwq->thread == NULL) | ||
1126 | return; | ||
1127 | |||
1128 | lock_map_acquire(&cwq->wq->lockdep_map); | ||
1129 | lock_map_release(&cwq->wq->lockdep_map); | ||
1130 | |||
1131 | flush_cpu_workqueue(cwq); | ||
1132 | /* | ||
1133 | * If the caller is CPU_POST_DEAD and cwq->worklist was not empty, | ||
1134 | * a concurrent flush_workqueue() can insert a barrier after us. | ||
1135 | * However, in that case run_workqueue() won't return and check | ||
1136 | * kthread_should_stop() until it flushes all work_struct's. | ||
1137 | * When ->worklist becomes empty it is safe to exit because no | ||
1138 | * more work_structs can be queued on this cwq: flush_workqueue | ||
1139 | * checks list_empty(), and a "normal" queue_work() can't use | ||
1140 | * a dead CPU. | ||
1141 | */ | ||
1142 | kthread_stop(cwq->thread); | ||
1143 | cwq->thread = NULL; | ||
1144 | } | ||
1145 | |||
1146 | /** | 1387 | /** |
1147 | * destroy_workqueue - safely terminate a workqueue | 1388 | * destroy_workqueue - safely terminate a workqueue |
1148 | * @wq: target workqueue | 1389 | * @wq: target workqueue |
@@ -1159,8 +1400,20 @@ void destroy_workqueue(struct workqueue_struct *wq) | |||
1159 | spin_unlock(&workqueue_lock); | 1400 | spin_unlock(&workqueue_lock); |
1160 | cpu_maps_update_done(); | 1401 | cpu_maps_update_done(); |
1161 | 1402 | ||
1162 | for_each_possible_cpu(cpu) | 1403 | flush_workqueue(wq); |
1163 | cleanup_workqueue_thread(get_cwq(cpu, wq)); | 1404 | |
1405 | for_each_possible_cpu(cpu) { | ||
1406 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); | ||
1407 | int i; | ||
1408 | |||
1409 | if (cwq->thread) { | ||
1410 | kthread_stop(cwq->thread); | ||
1411 | cwq->thread = NULL; | ||
1412 | } | ||
1413 | |||
1414 | for (i = 0; i < WORK_NR_COLORS; i++) | ||
1415 | BUG_ON(cwq->nr_in_flight[i]); | ||
1416 | } | ||
1164 | 1417 | ||
1165 | free_cwqs(wq->cpu_wq); | 1418 | free_cwqs(wq->cpu_wq); |
1166 | kfree(wq); | 1419 | kfree(wq); |
@@ -1185,9 +1438,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb, | |||
1185 | 1438 | ||
1186 | switch (action) { | 1439 | switch (action) { |
1187 | case CPU_POST_DEAD: | 1440 | case CPU_POST_DEAD: |
1188 | lock_map_acquire(&cwq->wq->lockdep_map); | 1441 | flush_workqueue(wq); |
1189 | lock_map_release(&cwq->wq->lockdep_map); | ||
1190 | flush_cpu_workqueue(cwq); | ||
1191 | break; | 1442 | break; |
1192 | } | 1443 | } |
1193 | } | 1444 | } |