diff options
Diffstat (limited to 'kernel/workqueue.c')
| -rw-r--r-- | kernel/workqueue.c | 45 |
1 files changed, 40 insertions, 5 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 5bfb213984b2..327d2deb4451 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
| @@ -229,6 +229,16 @@ static inline void set_wq_data(struct work_struct *work, | |||
| 229 | atomic_long_set(&work->data, new); | 229 | atomic_long_set(&work->data, new); |
| 230 | } | 230 | } |
| 231 | 231 | ||
| 232 | /* | ||
| 233 | * Clear WORK_STRUCT_PENDING and the workqueue on which it was queued. | ||
| 234 | */ | ||
| 235 | static inline void clear_wq_data(struct work_struct *work) | ||
| 236 | { | ||
| 237 | unsigned long flags = *work_data_bits(work) & | ||
| 238 | (1UL << WORK_STRUCT_STATIC); | ||
| 239 | atomic_long_set(&work->data, flags); | ||
| 240 | } | ||
| 241 | |||
| 232 | static inline | 242 | static inline |
| 233 | struct cpu_workqueue_struct *get_wq_data(struct work_struct *work) | 243 | struct cpu_workqueue_struct *get_wq_data(struct work_struct *work) |
| 234 | { | 244 | { |
| @@ -671,7 +681,7 @@ static int __cancel_work_timer(struct work_struct *work, | |||
| 671 | wait_on_work(work); | 681 | wait_on_work(work); |
| 672 | } while (unlikely(ret < 0)); | 682 | } while (unlikely(ret < 0)); |
| 673 | 683 | ||
| 674 | work_clear_pending(work); | 684 | clear_wq_data(work); |
| 675 | return ret; | 685 | return ret; |
| 676 | } | 686 | } |
| 677 | 687 | ||
| @@ -845,6 +855,30 @@ int schedule_on_each_cpu(work_func_t func) | |||
| 845 | return 0; | 855 | return 0; |
| 846 | } | 856 | } |
| 847 | 857 | ||
| 858 | /** | ||
| 859 | * flush_scheduled_work - ensure that any scheduled work has run to completion. | ||
| 860 | * | ||
| 861 | * Forces execution of the kernel-global workqueue and blocks until its | ||
| 862 | * completion. | ||
| 863 | * | ||
| 864 | * Think twice before calling this function! It's very easy to get into | ||
| 865 | * trouble if you don't take great care. Either of the following situations | ||
| 866 | * will lead to deadlock: | ||
| 867 | * | ||
| 868 | * One of the work items currently on the workqueue needs to acquire | ||
| 869 | * a lock held by your code or its caller. | ||
| 870 | * | ||
| 871 | * Your code is running in the context of a work routine. | ||
| 872 | * | ||
| 873 | * They will be detected by lockdep when they occur, but the first might not | ||
| 874 | * occur very often. It depends on what work items are on the workqueue and | ||
| 875 | * what locks they need, which you have no control over. | ||
| 876 | * | ||
| 877 | * In most situations flushing the entire workqueue is overkill; you merely | ||
| 878 | * need to know that a particular work item isn't queued and isn't running. | ||
| 879 | * In such cases you should use cancel_delayed_work_sync() or | ||
| 880 | * cancel_work_sync() instead. | ||
| 881 | */ | ||
| 848 | void flush_scheduled_work(void) | 882 | void flush_scheduled_work(void) |
| 849 | { | 883 | { |
| 850 | flush_workqueue(keventd_wq); | 884 | flush_workqueue(keventd_wq); |
| @@ -1076,7 +1110,7 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb, | |||
| 1076 | unsigned int cpu = (unsigned long)hcpu; | 1110 | unsigned int cpu = (unsigned long)hcpu; |
| 1077 | struct cpu_workqueue_struct *cwq; | 1111 | struct cpu_workqueue_struct *cwq; |
| 1078 | struct workqueue_struct *wq; | 1112 | struct workqueue_struct *wq; |
| 1079 | int ret = NOTIFY_OK; | 1113 | int err = 0; |
| 1080 | 1114 | ||
| 1081 | action &= ~CPU_TASKS_FROZEN; | 1115 | action &= ~CPU_TASKS_FROZEN; |
| 1082 | 1116 | ||
| @@ -1090,12 +1124,13 @@ undo: | |||
| 1090 | 1124 | ||
| 1091 | switch (action) { | 1125 | switch (action) { |
| 1092 | case CPU_UP_PREPARE: | 1126 | case CPU_UP_PREPARE: |
| 1093 | if (!create_workqueue_thread(cwq, cpu)) | 1127 | err = create_workqueue_thread(cwq, cpu); |
| 1128 | if (!err) | ||
| 1094 | break; | 1129 | break; |
| 1095 | printk(KERN_ERR "workqueue [%s] for %i failed\n", | 1130 | printk(KERN_ERR "workqueue [%s] for %i failed\n", |
| 1096 | wq->name, cpu); | 1131 | wq->name, cpu); |
| 1097 | action = CPU_UP_CANCELED; | 1132 | action = CPU_UP_CANCELED; |
| 1098 | ret = NOTIFY_BAD; | 1133 | err = -ENOMEM; |
| 1099 | goto undo; | 1134 | goto undo; |
| 1100 | 1135 | ||
| 1101 | case CPU_ONLINE: | 1136 | case CPU_ONLINE: |
| @@ -1116,7 +1151,7 @@ undo: | |||
| 1116 | cpumask_clear_cpu(cpu, cpu_populated_map); | 1151 | cpumask_clear_cpu(cpu, cpu_populated_map); |
| 1117 | } | 1152 | } |
| 1118 | 1153 | ||
| 1119 | return ret; | 1154 | return notifier_from_errno(err); |
| 1120 | } | 1155 | } |
| 1121 | 1156 | ||
| 1122 | #ifdef CONFIG_SMP | 1157 | #ifdef CONFIG_SMP |
