diff options
-rw-r--r-- | kernel/workqueue.c | 173 |
1 files changed, 59 insertions, 114 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index f7ab703285a6..dc78956ccf03 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
@@ -55,6 +55,7 @@ struct cpu_workqueue_struct { | |||
55 | struct list_head worklist; | 55 | struct list_head worklist; |
56 | wait_queue_head_t more_work; | 56 | wait_queue_head_t more_work; |
57 | struct work_struct *current_work; | 57 | struct work_struct *current_work; |
58 | unsigned int cpu; | ||
58 | 59 | ||
59 | struct workqueue_struct *wq; /* I: the owning workqueue */ | 60 | struct workqueue_struct *wq; /* I: the owning workqueue */ |
60 | struct task_struct *thread; | 61 | struct task_struct *thread; |
@@ -189,34 +190,19 @@ static DEFINE_SPINLOCK(workqueue_lock); | |||
189 | static LIST_HEAD(workqueues); | 190 | static LIST_HEAD(workqueues); |
190 | 191 | ||
191 | static int singlethread_cpu __read_mostly; | 192 | static int singlethread_cpu __read_mostly; |
192 | static const struct cpumask *cpu_singlethread_map __read_mostly; | ||
193 | /* | ||
194 | * _cpu_down() first removes CPU from cpu_online_map, then CPU_DEAD | ||
195 | * flushes cwq->worklist. This means that flush_workqueue/wait_on_work | ||
196 | * which comes in between can't use for_each_online_cpu(). We could | ||
197 | * use cpu_possible_map, the cpumask below is more a documentation | ||
198 | * than optimization. | ||
199 | */ | ||
200 | static cpumask_var_t cpu_populated_map __read_mostly; | ||
201 | |||
202 | /* If it's single threaded, it isn't in the list of workqueues. */ | ||
203 | static inline bool is_wq_single_threaded(struct workqueue_struct *wq) | ||
204 | { | ||
205 | return wq->flags & WQ_SINGLE_THREAD; | ||
206 | } | ||
207 | 193 | ||
208 | static const struct cpumask *wq_cpu_map(struct workqueue_struct *wq) | 194 | static struct cpu_workqueue_struct *get_cwq(unsigned int cpu, |
195 | struct workqueue_struct *wq) | ||
209 | { | 196 | { |
210 | return is_wq_single_threaded(wq) | 197 | return per_cpu_ptr(wq->cpu_wq, cpu); |
211 | ? cpu_singlethread_map : cpu_populated_map; | ||
212 | } | 198 | } |
213 | 199 | ||
214 | static struct cpu_workqueue_struct *get_cwq(unsigned int cpu, | 200 | static struct cpu_workqueue_struct *target_cwq(unsigned int cpu, |
215 | struct workqueue_struct *wq) | 201 | struct workqueue_struct *wq) |
216 | { | 202 | { |
217 | if (unlikely(is_wq_single_threaded(wq))) | 203 | if (unlikely(wq->flags & WQ_SINGLE_THREAD)) |
218 | cpu = singlethread_cpu; | 204 | cpu = singlethread_cpu; |
219 | return per_cpu_ptr(wq->cpu_wq, cpu); | 205 | return get_cwq(cpu, wq); |
220 | } | 206 | } |
221 | 207 | ||
222 | /* | 208 | /* |
@@ -279,7 +265,7 @@ static void insert_work(struct cpu_workqueue_struct *cwq, | |||
279 | static void __queue_work(unsigned int cpu, struct workqueue_struct *wq, | 265 | static void __queue_work(unsigned int cpu, struct workqueue_struct *wq, |
280 | struct work_struct *work) | 266 | struct work_struct *work) |
281 | { | 267 | { |
282 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); | 268 | struct cpu_workqueue_struct *cwq = target_cwq(cpu, wq); |
283 | unsigned long flags; | 269 | unsigned long flags; |
284 | 270 | ||
285 | debug_work_activate(work); | 271 | debug_work_activate(work); |
@@ -383,7 +369,7 @@ int queue_delayed_work_on(int cpu, struct workqueue_struct *wq, | |||
383 | timer_stats_timer_set_start_info(&dwork->timer); | 369 | timer_stats_timer_set_start_info(&dwork->timer); |
384 | 370 | ||
385 | /* This stores cwq for the moment, for the timer_fn */ | 371 | /* This stores cwq for the moment, for the timer_fn */ |
386 | set_wq_data(work, get_cwq(raw_smp_processor_id(), wq), 0); | 372 | set_wq_data(work, target_cwq(raw_smp_processor_id(), wq), 0); |
387 | timer->expires = jiffies + delay; | 373 | timer->expires = jiffies + delay; |
388 | timer->data = (unsigned long)dwork; | 374 | timer->data = (unsigned long)dwork; |
389 | timer->function = delayed_work_timer_fn; | 375 | timer->function = delayed_work_timer_fn; |
@@ -495,6 +481,10 @@ static int worker_thread(void *__cwq) | |||
495 | if (kthread_should_stop()) | 481 | if (kthread_should_stop()) |
496 | break; | 482 | break; |
497 | 483 | ||
484 | if (unlikely(!cpumask_equal(&cwq->thread->cpus_allowed, | ||
485 | get_cpu_mask(cwq->cpu)))) | ||
486 | set_cpus_allowed_ptr(cwq->thread, | ||
487 | get_cpu_mask(cwq->cpu)); | ||
498 | run_workqueue(cwq); | 488 | run_workqueue(cwq); |
499 | } | 489 | } |
500 | 490 | ||
@@ -574,14 +564,13 @@ static int flush_cpu_workqueue(struct cpu_workqueue_struct *cwq) | |||
574 | */ | 564 | */ |
575 | void flush_workqueue(struct workqueue_struct *wq) | 565 | void flush_workqueue(struct workqueue_struct *wq) |
576 | { | 566 | { |
577 | const struct cpumask *cpu_map = wq_cpu_map(wq); | ||
578 | int cpu; | 567 | int cpu; |
579 | 568 | ||
580 | might_sleep(); | 569 | might_sleep(); |
581 | lock_map_acquire(&wq->lockdep_map); | 570 | lock_map_acquire(&wq->lockdep_map); |
582 | lock_map_release(&wq->lockdep_map); | 571 | lock_map_release(&wq->lockdep_map); |
583 | for_each_cpu(cpu, cpu_map) | 572 | for_each_possible_cpu(cpu) |
584 | flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, cpu)); | 573 | flush_cpu_workqueue(get_cwq(cpu, wq)); |
585 | } | 574 | } |
586 | EXPORT_SYMBOL_GPL(flush_workqueue); | 575 | EXPORT_SYMBOL_GPL(flush_workqueue); |
587 | 576 | ||
@@ -699,7 +688,6 @@ static void wait_on_work(struct work_struct *work) | |||
699 | { | 688 | { |
700 | struct cpu_workqueue_struct *cwq; | 689 | struct cpu_workqueue_struct *cwq; |
701 | struct workqueue_struct *wq; | 690 | struct workqueue_struct *wq; |
702 | const struct cpumask *cpu_map; | ||
703 | int cpu; | 691 | int cpu; |
704 | 692 | ||
705 | might_sleep(); | 693 | might_sleep(); |
@@ -712,9 +700,8 @@ static void wait_on_work(struct work_struct *work) | |||
712 | return; | 700 | return; |
713 | 701 | ||
714 | wq = cwq->wq; | 702 | wq = cwq->wq; |
715 | cpu_map = wq_cpu_map(wq); | ||
716 | 703 | ||
717 | for_each_cpu(cpu, cpu_map) | 704 | for_each_possible_cpu(cpu) |
718 | wait_on_cpu_work(get_cwq(cpu, wq), work); | 705 | wait_on_cpu_work(get_cwq(cpu, wq), work); |
719 | } | 706 | } |
720 | 707 | ||
@@ -972,7 +959,7 @@ int current_is_keventd(void) | |||
972 | 959 | ||
973 | BUG_ON(!keventd_wq); | 960 | BUG_ON(!keventd_wq); |
974 | 961 | ||
975 | cwq = per_cpu_ptr(keventd_wq->cpu_wq, cpu); | 962 | cwq = get_cwq(cpu, keventd_wq); |
976 | if (current == cwq->thread) | 963 | if (current == cwq->thread) |
977 | ret = 1; | 964 | ret = 1; |
978 | 965 | ||
@@ -980,26 +967,12 @@ int current_is_keventd(void) | |||
980 | 967 | ||
981 | } | 968 | } |
982 | 969 | ||
983 | static struct cpu_workqueue_struct * | ||
984 | init_cpu_workqueue(struct workqueue_struct *wq, int cpu) | ||
985 | { | ||
986 | struct cpu_workqueue_struct *cwq = per_cpu_ptr(wq->cpu_wq, cpu); | ||
987 | |||
988 | cwq->wq = wq; | ||
989 | spin_lock_init(&cwq->lock); | ||
990 | INIT_LIST_HEAD(&cwq->worklist); | ||
991 | init_waitqueue_head(&cwq->more_work); | ||
992 | |||
993 | return cwq; | ||
994 | } | ||
995 | |||
996 | static int create_workqueue_thread(struct cpu_workqueue_struct *cwq, int cpu) | 970 | static int create_workqueue_thread(struct cpu_workqueue_struct *cwq, int cpu) |
997 | { | 971 | { |
998 | struct workqueue_struct *wq = cwq->wq; | 972 | struct workqueue_struct *wq = cwq->wq; |
999 | const char *fmt = is_wq_single_threaded(wq) ? "%s" : "%s/%d"; | ||
1000 | struct task_struct *p; | 973 | struct task_struct *p; |
1001 | 974 | ||
1002 | p = kthread_create(worker_thread, cwq, fmt, wq->name, cpu); | 975 | p = kthread_create(worker_thread, cwq, "%s/%d", wq->name, cpu); |
1003 | /* | 976 | /* |
1004 | * Nobody can add the work_struct to this cwq, | 977 | * Nobody can add the work_struct to this cwq, |
1005 | * if (caller is __create_workqueue) | 978 | * if (caller is __create_workqueue) |
@@ -1031,8 +1004,8 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
1031 | struct lock_class_key *key, | 1004 | struct lock_class_key *key, |
1032 | const char *lock_name) | 1005 | const char *lock_name) |
1033 | { | 1006 | { |
1007 | bool singlethread = flags & WQ_SINGLE_THREAD; | ||
1034 | struct workqueue_struct *wq; | 1008 | struct workqueue_struct *wq; |
1035 | struct cpu_workqueue_struct *cwq; | ||
1036 | int err = 0, cpu; | 1009 | int err = 0, cpu; |
1037 | 1010 | ||
1038 | wq = kzalloc(sizeof(*wq), GFP_KERNEL); | 1011 | wq = kzalloc(sizeof(*wq), GFP_KERNEL); |
@@ -1048,37 +1021,37 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
1048 | lockdep_init_map(&wq->lockdep_map, lock_name, key, 0); | 1021 | lockdep_init_map(&wq->lockdep_map, lock_name, key, 0); |
1049 | INIT_LIST_HEAD(&wq->list); | 1022 | INIT_LIST_HEAD(&wq->list); |
1050 | 1023 | ||
1051 | if (flags & WQ_SINGLE_THREAD) { | 1024 | cpu_maps_update_begin(); |
1052 | cwq = init_cpu_workqueue(wq, singlethread_cpu); | 1025 | /* |
1053 | err = create_workqueue_thread(cwq, singlethread_cpu); | 1026 | * We must initialize cwqs for each possible cpu even if we |
1054 | start_workqueue_thread(cwq, -1); | 1027 | * are going to call destroy_workqueue() finally. Otherwise |
1055 | } else { | 1028 | * cpu_up() can hit the uninitialized cwq once we drop the |
1056 | cpu_maps_update_begin(); | 1029 | * lock. |
1057 | /* | 1030 | */ |
1058 | * We must place this wq on list even if the code below fails. | 1031 | for_each_possible_cpu(cpu) { |
1059 | * cpu_down(cpu) can remove cpu from cpu_populated_map before | 1032 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); |
1060 | * destroy_workqueue() takes the lock, in that case we leak | 1033 | |
1061 | * cwq[cpu]->thread. | 1034 | cwq->wq = wq; |
1062 | */ | 1035 | cwq->cpu = cpu; |
1063 | spin_lock(&workqueue_lock); | 1036 | spin_lock_init(&cwq->lock); |
1064 | list_add(&wq->list, &workqueues); | 1037 | INIT_LIST_HEAD(&cwq->worklist); |
1065 | spin_unlock(&workqueue_lock); | 1038 | init_waitqueue_head(&cwq->more_work); |
1066 | /* | 1039 | |
1067 | * We must initialize cwqs for each possible cpu even if we | 1040 | if (err) |
1068 | * are going to call destroy_workqueue() finally. Otherwise | 1041 | continue; |
1069 | * cpu_up() can hit the uninitialized cwq once we drop the | 1042 | err = create_workqueue_thread(cwq, cpu); |
1070 | * lock. | 1043 | if (cpu_online(cpu) && !singlethread) |
1071 | */ | ||
1072 | for_each_possible_cpu(cpu) { | ||
1073 | cwq = init_cpu_workqueue(wq, cpu); | ||
1074 | if (err || !cpu_online(cpu)) | ||
1075 | continue; | ||
1076 | err = create_workqueue_thread(cwq, cpu); | ||
1077 | start_workqueue_thread(cwq, cpu); | 1044 | start_workqueue_thread(cwq, cpu); |
1078 | } | 1045 | else |
1079 | cpu_maps_update_done(); | 1046 | start_workqueue_thread(cwq, -1); |
1080 | } | 1047 | } |
1081 | 1048 | ||
1049 | spin_lock(&workqueue_lock); | ||
1050 | list_add(&wq->list, &workqueues); | ||
1051 | spin_unlock(&workqueue_lock); | ||
1052 | |||
1053 | cpu_maps_update_done(); | ||
1054 | |||
1082 | if (err) { | 1055 | if (err) { |
1083 | destroy_workqueue(wq); | 1056 | destroy_workqueue(wq); |
1084 | wq = NULL; | 1057 | wq = NULL; |
@@ -1128,17 +1101,16 @@ static void cleanup_workqueue_thread(struct cpu_workqueue_struct *cwq) | |||
1128 | */ | 1101 | */ |
1129 | void destroy_workqueue(struct workqueue_struct *wq) | 1102 | void destroy_workqueue(struct workqueue_struct *wq) |
1130 | { | 1103 | { |
1131 | const struct cpumask *cpu_map = wq_cpu_map(wq); | ||
1132 | int cpu; | 1104 | int cpu; |
1133 | 1105 | ||
1134 | cpu_maps_update_begin(); | 1106 | cpu_maps_update_begin(); |
1135 | spin_lock(&workqueue_lock); | 1107 | spin_lock(&workqueue_lock); |
1136 | list_del(&wq->list); | 1108 | list_del(&wq->list); |
1137 | spin_unlock(&workqueue_lock); | 1109 | spin_unlock(&workqueue_lock); |
1110 | cpu_maps_update_done(); | ||
1138 | 1111 | ||
1139 | for_each_cpu(cpu, cpu_map) | 1112 | for_each_possible_cpu(cpu) |
1140 | cleanup_workqueue_thread(per_cpu_ptr(wq->cpu_wq, cpu)); | 1113 | cleanup_workqueue_thread(get_cwq(cpu, wq)); |
1141 | cpu_maps_update_done(); | ||
1142 | 1114 | ||
1143 | free_percpu(wq->cpu_wq); | 1115 | free_percpu(wq->cpu_wq); |
1144 | kfree(wq); | 1116 | kfree(wq); |
@@ -1152,48 +1124,25 @@ static int __devinit workqueue_cpu_callback(struct notifier_block *nfb, | |||
1152 | unsigned int cpu = (unsigned long)hcpu; | 1124 | unsigned int cpu = (unsigned long)hcpu; |
1153 | struct cpu_workqueue_struct *cwq; | 1125 | struct cpu_workqueue_struct *cwq; |
1154 | struct workqueue_struct *wq; | 1126 | struct workqueue_struct *wq; |
1155 | int err = 0; | ||
1156 | 1127 | ||
1157 | action &= ~CPU_TASKS_FROZEN; | 1128 | action &= ~CPU_TASKS_FROZEN; |
1158 | 1129 | ||
1159 | switch (action) { | ||
1160 | case CPU_UP_PREPARE: | ||
1161 | cpumask_set_cpu(cpu, cpu_populated_map); | ||
1162 | } | ||
1163 | undo: | ||
1164 | list_for_each_entry(wq, &workqueues, list) { | 1130 | list_for_each_entry(wq, &workqueues, list) { |
1165 | cwq = per_cpu_ptr(wq->cpu_wq, cpu); | 1131 | if (wq->flags & WQ_SINGLE_THREAD) |
1132 | continue; | ||
1166 | 1133 | ||
1167 | switch (action) { | 1134 | cwq = get_cwq(cpu, wq); |
1168 | case CPU_UP_PREPARE: | ||
1169 | err = create_workqueue_thread(cwq, cpu); | ||
1170 | if (!err) | ||
1171 | break; | ||
1172 | printk(KERN_ERR "workqueue [%s] for %i failed\n", | ||
1173 | wq->name, cpu); | ||
1174 | action = CPU_UP_CANCELED; | ||
1175 | err = -ENOMEM; | ||
1176 | goto undo; | ||
1177 | |||
1178 | case CPU_ONLINE: | ||
1179 | start_workqueue_thread(cwq, cpu); | ||
1180 | break; | ||
1181 | 1135 | ||
1182 | case CPU_UP_CANCELED: | 1136 | switch (action) { |
1183 | start_workqueue_thread(cwq, -1); | ||
1184 | case CPU_POST_DEAD: | 1137 | case CPU_POST_DEAD: |
1185 | cleanup_workqueue_thread(cwq); | 1138 | lock_map_acquire(&cwq->wq->lockdep_map); |
1139 | lock_map_release(&cwq->wq->lockdep_map); | ||
1140 | flush_cpu_workqueue(cwq); | ||
1186 | break; | 1141 | break; |
1187 | } | 1142 | } |
1188 | } | 1143 | } |
1189 | 1144 | ||
1190 | switch (action) { | 1145 | return notifier_from_errno(0); |
1191 | case CPU_UP_CANCELED: | ||
1192 | case CPU_POST_DEAD: | ||
1193 | cpumask_clear_cpu(cpu, cpu_populated_map); | ||
1194 | } | ||
1195 | |||
1196 | return notifier_from_errno(err); | ||
1197 | } | 1146 | } |
1198 | 1147 | ||
1199 | #ifdef CONFIG_SMP | 1148 | #ifdef CONFIG_SMP |
@@ -1245,11 +1194,7 @@ EXPORT_SYMBOL_GPL(work_on_cpu); | |||
1245 | 1194 | ||
1246 | void __init init_workqueues(void) | 1195 | void __init init_workqueues(void) |
1247 | { | 1196 | { |
1248 | alloc_cpumask_var(&cpu_populated_map, GFP_KERNEL); | ||
1249 | |||
1250 | cpumask_copy(cpu_populated_map, cpu_online_mask); | ||
1251 | singlethread_cpu = cpumask_first(cpu_possible_mask); | 1197 | singlethread_cpu = cpumask_first(cpu_possible_mask); |
1252 | cpu_singlethread_map = cpumask_of(singlethread_cpu); | ||
1253 | hotcpu_notifier(workqueue_cpu_callback, 0); | 1198 | hotcpu_notifier(workqueue_cpu_callback, 0); |
1254 | keventd_wq = create_workqueue("events"); | 1199 | keventd_wq = create_workqueue("events"); |
1255 | BUG_ON(!keventd_wq); | 1200 | BUG_ON(!keventd_wq); |