diff options
Diffstat (limited to 'kernel/workqueue.c')
| -rw-r--r-- | kernel/workqueue.c | 18 |
1 files changed, 8 insertions, 10 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index ea422254f8bf..74f3f7825229 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
| @@ -227,13 +227,7 @@ EXPORT_SYMBOL_GPL(queue_delayed_work_on); | |||
| 227 | 227 | ||
| 228 | static void run_workqueue(struct cpu_workqueue_struct *cwq) | 228 | static void run_workqueue(struct cpu_workqueue_struct *cwq) |
| 229 | { | 229 | { |
| 230 | unsigned long flags; | 230 | spin_lock_irq(&cwq->lock); |
| 231 | |||
| 232 | /* | ||
| 233 | * Keep taking off work from the queue until | ||
| 234 | * done. | ||
| 235 | */ | ||
| 236 | spin_lock_irqsave(&cwq->lock, flags); | ||
| 237 | cwq->run_depth++; | 231 | cwq->run_depth++; |
| 238 | if (cwq->run_depth > 3) { | 232 | if (cwq->run_depth > 3) { |
| 239 | /* morton gets to eat his hat */ | 233 | /* morton gets to eat his hat */ |
| @@ -248,7 +242,7 @@ static void run_workqueue(struct cpu_workqueue_struct *cwq) | |||
| 248 | 242 | ||
| 249 | cwq->current_work = work; | 243 | cwq->current_work = work; |
| 250 | list_del_init(cwq->worklist.next); | 244 | list_del_init(cwq->worklist.next); |
| 251 | spin_unlock_irqrestore(&cwq->lock, flags); | 245 | spin_unlock_irq(&cwq->lock); |
| 252 | 246 | ||
| 253 | BUG_ON(get_wq_data(work) != cwq); | 247 | BUG_ON(get_wq_data(work) != cwq); |
| 254 | if (!test_bit(WORK_STRUCT_NOAUTOREL, work_data_bits(work))) | 248 | if (!test_bit(WORK_STRUCT_NOAUTOREL, work_data_bits(work))) |
| @@ -266,11 +260,11 @@ static void run_workqueue(struct cpu_workqueue_struct *cwq) | |||
| 266 | dump_stack(); | 260 | dump_stack(); |
| 267 | } | 261 | } |
| 268 | 262 | ||
| 269 | spin_lock_irqsave(&cwq->lock, flags); | 263 | spin_lock_irq(&cwq->lock); |
| 270 | cwq->current_work = NULL; | 264 | cwq->current_work = NULL; |
| 271 | } | 265 | } |
| 272 | cwq->run_depth--; | 266 | cwq->run_depth--; |
| 273 | spin_unlock_irqrestore(&cwq->lock, flags); | 267 | spin_unlock_irq(&cwq->lock); |
| 274 | } | 268 | } |
| 275 | 269 | ||
| 276 | /* | 270 | /* |
| @@ -399,6 +393,8 @@ static void flush_cpu_workqueue(struct cpu_workqueue_struct *cwq) | |||
| 399 | */ | 393 | */ |
| 400 | void fastcall flush_workqueue(struct workqueue_struct *wq) | 394 | void fastcall flush_workqueue(struct workqueue_struct *wq) |
| 401 | { | 395 | { |
| 396 | might_sleep(); | ||
| 397 | |||
| 402 | if (is_single_threaded(wq)) | 398 | if (is_single_threaded(wq)) |
| 403 | flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, singlethread_cpu)); | 399 | flush_cpu_workqueue(per_cpu_ptr(wq->cpu_wq, singlethread_cpu)); |
| 404 | else { | 400 | else { |
| @@ -445,6 +441,8 @@ void flush_work(struct workqueue_struct *wq, struct work_struct *work) | |||
| 445 | { | 441 | { |
| 446 | struct cpu_workqueue_struct *cwq; | 442 | struct cpu_workqueue_struct *cwq; |
| 447 | 443 | ||
| 444 | might_sleep(); | ||
| 445 | |||
| 448 | cwq = get_wq_data(work); | 446 | cwq = get_wq_data(work); |
| 449 | /* Was it ever queued ? */ | 447 | /* Was it ever queued ? */ |
| 450 | if (!cwq) | 448 | if (!cwq) |
