diff options
| -rw-r--r-- | kernel/workqueue.c | 211 |
1 files changed, 150 insertions, 61 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index 56e47c59d73b..600db10a4dbf 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
| @@ -33,6 +33,7 @@ | |||
| 33 | #include <linux/kallsyms.h> | 33 | #include <linux/kallsyms.h> |
| 34 | #include <linux/debug_locks.h> | 34 | #include <linux/debug_locks.h> |
| 35 | #include <linux/lockdep.h> | 35 | #include <linux/lockdep.h> |
| 36 | #include <linux/idr.h> | ||
| 36 | 37 | ||
| 37 | /* | 38 | /* |
| 38 | * Structure fields follow one of the following exclusion rules. | 39 | * Structure fields follow one of the following exclusion rules. |
| @@ -46,6 +47,15 @@ | |||
| 46 | * W: workqueue_lock protected. | 47 | * W: workqueue_lock protected. |
| 47 | */ | 48 | */ |
| 48 | 49 | ||
| 50 | struct cpu_workqueue_struct; | ||
| 51 | |||
| 52 | struct worker { | ||
| 53 | struct work_struct *current_work; /* L: work being processed */ | ||
| 54 | struct task_struct *task; /* I: worker task */ | ||
| 55 | struct cpu_workqueue_struct *cwq; /* I: the associated cwq */ | ||
| 56 | int id; /* I: worker id */ | ||
| 57 | }; | ||
| 58 | |||
| 49 | /* | 59 | /* |
| 50 | * The per-CPU workqueue (if single thread, we always use the first | 60 | * The per-CPU workqueue (if single thread, we always use the first |
| 51 | * possible cpu). The lower WORK_STRUCT_FLAG_BITS of | 61 | * possible cpu). The lower WORK_STRUCT_FLAG_BITS of |
| @@ -58,15 +68,14 @@ struct cpu_workqueue_struct { | |||
| 58 | 68 | ||
| 59 | struct list_head worklist; | 69 | struct list_head worklist; |
| 60 | wait_queue_head_t more_work; | 70 | wait_queue_head_t more_work; |
| 61 | struct work_struct *current_work; | ||
| 62 | unsigned int cpu; | 71 | unsigned int cpu; |
| 72 | struct worker *worker; | ||
| 63 | 73 | ||
| 64 | struct workqueue_struct *wq; /* I: the owning workqueue */ | 74 | struct workqueue_struct *wq; /* I: the owning workqueue */ |
| 65 | int work_color; /* L: current color */ | 75 | int work_color; /* L: current color */ |
| 66 | int flush_color; /* L: flushing color */ | 76 | int flush_color; /* L: flushing color */ |
| 67 | int nr_in_flight[WORK_NR_COLORS]; | 77 | int nr_in_flight[WORK_NR_COLORS]; |
| 68 | /* L: nr of in_flight works */ | 78 | /* L: nr of in_flight works */ |
| 69 | struct task_struct *thread; | ||
| 70 | }; | 79 | }; |
| 71 | 80 | ||
| 72 | /* | 81 | /* |
| @@ -214,6 +223,9 @@ static inline void debug_work_deactivate(struct work_struct *work) { } | |||
| 214 | /* Serializes the accesses to the list of workqueues. */ | 223 | /* Serializes the accesses to the list of workqueues. */ |
| 215 | static DEFINE_SPINLOCK(workqueue_lock); | 224 | static DEFINE_SPINLOCK(workqueue_lock); |
| 216 | static LIST_HEAD(workqueues); | 225 | static LIST_HEAD(workqueues); |
| 226 | static DEFINE_PER_CPU(struct ida, worker_ida); | ||
| 227 | |||
| 228 | static int worker_thread(void *__worker); | ||
| 217 | 229 | ||
| 218 | static int singlethread_cpu __read_mostly; | 230 | static int singlethread_cpu __read_mostly; |
| 219 | 231 | ||
| @@ -428,6 +440,105 @@ int queue_delayed_work_on(int cpu, struct workqueue_struct *wq, | |||
| 428 | } | 440 | } |
| 429 | EXPORT_SYMBOL_GPL(queue_delayed_work_on); | 441 | EXPORT_SYMBOL_GPL(queue_delayed_work_on); |
| 430 | 442 | ||
| 443 | static struct worker *alloc_worker(void) | ||
| 444 | { | ||
| 445 | struct worker *worker; | ||
| 446 | |||
| 447 | worker = kzalloc(sizeof(*worker), GFP_KERNEL); | ||
| 448 | return worker; | ||
| 449 | } | ||
| 450 | |||
| 451 | /** | ||
| 452 | * create_worker - create a new workqueue worker | ||
| 453 | * @cwq: cwq the new worker will belong to | ||
| 454 | * @bind: whether to set affinity to @cpu or not | ||
| 455 | * | ||
| 456 | * Create a new worker which is bound to @cwq. The returned worker | ||
| 457 | * can be started by calling start_worker() or destroyed using | ||
| 458 | * destroy_worker(). | ||
| 459 | * | ||
| 460 | * CONTEXT: | ||
| 461 | * Might sleep. Does GFP_KERNEL allocations. | ||
| 462 | * | ||
| 463 | * RETURNS: | ||
| 464 | * Pointer to the newly created worker. | ||
| 465 | */ | ||
| 466 | static struct worker *create_worker(struct cpu_workqueue_struct *cwq, bool bind) | ||
| 467 | { | ||
| 468 | int id = -1; | ||
| 469 | struct worker *worker = NULL; | ||
| 470 | |||
| 471 | spin_lock(&workqueue_lock); | ||
| 472 | while (ida_get_new(&per_cpu(worker_ida, cwq->cpu), &id)) { | ||
| 473 | spin_unlock(&workqueue_lock); | ||
| 474 | if (!ida_pre_get(&per_cpu(worker_ida, cwq->cpu), GFP_KERNEL)) | ||
| 475 | goto fail; | ||
| 476 | spin_lock(&workqueue_lock); | ||
| 477 | } | ||
| 478 | spin_unlock(&workqueue_lock); | ||
| 479 | |||
| 480 | worker = alloc_worker(); | ||
| 481 | if (!worker) | ||
| 482 | goto fail; | ||
| 483 | |||
| 484 | worker->cwq = cwq; | ||
| 485 | worker->id = id; | ||
| 486 | |||
| 487 | worker->task = kthread_create(worker_thread, worker, "kworker/%u:%d", | ||
| 488 | cwq->cpu, id); | ||
| 489 | if (IS_ERR(worker->task)) | ||
| 490 | goto fail; | ||
| 491 | |||
| 492 | if (bind) | ||
| 493 | kthread_bind(worker->task, cwq->cpu); | ||
| 494 | |||
| 495 | return worker; | ||
| 496 | fail: | ||
| 497 | if (id >= 0) { | ||
| 498 | spin_lock(&workqueue_lock); | ||
| 499 | ida_remove(&per_cpu(worker_ida, cwq->cpu), id); | ||
| 500 | spin_unlock(&workqueue_lock); | ||
| 501 | } | ||
| 502 | kfree(worker); | ||
| 503 | return NULL; | ||
| 504 | } | ||
| 505 | |||
| 506 | /** | ||
| 507 | * start_worker - start a newly created worker | ||
| 508 | * @worker: worker to start | ||
| 509 | * | ||
| 510 | * Start @worker. | ||
| 511 | * | ||
| 512 | * CONTEXT: | ||
| 513 | * spin_lock_irq(cwq->lock). | ||
| 514 | */ | ||
| 515 | static void start_worker(struct worker *worker) | ||
| 516 | { | ||
| 517 | wake_up_process(worker->task); | ||
| 518 | } | ||
| 519 | |||
| 520 | /** | ||
| 521 | * destroy_worker - destroy a workqueue worker | ||
| 522 | * @worker: worker to be destroyed | ||
| 523 | * | ||
| 524 | * Destroy @worker. | ||
| 525 | */ | ||
| 526 | static void destroy_worker(struct worker *worker) | ||
| 527 | { | ||
| 528 | int cpu = worker->cwq->cpu; | ||
| 529 | int id = worker->id; | ||
| 530 | |||
| 531 | /* sanity check frenzy */ | ||
| 532 | BUG_ON(worker->current_work); | ||
| 533 | |||
| 534 | kthread_stop(worker->task); | ||
| 535 | kfree(worker); | ||
| 536 | |||
| 537 | spin_lock(&workqueue_lock); | ||
| 538 | ida_remove(&per_cpu(worker_ida, cpu), id); | ||
| 539 | spin_unlock(&workqueue_lock); | ||
| 540 | } | ||
| 541 | |||
| 431 | /** | 542 | /** |
| 432 | * cwq_dec_nr_in_flight - decrement cwq's nr_in_flight | 543 | * cwq_dec_nr_in_flight - decrement cwq's nr_in_flight |
| 433 | * @cwq: cwq of interest | 544 | * @cwq: cwq of interest |
| @@ -468,7 +579,7 @@ static void cwq_dec_nr_in_flight(struct cpu_workqueue_struct *cwq, int color) | |||
| 468 | 579 | ||
| 469 | /** | 580 | /** |
| 470 | * process_one_work - process single work | 581 | * process_one_work - process single work |
| 471 | * @cwq: cwq to process work for | 582 | * @worker: self |
| 472 | * @work: work to process | 583 | * @work: work to process |
| 473 | * | 584 | * |
| 474 | * Process @work. This function contains all the logics necessary to | 585 | * Process @work. This function contains all the logics necessary to |
| @@ -480,9 +591,9 @@ static void cwq_dec_nr_in_flight(struct cpu_workqueue_struct *cwq, int color) | |||
| 480 | * CONTEXT: | 591 | * CONTEXT: |
| 481 | * spin_lock_irq(cwq->lock) which is released and regrabbed. | 592 | * spin_lock_irq(cwq->lock) which is released and regrabbed. |
| 482 | */ | 593 | */ |
| 483 | static void process_one_work(struct cpu_workqueue_struct *cwq, | 594 | static void process_one_work(struct worker *worker, struct work_struct *work) |
| 484 | struct work_struct *work) | ||
| 485 | { | 595 | { |
| 596 | struct cpu_workqueue_struct *cwq = worker->cwq; | ||
| 486 | work_func_t f = work->func; | 597 | work_func_t f = work->func; |
| 487 | int work_color; | 598 | int work_color; |
| 488 | #ifdef CONFIG_LOCKDEP | 599 | #ifdef CONFIG_LOCKDEP |
| @@ -497,7 +608,7 @@ static void process_one_work(struct cpu_workqueue_struct *cwq, | |||
| 497 | #endif | 608 | #endif |
| 498 | /* claim and process */ | 609 | /* claim and process */ |
| 499 | debug_work_deactivate(work); | 610 | debug_work_deactivate(work); |
| 500 | cwq->current_work = work; | 611 | worker->current_work = work; |
| 501 | work_color = get_work_color(work); | 612 | work_color = get_work_color(work); |
| 502 | list_del_init(&work->entry); | 613 | list_del_init(&work->entry); |
| 503 | 614 | ||
| @@ -524,30 +635,33 @@ static void process_one_work(struct cpu_workqueue_struct *cwq, | |||
| 524 | spin_lock_irq(&cwq->lock); | 635 | spin_lock_irq(&cwq->lock); |
| 525 | 636 | ||
| 526 | /* we're done with it, release */ | 637 | /* we're done with it, release */ |
| 527 | cwq->current_work = NULL; | 638 | worker->current_work = NULL; |
| 528 | cwq_dec_nr_in_flight(cwq, work_color); | 639 | cwq_dec_nr_in_flight(cwq, work_color); |
| 529 | } | 640 | } |
| 530 | 641 | ||
| 531 | static void run_workqueue(struct cpu_workqueue_struct *cwq) | 642 | static void run_workqueue(struct worker *worker) |
| 532 | { | 643 | { |
| 644 | struct cpu_workqueue_struct *cwq = worker->cwq; | ||
| 645 | |||
| 533 | spin_lock_irq(&cwq->lock); | 646 | spin_lock_irq(&cwq->lock); |
| 534 | while (!list_empty(&cwq->worklist)) { | 647 | while (!list_empty(&cwq->worklist)) { |
| 535 | struct work_struct *work = list_entry(cwq->worklist.next, | 648 | struct work_struct *work = list_entry(cwq->worklist.next, |
| 536 | struct work_struct, entry); | 649 | struct work_struct, entry); |
| 537 | process_one_work(cwq, work); | 650 | process_one_work(worker, work); |
| 538 | } | 651 | } |
| 539 | spin_unlock_irq(&cwq->lock); | 652 | spin_unlock_irq(&cwq->lock); |
| 540 | } | 653 | } |
| 541 | 654 | ||
| 542 | /** | 655 | /** |
| 543 | * worker_thread - the worker thread function | 656 | * worker_thread - the worker thread function |
| 544 | * @__cwq: cwq to serve | 657 | * @__worker: self |
| 545 | * | 658 | * |
| 546 | * The cwq worker thread function. | 659 | * The cwq worker thread function. |
| 547 | */ | 660 | */ |
| 548 | static int worker_thread(void *__cwq) | 661 | static int worker_thread(void *__worker) |
| 549 | { | 662 | { |
| 550 | struct cpu_workqueue_struct *cwq = __cwq; | 663 | struct worker *worker = __worker; |
| 664 | struct cpu_workqueue_struct *cwq = worker->cwq; | ||
| 551 | DEFINE_WAIT(wait); | 665 | DEFINE_WAIT(wait); |
| 552 | 666 | ||
| 553 | if (cwq->wq->flags & WQ_FREEZEABLE) | 667 | if (cwq->wq->flags & WQ_FREEZEABLE) |
| @@ -566,11 +680,11 @@ static int worker_thread(void *__cwq) | |||
| 566 | if (kthread_should_stop()) | 680 | if (kthread_should_stop()) |
| 567 | break; | 681 | break; |
| 568 | 682 | ||
| 569 | if (unlikely(!cpumask_equal(&cwq->thread->cpus_allowed, | 683 | if (unlikely(!cpumask_equal(&worker->task->cpus_allowed, |
| 570 | get_cpu_mask(cwq->cpu)))) | 684 | get_cpu_mask(cwq->cpu)))) |
| 571 | set_cpus_allowed_ptr(cwq->thread, | 685 | set_cpus_allowed_ptr(worker->task, |
| 572 | get_cpu_mask(cwq->cpu)); | 686 | get_cpu_mask(cwq->cpu)); |
| 573 | run_workqueue(cwq); | 687 | run_workqueue(worker); |
| 574 | } | 688 | } |
| 575 | 689 | ||
| 576 | return 0; | 690 | return 0; |
| @@ -873,7 +987,7 @@ int flush_work(struct work_struct *work) | |||
| 873 | goto already_gone; | 987 | goto already_gone; |
| 874 | prev = &work->entry; | 988 | prev = &work->entry; |
| 875 | } else { | 989 | } else { |
| 876 | if (cwq->current_work != work) | 990 | if (!cwq->worker || cwq->worker->current_work != work) |
| 877 | goto already_gone; | 991 | goto already_gone; |
| 878 | prev = &cwq->worklist; | 992 | prev = &cwq->worklist; |
| 879 | } | 993 | } |
| @@ -937,7 +1051,7 @@ static void wait_on_cpu_work(struct cpu_workqueue_struct *cwq, | |||
| 937 | int running = 0; | 1051 | int running = 0; |
| 938 | 1052 | ||
| 939 | spin_lock_irq(&cwq->lock); | 1053 | spin_lock_irq(&cwq->lock); |
| 940 | if (unlikely(cwq->current_work == work)) { | 1054 | if (unlikely(cwq->worker && cwq->worker->current_work == work)) { |
| 941 | insert_wq_barrier(cwq, &barr, cwq->worklist.next); | 1055 | insert_wq_barrier(cwq, &barr, cwq->worklist.next); |
| 942 | running = 1; | 1056 | running = 1; |
| 943 | } | 1057 | } |
| @@ -1225,7 +1339,7 @@ int current_is_keventd(void) | |||
| 1225 | BUG_ON(!keventd_wq); | 1339 | BUG_ON(!keventd_wq); |
| 1226 | 1340 | ||
| 1227 | cwq = get_cwq(cpu, keventd_wq); | 1341 | cwq = get_cwq(cpu, keventd_wq); |
| 1228 | if (current == cwq->thread) | 1342 | if (current == cwq->worker->task) |
| 1229 | ret = 1; | 1343 | ret = 1; |
| 1230 | 1344 | ||
| 1231 | return ret; | 1345 | return ret; |
| @@ -1279,38 +1393,6 @@ static void free_cwqs(struct cpu_workqueue_struct *cwqs) | |||
| 1279 | #endif | 1393 | #endif |
| 1280 | } | 1394 | } |
| 1281 | 1395 | ||
| 1282 | static int create_workqueue_thread(struct cpu_workqueue_struct *cwq, int cpu) | ||
| 1283 | { | ||
| 1284 | struct workqueue_struct *wq = cwq->wq; | ||
| 1285 | struct task_struct *p; | ||
| 1286 | |||
| 1287 | p = kthread_create(worker_thread, cwq, "%s/%d", wq->name, cpu); | ||
| 1288 | /* | ||
| 1289 | * Nobody can add the work_struct to this cwq, | ||
| 1290 | * if (caller is __create_workqueue) | ||
| 1291 | * nobody should see this wq | ||
| 1292 | * else // caller is CPU_UP_PREPARE | ||
| 1293 | * cpu is not on cpu_online_map | ||
| 1294 | * so we can abort safely. | ||
| 1295 | */ | ||
| 1296 | if (IS_ERR(p)) | ||
| 1297 | return PTR_ERR(p); | ||
| 1298 | cwq->thread = p; | ||
| 1299 | |||
| 1300 | return 0; | ||
| 1301 | } | ||
| 1302 | |||
| 1303 | static void start_workqueue_thread(struct cpu_workqueue_struct *cwq, int cpu) | ||
| 1304 | { | ||
| 1305 | struct task_struct *p = cwq->thread; | ||
| 1306 | |||
| 1307 | if (p != NULL) { | ||
| 1308 | if (cpu >= 0) | ||
| 1309 | kthread_bind(p, cpu); | ||
| 1310 | wake_up_process(p); | ||
| 1311 | } | ||
| 1312 | } | ||
| 1313 | |||
| 1314 | struct workqueue_struct *__create_workqueue_key(const char *name, | 1396 | struct workqueue_struct *__create_workqueue_key(const char *name, |
| 1315 | unsigned int flags, | 1397 | unsigned int flags, |
| 1316 | struct lock_class_key *key, | 1398 | struct lock_class_key *key, |
| @@ -1318,7 +1400,8 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
| 1318 | { | 1400 | { |
| 1319 | bool singlethread = flags & WQ_SINGLE_THREAD; | 1401 | bool singlethread = flags & WQ_SINGLE_THREAD; |
| 1320 | struct workqueue_struct *wq; | 1402 | struct workqueue_struct *wq; |
| 1321 | int err = 0, cpu; | 1403 | bool failed = false; |
| 1404 | unsigned int cpu; | ||
| 1322 | 1405 | ||
| 1323 | wq = kzalloc(sizeof(*wq), GFP_KERNEL); | 1406 | wq = kzalloc(sizeof(*wq), GFP_KERNEL); |
| 1324 | if (!wq) | 1407 | if (!wq) |
| @@ -1348,20 +1431,21 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
| 1348 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); | 1431 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); |
| 1349 | 1432 | ||
| 1350 | BUG_ON((unsigned long)cwq & WORK_STRUCT_FLAG_MASK); | 1433 | BUG_ON((unsigned long)cwq & WORK_STRUCT_FLAG_MASK); |
| 1351 | cwq->wq = wq; | ||
| 1352 | cwq->cpu = cpu; | 1434 | cwq->cpu = cpu; |
| 1435 | cwq->wq = wq; | ||
| 1353 | cwq->flush_color = -1; | 1436 | cwq->flush_color = -1; |
| 1354 | spin_lock_init(&cwq->lock); | 1437 | spin_lock_init(&cwq->lock); |
| 1355 | INIT_LIST_HEAD(&cwq->worklist); | 1438 | INIT_LIST_HEAD(&cwq->worklist); |
| 1356 | init_waitqueue_head(&cwq->more_work); | 1439 | init_waitqueue_head(&cwq->more_work); |
| 1357 | 1440 | ||
| 1358 | if (err) | 1441 | if (failed) |
| 1359 | continue; | 1442 | continue; |
| 1360 | err = create_workqueue_thread(cwq, cpu); | 1443 | cwq->worker = create_worker(cwq, |
| 1361 | if (cpu_online(cpu) && !singlethread) | 1444 | cpu_online(cpu) && !singlethread); |
| 1362 | start_workqueue_thread(cwq, cpu); | 1445 | if (cwq->worker) |
| 1446 | start_worker(cwq->worker); | ||
| 1363 | else | 1447 | else |
| 1364 | start_workqueue_thread(cwq, -1); | 1448 | failed = true; |
| 1365 | } | 1449 | } |
| 1366 | 1450 | ||
| 1367 | spin_lock(&workqueue_lock); | 1451 | spin_lock(&workqueue_lock); |
| @@ -1370,7 +1454,7 @@ struct workqueue_struct *__create_workqueue_key(const char *name, | |||
| 1370 | 1454 | ||
| 1371 | cpu_maps_update_done(); | 1455 | cpu_maps_update_done(); |
| 1372 | 1456 | ||
| 1373 | if (err) { | 1457 | if (failed) { |
| 1374 | destroy_workqueue(wq); | 1458 | destroy_workqueue(wq); |
| 1375 | wq = NULL; | 1459 | wq = NULL; |
| 1376 | } | 1460 | } |
| @@ -1406,9 +1490,9 @@ void destroy_workqueue(struct workqueue_struct *wq) | |||
| 1406 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); | 1490 | struct cpu_workqueue_struct *cwq = get_cwq(cpu, wq); |
| 1407 | int i; | 1491 | int i; |
| 1408 | 1492 | ||
| 1409 | if (cwq->thread) { | 1493 | if (cwq->worker) { |
| 1410 | kthread_stop(cwq->thread); | 1494 | destroy_worker(cwq->worker); |
| 1411 | cwq->thread = NULL; | 1495 | cwq->worker = NULL; |
| 1412 | } | 1496 | } |
| 1413 | 1497 | ||
| 1414 | for (i = 0; i < WORK_NR_COLORS; i++) | 1498 | for (i = 0; i < WORK_NR_COLORS; i++) |
| @@ -1495,6 +1579,11 @@ EXPORT_SYMBOL_GPL(work_on_cpu); | |||
| 1495 | 1579 | ||
| 1496 | void __init init_workqueues(void) | 1580 | void __init init_workqueues(void) |
| 1497 | { | 1581 | { |
| 1582 | unsigned int cpu; | ||
| 1583 | |||
| 1584 | for_each_possible_cpu(cpu) | ||
| 1585 | ida_init(&per_cpu(worker_ida, cpu)); | ||
| 1586 | |||
| 1498 | singlethread_cpu = cpumask_first(cpu_possible_mask); | 1587 | singlethread_cpu = cpumask_first(cpu_possible_mask); |
| 1499 | hotcpu_notifier(workqueue_cpu_callback, 0); | 1588 | hotcpu_notifier(workqueue_cpu_callback, 0); |
| 1500 | keventd_wq = create_workqueue("events"); | 1589 | keventd_wq = create_workqueue("events"); |
