diff options
author | Andrew Morton <akpm@linux-foundation.org> | 2009-04-09 11:50:37 -0400 |
---|---|---|
committer | Rusty Russell <rusty@rustcorp.com.au> | 2009-04-08 20:20:37 -0400 |
commit | 6b44003e5ca66a3fffeb5bc90f40ada2c4340896 (patch) | |
tree | d9dce0a39b5e66d8e760344a51ffb6de9594cd2f /kernel | |
parent | 1c99315bb36b5d776210546d438ca928dc9b1f22 (diff) |
work_on_cpu(): rewrite it to create a kernel thread on demand
Impact: circular locking bugfix
The various implemetnations and proposed implemetnations of work_on_cpu()
are vulnerable to various deadlocks because they all used queues of some
form.
Unrelated pieces of kernel code thus gained dependencies wherein if one
work_on_cpu() caller holds a lock which some other work_on_cpu() callback
also takes, the kernel could rarely deadlock.
Fix this by creating a short-lived kernel thread for each work_on_cpu()
invokation.
This is not terribly fast, but the only current caller of work_on_cpu() is
pci_call_probe().
It would be nice to find some other way of doing the node-local
allocations in the PCI probe code so that we can zap work_on_cpu()
altogether. The code there is rather nasty. I can't think of anything
simple at this time...
Cc: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/workqueue.c | 36 |
1 files changed, 19 insertions, 17 deletions
diff --git a/kernel/workqueue.c b/kernel/workqueue.c index b6b966ce1451..f71fb2a08950 100644 --- a/kernel/workqueue.c +++ b/kernel/workqueue.c | |||
@@ -966,20 +966,20 @@ undo: | |||
966 | } | 966 | } |
967 | 967 | ||
968 | #ifdef CONFIG_SMP | 968 | #ifdef CONFIG_SMP |
969 | static struct workqueue_struct *work_on_cpu_wq __read_mostly; | ||
970 | 969 | ||
971 | struct work_for_cpu { | 970 | struct work_for_cpu { |
972 | struct work_struct work; | 971 | struct completion completion; |
973 | long (*fn)(void *); | 972 | long (*fn)(void *); |
974 | void *arg; | 973 | void *arg; |
975 | long ret; | 974 | long ret; |
976 | }; | 975 | }; |
977 | 976 | ||
978 | static void do_work_for_cpu(struct work_struct *w) | 977 | static int do_work_for_cpu(void *_wfc) |
979 | { | 978 | { |
980 | struct work_for_cpu *wfc = container_of(w, struct work_for_cpu, work); | 979 | struct work_for_cpu *wfc = _wfc; |
981 | |||
982 | wfc->ret = wfc->fn(wfc->arg); | 980 | wfc->ret = wfc->fn(wfc->arg); |
981 | complete(&wfc->completion); | ||
982 | return 0; | ||
983 | } | 983 | } |
984 | 984 | ||
985 | /** | 985 | /** |
@@ -990,17 +990,23 @@ static void do_work_for_cpu(struct work_struct *w) | |||
990 | * | 990 | * |
991 | * This will return the value @fn returns. | 991 | * This will return the value @fn returns. |
992 | * It is up to the caller to ensure that the cpu doesn't go offline. | 992 | * It is up to the caller to ensure that the cpu doesn't go offline. |
993 | * The caller must not hold any locks which would prevent @fn from completing. | ||
993 | */ | 994 | */ |
994 | long work_on_cpu(unsigned int cpu, long (*fn)(void *), void *arg) | 995 | long work_on_cpu(unsigned int cpu, long (*fn)(void *), void *arg) |
995 | { | 996 | { |
996 | struct work_for_cpu wfc; | 997 | struct task_struct *sub_thread; |
997 | 998 | struct work_for_cpu wfc = { | |
998 | INIT_WORK(&wfc.work, do_work_for_cpu); | 999 | .completion = COMPLETION_INITIALIZER_ONSTACK(wfc.completion), |
999 | wfc.fn = fn; | 1000 | .fn = fn, |
1000 | wfc.arg = arg; | 1001 | .arg = arg, |
1001 | queue_work_on(cpu, work_on_cpu_wq, &wfc.work); | 1002 | }; |
1002 | flush_work(&wfc.work); | 1003 | |
1003 | 1004 | sub_thread = kthread_create(do_work_for_cpu, &wfc, "work_for_cpu"); | |
1005 | if (IS_ERR(sub_thread)) | ||
1006 | return PTR_ERR(sub_thread); | ||
1007 | kthread_bind(sub_thread, cpu); | ||
1008 | wake_up_process(sub_thread); | ||
1009 | wait_for_completion(&wfc.completion); | ||
1004 | return wfc.ret; | 1010 | return wfc.ret; |
1005 | } | 1011 | } |
1006 | EXPORT_SYMBOL_GPL(work_on_cpu); | 1012 | EXPORT_SYMBOL_GPL(work_on_cpu); |
@@ -1016,8 +1022,4 @@ void __init init_workqueues(void) | |||
1016 | hotcpu_notifier(workqueue_cpu_callback, 0); | 1022 | hotcpu_notifier(workqueue_cpu_callback, 0); |
1017 | keventd_wq = create_workqueue("events"); | 1023 | keventd_wq = create_workqueue("events"); |
1018 | BUG_ON(!keventd_wq); | 1024 | BUG_ON(!keventd_wq); |
1019 | #ifdef CONFIG_SMP | ||
1020 | work_on_cpu_wq = create_workqueue("work_on_cpu"); | ||
1021 | BUG_ON(!work_on_cpu_wq); | ||
1022 | #endif | ||
1023 | } | 1025 | } |