aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--include/linux/task_work.h3
-rw-r--r--kernel/task_work.c22
2 files changed, 17 insertions, 8 deletions
diff --git a/include/linux/task_work.h b/include/linux/task_work.h
index fb46b03b1852..ca5a1cf27dae 100644
--- a/include/linux/task_work.h
+++ b/include/linux/task_work.h
@@ -18,8 +18,7 @@ void task_work_run(void);
18 18
19static inline void exit_task_work(struct task_struct *task) 19static inline void exit_task_work(struct task_struct *task)
20{ 20{
21 if (unlikely(task->task_works)) 21 task_work_run();
22 task_work_run();
23} 22}
24 23
25#endif /* _LINUX_TASK_WORK_H */ 24#endif /* _LINUX_TASK_WORK_H */
diff --git a/kernel/task_work.c b/kernel/task_work.c
index f13ec0bda1d5..65bd3c92d6f3 100644
--- a/kernel/task_work.c
+++ b/kernel/task_work.c
@@ -2,16 +2,17 @@
2#include <linux/task_work.h> 2#include <linux/task_work.h>
3#include <linux/tracehook.h> 3#include <linux/tracehook.h>
4 4
5static struct callback_head work_exited; /* all we need is ->next == NULL */
6
5int 7int
6task_work_add(struct task_struct *task, struct callback_head *work, bool notify) 8task_work_add(struct task_struct *task, struct callback_head *work, bool notify)
7{ 9{
8 struct callback_head *head; 10 struct callback_head *head;
9 /* 11
10 * Not inserting the new work if the task has already passed
11 * exit_task_work() is the responisbility of callers.
12 */
13 do { 12 do {
14 head = ACCESS_ONCE(task->task_works); 13 head = ACCESS_ONCE(task->task_works);
14 if (unlikely(head == &work_exited))
15 return -ESRCH;
15 work->next = head; 16 work->next = head;
16 } while (cmpxchg(&task->task_works, head, work) != head); 17 } while (cmpxchg(&task->task_works, head, work) != head);
17 18
@@ -30,7 +31,7 @@ task_work_cancel(struct task_struct *task, task_work_func_t func)
30 * If cmpxchg() fails we continue without updating pprev. 31 * If cmpxchg() fails we continue without updating pprev.
31 * Either we raced with task_work_add() which added the 32 * Either we raced with task_work_add() which added the
32 * new entry before this work, we will find it again. Or 33 * new entry before this work, we will find it again. Or
33 * we raced with task_work_run(), *pprev == NULL. 34 * we raced with task_work_run(), *pprev == NULL/exited.
34 */ 35 */
35 raw_spin_lock_irqsave(&task->pi_lock, flags); 36 raw_spin_lock_irqsave(&task->pi_lock, flags);
36 while ((work = ACCESS_ONCE(*pprev))) { 37 while ((work = ACCESS_ONCE(*pprev))) {
@@ -51,7 +52,16 @@ void task_work_run(void)
51 struct callback_head *work, *head, *next; 52 struct callback_head *work, *head, *next;
52 53
53 for (;;) { 54 for (;;) {
54 work = xchg(&task->task_works, NULL); 55 /*
56 * work->func() can do task_work_add(), do not set
57 * work_exited unless the list is empty.
58 */
59 do {
60 work = ACCESS_ONCE(task->task_works);
61 head = !work && (task->flags & PF_EXITING) ?
62 &work_exited : NULL;
63 } while (cmpxchg(&task->task_works, work, head) != work);
64
55 if (!work) 65 if (!work)
56 break; 66 break;
57 /* 67 /*