diff options
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/audit_tree.c | 10 | ||||
-rw-r--r-- | kernel/audit_watch.c | 25 | ||||
-rw-r--r-- | kernel/cgroup.c | 6 | ||||
-rw-r--r-- | kernel/exit.c | 6 | ||||
-rw-r--r-- | kernel/fork.c | 2 | ||||
-rw-r--r-- | kernel/irq/manage.c | 6 | ||||
-rw-r--r-- | kernel/signal.c | 15 | ||||
-rw-r--r-- | kernel/task_work.c | 94 |
8 files changed, 77 insertions, 87 deletions
diff --git a/kernel/audit_tree.c b/kernel/audit_tree.c index 5bf0790497e7..3a5ca582ba1e 100644 --- a/kernel/audit_tree.c +++ b/kernel/audit_tree.c | |||
@@ -595,7 +595,7 @@ void audit_trim_trees(void) | |||
595 | 595 | ||
596 | root_mnt = collect_mounts(&path); | 596 | root_mnt = collect_mounts(&path); |
597 | path_put(&path); | 597 | path_put(&path); |
598 | if (!root_mnt) | 598 | if (IS_ERR(root_mnt)) |
599 | goto skip_it; | 599 | goto skip_it; |
600 | 600 | ||
601 | spin_lock(&hash_lock); | 601 | spin_lock(&hash_lock); |
@@ -669,8 +669,8 @@ int audit_add_tree_rule(struct audit_krule *rule) | |||
669 | goto Err; | 669 | goto Err; |
670 | mnt = collect_mounts(&path); | 670 | mnt = collect_mounts(&path); |
671 | path_put(&path); | 671 | path_put(&path); |
672 | if (!mnt) { | 672 | if (IS_ERR(mnt)) { |
673 | err = -ENOMEM; | 673 | err = PTR_ERR(mnt); |
674 | goto Err; | 674 | goto Err; |
675 | } | 675 | } |
676 | 676 | ||
@@ -719,8 +719,8 @@ int audit_tag_tree(char *old, char *new) | |||
719 | return err; | 719 | return err; |
720 | tagged = collect_mounts(&path2); | 720 | tagged = collect_mounts(&path2); |
721 | path_put(&path2); | 721 | path_put(&path2); |
722 | if (!tagged) | 722 | if (IS_ERR(tagged)) |
723 | return -ENOMEM; | 723 | return PTR_ERR(tagged); |
724 | 724 | ||
725 | err = kern_path(old, 0, &path1); | 725 | err = kern_path(old, 0, &path1); |
726 | if (err) { | 726 | if (err) { |
diff --git a/kernel/audit_watch.c b/kernel/audit_watch.c index e683869365d9..3823281401b5 100644 --- a/kernel/audit_watch.c +++ b/kernel/audit_watch.c | |||
@@ -355,34 +355,15 @@ static void audit_remove_parent_watches(struct audit_parent *parent) | |||
355 | /* Get path information necessary for adding watches. */ | 355 | /* Get path information necessary for adding watches. */ |
356 | static int audit_get_nd(struct audit_watch *watch, struct path *parent) | 356 | static int audit_get_nd(struct audit_watch *watch, struct path *parent) |
357 | { | 357 | { |
358 | struct nameidata nd; | 358 | struct dentry *d = kern_path_locked(watch->path, parent); |
359 | struct dentry *d; | 359 | if (IS_ERR(d)) |
360 | int err; | ||
361 | |||
362 | err = kern_path_parent(watch->path, &nd); | ||
363 | if (err) | ||
364 | return err; | ||
365 | |||
366 | if (nd.last_type != LAST_NORM) { | ||
367 | path_put(&nd.path); | ||
368 | return -EINVAL; | ||
369 | } | ||
370 | |||
371 | mutex_lock_nested(&nd.path.dentry->d_inode->i_mutex, I_MUTEX_PARENT); | ||
372 | d = lookup_one_len(nd.last.name, nd.path.dentry, nd.last.len); | ||
373 | if (IS_ERR(d)) { | ||
374 | mutex_unlock(&nd.path.dentry->d_inode->i_mutex); | ||
375 | path_put(&nd.path); | ||
376 | return PTR_ERR(d); | 360 | return PTR_ERR(d); |
377 | } | 361 | mutex_unlock(&parent->dentry->d_inode->i_mutex); |
378 | if (d->d_inode) { | 362 | if (d->d_inode) { |
379 | /* update watch filter fields */ | 363 | /* update watch filter fields */ |
380 | watch->dev = d->d_inode->i_sb->s_dev; | 364 | watch->dev = d->d_inode->i_sb->s_dev; |
381 | watch->ino = d->d_inode->i_ino; | 365 | watch->ino = d->d_inode->i_ino; |
382 | } | 366 | } |
383 | mutex_unlock(&nd.path.dentry->d_inode->i_mutex); | ||
384 | |||
385 | *parent = nd.path; | ||
386 | dput(d); | 367 | dput(d); |
387 | return 0; | 368 | return 0; |
388 | } | 369 | } |
diff --git a/kernel/cgroup.c b/kernel/cgroup.c index b303dfc7dce0..af2b5641fc8b 100644 --- a/kernel/cgroup.c +++ b/kernel/cgroup.c | |||
@@ -822,7 +822,7 @@ EXPORT_SYMBOL_GPL(cgroup_unlock); | |||
822 | */ | 822 | */ |
823 | 823 | ||
824 | static int cgroup_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode); | 824 | static int cgroup_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode); |
825 | static struct dentry *cgroup_lookup(struct inode *, struct dentry *, struct nameidata *); | 825 | static struct dentry *cgroup_lookup(struct inode *, struct dentry *, unsigned int); |
826 | static int cgroup_rmdir(struct inode *unused_dir, struct dentry *dentry); | 826 | static int cgroup_rmdir(struct inode *unused_dir, struct dentry *dentry); |
827 | static int cgroup_populate_dir(struct cgroup *cgrp); | 827 | static int cgroup_populate_dir(struct cgroup *cgrp); |
828 | static const struct inode_operations cgroup_dir_inode_operations; | 828 | static const struct inode_operations cgroup_dir_inode_operations; |
@@ -1587,7 +1587,7 @@ static struct dentry *cgroup_mount(struct file_system_type *fs_type, | |||
1587 | opts.new_root = new_root; | 1587 | opts.new_root = new_root; |
1588 | 1588 | ||
1589 | /* Locate an existing or new sb for this hierarchy */ | 1589 | /* Locate an existing or new sb for this hierarchy */ |
1590 | sb = sget(fs_type, cgroup_test_super, cgroup_set_super, &opts); | 1590 | sb = sget(fs_type, cgroup_test_super, cgroup_set_super, 0, &opts); |
1591 | if (IS_ERR(sb)) { | 1591 | if (IS_ERR(sb)) { |
1592 | ret = PTR_ERR(sb); | 1592 | ret = PTR_ERR(sb); |
1593 | cgroup_drop_root(opts.new_root); | 1593 | cgroup_drop_root(opts.new_root); |
@@ -2570,7 +2570,7 @@ static const struct inode_operations cgroup_dir_inode_operations = { | |||
2570 | .rename = cgroup_rename, | 2570 | .rename = cgroup_rename, |
2571 | }; | 2571 | }; |
2572 | 2572 | ||
2573 | static struct dentry *cgroup_lookup(struct inode *dir, struct dentry *dentry, struct nameidata *nd) | 2573 | static struct dentry *cgroup_lookup(struct inode *dir, struct dentry *dentry, unsigned int flags) |
2574 | { | 2574 | { |
2575 | if (dentry->d_name.len > NAME_MAX) | 2575 | if (dentry->d_name.len > NAME_MAX) |
2576 | return ERR_PTR(-ENAMETOOLONG); | 2576 | return ERR_PTR(-ENAMETOOLONG); |
diff --git a/kernel/exit.c b/kernel/exit.c index 2f59cc334516..d17f6c4ddfa9 100644 --- a/kernel/exit.c +++ b/kernel/exit.c | |||
@@ -953,14 +953,11 @@ void do_exit(long code) | |||
953 | exit_signals(tsk); /* sets PF_EXITING */ | 953 | exit_signals(tsk); /* sets PF_EXITING */ |
954 | /* | 954 | /* |
955 | * tsk->flags are checked in the futex code to protect against | 955 | * tsk->flags are checked in the futex code to protect against |
956 | * an exiting task cleaning up the robust pi futexes, and in | 956 | * an exiting task cleaning up the robust pi futexes. |
957 | * task_work_add() to avoid the race with exit_task_work(). | ||
958 | */ | 957 | */ |
959 | smp_mb(); | 958 | smp_mb(); |
960 | raw_spin_unlock_wait(&tsk->pi_lock); | 959 | raw_spin_unlock_wait(&tsk->pi_lock); |
961 | 960 | ||
962 | exit_task_work(tsk); | ||
963 | |||
964 | if (unlikely(in_atomic())) | 961 | if (unlikely(in_atomic())) |
965 | printk(KERN_INFO "note: %s[%d] exited with preempt_count %d\n", | 962 | printk(KERN_INFO "note: %s[%d] exited with preempt_count %d\n", |
966 | current->comm, task_pid_nr(current), | 963 | current->comm, task_pid_nr(current), |
@@ -995,6 +992,7 @@ void do_exit(long code) | |||
995 | exit_shm(tsk); | 992 | exit_shm(tsk); |
996 | exit_files(tsk); | 993 | exit_files(tsk); |
997 | exit_fs(tsk); | 994 | exit_fs(tsk); |
995 | exit_task_work(tsk); | ||
998 | check_stack_usage(); | 996 | check_stack_usage(); |
999 | exit_thread(); | 997 | exit_thread(); |
1000 | 998 | ||
diff --git a/kernel/fork.c b/kernel/fork.c index f00e319d8376..ff1cad3b7bdc 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
@@ -1420,7 +1420,7 @@ static struct task_struct *copy_process(unsigned long clone_flags, | |||
1420 | */ | 1420 | */ |
1421 | p->group_leader = p; | 1421 | p->group_leader = p; |
1422 | INIT_LIST_HEAD(&p->thread_group); | 1422 | INIT_LIST_HEAD(&p->thread_group); |
1423 | INIT_HLIST_HEAD(&p->task_works); | 1423 | p->task_works = NULL; |
1424 | 1424 | ||
1425 | /* Now that the task is set up, run cgroup callbacks if | 1425 | /* Now that the task is set up, run cgroup callbacks if |
1426 | * necessary. We need to run them before the task is visible | 1426 | * necessary. We need to run them before the task is visible |
diff --git a/kernel/irq/manage.c b/kernel/irq/manage.c index 8c548232ba39..814c9ef6bba1 100644 --- a/kernel/irq/manage.c +++ b/kernel/irq/manage.c | |||
@@ -781,7 +781,7 @@ static void wake_threads_waitq(struct irq_desc *desc) | |||
781 | wake_up(&desc->wait_for_threads); | 781 | wake_up(&desc->wait_for_threads); |
782 | } | 782 | } |
783 | 783 | ||
784 | static void irq_thread_dtor(struct task_work *unused) | 784 | static void irq_thread_dtor(struct callback_head *unused) |
785 | { | 785 | { |
786 | struct task_struct *tsk = current; | 786 | struct task_struct *tsk = current; |
787 | struct irq_desc *desc; | 787 | struct irq_desc *desc; |
@@ -813,7 +813,7 @@ static void irq_thread_dtor(struct task_work *unused) | |||
813 | */ | 813 | */ |
814 | static int irq_thread(void *data) | 814 | static int irq_thread(void *data) |
815 | { | 815 | { |
816 | struct task_work on_exit_work; | 816 | struct callback_head on_exit_work; |
817 | static const struct sched_param param = { | 817 | static const struct sched_param param = { |
818 | .sched_priority = MAX_USER_RT_PRIO/2, | 818 | .sched_priority = MAX_USER_RT_PRIO/2, |
819 | }; | 819 | }; |
@@ -830,7 +830,7 @@ static int irq_thread(void *data) | |||
830 | 830 | ||
831 | sched_setscheduler(current, SCHED_FIFO, ¶m); | 831 | sched_setscheduler(current, SCHED_FIFO, ¶m); |
832 | 832 | ||
833 | init_task_work(&on_exit_work, irq_thread_dtor, NULL); | 833 | init_task_work(&on_exit_work, irq_thread_dtor); |
834 | task_work_add(current, &on_exit_work, false); | 834 | task_work_add(current, &on_exit_work, false); |
835 | 835 | ||
836 | while (!irq_wait_for_interrupt(action)) { | 836 | while (!irq_wait_for_interrupt(action)) { |
diff --git a/kernel/signal.c b/kernel/signal.c index 677102789cf2..be4f856d52f8 100644 --- a/kernel/signal.c +++ b/kernel/signal.c | |||
@@ -1971,6 +1971,13 @@ static void ptrace_do_notify(int signr, int exit_code, int why) | |||
1971 | void ptrace_notify(int exit_code) | 1971 | void ptrace_notify(int exit_code) |
1972 | { | 1972 | { |
1973 | BUG_ON((exit_code & (0x7f | ~0xffff)) != SIGTRAP); | 1973 | BUG_ON((exit_code & (0x7f | ~0xffff)) != SIGTRAP); |
1974 | if (unlikely(current->task_works)) { | ||
1975 | if (test_and_clear_ti_thread_flag(current_thread_info(), | ||
1976 | TIF_NOTIFY_RESUME)) { | ||
1977 | smp_mb__after_clear_bit(); | ||
1978 | task_work_run(); | ||
1979 | } | ||
1980 | } | ||
1974 | 1981 | ||
1975 | spin_lock_irq(¤t->sighand->siglock); | 1982 | spin_lock_irq(¤t->sighand->siglock); |
1976 | ptrace_do_notify(SIGTRAP, exit_code, CLD_TRAPPED); | 1983 | ptrace_do_notify(SIGTRAP, exit_code, CLD_TRAPPED); |
@@ -2191,6 +2198,14 @@ int get_signal_to_deliver(siginfo_t *info, struct k_sigaction *return_ka, | |||
2191 | struct signal_struct *signal = current->signal; | 2198 | struct signal_struct *signal = current->signal; |
2192 | int signr; | 2199 | int signr; |
2193 | 2200 | ||
2201 | if (unlikely(current->task_works)) { | ||
2202 | if (test_and_clear_ti_thread_flag(current_thread_info(), | ||
2203 | TIF_NOTIFY_RESUME)) { | ||
2204 | smp_mb__after_clear_bit(); | ||
2205 | task_work_run(); | ||
2206 | } | ||
2207 | } | ||
2208 | |||
2194 | if (unlikely(uprobe_deny_signal())) | 2209 | if (unlikely(uprobe_deny_signal())) |
2195 | return 0; | 2210 | return 0; |
2196 | 2211 | ||
diff --git a/kernel/task_work.c b/kernel/task_work.c index 82d1c794066d..91d4e1742a0c 100644 --- a/kernel/task_work.c +++ b/kernel/task_work.c | |||
@@ -3,82 +3,78 @@ | |||
3 | #include <linux/tracehook.h> | 3 | #include <linux/tracehook.h> |
4 | 4 | ||
5 | int | 5 | int |
6 | task_work_add(struct task_struct *task, struct task_work *twork, bool notify) | 6 | task_work_add(struct task_struct *task, struct callback_head *twork, bool notify) |
7 | { | 7 | { |
8 | struct callback_head *last, *first; | ||
8 | unsigned long flags; | 9 | unsigned long flags; |
9 | int err = -ESRCH; | ||
10 | 10 | ||
11 | #ifndef TIF_NOTIFY_RESUME | ||
12 | if (notify) | ||
13 | return -ENOTSUPP; | ||
14 | #endif | ||
15 | /* | 11 | /* |
16 | * We must not insert the new work if the task has already passed | 12 | * Not inserting the new work if the task has already passed |
17 | * exit_task_work(). We rely on do_exit()->raw_spin_unlock_wait() | 13 | * exit_task_work() is the responisbility of callers. |
18 | * and check PF_EXITING under pi_lock. | ||
19 | */ | 14 | */ |
20 | raw_spin_lock_irqsave(&task->pi_lock, flags); | 15 | raw_spin_lock_irqsave(&task->pi_lock, flags); |
21 | if (likely(!(task->flags & PF_EXITING))) { | 16 | last = task->task_works; |
22 | hlist_add_head(&twork->hlist, &task->task_works); | 17 | first = last ? last->next : twork; |
23 | err = 0; | 18 | twork->next = first; |
24 | } | 19 | if (last) |
20 | last->next = twork; | ||
21 | task->task_works = twork; | ||
25 | raw_spin_unlock_irqrestore(&task->pi_lock, flags); | 22 | raw_spin_unlock_irqrestore(&task->pi_lock, flags); |
26 | 23 | ||
27 | /* test_and_set_bit() implies mb(), see tracehook_notify_resume(). */ | 24 | /* test_and_set_bit() implies mb(), see tracehook_notify_resume(). */ |
28 | if (likely(!err) && notify) | 25 | if (notify) |
29 | set_notify_resume(task); | 26 | set_notify_resume(task); |
30 | return err; | 27 | return 0; |
31 | } | 28 | } |
32 | 29 | ||
33 | struct task_work * | 30 | struct callback_head * |
34 | task_work_cancel(struct task_struct *task, task_work_func_t func) | 31 | task_work_cancel(struct task_struct *task, task_work_func_t func) |
35 | { | 32 | { |
36 | unsigned long flags; | 33 | unsigned long flags; |
37 | struct task_work *twork; | 34 | struct callback_head *last, *res = NULL; |
38 | struct hlist_node *pos; | ||
39 | 35 | ||
40 | raw_spin_lock_irqsave(&task->pi_lock, flags); | 36 | raw_spin_lock_irqsave(&task->pi_lock, flags); |
41 | hlist_for_each_entry(twork, pos, &task->task_works, hlist) { | 37 | last = task->task_works; |
42 | if (twork->func == func) { | 38 | if (last) { |
43 | hlist_del(&twork->hlist); | 39 | struct callback_head *q = last, *p = q->next; |
44 | goto found; | 40 | while (1) { |
41 | if (p->func == func) { | ||
42 | q->next = p->next; | ||
43 | if (p == last) | ||
44 | task->task_works = q == p ? NULL : q; | ||
45 | res = p; | ||
46 | break; | ||
47 | } | ||
48 | if (p == last) | ||
49 | break; | ||
50 | q = p; | ||
51 | p = q->next; | ||
45 | } | 52 | } |
46 | } | 53 | } |
47 | twork = NULL; | ||
48 | found: | ||
49 | raw_spin_unlock_irqrestore(&task->pi_lock, flags); | 54 | raw_spin_unlock_irqrestore(&task->pi_lock, flags); |
50 | 55 | return res; | |
51 | return twork; | ||
52 | } | 56 | } |
53 | 57 | ||
54 | void task_work_run(void) | 58 | void task_work_run(void) |
55 | { | 59 | { |
56 | struct task_struct *task = current; | 60 | struct task_struct *task = current; |
57 | struct hlist_head task_works; | 61 | struct callback_head *p, *q; |
58 | struct hlist_node *pos; | ||
59 | 62 | ||
60 | raw_spin_lock_irq(&task->pi_lock); | 63 | while (1) { |
61 | hlist_move_list(&task->task_works, &task_works); | 64 | raw_spin_lock_irq(&task->pi_lock); |
62 | raw_spin_unlock_irq(&task->pi_lock); | 65 | p = task->task_works; |
66 | task->task_works = NULL; | ||
67 | raw_spin_unlock_irq(&task->pi_lock); | ||
63 | 68 | ||
64 | if (unlikely(hlist_empty(&task_works))) | 69 | if (unlikely(!p)) |
65 | return; | 70 | return; |
66 | /* | ||
67 | * We use hlist to save the space in task_struct, but we want fifo. | ||
68 | * Find the last entry, the list should be short, then process them | ||
69 | * in reverse order. | ||
70 | */ | ||
71 | for (pos = task_works.first; pos->next; pos = pos->next) | ||
72 | ; | ||
73 | 71 | ||
74 | for (;;) { | 72 | q = p->next; /* head */ |
75 | struct hlist_node **pprev = pos->pprev; | 73 | p->next = NULL; /* cut it */ |
76 | struct task_work *twork = container_of(pos, struct task_work, | 74 | while (q) { |
77 | hlist); | 75 | p = q->next; |
78 | twork->func(twork); | 76 | q->func(q); |
79 | 77 | q = p; | |
80 | if (pprev == &task_works.first) | 78 | } |
81 | break; | ||
82 | pos = container_of(pprev, struct hlist_node, next); | ||
83 | } | 79 | } |
84 | } | 80 | } |