diff options
author | Kirill Tkhai <ktkhai@parallels.com> | 2014-10-22 03:17:11 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2014-10-28 05:46:02 -0400 |
commit | 1effd9f19324efb05fccc7421530e11a52db0278 (patch) | |
tree | 8f7c0ec7af6435d605504ab5094e95ffab7d1001 /kernel | |
parent | aee38ea95419c818dfdde52b115aeffe9cbb259b (diff) |
sched/numa: Fix unsafe get_task_struct() in task_numa_assign()
Unlocked access to dst_rq->curr in task_numa_compare() is racy.
If curr task is exiting this may be a reason of use-after-free:
task_numa_compare() do_exit()
... current->flags |= PF_EXITING;
... release_task()
... ~~delayed_put_task_struct()~~
... schedule()
rcu_read_lock() ...
cur = ACCESS_ONCE(dst_rq->curr) ...
... rq->curr = next;
... context_switch()
... finish_task_switch()
... put_task_struct()
... __put_task_struct()
... free_task_struct()
task_numa_assign() ...
get_task_struct() ...
As noted by Oleg:
<<The lockless get_task_struct(tsk) is only safe if tsk == current
and didn't pass exit_notify(), or if this tsk was found on a rcu
protected list (say, for_each_process() or find_task_by_vpid()).
IOW, it is only safe if release_task() was not called before we
take rcu_read_lock(), in this case we can rely on the fact that
delayed_put_pid() can not drop the (potentially) last reference
until rcu_read_unlock().
And as Kirill pointed out task_numa_compare()->task_numa_assign()
path does get_task_struct(dst_rq->curr) and this is not safe. The
task_struct itself can't go away, but rcu_read_lock() can't save
us from the final put_task_struct() in finish_task_switch(); this
reference goes away without rcu gp>>
The patch provides simple check of PF_EXITING flag. If it's not set,
this guarantees that call_rcu() of delayed_put_task_struct() callback
hasn't happened yet, so we can safely do get_task_struct() in
task_numa_assign().
Locked dst_rq->lock protects from concurrency with the last schedule().
Reusing or unmapping of cur's memory may happen without it.
Suggested-by: Oleg Nesterov <oleg@redhat.com>
Signed-off-by: Kirill Tkhai <ktkhai@parallels.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Link: http://lkml.kernel.org/r/1413962231.19914.130.camel@tkhai
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/sched/fair.c | 14 |
1 files changed, 12 insertions, 2 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 0b069bf3e708..fbc0b8214af0 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -1164,9 +1164,19 @@ static void task_numa_compare(struct task_numa_env *env, | |||
1164 | long moveimp = imp; | 1164 | long moveimp = imp; |
1165 | 1165 | ||
1166 | rcu_read_lock(); | 1166 | rcu_read_lock(); |
1167 | cur = ACCESS_ONCE(dst_rq->curr); | 1167 | |
1168 | if (cur->pid == 0) /* idle */ | 1168 | raw_spin_lock_irq(&dst_rq->lock); |
1169 | cur = dst_rq->curr; | ||
1170 | /* | ||
1171 | * No need to move the exiting task, and this ensures that ->curr | ||
1172 | * wasn't reaped and thus get_task_struct() in task_numa_assign() | ||
1173 | * is safe under RCU read lock. | ||
1174 | * Note that rcu_read_lock() itself can't protect from the final | ||
1175 | * put_task_struct() after the last schedule(). | ||
1176 | */ | ||
1177 | if ((cur->flags & PF_EXITING) || is_idle_task(cur)) | ||
1169 | cur = NULL; | 1178 | cur = NULL; |
1179 | raw_spin_unlock_irq(&dst_rq->lock); | ||
1170 | 1180 | ||
1171 | /* | 1181 | /* |
1172 | * "imp" is the fault differential for the source task between the | 1182 | * "imp" is the fault differential for the source task between the |