aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
authorPeter Zijlstra <peterz@infradead.org>2017-01-11 15:09:50 -0500
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>2017-02-01 02:33:14 -0500
commit922813f4d66fb317e8602d058d03a1619af1ffd0 (patch)
tree627cf3a45925f8b01d8bdc12c2fda2f85a19703e /kernel
parentf5f415c13209c5a7a71da85462815288321483df (diff)
perf/core: Fix concurrent sys_perf_event_open() vs. 'move_group' race
commit 321027c1fe77f892f4ea07846aeae08cefbbb290 upstream. Di Shen reported a race between two concurrent sys_perf_event_open() calls where both try and move the same pre-existing software group into a hardware context. The problem is exactly that described in commit: f63a8daa5812 ("perf: Fix event->ctx locking") ... where, while we wait for a ctx->mutex acquisition, the event->ctx relation can have changed under us. That very same commit failed to recognise sys_perf_event_context() as an external access vector to the events and thereby didn't apply the established locking rules correctly. So while one sys_perf_event_open() call is stuck waiting on mutex_lock_double(), the other (which owns said locks) moves the group about. So by the time the former sys_perf_event_open() acquires the locks, the context we've acquired is stale (and possibly dead). Apply the established locking rules as per perf_event_ctx_lock_nested() to the mutex_lock_double() for the 'move_group' case. This obviously means we need to validate state after we acquire the locks. Reported-by: Di Shen (Keen Lab) Tested-by: John Dias <joaodias@google.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com> Cc: Arnaldo Carvalho de Melo <acme@kernel.org> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Cc: Jiri Olsa <jolsa@redhat.com> Cc: Kees Cook <keescook@chromium.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Min Chong <mchong@google.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Stephane Eranian <eranian@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vince Weaver <vincent.weaver@maine.edu> Fixes: f63a8daa5812 ("perf: Fix event->ctx locking") Link: http://lkml.kernel.org/r/20170106131444.GZ3174@twins.programming.kicks-ass.net Signed-off-by: Ingo Molnar <mingo@kernel.org> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/events/core.c58
1 files changed, 54 insertions, 4 deletions
diff --git a/kernel/events/core.c b/kernel/events/core.c
index 02c8421f8c01..e5a8839e7076 100644
--- a/kernel/events/core.c
+++ b/kernel/events/core.c
@@ -9503,6 +9503,37 @@ static int perf_event_set_clock(struct perf_event *event, clockid_t clk_id)
9503 return 0; 9503 return 0;
9504} 9504}
9505 9505
9506/*
9507 * Variation on perf_event_ctx_lock_nested(), except we take two context
9508 * mutexes.
9509 */
9510static struct perf_event_context *
9511__perf_event_ctx_lock_double(struct perf_event *group_leader,
9512 struct perf_event_context *ctx)
9513{
9514 struct perf_event_context *gctx;
9515
9516again:
9517 rcu_read_lock();
9518 gctx = READ_ONCE(group_leader->ctx);
9519 if (!atomic_inc_not_zero(&gctx->refcount)) {
9520 rcu_read_unlock();
9521 goto again;
9522 }
9523 rcu_read_unlock();
9524
9525 mutex_lock_double(&gctx->mutex, &ctx->mutex);
9526
9527 if (group_leader->ctx != gctx) {
9528 mutex_unlock(&ctx->mutex);
9529 mutex_unlock(&gctx->mutex);
9530 put_ctx(gctx);
9531 goto again;
9532 }
9533
9534 return gctx;
9535}
9536
9506/** 9537/**
9507 * sys_perf_event_open - open a performance event, associate it to a task/cpu 9538 * sys_perf_event_open - open a performance event, associate it to a task/cpu
9508 * 9539 *
@@ -9746,12 +9777,31 @@ SYSCALL_DEFINE5(perf_event_open,
9746 } 9777 }
9747 9778
9748 if (move_group) { 9779 if (move_group) {
9749 gctx = group_leader->ctx; 9780 gctx = __perf_event_ctx_lock_double(group_leader, ctx);
9750 mutex_lock_double(&gctx->mutex, &ctx->mutex); 9781
9751 if (gctx->task == TASK_TOMBSTONE) { 9782 if (gctx->task == TASK_TOMBSTONE) {
9752 err = -ESRCH; 9783 err = -ESRCH;
9753 goto err_locked; 9784 goto err_locked;
9754 } 9785 }
9786
9787 /*
9788 * Check if we raced against another sys_perf_event_open() call
9789 * moving the software group underneath us.
9790 */
9791 if (!(group_leader->group_caps & PERF_EV_CAP_SOFTWARE)) {
9792 /*
9793 * If someone moved the group out from under us, check
9794 * if this new event wound up on the same ctx, if so
9795 * its the regular !move_group case, otherwise fail.
9796 */
9797 if (gctx != ctx) {
9798 err = -EINVAL;
9799 goto err_locked;
9800 } else {
9801 perf_event_ctx_unlock(group_leader, gctx);
9802 move_group = 0;
9803 }
9804 }
9755 } else { 9805 } else {
9756 mutex_lock(&ctx->mutex); 9806 mutex_lock(&ctx->mutex);
9757 } 9807 }
@@ -9853,7 +9903,7 @@ SYSCALL_DEFINE5(perf_event_open,
9853 perf_unpin_context(ctx); 9903 perf_unpin_context(ctx);
9854 9904
9855 if (move_group) 9905 if (move_group)
9856 mutex_unlock(&gctx->mutex); 9906 perf_event_ctx_unlock(group_leader, gctx);
9857 mutex_unlock(&ctx->mutex); 9907 mutex_unlock(&ctx->mutex);
9858 9908
9859 if (task) { 9909 if (task) {
@@ -9879,7 +9929,7 @@ SYSCALL_DEFINE5(perf_event_open,
9879 9929
9880err_locked: 9930err_locked:
9881 if (move_group) 9931 if (move_group)
9882 mutex_unlock(&gctx->mutex); 9932 perf_event_ctx_unlock(group_leader, gctx);
9883 mutex_unlock(&ctx->mutex); 9933 mutex_unlock(&ctx->mutex);
9884/* err_file: */ 9934/* err_file: */
9885 fput(event_file); 9935 fput(event_file);