aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
authorBen Segall <bsegall@google.com>2013-10-16 14:16:22 -0400
committerIngo Molnar <mingo@kernel.org>2013-10-29 07:02:21 -0400
commit927b54fccbf04207ec92f669dce6806848cbec7d (patch)
tree4ee5a37f4796df9883c6beb9b7a91ca8c8534049 /kernel
parentdb06e78cc13d70f10877e0557becc88ab3ad2be8 (diff)
sched: Fix hrtimer_cancel()/rq->lock deadlock
__start_cfs_bandwidth calls hrtimer_cancel while holding rq->lock, waiting for the hrtimer to finish. However, if sched_cfs_period_timer runs for another loop iteration, the hrtimer can attempt to take rq->lock, resulting in deadlock. Fix this by ensuring that cfs_b->timer_active is cleared only if the _latest_ call to do_sched_cfs_period_timer is returning as idle. Then __start_cfs_bandwidth can just call hrtimer_try_to_cancel and wait for that to succeed or timer_active == 1. Signed-off-by: Ben Segall <bsegall@google.com> Signed-off-by: Peter Zijlstra <peterz@infradead.org> Cc: pjt@google.com Link: http://lkml.kernel.org/r/20131016181622.22647.16643.stgit@sword-of-the-dawn.mtv.corp.google.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched/fair.c15
1 files changed, 11 insertions, 4 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 897d97762d8a..f6308cb44d09 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -3225,6 +3225,13 @@ static int do_sched_cfs_period_timer(struct cfs_bandwidth *cfs_b, int overrun)
3225 if (idle) 3225 if (idle)
3226 goto out_unlock; 3226 goto out_unlock;
3227 3227
3228 /*
3229 * if we have relooped after returning idle once, we need to update our
3230 * status as actually running, so that other cpus doing
3231 * __start_cfs_bandwidth will stop trying to cancel us.
3232 */
3233 cfs_b->timer_active = 1;
3234
3228 __refill_cfs_bandwidth_runtime(cfs_b); 3235 __refill_cfs_bandwidth_runtime(cfs_b);
3229 3236
3230 if (!throttled) { 3237 if (!throttled) {
@@ -3493,11 +3500,11 @@ void __start_cfs_bandwidth(struct cfs_bandwidth *cfs_b)
3493 * (timer_active==0 becomes visible before the hrtimer call-back 3500 * (timer_active==0 becomes visible before the hrtimer call-back
3494 * terminates). In either case we ensure that it's re-programmed 3501 * terminates). In either case we ensure that it's re-programmed
3495 */ 3502 */
3496 while (unlikely(hrtimer_active(&cfs_b->period_timer))) { 3503 while (unlikely(hrtimer_active(&cfs_b->period_timer)) &&
3504 hrtimer_try_to_cancel(&cfs_b->period_timer) < 0) {
3505 /* bounce the lock to allow do_sched_cfs_period_timer to run */
3497 raw_spin_unlock(&cfs_b->lock); 3506 raw_spin_unlock(&cfs_b->lock);
3498 /* ensure cfs_b->lock is available while we wait */ 3507 cpu_relax();
3499 hrtimer_cancel(&cfs_b->period_timer);
3500
3501 raw_spin_lock(&cfs_b->lock); 3508 raw_spin_lock(&cfs_b->lock);
3502 /* if someone else restarted the timer then we're done */ 3509 /* if someone else restarted the timer then we're done */
3503 if (cfs_b->timer_active) 3510 if (cfs_b->timer_active)