diff options
author | Thomas Gleixner <tglx@linutronix.de> | 2009-11-17 08:28:38 -0500 |
---|---|---|
committer | Thomas Gleixner <tglx@linutronix.de> | 2009-12-14 17:55:33 -0500 |
commit | 05fa785cf80c9b7c0254c3056037147aed3ea16b (patch) | |
tree | 3d5c69d449b9240dc6d1005dddf344e467de4f34 /kernel/sched_rt.c | |
parent | a26724591edba5acc528d41f3906a972590e8f54 (diff) |
sched: Convert rq->lock to raw_spinlock
Convert locks which cannot be sleeping locks in preempt-rt to
raw_spinlocks.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_rt.c')
-rw-r--r-- | kernel/sched_rt.c | 14 |
1 files changed, 7 insertions, 7 deletions
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c index aecbd9c6b20c..a8325a7ff94c 100644 --- a/kernel/sched_rt.c +++ b/kernel/sched_rt.c | |||
@@ -454,9 +454,9 @@ static void disable_runtime(struct rq *rq) | |||
454 | { | 454 | { |
455 | unsigned long flags; | 455 | unsigned long flags; |
456 | 456 | ||
457 | spin_lock_irqsave(&rq->lock, flags); | 457 | raw_spin_lock_irqsave(&rq->lock, flags); |
458 | __disable_runtime(rq); | 458 | __disable_runtime(rq); |
459 | spin_unlock_irqrestore(&rq->lock, flags); | 459 | raw_spin_unlock_irqrestore(&rq->lock, flags); |
460 | } | 460 | } |
461 | 461 | ||
462 | static void __enable_runtime(struct rq *rq) | 462 | static void __enable_runtime(struct rq *rq) |
@@ -486,9 +486,9 @@ static void enable_runtime(struct rq *rq) | |||
486 | { | 486 | { |
487 | unsigned long flags; | 487 | unsigned long flags; |
488 | 488 | ||
489 | spin_lock_irqsave(&rq->lock, flags); | 489 | raw_spin_lock_irqsave(&rq->lock, flags); |
490 | __enable_runtime(rq); | 490 | __enable_runtime(rq); |
491 | spin_unlock_irqrestore(&rq->lock, flags); | 491 | raw_spin_unlock_irqrestore(&rq->lock, flags); |
492 | } | 492 | } |
493 | 493 | ||
494 | static int balance_runtime(struct rt_rq *rt_rq) | 494 | static int balance_runtime(struct rt_rq *rt_rq) |
@@ -524,7 +524,7 @@ static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun) | |||
524 | struct rt_rq *rt_rq = sched_rt_period_rt_rq(rt_b, i); | 524 | struct rt_rq *rt_rq = sched_rt_period_rt_rq(rt_b, i); |
525 | struct rq *rq = rq_of_rt_rq(rt_rq); | 525 | struct rq *rq = rq_of_rt_rq(rt_rq); |
526 | 526 | ||
527 | spin_lock(&rq->lock); | 527 | raw_spin_lock(&rq->lock); |
528 | if (rt_rq->rt_time) { | 528 | if (rt_rq->rt_time) { |
529 | u64 runtime; | 529 | u64 runtime; |
530 | 530 | ||
@@ -545,7 +545,7 @@ static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun) | |||
545 | 545 | ||
546 | if (enqueue) | 546 | if (enqueue) |
547 | sched_rt_rq_enqueue(rt_rq); | 547 | sched_rt_rq_enqueue(rt_rq); |
548 | spin_unlock(&rq->lock); | 548 | raw_spin_unlock(&rq->lock); |
549 | } | 549 | } |
550 | 550 | ||
551 | return idle; | 551 | return idle; |
@@ -1246,7 +1246,7 @@ static struct rq *find_lock_lowest_rq(struct task_struct *task, struct rq *rq) | |||
1246 | task_running(rq, task) || | 1246 | task_running(rq, task) || |
1247 | !task->se.on_rq)) { | 1247 | !task->se.on_rq)) { |
1248 | 1248 | ||
1249 | spin_unlock(&lowest_rq->lock); | 1249 | raw_spin_unlock(&lowest_rq->lock); |
1250 | lowest_rq = NULL; | 1250 | lowest_rq = NULL; |
1251 | break; | 1251 | break; |
1252 | } | 1252 | } |