diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2009-03-13 07:21:27 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-03-31 08:52:52 -0400 |
commit | 7f1e2ca9f04b02794597f60e7b1d43f0a1317939 (patch) | |
tree | dad1d80ab4232f3ca12d7ff9a58f0dc76133dab1 /kernel/softirq.c | |
parent | 7bee946358c3cb957d4aa648fc5ab3cad0b232d0 (diff) |
hrtimer: fix rq->lock inversion (again)
It appears I inadvertly introduced rq->lock recursion to the
hrtimer_start() path when I delegated running already expired
timers to softirq context.
This patch fixes it by introducing a __hrtimer_start_range_ns()
method that will not use raise_softirq_irqoff() but
__raise_softirq_irqoff() which avoids the wakeup.
It then also changes schedule() to check for pending softirqs and
do the wakeup then, I'm not quite sure I like this last bit, nor
am I convinced its really needed.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: paulus@samba.org
LKML-Reference: <20090313112301.096138802@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/softirq.c')
-rw-r--r-- | kernel/softirq.c | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/kernel/softirq.c b/kernel/softirq.c index 487751604300..accc85197c49 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c | |||
@@ -58,7 +58,7 @@ static DEFINE_PER_CPU(struct task_struct *, ksoftirqd); | |||
58 | * to the pending events, so lets the scheduler to balance | 58 | * to the pending events, so lets the scheduler to balance |
59 | * the softirq load for us. | 59 | * the softirq load for us. |
60 | */ | 60 | */ |
61 | static inline void wakeup_softirqd(void) | 61 | void wakeup_softirqd(void) |
62 | { | 62 | { |
63 | /* Interrupts are disabled: no need to stop preemption */ | 63 | /* Interrupts are disabled: no need to stop preemption */ |
64 | struct task_struct *tsk = __get_cpu_var(ksoftirqd); | 64 | struct task_struct *tsk = __get_cpu_var(ksoftirqd); |