diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2009-01-14 09:36:26 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-01-14 12:09:00 -0500 |
commit | 41719b03091911028116155deddc5eedf8c45e37 (patch) | |
tree | 20a699807d78bc0af86b19443dc751415c0cc6f7 | |
parent | 93d81d1aca26e64a75d06a85f7e128b5f49053e7 (diff) |
mutex: preemption fixes
The problem is that dropping the spinlock right before schedule is a voluntary
preemption point and can cause a schedule, right after which we schedule again.
Fix this inefficiency by keeping preemption disabled until we schedule, do this
by explicity disabling preemption and providing a schedule() variant that
assumes preemption is already disabled.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r-- | include/linux/sched.h | 1 | ||||
-rw-r--r-- | kernel/mutex.c | 5 | ||||
-rw-r--r-- | kernel/sched.c | 10 |
3 files changed, 12 insertions, 4 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index 4cae9b81a1f8..9f0b372cfa6f 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
@@ -328,6 +328,7 @@ extern signed long schedule_timeout(signed long timeout); | |||
328 | extern signed long schedule_timeout_interruptible(signed long timeout); | 328 | extern signed long schedule_timeout_interruptible(signed long timeout); |
329 | extern signed long schedule_timeout_killable(signed long timeout); | 329 | extern signed long schedule_timeout_killable(signed long timeout); |
330 | extern signed long schedule_timeout_uninterruptible(signed long timeout); | 330 | extern signed long schedule_timeout_uninterruptible(signed long timeout); |
331 | asmlinkage void __schedule(void); | ||
331 | asmlinkage void schedule(void); | 332 | asmlinkage void schedule(void); |
332 | 333 | ||
333 | struct nsproxy; | 334 | struct nsproxy; |
diff --git a/kernel/mutex.c b/kernel/mutex.c index 357c6d221efe..524ffc33dc05 100644 --- a/kernel/mutex.c +++ b/kernel/mutex.c | |||
@@ -131,6 +131,7 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass, | |||
131 | struct mutex_waiter waiter; | 131 | struct mutex_waiter waiter; |
132 | unsigned long flags; | 132 | unsigned long flags; |
133 | 133 | ||
134 | preempt_disable(); | ||
134 | spin_lock_mutex(&lock->wait_lock, flags); | 135 | spin_lock_mutex(&lock->wait_lock, flags); |
135 | 136 | ||
136 | debug_mutex_lock_common(lock, &waiter); | 137 | debug_mutex_lock_common(lock, &waiter); |
@@ -170,13 +171,14 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass, | |||
170 | spin_unlock_mutex(&lock->wait_lock, flags); | 171 | spin_unlock_mutex(&lock->wait_lock, flags); |
171 | 172 | ||
172 | debug_mutex_free_waiter(&waiter); | 173 | debug_mutex_free_waiter(&waiter); |
174 | preempt_enable(); | ||
173 | return -EINTR; | 175 | return -EINTR; |
174 | } | 176 | } |
175 | __set_task_state(task, state); | 177 | __set_task_state(task, state); |
176 | 178 | ||
177 | /* didnt get the lock, go to sleep: */ | 179 | /* didnt get the lock, go to sleep: */ |
178 | spin_unlock_mutex(&lock->wait_lock, flags); | 180 | spin_unlock_mutex(&lock->wait_lock, flags); |
179 | schedule(); | 181 | __schedule(); |
180 | spin_lock_mutex(&lock->wait_lock, flags); | 182 | spin_lock_mutex(&lock->wait_lock, flags); |
181 | } | 183 | } |
182 | 184 | ||
@@ -193,6 +195,7 @@ done: | |||
193 | spin_unlock_mutex(&lock->wait_lock, flags); | 195 | spin_unlock_mutex(&lock->wait_lock, flags); |
194 | 196 | ||
195 | debug_mutex_free_waiter(&waiter); | 197 | debug_mutex_free_waiter(&waiter); |
198 | preempt_enable(); | ||
196 | 199 | ||
197 | return 0; | 200 | return 0; |
198 | } | 201 | } |
diff --git a/kernel/sched.c b/kernel/sched.c index 8be2c13b50d0..b001c133c359 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -4538,15 +4538,13 @@ pick_next_task(struct rq *rq, struct task_struct *prev) | |||
4538 | /* | 4538 | /* |
4539 | * schedule() is the main scheduler function. | 4539 | * schedule() is the main scheduler function. |
4540 | */ | 4540 | */ |
4541 | asmlinkage void __sched schedule(void) | 4541 | asmlinkage void __sched __schedule(void) |
4542 | { | 4542 | { |
4543 | struct task_struct *prev, *next; | 4543 | struct task_struct *prev, *next; |
4544 | unsigned long *switch_count; | 4544 | unsigned long *switch_count; |
4545 | struct rq *rq; | 4545 | struct rq *rq; |
4546 | int cpu; | 4546 | int cpu; |
4547 | 4547 | ||
4548 | need_resched: | ||
4549 | preempt_disable(); | ||
4550 | cpu = smp_processor_id(); | 4548 | cpu = smp_processor_id(); |
4551 | rq = cpu_rq(cpu); | 4549 | rq = cpu_rq(cpu); |
4552 | rcu_qsctr_inc(cpu); | 4550 | rcu_qsctr_inc(cpu); |
@@ -4603,7 +4601,13 @@ need_resched_nonpreemptible: | |||
4603 | 4601 | ||
4604 | if (unlikely(reacquire_kernel_lock(current) < 0)) | 4602 | if (unlikely(reacquire_kernel_lock(current) < 0)) |
4605 | goto need_resched_nonpreemptible; | 4603 | goto need_resched_nonpreemptible; |
4604 | } | ||
4606 | 4605 | ||
4606 | asmlinkage void __sched schedule(void) | ||
4607 | { | ||
4608 | need_resched: | ||
4609 | preempt_disable(); | ||
4610 | __schedule(); | ||
4607 | preempt_enable_no_resched(); | 4611 | preempt_enable_no_resched(); |
4608 | if (unlikely(test_thread_flag(TIF_NEED_RESCHED))) | 4612 | if (unlikely(test_thread_flag(TIF_NEED_RESCHED))) |
4609 | goto need_resched; | 4613 | goto need_resched; |