diff options
author | Jeff Garzik <jgarzik@pretzel.yyz.us> | 2005-06-22 21:50:57 -0400 |
---|---|---|
committer | Jeff Garzik <jgarzik@pobox.com> | 2005-06-22 21:50:57 -0400 |
commit | a5324343955997d1439f26518ddac567cd5d134b (patch) | |
tree | f43558389c41e3a0f076c4ee55d77c4aa1561779 /kernel/sched.c | |
parent | 8199d3a79c224bbe5943fa08684e1f93a17881b0 (diff) | |
parent | a4936044001694f033fe4ea94d6034d51a6b465c (diff) |
Merge /spare/repo/linux-2.6/
Diffstat (limited to 'kernel/sched.c')
-rw-r--r-- | kernel/sched.c | 13 |
1 files changed, 8 insertions, 5 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index 0dc3158667a2..deca041fc364 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -3755,19 +3755,22 @@ EXPORT_SYMBOL(cond_resched); | |||
3755 | */ | 3755 | */ |
3756 | int cond_resched_lock(spinlock_t * lock) | 3756 | int cond_resched_lock(spinlock_t * lock) |
3757 | { | 3757 | { |
3758 | int ret = 0; | ||
3759 | |||
3758 | if (need_lockbreak(lock)) { | 3760 | if (need_lockbreak(lock)) { |
3759 | spin_unlock(lock); | 3761 | spin_unlock(lock); |
3760 | cpu_relax(); | 3762 | cpu_relax(); |
3763 | ret = 1; | ||
3761 | spin_lock(lock); | 3764 | spin_lock(lock); |
3762 | } | 3765 | } |
3763 | if (need_resched()) { | 3766 | if (need_resched()) { |
3764 | _raw_spin_unlock(lock); | 3767 | _raw_spin_unlock(lock); |
3765 | preempt_enable_no_resched(); | 3768 | preempt_enable_no_resched(); |
3766 | __cond_resched(); | 3769 | __cond_resched(); |
3770 | ret = 1; | ||
3767 | spin_lock(lock); | 3771 | spin_lock(lock); |
3768 | return 1; | ||
3769 | } | 3772 | } |
3770 | return 0; | 3773 | return ret; |
3771 | } | 3774 | } |
3772 | 3775 | ||
3773 | EXPORT_SYMBOL(cond_resched_lock); | 3776 | EXPORT_SYMBOL(cond_resched_lock); |
@@ -3811,7 +3814,7 @@ EXPORT_SYMBOL(yield); | |||
3811 | */ | 3814 | */ |
3812 | void __sched io_schedule(void) | 3815 | void __sched io_schedule(void) |
3813 | { | 3816 | { |
3814 | struct runqueue *rq = &per_cpu(runqueues, _smp_processor_id()); | 3817 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); |
3815 | 3818 | ||
3816 | atomic_inc(&rq->nr_iowait); | 3819 | atomic_inc(&rq->nr_iowait); |
3817 | schedule(); | 3820 | schedule(); |
@@ -3822,7 +3825,7 @@ EXPORT_SYMBOL(io_schedule); | |||
3822 | 3825 | ||
3823 | long __sched io_schedule_timeout(long timeout) | 3826 | long __sched io_schedule_timeout(long timeout) |
3824 | { | 3827 | { |
3825 | struct runqueue *rq = &per_cpu(runqueues, _smp_processor_id()); | 3828 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); |
3826 | long ret; | 3829 | long ret; |
3827 | 3830 | ||
3828 | atomic_inc(&rq->nr_iowait); | 3831 | atomic_inc(&rq->nr_iowait); |
@@ -4243,7 +4246,7 @@ static void move_task_off_dead_cpu(int dead_cpu, struct task_struct *tsk) | |||
4243 | 4246 | ||
4244 | /* No more Mr. Nice Guy. */ | 4247 | /* No more Mr. Nice Guy. */ |
4245 | if (dest_cpu == NR_CPUS) { | 4248 | if (dest_cpu == NR_CPUS) { |
4246 | tsk->cpus_allowed = cpuset_cpus_allowed(tsk); | 4249 | cpus_setall(tsk->cpus_allowed); |
4247 | dest_cpu = any_online_cpu(tsk->cpus_allowed); | 4250 | dest_cpu = any_online_cpu(tsk->cpus_allowed); |
4248 | 4251 | ||
4249 | /* | 4252 | /* |