diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2008-06-27 07:41:33 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-06-27 08:31:44 -0400 |
commit | 93b75217df39e6d75889cc6f8050343286aff4a5 (patch) | |
tree | d3091afd2b88aa9732ff63b3a839bdec358aef7d /kernel/sched.c | |
parent | cb5ef42a03a13f95a9ea94e6cda4f7a47497871f (diff) |
sched: disable source/target_load bias
The bias given by source/target_load functions can be very large, disable
it by default to get faster convergence.
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Cc: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com>
Cc: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched.c')
-rw-r--r-- | kernel/sched.c | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index 10d43f5bf0fc..6c5eb3bc37e0 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -2000,7 +2000,7 @@ static unsigned long source_load(int cpu, int type) | |||
2000 | struct rq *rq = cpu_rq(cpu); | 2000 | struct rq *rq = cpu_rq(cpu); |
2001 | unsigned long total = weighted_cpuload(cpu); | 2001 | unsigned long total = weighted_cpuload(cpu); |
2002 | 2002 | ||
2003 | if (type == 0) | 2003 | if (type == 0 || !sched_feat(LB_BIAS)) |
2004 | return total; | 2004 | return total; |
2005 | 2005 | ||
2006 | return min(rq->cpu_load[type-1], total); | 2006 | return min(rq->cpu_load[type-1], total); |
@@ -2015,7 +2015,7 @@ static unsigned long target_load(int cpu, int type) | |||
2015 | struct rq *rq = cpu_rq(cpu); | 2015 | struct rq *rq = cpu_rq(cpu); |
2016 | unsigned long total = weighted_cpuload(cpu); | 2016 | unsigned long total = weighted_cpuload(cpu); |
2017 | 2017 | ||
2018 | if (type == 0) | 2018 | if (type == 0 || !sched_feat(LB_BIAS)) |
2019 | return total; | 2019 | return total; |
2020 | 2020 | ||
2021 | return max(rq->cpu_load[type-1], total); | 2021 | return max(rq->cpu_load[type-1], total); |