aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
authorIngo Molnar <mingo@elte.hu>2007-08-10 17:05:11 -0400
committerIngo Molnar <mingo@elte.hu>2007-08-10 17:05:11 -0400
commite56f31aad9d8c0102bc074cdab4e3ee76b38600d (patch)
tree3ff937cbd564cc57249f736c2a58ae477e4fbef5 /kernel
parent529c77261bccd9d37f110f58b0753d95beaa9fa2 (diff)
sched: fix typo in the FAIR_GROUP_SCHED branch
while there's no in-tree way to turn group scheduling at the moment, fix a typo in it nevertheless. Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched_fair.c7
1 files changed, 3 insertions, 4 deletions
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index e91db32cadfd..c5af38948a1e 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -959,13 +959,12 @@ load_balance_fair(struct rq *this_rq, int this_cpu, struct rq *busiest,
959 for_each_leaf_cfs_rq(busiest, busy_cfs_rq) { 959 for_each_leaf_cfs_rq(busiest, busy_cfs_rq) {
960#ifdef CONFIG_FAIR_GROUP_SCHED 960#ifdef CONFIG_FAIR_GROUP_SCHED
961 struct cfs_rq *this_cfs_rq; 961 struct cfs_rq *this_cfs_rq;
962 long imbalances; 962 long imbalance;
963 unsigned long maxload; 963 unsigned long maxload;
964 964
965 this_cfs_rq = cpu_cfs_rq(busy_cfs_rq, this_cpu); 965 this_cfs_rq = cpu_cfs_rq(busy_cfs_rq, this_cpu);
966 966
967 imbalance = busy_cfs_rq->load.weight - 967 imbalance = busy_cfs_rq->load.weight - this_cfs_rq->load.weight;
968 this_cfs_rq->load.weight;
969 /* Don't pull if this_cfs_rq has more load than busy_cfs_rq */ 968 /* Don't pull if this_cfs_rq has more load than busy_cfs_rq */
970 if (imbalance <= 0) 969 if (imbalance <= 0)
971 continue; 970 continue;
@@ -976,7 +975,7 @@ load_balance_fair(struct rq *this_rq, int this_cpu, struct rq *busiest,
976 975
977 *this_best_prio = cfs_rq_best_prio(this_cfs_rq); 976 *this_best_prio = cfs_rq_best_prio(this_cfs_rq);
978#else 977#else
979#define maxload rem_load_move 978# define maxload rem_load_move
980#endif 979#endif
981 /* pass busy_cfs_rq argument into 980 /* pass busy_cfs_rq argument into
982 * load_balance_[start|next]_fair iterators 981 * load_balance_[start|next]_fair iterators