aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2009-09-14 13:37:39 -0400
committerIngo Molnar <mingo@elte.hu>2009-09-15 10:51:29 -0400
commit0763a660a84220cc3900fd32abdd7ad109e2278d (patch)
tree1b745884fb79627e05c026ee5bfb115b1f722db8
parent8e6598af3f35629c37249a610cf13e73f70db279 (diff)
sched: Rename select_task_rq() argument
In order to be able to rename the sync argument, we need to rename the current flag argument. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> LKML-Reference: <new-submission> Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r--include/linux/sched.h2
-rw-r--r--kernel/sched_fair.c14
-rw-r--r--kernel/sched_idletask.c2
-rw-r--r--kernel/sched_rt.c4
4 files changed, 11 insertions, 11 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h
index fc4c0f9393d2..5c116f03d74c 100644
--- a/include/linux/sched.h
+++ b/include/linux/sched.h
@@ -1037,7 +1037,7 @@ struct sched_class {
1037 void (*put_prev_task) (struct rq *rq, struct task_struct *p); 1037 void (*put_prev_task) (struct rq *rq, struct task_struct *p);
1038 1038
1039#ifdef CONFIG_SMP 1039#ifdef CONFIG_SMP
1040 int (*select_task_rq)(struct task_struct *p, int flag, int sync); 1040 int (*select_task_rq)(struct task_struct *p, int sd_flag, int sync);
1041 1041
1042 unsigned long (*load_balance) (struct rq *this_rq, int this_cpu, 1042 unsigned long (*load_balance) (struct rq *this_rq, int this_cpu,
1043 struct rq *busiest, unsigned long max_load_move, 1043 struct rq *busiest, unsigned long max_load_move,
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index 19593568031a..b554e63c521a 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -1331,7 +1331,7 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
1331 * 1331 *
1332 * preempt must be disabled. 1332 * preempt must be disabled.
1333 */ 1333 */
1334static int select_task_rq_fair(struct task_struct *p, int flag, int sync) 1334static int select_task_rq_fair(struct task_struct *p, int sd_flag, int sync)
1335{ 1335{
1336 struct sched_domain *tmp, *sd = NULL; 1336 struct sched_domain *tmp, *sd = NULL;
1337 int cpu = smp_processor_id(); 1337 int cpu = smp_processor_id();
@@ -1339,7 +1339,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
1339 int new_cpu = cpu; 1339 int new_cpu = cpu;
1340 int want_affine = 0; 1340 int want_affine = 0;
1341 1341
1342 if (flag & SD_BALANCE_WAKE) { 1342 if (sd_flag & SD_BALANCE_WAKE) {
1343 if (sched_feat(AFFINE_WAKEUPS)) 1343 if (sched_feat(AFFINE_WAKEUPS))
1344 want_affine = 1; 1344 want_affine = 1;
1345 new_cpu = prev_cpu; 1345 new_cpu = prev_cpu;
@@ -1368,7 +1368,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
1368 break; 1368 break;
1369 } 1369 }
1370 1370
1371 switch (flag) { 1371 switch (sd_flag) {
1372 case SD_BALANCE_WAKE: 1372 case SD_BALANCE_WAKE:
1373 if (!sched_feat(LB_WAKEUP_UPDATE)) 1373 if (!sched_feat(LB_WAKEUP_UPDATE))
1374 break; 1374 break;
@@ -1392,7 +1392,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
1392 want_affine = 0; 1392 want_affine = 0;
1393 } 1393 }
1394 1394
1395 if (!(tmp->flags & flag)) 1395 if (!(tmp->flags & sd_flag))
1396 continue; 1396 continue;
1397 1397
1398 sd = tmp; 1398 sd = tmp;
@@ -1402,12 +1402,12 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
1402 struct sched_group *group; 1402 struct sched_group *group;
1403 int weight; 1403 int weight;
1404 1404
1405 if (!(sd->flags & flag)) { 1405 if (!(sd->flags & sd_flag)) {
1406 sd = sd->child; 1406 sd = sd->child;
1407 continue; 1407 continue;
1408 } 1408 }
1409 1409
1410 group = find_idlest_group(sd, p, cpu, flag); 1410 group = find_idlest_group(sd, p, cpu, sd_flag);
1411 if (!group) { 1411 if (!group) {
1412 sd = sd->child; 1412 sd = sd->child;
1413 continue; 1413 continue;
@@ -1427,7 +1427,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync)
1427 for_each_domain(cpu, tmp) { 1427 for_each_domain(cpu, tmp) {
1428 if (weight <= cpumask_weight(sched_domain_span(tmp))) 1428 if (weight <= cpumask_weight(sched_domain_span(tmp)))
1429 break; 1429 break;
1430 if (tmp->flags & flag) 1430 if (tmp->flags & sd_flag)
1431 sd = tmp; 1431 sd = tmp;
1432 } 1432 }
1433 /* while loop will break here if sd == NULL */ 1433 /* while loop will break here if sd == NULL */
diff --git a/kernel/sched_idletask.c b/kernel/sched_idletask.c
index 99b2f0337609..9ff7697e5dc4 100644
--- a/kernel/sched_idletask.c
+++ b/kernel/sched_idletask.c
@@ -6,7 +6,7 @@
6 */ 6 */
7 7
8#ifdef CONFIG_SMP 8#ifdef CONFIG_SMP
9static int select_task_rq_idle(struct task_struct *p, int flag, int sync) 9static int select_task_rq_idle(struct task_struct *p, int sd_flag, int sync)
10{ 10{
11 return task_cpu(p); /* IDLE tasks as never migrated */ 11 return task_cpu(p); /* IDLE tasks as never migrated */
12} 12}
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c
index 438380810ac4..97c53f3f51a7 100644
--- a/kernel/sched_rt.c
+++ b/kernel/sched_rt.c
@@ -938,11 +938,11 @@ static void yield_task_rt(struct rq *rq)
938#ifdef CONFIG_SMP 938#ifdef CONFIG_SMP
939static int find_lowest_rq(struct task_struct *task); 939static int find_lowest_rq(struct task_struct *task);
940 940
941static int select_task_rq_rt(struct task_struct *p, int flag, int sync) 941static int select_task_rq_rt(struct task_struct *p, int sd_flag, int sync)
942{ 942{
943 struct rq *rq = task_rq(p); 943 struct rq *rq = task_rq(p);
944 944
945 if (flag != SD_BALANCE_WAKE) 945 if (sd_flag != SD_BALANCE_WAKE)
946 return smp_processor_id(); 946 return smp_processor_id();
947 947
948 /* 948 /*