diff options
-rw-r--r-- | include/linux/sched.h | 2 | ||||
-rw-r--r-- | kernel/sched_fair.c | 14 | ||||
-rw-r--r-- | kernel/sched_idletask.c | 2 | ||||
-rw-r--r-- | kernel/sched_rt.c | 4 |
4 files changed, 11 insertions, 11 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index fc4c0f9393d2..5c116f03d74c 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
@@ -1037,7 +1037,7 @@ struct sched_class { | |||
1037 | void (*put_prev_task) (struct rq *rq, struct task_struct *p); | 1037 | void (*put_prev_task) (struct rq *rq, struct task_struct *p); |
1038 | 1038 | ||
1039 | #ifdef CONFIG_SMP | 1039 | #ifdef CONFIG_SMP |
1040 | int (*select_task_rq)(struct task_struct *p, int flag, int sync); | 1040 | int (*select_task_rq)(struct task_struct *p, int sd_flag, int sync); |
1041 | 1041 | ||
1042 | unsigned long (*load_balance) (struct rq *this_rq, int this_cpu, | 1042 | unsigned long (*load_balance) (struct rq *this_rq, int this_cpu, |
1043 | struct rq *busiest, unsigned long max_load_move, | 1043 | struct rq *busiest, unsigned long max_load_move, |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 19593568031a..b554e63c521a 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
@@ -1331,7 +1331,7 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu) | |||
1331 | * | 1331 | * |
1332 | * preempt must be disabled. | 1332 | * preempt must be disabled. |
1333 | */ | 1333 | */ |
1334 | static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | 1334 | static int select_task_rq_fair(struct task_struct *p, int sd_flag, int sync) |
1335 | { | 1335 | { |
1336 | struct sched_domain *tmp, *sd = NULL; | 1336 | struct sched_domain *tmp, *sd = NULL; |
1337 | int cpu = smp_processor_id(); | 1337 | int cpu = smp_processor_id(); |
@@ -1339,7 +1339,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | |||
1339 | int new_cpu = cpu; | 1339 | int new_cpu = cpu; |
1340 | int want_affine = 0; | 1340 | int want_affine = 0; |
1341 | 1341 | ||
1342 | if (flag & SD_BALANCE_WAKE) { | 1342 | if (sd_flag & SD_BALANCE_WAKE) { |
1343 | if (sched_feat(AFFINE_WAKEUPS)) | 1343 | if (sched_feat(AFFINE_WAKEUPS)) |
1344 | want_affine = 1; | 1344 | want_affine = 1; |
1345 | new_cpu = prev_cpu; | 1345 | new_cpu = prev_cpu; |
@@ -1368,7 +1368,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | |||
1368 | break; | 1368 | break; |
1369 | } | 1369 | } |
1370 | 1370 | ||
1371 | switch (flag) { | 1371 | switch (sd_flag) { |
1372 | case SD_BALANCE_WAKE: | 1372 | case SD_BALANCE_WAKE: |
1373 | if (!sched_feat(LB_WAKEUP_UPDATE)) | 1373 | if (!sched_feat(LB_WAKEUP_UPDATE)) |
1374 | break; | 1374 | break; |
@@ -1392,7 +1392,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | |||
1392 | want_affine = 0; | 1392 | want_affine = 0; |
1393 | } | 1393 | } |
1394 | 1394 | ||
1395 | if (!(tmp->flags & flag)) | 1395 | if (!(tmp->flags & sd_flag)) |
1396 | continue; | 1396 | continue; |
1397 | 1397 | ||
1398 | sd = tmp; | 1398 | sd = tmp; |
@@ -1402,12 +1402,12 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | |||
1402 | struct sched_group *group; | 1402 | struct sched_group *group; |
1403 | int weight; | 1403 | int weight; |
1404 | 1404 | ||
1405 | if (!(sd->flags & flag)) { | 1405 | if (!(sd->flags & sd_flag)) { |
1406 | sd = sd->child; | 1406 | sd = sd->child; |
1407 | continue; | 1407 | continue; |
1408 | } | 1408 | } |
1409 | 1409 | ||
1410 | group = find_idlest_group(sd, p, cpu, flag); | 1410 | group = find_idlest_group(sd, p, cpu, sd_flag); |
1411 | if (!group) { | 1411 | if (!group) { |
1412 | sd = sd->child; | 1412 | sd = sd->child; |
1413 | continue; | 1413 | continue; |
@@ -1427,7 +1427,7 @@ static int select_task_rq_fair(struct task_struct *p, int flag, int sync) | |||
1427 | for_each_domain(cpu, tmp) { | 1427 | for_each_domain(cpu, tmp) { |
1428 | if (weight <= cpumask_weight(sched_domain_span(tmp))) | 1428 | if (weight <= cpumask_weight(sched_domain_span(tmp))) |
1429 | break; | 1429 | break; |
1430 | if (tmp->flags & flag) | 1430 | if (tmp->flags & sd_flag) |
1431 | sd = tmp; | 1431 | sd = tmp; |
1432 | } | 1432 | } |
1433 | /* while loop will break here if sd == NULL */ | 1433 | /* while loop will break here if sd == NULL */ |
diff --git a/kernel/sched_idletask.c b/kernel/sched_idletask.c index 99b2f0337609..9ff7697e5dc4 100644 --- a/kernel/sched_idletask.c +++ b/kernel/sched_idletask.c | |||
@@ -6,7 +6,7 @@ | |||
6 | */ | 6 | */ |
7 | 7 | ||
8 | #ifdef CONFIG_SMP | 8 | #ifdef CONFIG_SMP |
9 | static int select_task_rq_idle(struct task_struct *p, int flag, int sync) | 9 | static int select_task_rq_idle(struct task_struct *p, int sd_flag, int sync) |
10 | { | 10 | { |
11 | return task_cpu(p); /* IDLE tasks as never migrated */ | 11 | return task_cpu(p); /* IDLE tasks as never migrated */ |
12 | } | 12 | } |
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c index 438380810ac4..97c53f3f51a7 100644 --- a/kernel/sched_rt.c +++ b/kernel/sched_rt.c | |||
@@ -938,11 +938,11 @@ static void yield_task_rt(struct rq *rq) | |||
938 | #ifdef CONFIG_SMP | 938 | #ifdef CONFIG_SMP |
939 | static int find_lowest_rq(struct task_struct *task); | 939 | static int find_lowest_rq(struct task_struct *task); |
940 | 940 | ||
941 | static int select_task_rq_rt(struct task_struct *p, int flag, int sync) | 941 | static int select_task_rq_rt(struct task_struct *p, int sd_flag, int sync) |
942 | { | 942 | { |
943 | struct rq *rq = task_rq(p); | 943 | struct rq *rq = task_rq(p); |
944 | 944 | ||
945 | if (flag != SD_BALANCE_WAKE) | 945 | if (sd_flag != SD_BALANCE_WAKE) |
946 | return smp_processor_id(); | 946 | return smp_processor_id(); |
947 | 947 | ||
948 | /* | 948 | /* |