diff options
author | Rusty Russell <rusty@rustcorp.com.au> | 2008-11-24 11:05:02 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-11-24 11:49:47 -0500 |
commit | abcd083a1a658d2bc1f7fced02632bfe03918002 (patch) | |
tree | ba9fe810b1fc62c7f66f9b5fc17c6f3e6bcbd35b /kernel | |
parent | ea6f18ed5a1531caf678374f30a0990c9e6742f3 (diff) |
sched: convert sched.c from for_each_cpu_mask to for_each_cpu.
Impact: trivial API conversion
This is a simple conversion, but note that for_each_cpu() terminates
with i >= nr_cpu_ids, not i == NR_CPUS like for_each_cpu_mask() did.
I don't convert all of them: sd->span changes in a later patch, so
change those iterators there rather than here.
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/sched.c | 36 |
1 files changed, 18 insertions, 18 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index dd22cec499b8..e59978eead17 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -2061,7 +2061,7 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu) | |||
2061 | /* Tally up the load of all CPUs in the group */ | 2061 | /* Tally up the load of all CPUs in the group */ |
2062 | avg_load = 0; | 2062 | avg_load = 0; |
2063 | 2063 | ||
2064 | for_each_cpu_mask_nr(i, group->cpumask) { | 2064 | for_each_cpu(i, &group->cpumask) { |
2065 | /* Bias balancing toward cpus of our domain */ | 2065 | /* Bias balancing toward cpus of our domain */ |
2066 | if (local_group) | 2066 | if (local_group) |
2067 | load = source_load(i, load_idx); | 2067 | load = source_load(i, load_idx); |
@@ -2103,7 +2103,7 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu, | |||
2103 | /* Traverse only the allowed CPUs */ | 2103 | /* Traverse only the allowed CPUs */ |
2104 | cpus_and(*tmp, group->cpumask, p->cpus_allowed); | 2104 | cpus_and(*tmp, group->cpumask, p->cpus_allowed); |
2105 | 2105 | ||
2106 | for_each_cpu_mask_nr(i, *tmp) { | 2106 | for_each_cpu(i, tmp) { |
2107 | load = weighted_cpuload(i); | 2107 | load = weighted_cpuload(i); |
2108 | 2108 | ||
2109 | if (load < min_load || (load == min_load && i == this_cpu)) { | 2109 | if (load < min_load || (load == min_load && i == this_cpu)) { |
@@ -3121,7 +3121,7 @@ find_busiest_group(struct sched_domain *sd, int this_cpu, | |||
3121 | max_cpu_load = 0; | 3121 | max_cpu_load = 0; |
3122 | min_cpu_load = ~0UL; | 3122 | min_cpu_load = ~0UL; |
3123 | 3123 | ||
3124 | for_each_cpu_mask_nr(i, group->cpumask) { | 3124 | for_each_cpu(i, &group->cpumask) { |
3125 | struct rq *rq; | 3125 | struct rq *rq; |
3126 | 3126 | ||
3127 | if (!cpu_isset(i, *cpus)) | 3127 | if (!cpu_isset(i, *cpus)) |
@@ -3400,7 +3400,7 @@ find_busiest_queue(struct sched_group *group, enum cpu_idle_type idle, | |||
3400 | unsigned long max_load = 0; | 3400 | unsigned long max_load = 0; |
3401 | int i; | 3401 | int i; |
3402 | 3402 | ||
3403 | for_each_cpu_mask_nr(i, group->cpumask) { | 3403 | for_each_cpu(i, &group->cpumask) { |
3404 | unsigned long wl; | 3404 | unsigned long wl; |
3405 | 3405 | ||
3406 | if (!cpu_isset(i, *cpus)) | 3406 | if (!cpu_isset(i, *cpus)) |
@@ -3942,7 +3942,7 @@ static void run_rebalance_domains(struct softirq_action *h) | |||
3942 | int balance_cpu; | 3942 | int balance_cpu; |
3943 | 3943 | ||
3944 | cpu_clear(this_cpu, cpus); | 3944 | cpu_clear(this_cpu, cpus); |
3945 | for_each_cpu_mask_nr(balance_cpu, cpus) { | 3945 | for_each_cpu(balance_cpu, &cpus) { |
3946 | /* | 3946 | /* |
3947 | * If this cpu gets work to do, stop the load balancing | 3947 | * If this cpu gets work to do, stop the load balancing |
3948 | * work being done for other cpus. Next load | 3948 | * work being done for other cpus. Next load |
@@ -6906,7 +6906,7 @@ init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map, | |||
6906 | 6906 | ||
6907 | cpus_clear(*covered); | 6907 | cpus_clear(*covered); |
6908 | 6908 | ||
6909 | for_each_cpu_mask_nr(i, *span) { | 6909 | for_each_cpu(i, span) { |
6910 | struct sched_group *sg; | 6910 | struct sched_group *sg; |
6911 | int group = group_fn(i, cpu_map, &sg, tmpmask); | 6911 | int group = group_fn(i, cpu_map, &sg, tmpmask); |
6912 | int j; | 6912 | int j; |
@@ -6917,7 +6917,7 @@ init_sched_build_groups(const cpumask_t *span, const cpumask_t *cpu_map, | |||
6917 | cpus_clear(sg->cpumask); | 6917 | cpus_clear(sg->cpumask); |
6918 | sg->__cpu_power = 0; | 6918 | sg->__cpu_power = 0; |
6919 | 6919 | ||
6920 | for_each_cpu_mask_nr(j, *span) { | 6920 | for_each_cpu(j, span) { |
6921 | if (group_fn(j, cpu_map, NULL, tmpmask) != group) | 6921 | if (group_fn(j, cpu_map, NULL, tmpmask) != group) |
6922 | continue; | 6922 | continue; |
6923 | 6923 | ||
@@ -7117,7 +7117,7 @@ static void init_numa_sched_groups_power(struct sched_group *group_head) | |||
7117 | if (!sg) | 7117 | if (!sg) |
7118 | return; | 7118 | return; |
7119 | do { | 7119 | do { |
7120 | for_each_cpu_mask_nr(j, sg->cpumask) { | 7120 | for_each_cpu(j, &sg->cpumask) { |
7121 | struct sched_domain *sd; | 7121 | struct sched_domain *sd; |
7122 | 7122 | ||
7123 | sd = &per_cpu(phys_domains, j); | 7123 | sd = &per_cpu(phys_domains, j); |
@@ -7142,7 +7142,7 @@ static void free_sched_groups(const cpumask_t *cpu_map, cpumask_t *nodemask) | |||
7142 | { | 7142 | { |
7143 | int cpu, i; | 7143 | int cpu, i; |
7144 | 7144 | ||
7145 | for_each_cpu_mask_nr(cpu, *cpu_map) { | 7145 | for_each_cpu(cpu, cpu_map) { |
7146 | struct sched_group **sched_group_nodes | 7146 | struct sched_group **sched_group_nodes |
7147 | = sched_group_nodes_bycpu[cpu]; | 7147 | = sched_group_nodes_bycpu[cpu]; |
7148 | 7148 | ||
@@ -7396,7 +7396,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7396 | /* | 7396 | /* |
7397 | * Set up domains for cpus specified by the cpu_map. | 7397 | * Set up domains for cpus specified by the cpu_map. |
7398 | */ | 7398 | */ |
7399 | for_each_cpu_mask_nr(i, *cpu_map) { | 7399 | for_each_cpu(i, cpu_map) { |
7400 | struct sched_domain *sd = NULL, *p; | 7400 | struct sched_domain *sd = NULL, *p; |
7401 | SCHED_CPUMASK_VAR(nodemask, allmasks); | 7401 | SCHED_CPUMASK_VAR(nodemask, allmasks); |
7402 | 7402 | ||
@@ -7463,7 +7463,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7463 | 7463 | ||
7464 | #ifdef CONFIG_SCHED_SMT | 7464 | #ifdef CONFIG_SCHED_SMT |
7465 | /* Set up CPU (sibling) groups */ | 7465 | /* Set up CPU (sibling) groups */ |
7466 | for_each_cpu_mask_nr(i, *cpu_map) { | 7466 | for_each_cpu(i, cpu_map) { |
7467 | SCHED_CPUMASK_VAR(this_sibling_map, allmasks); | 7467 | SCHED_CPUMASK_VAR(this_sibling_map, allmasks); |
7468 | SCHED_CPUMASK_VAR(send_covered, allmasks); | 7468 | SCHED_CPUMASK_VAR(send_covered, allmasks); |
7469 | 7469 | ||
@@ -7480,7 +7480,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7480 | 7480 | ||
7481 | #ifdef CONFIG_SCHED_MC | 7481 | #ifdef CONFIG_SCHED_MC |
7482 | /* Set up multi-core groups */ | 7482 | /* Set up multi-core groups */ |
7483 | for_each_cpu_mask_nr(i, *cpu_map) { | 7483 | for_each_cpu(i, cpu_map) { |
7484 | SCHED_CPUMASK_VAR(this_core_map, allmasks); | 7484 | SCHED_CPUMASK_VAR(this_core_map, allmasks); |
7485 | SCHED_CPUMASK_VAR(send_covered, allmasks); | 7485 | SCHED_CPUMASK_VAR(send_covered, allmasks); |
7486 | 7486 | ||
@@ -7547,7 +7547,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7547 | goto error; | 7547 | goto error; |
7548 | } | 7548 | } |
7549 | sched_group_nodes[i] = sg; | 7549 | sched_group_nodes[i] = sg; |
7550 | for_each_cpu_mask_nr(j, *nodemask) { | 7550 | for_each_cpu(j, nodemask) { |
7551 | struct sched_domain *sd; | 7551 | struct sched_domain *sd; |
7552 | 7552 | ||
7553 | sd = &per_cpu(node_domains, j); | 7553 | sd = &per_cpu(node_domains, j); |
@@ -7593,21 +7593,21 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7593 | 7593 | ||
7594 | /* Calculate CPU power for physical packages and nodes */ | 7594 | /* Calculate CPU power for physical packages and nodes */ |
7595 | #ifdef CONFIG_SCHED_SMT | 7595 | #ifdef CONFIG_SCHED_SMT |
7596 | for_each_cpu_mask_nr(i, *cpu_map) { | 7596 | for_each_cpu(i, cpu_map) { |
7597 | struct sched_domain *sd = &per_cpu(cpu_domains, i); | 7597 | struct sched_domain *sd = &per_cpu(cpu_domains, i); |
7598 | 7598 | ||
7599 | init_sched_groups_power(i, sd); | 7599 | init_sched_groups_power(i, sd); |
7600 | } | 7600 | } |
7601 | #endif | 7601 | #endif |
7602 | #ifdef CONFIG_SCHED_MC | 7602 | #ifdef CONFIG_SCHED_MC |
7603 | for_each_cpu_mask_nr(i, *cpu_map) { | 7603 | for_each_cpu(i, cpu_map) { |
7604 | struct sched_domain *sd = &per_cpu(core_domains, i); | 7604 | struct sched_domain *sd = &per_cpu(core_domains, i); |
7605 | 7605 | ||
7606 | init_sched_groups_power(i, sd); | 7606 | init_sched_groups_power(i, sd); |
7607 | } | 7607 | } |
7608 | #endif | 7608 | #endif |
7609 | 7609 | ||
7610 | for_each_cpu_mask_nr(i, *cpu_map) { | 7610 | for_each_cpu(i, cpu_map) { |
7611 | struct sched_domain *sd = &per_cpu(phys_domains, i); | 7611 | struct sched_domain *sd = &per_cpu(phys_domains, i); |
7612 | 7612 | ||
7613 | init_sched_groups_power(i, sd); | 7613 | init_sched_groups_power(i, sd); |
@@ -7627,7 +7627,7 @@ static int __build_sched_domains(const cpumask_t *cpu_map, | |||
7627 | #endif | 7627 | #endif |
7628 | 7628 | ||
7629 | /* Attach the domains */ | 7629 | /* Attach the domains */ |
7630 | for_each_cpu_mask_nr(i, *cpu_map) { | 7630 | for_each_cpu(i, cpu_map) { |
7631 | struct sched_domain *sd; | 7631 | struct sched_domain *sd; |
7632 | #ifdef CONFIG_SCHED_SMT | 7632 | #ifdef CONFIG_SCHED_SMT |
7633 | sd = &per_cpu(cpu_domains, i); | 7633 | sd = &per_cpu(cpu_domains, i); |
@@ -7709,7 +7709,7 @@ static void detach_destroy_domains(const cpumask_t *cpu_map) | |||
7709 | cpumask_t tmpmask; | 7709 | cpumask_t tmpmask; |
7710 | int i; | 7710 | int i; |
7711 | 7711 | ||
7712 | for_each_cpu_mask_nr(i, *cpu_map) | 7712 | for_each_cpu(i, cpu_map) |
7713 | cpu_attach_domain(NULL, &def_root_domain, i); | 7713 | cpu_attach_domain(NULL, &def_root_domain, i); |
7714 | synchronize_sched(); | 7714 | synchronize_sched(); |
7715 | arch_destroy_sched_domains(cpu_map, &tmpmask); | 7715 | arch_destroy_sched_domains(cpu_map, &tmpmask); |