diff options
author | Vincent Guittot <vincent.guittot@linaro.org> | 2019-06-17 11:00:17 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2019-06-24 13:23:39 -0400 |
commit | 8ec59c0f5f4966f89f4e3e3cab81710c7fa959d0 (patch) | |
tree | ff6a6d519b089e759a168302f917ebe8d28a8046 | |
parent | d2abae71ebcc409828b24ce9da402548ecdf1311 (diff) |
sched/topology: Remove unused 'sd' parameter from arch_scale_cpu_capacity()
The 'struct sched_domain *sd' parameter to arch_scale_cpu_capacity() is
unused since commit:
765d0af19f5f ("sched/topology: Remove the ::smt_gain field from 'struct sched_domain'")
Remove it.
Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Viresh Kumar <viresh.kumar@linaro.org>
Reviewed-by: Valentin Schneider <valentin.schneider@arm.com>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: gregkh@linuxfoundation.org
Cc: linux@armlinux.org.uk
Cc: quentin.perret@arm.com
Cc: rafael@kernel.org
Link: https://lkml.kernel.org/r/1560783617-5827-1-git-send-email-vincent.guittot@linaro.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
-rw-r--r-- | arch/arm/kernel/topology.c | 2 | ||||
-rw-r--r-- | drivers/base/arch_topology.c | 6 | ||||
-rw-r--r-- | include/linux/arch_topology.h | 2 | ||||
-rw-r--r-- | include/linux/energy_model.h | 2 | ||||
-rw-r--r-- | include/linux/sched/topology.h | 14 | ||||
-rw-r--r-- | kernel/power/energy_model.c | 2 | ||||
-rw-r--r-- | kernel/sched/cpufreq_schedutil.c | 2 | ||||
-rw-r--r-- | kernel/sched/deadline.c | 2 | ||||
-rw-r--r-- | kernel/sched/fair.c | 6 | ||||
-rw-r--r-- | kernel/sched/pelt.c | 2 | ||||
-rw-r--r-- | kernel/sched/pelt.h | 2 | ||||
-rw-r--r-- | kernel/sched/sched.h | 2 | ||||
-rw-r--r-- | kernel/sched/topology.c | 8 |
13 files changed, 22 insertions, 30 deletions
diff --git a/arch/arm/kernel/topology.c b/arch/arm/kernel/topology.c index 60e375ce1ab2..d17cb1e6d679 100644 --- a/arch/arm/kernel/topology.c +++ b/arch/arm/kernel/topology.c | |||
@@ -169,7 +169,7 @@ static void update_cpu_capacity(unsigned int cpu) | |||
169 | topology_set_cpu_scale(cpu, cpu_capacity(cpu) / middle_capacity); | 169 | topology_set_cpu_scale(cpu, cpu_capacity(cpu) / middle_capacity); |
170 | 170 | ||
171 | pr_info("CPU%u: update cpu_capacity %lu\n", | 171 | pr_info("CPU%u: update cpu_capacity %lu\n", |
172 | cpu, topology_get_cpu_scale(NULL, cpu)); | 172 | cpu, topology_get_cpu_scale(cpu)); |
173 | } | 173 | } |
174 | 174 | ||
175 | #else | 175 | #else |
diff --git a/drivers/base/arch_topology.c b/drivers/base/arch_topology.c index 1739d7e1952a..9b09e31ae82f 100644 --- a/drivers/base/arch_topology.c +++ b/drivers/base/arch_topology.c | |||
@@ -43,7 +43,7 @@ static ssize_t cpu_capacity_show(struct device *dev, | |||
43 | { | 43 | { |
44 | struct cpu *cpu = container_of(dev, struct cpu, dev); | 44 | struct cpu *cpu = container_of(dev, struct cpu, dev); |
45 | 45 | ||
46 | return sprintf(buf, "%lu\n", topology_get_cpu_scale(NULL, cpu->dev.id)); | 46 | return sprintf(buf, "%lu\n", topology_get_cpu_scale(cpu->dev.id)); |
47 | } | 47 | } |
48 | 48 | ||
49 | static void update_topology_flags_workfn(struct work_struct *work); | 49 | static void update_topology_flags_workfn(struct work_struct *work); |
@@ -116,7 +116,7 @@ void topology_normalize_cpu_scale(void) | |||
116 | / capacity_scale; | 116 | / capacity_scale; |
117 | topology_set_cpu_scale(cpu, capacity); | 117 | topology_set_cpu_scale(cpu, capacity); |
118 | pr_debug("cpu_capacity: CPU%d cpu_capacity=%lu\n", | 118 | pr_debug("cpu_capacity: CPU%d cpu_capacity=%lu\n", |
119 | cpu, topology_get_cpu_scale(NULL, cpu)); | 119 | cpu, topology_get_cpu_scale(cpu)); |
120 | } | 120 | } |
121 | } | 121 | } |
122 | 122 | ||
@@ -185,7 +185,7 @@ init_cpu_capacity_callback(struct notifier_block *nb, | |||
185 | cpumask_andnot(cpus_to_visit, cpus_to_visit, policy->related_cpus); | 185 | cpumask_andnot(cpus_to_visit, cpus_to_visit, policy->related_cpus); |
186 | 186 | ||
187 | for_each_cpu(cpu, policy->related_cpus) { | 187 | for_each_cpu(cpu, policy->related_cpus) { |
188 | raw_capacity[cpu] = topology_get_cpu_scale(NULL, cpu) * | 188 | raw_capacity[cpu] = topology_get_cpu_scale(cpu) * |
189 | policy->cpuinfo.max_freq / 1000UL; | 189 | policy->cpuinfo.max_freq / 1000UL; |
190 | capacity_scale = max(raw_capacity[cpu], capacity_scale); | 190 | capacity_scale = max(raw_capacity[cpu], capacity_scale); |
191 | } | 191 | } |
diff --git a/include/linux/arch_topology.h b/include/linux/arch_topology.h index d9bdc1a7f4e7..1cfe05ea1d89 100644 --- a/include/linux/arch_topology.h +++ b/include/linux/arch_topology.h | |||
@@ -18,7 +18,7 @@ DECLARE_PER_CPU(unsigned long, cpu_scale); | |||
18 | 18 | ||
19 | struct sched_domain; | 19 | struct sched_domain; |
20 | static inline | 20 | static inline |
21 | unsigned long topology_get_cpu_scale(struct sched_domain *sd, int cpu) | 21 | unsigned long topology_get_cpu_scale(int cpu) |
22 | { | 22 | { |
23 | return per_cpu(cpu_scale, cpu); | 23 | return per_cpu(cpu_scale, cpu); |
24 | } | 24 | } |
diff --git a/include/linux/energy_model.h b/include/linux/energy_model.h index aa027f7bcb3e..73f8c3cb9588 100644 --- a/include/linux/energy_model.h +++ b/include/linux/energy_model.h | |||
@@ -89,7 +89,7 @@ static inline unsigned long em_pd_energy(struct em_perf_domain *pd, | |||
89 | * like schedutil. | 89 | * like schedutil. |
90 | */ | 90 | */ |
91 | cpu = cpumask_first(to_cpumask(pd->cpus)); | 91 | cpu = cpumask_first(to_cpumask(pd->cpus)); |
92 | scale_cpu = arch_scale_cpu_capacity(NULL, cpu); | 92 | scale_cpu = arch_scale_cpu_capacity(cpu); |
93 | cs = &pd->table[pd->nr_cap_states - 1]; | 93 | cs = &pd->table[pd->nr_cap_states - 1]; |
94 | freq = map_util_freq(max_util, cs->frequency, scale_cpu); | 94 | freq = map_util_freq(max_util, cs->frequency, scale_cpu); |
95 | 95 | ||
diff --git a/include/linux/sched/topology.h b/include/linux/sched/topology.h index 53afbe07354a..e445d3767cdd 100644 --- a/include/linux/sched/topology.h +++ b/include/linux/sched/topology.h | |||
@@ -196,14 +196,6 @@ extern void set_sched_topology(struct sched_domain_topology_level *tl); | |||
196 | # define SD_INIT_NAME(type) | 196 | # define SD_INIT_NAME(type) |
197 | #endif | 197 | #endif |
198 | 198 | ||
199 | #ifndef arch_scale_cpu_capacity | ||
200 | static __always_inline | ||
201 | unsigned long arch_scale_cpu_capacity(struct sched_domain *sd, int cpu) | ||
202 | { | ||
203 | return SCHED_CAPACITY_SCALE; | ||
204 | } | ||
205 | #endif | ||
206 | |||
207 | #else /* CONFIG_SMP */ | 199 | #else /* CONFIG_SMP */ |
208 | 200 | ||
209 | struct sched_domain_attr; | 201 | struct sched_domain_attr; |
@@ -219,16 +211,16 @@ static inline bool cpus_share_cache(int this_cpu, int that_cpu) | |||
219 | return true; | 211 | return true; |
220 | } | 212 | } |
221 | 213 | ||
214 | #endif /* !CONFIG_SMP */ | ||
215 | |||
222 | #ifndef arch_scale_cpu_capacity | 216 | #ifndef arch_scale_cpu_capacity |
223 | static __always_inline | 217 | static __always_inline |
224 | unsigned long arch_scale_cpu_capacity(void __always_unused *sd, int cpu) | 218 | unsigned long arch_scale_cpu_capacity(int cpu) |
225 | { | 219 | { |
226 | return SCHED_CAPACITY_SCALE; | 220 | return SCHED_CAPACITY_SCALE; |
227 | } | 221 | } |
228 | #endif | 222 | #endif |
229 | 223 | ||
230 | #endif /* !CONFIG_SMP */ | ||
231 | |||
232 | static inline int task_node(const struct task_struct *p) | 224 | static inline int task_node(const struct task_struct *p) |
233 | { | 225 | { |
234 | return cpu_to_node(task_cpu(p)); | 226 | return cpu_to_node(task_cpu(p)); |
diff --git a/kernel/power/energy_model.c b/kernel/power/energy_model.c index 7d66ee68aaaf..0a9326f5f421 100644 --- a/kernel/power/energy_model.c +++ b/kernel/power/energy_model.c | |||
@@ -223,7 +223,7 @@ int em_register_perf_domain(cpumask_t *span, unsigned int nr_states, | |||
223 | * All CPUs of a domain must have the same micro-architecture | 223 | * All CPUs of a domain must have the same micro-architecture |
224 | * since they all share the same table. | 224 | * since they all share the same table. |
225 | */ | 225 | */ |
226 | cap = arch_scale_cpu_capacity(NULL, cpu); | 226 | cap = arch_scale_cpu_capacity(cpu); |
227 | if (prev_cap && prev_cap != cap) { | 227 | if (prev_cap && prev_cap != cap) { |
228 | pr_err("CPUs of %*pbl must have the same capacity\n", | 228 | pr_err("CPUs of %*pbl must have the same capacity\n", |
229 | cpumask_pr_args(span)); | 229 | cpumask_pr_args(span)); |
diff --git a/kernel/sched/cpufreq_schedutil.c b/kernel/sched/cpufreq_schedutil.c index 962cf343f798..7c4ce69067c4 100644 --- a/kernel/sched/cpufreq_schedutil.c +++ b/kernel/sched/cpufreq_schedutil.c | |||
@@ -276,7 +276,7 @@ static unsigned long sugov_get_util(struct sugov_cpu *sg_cpu) | |||
276 | { | 276 | { |
277 | struct rq *rq = cpu_rq(sg_cpu->cpu); | 277 | struct rq *rq = cpu_rq(sg_cpu->cpu); |
278 | unsigned long util = cpu_util_cfs(rq); | 278 | unsigned long util = cpu_util_cfs(rq); |
279 | unsigned long max = arch_scale_cpu_capacity(NULL, sg_cpu->cpu); | 279 | unsigned long max = arch_scale_cpu_capacity(sg_cpu->cpu); |
280 | 280 | ||
281 | sg_cpu->max = max; | 281 | sg_cpu->max = max; |
282 | sg_cpu->bw_dl = cpu_bw_dl(rq); | 282 | sg_cpu->bw_dl = cpu_bw_dl(rq); |
diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c index c1ef30861068..8b5bb2ac16e2 100644 --- a/kernel/sched/deadline.c +++ b/kernel/sched/deadline.c | |||
@@ -1195,7 +1195,7 @@ static void update_curr_dl(struct rq *rq) | |||
1195 | &curr->dl); | 1195 | &curr->dl); |
1196 | } else { | 1196 | } else { |
1197 | unsigned long scale_freq = arch_scale_freq_capacity(cpu); | 1197 | unsigned long scale_freq = arch_scale_freq_capacity(cpu); |
1198 | unsigned long scale_cpu = arch_scale_cpu_capacity(NULL, cpu); | 1198 | unsigned long scale_cpu = arch_scale_cpu_capacity(cpu); |
1199 | 1199 | ||
1200 | scaled_delta_exec = cap_scale(delta_exec, scale_freq); | 1200 | scaled_delta_exec = cap_scale(delta_exec, scale_freq); |
1201 | scaled_delta_exec = cap_scale(scaled_delta_exec, scale_cpu); | 1201 | scaled_delta_exec = cap_scale(scaled_delta_exec, scale_cpu); |
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 3c11dcdedcbc..4f8754157763 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -764,7 +764,7 @@ void post_init_entity_util_avg(struct task_struct *p) | |||
764 | struct sched_entity *se = &p->se; | 764 | struct sched_entity *se = &p->se; |
765 | struct cfs_rq *cfs_rq = cfs_rq_of(se); | 765 | struct cfs_rq *cfs_rq = cfs_rq_of(se); |
766 | struct sched_avg *sa = &se->avg; | 766 | struct sched_avg *sa = &se->avg; |
767 | long cpu_scale = arch_scale_cpu_capacity(NULL, cpu_of(rq_of(cfs_rq))); | 767 | long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq))); |
768 | long cap = (long)(cpu_scale - cfs_rq->avg.util_avg) / 2; | 768 | long cap = (long)(cpu_scale - cfs_rq->avg.util_avg) / 2; |
769 | 769 | ||
770 | if (cap > 0) { | 770 | if (cap > 0) { |
@@ -7646,7 +7646,7 @@ static inline void init_sd_lb_stats(struct sd_lb_stats *sds) | |||
7646 | static unsigned long scale_rt_capacity(struct sched_domain *sd, int cpu) | 7646 | static unsigned long scale_rt_capacity(struct sched_domain *sd, int cpu) |
7647 | { | 7647 | { |
7648 | struct rq *rq = cpu_rq(cpu); | 7648 | struct rq *rq = cpu_rq(cpu); |
7649 | unsigned long max = arch_scale_cpu_capacity(sd, cpu); | 7649 | unsigned long max = arch_scale_cpu_capacity(cpu); |
7650 | unsigned long used, free; | 7650 | unsigned long used, free; |
7651 | unsigned long irq; | 7651 | unsigned long irq; |
7652 | 7652 | ||
@@ -7671,7 +7671,7 @@ static void update_cpu_capacity(struct sched_domain *sd, int cpu) | |||
7671 | unsigned long capacity = scale_rt_capacity(sd, cpu); | 7671 | unsigned long capacity = scale_rt_capacity(sd, cpu); |
7672 | struct sched_group *sdg = sd->groups; | 7672 | struct sched_group *sdg = sd->groups; |
7673 | 7673 | ||
7674 | cpu_rq(cpu)->cpu_capacity_orig = arch_scale_cpu_capacity(sd, cpu); | 7674 | cpu_rq(cpu)->cpu_capacity_orig = arch_scale_cpu_capacity(cpu); |
7675 | 7675 | ||
7676 | if (!capacity) | 7676 | if (!capacity) |
7677 | capacity = 1; | 7677 | capacity = 1; |
diff --git a/kernel/sched/pelt.c b/kernel/sched/pelt.c index befce29bd882..42ea66b07b1d 100644 --- a/kernel/sched/pelt.c +++ b/kernel/sched/pelt.c | |||
@@ -366,7 +366,7 @@ int update_irq_load_avg(struct rq *rq, u64 running) | |||
366 | * reflect the real amount of computation | 366 | * reflect the real amount of computation |
367 | */ | 367 | */ |
368 | running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); | 368 | running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); |
369 | running = cap_scale(running, arch_scale_cpu_capacity(NULL, cpu_of(rq))); | 369 | running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq))); |
370 | 370 | ||
371 | /* | 371 | /* |
372 | * We know the time that has been used by interrupt since last update | 372 | * We know the time that has been used by interrupt since last update |
diff --git a/kernel/sched/pelt.h b/kernel/sched/pelt.h index 7489d5f56960..afff644da065 100644 --- a/kernel/sched/pelt.h +++ b/kernel/sched/pelt.h | |||
@@ -79,7 +79,7 @@ static inline void update_rq_clock_pelt(struct rq *rq, s64 delta) | |||
79 | * Scale the elapsed time to reflect the real amount of | 79 | * Scale the elapsed time to reflect the real amount of |
80 | * computation | 80 | * computation |
81 | */ | 81 | */ |
82 | delta = cap_scale(delta, arch_scale_cpu_capacity(NULL, cpu_of(rq))); | 82 | delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq))); |
83 | delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); | 83 | delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); |
84 | 84 | ||
85 | rq->clock_pelt += delta; | 85 | rq->clock_pelt += delta; |
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index b08dee29ef5e..e58ab597ec88 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h | |||
@@ -2248,7 +2248,7 @@ unsigned long schedutil_freq_util(int cpu, unsigned long util_cfs, | |||
2248 | 2248 | ||
2249 | static inline unsigned long schedutil_energy_util(int cpu, unsigned long cfs) | 2249 | static inline unsigned long schedutil_energy_util(int cpu, unsigned long cfs) |
2250 | { | 2250 | { |
2251 | unsigned long max = arch_scale_cpu_capacity(NULL, cpu); | 2251 | unsigned long max = arch_scale_cpu_capacity(cpu); |
2252 | 2252 | ||
2253 | return schedutil_freq_util(cpu, cfs, max, ENERGY_UTIL); | 2253 | return schedutil_freq_util(cpu, cfs, max, ENERGY_UTIL); |
2254 | } | 2254 | } |
diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c index 63184cf0d0d7..f751ce0b783e 100644 --- a/kernel/sched/topology.c +++ b/kernel/sched/topology.c | |||
@@ -1874,10 +1874,10 @@ static struct sched_domain_topology_level | |||
1874 | unsigned long cap; | 1874 | unsigned long cap; |
1875 | 1875 | ||
1876 | /* Is there any asymmetry? */ | 1876 | /* Is there any asymmetry? */ |
1877 | cap = arch_scale_cpu_capacity(NULL, cpumask_first(cpu_map)); | 1877 | cap = arch_scale_cpu_capacity(cpumask_first(cpu_map)); |
1878 | 1878 | ||
1879 | for_each_cpu(i, cpu_map) { | 1879 | for_each_cpu(i, cpu_map) { |
1880 | if (arch_scale_cpu_capacity(NULL, i) != cap) { | 1880 | if (arch_scale_cpu_capacity(i) != cap) { |
1881 | asym = true; | 1881 | asym = true; |
1882 | break; | 1882 | break; |
1883 | } | 1883 | } |
@@ -1892,7 +1892,7 @@ static struct sched_domain_topology_level | |||
1892 | * to everyone. | 1892 | * to everyone. |
1893 | */ | 1893 | */ |
1894 | for_each_cpu(i, cpu_map) { | 1894 | for_each_cpu(i, cpu_map) { |
1895 | unsigned long max_capacity = arch_scale_cpu_capacity(NULL, i); | 1895 | unsigned long max_capacity = arch_scale_cpu_capacity(i); |
1896 | int tl_id = 0; | 1896 | int tl_id = 0; |
1897 | 1897 | ||
1898 | for_each_sd_topology(tl) { | 1898 | for_each_sd_topology(tl) { |
@@ -1902,7 +1902,7 @@ static struct sched_domain_topology_level | |||
1902 | for_each_cpu_and(j, tl->mask(i), cpu_map) { | 1902 | for_each_cpu_and(j, tl->mask(i), cpu_map) { |
1903 | unsigned long capacity; | 1903 | unsigned long capacity; |
1904 | 1904 | ||
1905 | capacity = arch_scale_cpu_capacity(NULL, j); | 1905 | capacity = arch_scale_cpu_capacity(j); |
1906 | 1906 | ||
1907 | if (capacity <= max_capacity) | 1907 | if (capacity <= max_capacity) |
1908 | continue; | 1908 | continue; |