aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/cpufreq/cpufreq_ondemand.c
diff options
context:
space:
mode:
authorvenkatesh.pallipadi@intel.com <venkatesh.pallipadi@intel.com>2008-08-04 14:59:08 -0400
committerDave Jones <davej@redhat.com>2008-10-09 13:52:43 -0400
commitc43aa3bd99a67009a167430e80c5fde6f37288d8 (patch)
tree41e0aa9b864aa6a3d214c1c5ad52bc1f934e1423 /drivers/cpufreq/cpufreq_ondemand.c
parentbf0b90e357c883e8efd72954432efe652de74c76 (diff)
[CPUFREQ][2/6] cpufreq: Change load calculation in ondemand for software coordination
Change the load calculation algorithm in ondemand to work well with software coordination of frequency across the dependent cpus. Multiply individual CPU utilization with the average freq of that logical CPU during the measurement interval (using getavg call). And find the max CPU utilization number in terms of CPU freq. That number is then used to get to the target freq for next sampling interval. Signed-off-by: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> Signed-off-by: Dave Jones <davej@redhat.com>
Diffstat (limited to 'drivers/cpufreq/cpufreq_ondemand.c')
-rw-r--r--drivers/cpufreq/cpufreq_ondemand.c65
1 files changed, 35 insertions, 30 deletions
diff --git a/drivers/cpufreq/cpufreq_ondemand.c b/drivers/cpufreq/cpufreq_ondemand.c
index f56debd9a8d7..42fcb146ba22 100644
--- a/drivers/cpufreq/cpufreq_ondemand.c
+++ b/drivers/cpufreq/cpufreq_ondemand.c
@@ -334,9 +334,7 @@ static struct attribute_group dbs_attr_group = {
334 334
335static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info) 335static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
336{ 336{
337 unsigned int idle_ticks, total_ticks; 337 unsigned int max_load_freq;
338 unsigned int load = 0;
339 cputime64_t cur_jiffies;
340 338
341 struct cpufreq_policy *policy; 339 struct cpufreq_policy *policy;
342 unsigned int j; 340 unsigned int j;
@@ -346,13 +344,7 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
346 344
347 this_dbs_info->freq_lo = 0; 345 this_dbs_info->freq_lo = 0;
348 policy = this_dbs_info->cur_policy; 346 policy = this_dbs_info->cur_policy;
349 cur_jiffies = jiffies64_to_cputime64(get_jiffies_64());
350 total_ticks = (unsigned int) cputime64_sub(cur_jiffies,
351 this_dbs_info->prev_cpu_wall);
352 this_dbs_info->prev_cpu_wall = get_jiffies_64();
353 347
354 if (!total_ticks)
355 return;
356 /* 348 /*
357 * Every sampling_rate, we check, if current idle time is less 349 * Every sampling_rate, we check, if current idle time is less
358 * than 20% (default), then we try to increase frequency 350 * than 20% (default), then we try to increase frequency
@@ -365,27 +357,46 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
365 * 5% (default) of current frequency 357 * 5% (default) of current frequency
366 */ 358 */
367 359
368 /* Get Idle Time */ 360 /* Get Absolute Load - in terms of freq */
369 idle_ticks = UINT_MAX; 361 max_load_freq = 0;
362
370 for_each_cpu_mask_nr(j, policy->cpus) { 363 for_each_cpu_mask_nr(j, policy->cpus) {
371 cputime64_t total_idle_ticks;
372 unsigned int tmp_idle_ticks;
373 struct cpu_dbs_info_s *j_dbs_info; 364 struct cpu_dbs_info_s *j_dbs_info;
365 cputime64_t cur_wall_time, cur_idle_time;
366 unsigned int idle_time, wall_time;
367 unsigned int load, load_freq;
368 int freq_avg;
374 369
375 j_dbs_info = &per_cpu(cpu_dbs_info, j); 370 j_dbs_info = &per_cpu(cpu_dbs_info, j);
376 total_idle_ticks = get_cpu_idle_time(j); 371 cur_wall_time = jiffies64_to_cputime64(get_jiffies_64());
377 tmp_idle_ticks = (unsigned int) cputime64_sub(total_idle_ticks, 372 wall_time = (unsigned int) cputime64_sub(cur_wall_time,
373 j_dbs_info->prev_cpu_wall);
374 j_dbs_info->prev_cpu_wall = cur_wall_time;
375
376 cur_idle_time = get_cpu_idle_time(j);
377 idle_time = (unsigned int) cputime64_sub(cur_idle_time,
378 j_dbs_info->prev_cpu_idle); 378 j_dbs_info->prev_cpu_idle);
379 j_dbs_info->prev_cpu_idle = total_idle_ticks; 379 j_dbs_info->prev_cpu_idle = cur_idle_time;
380 380
381 if (tmp_idle_ticks < idle_ticks) 381 if (unlikely(wall_time <= idle_time ||
382 idle_ticks = tmp_idle_ticks; 382 (cputime_to_msecs(wall_time) <
383 dbs_tuners_ins.sampling_rate / (2 * 1000)))) {
384 continue;
385 }
386
387 load = 100 * (wall_time - idle_time) / wall_time;
388
389 freq_avg = __cpufreq_driver_getavg(policy, j);
390 if (freq_avg <= 0)
391 freq_avg = policy->cur;
392
393 load_freq = load * freq_avg;
394 if (load_freq > max_load_freq)
395 max_load_freq = load_freq;
383 } 396 }
384 if (likely(total_ticks > idle_ticks))
385 load = (100 * (total_ticks - idle_ticks)) / total_ticks;
386 397
387 /* Check for frequency increase */ 398 /* Check for frequency increase */
388 if (load > dbs_tuners_ins.up_threshold) { 399 if (max_load_freq > dbs_tuners_ins.up_threshold * policy->cur) {
389 /* if we are already at full speed then break out early */ 400 /* if we are already at full speed then break out early */
390 if (!dbs_tuners_ins.powersave_bias) { 401 if (!dbs_tuners_ins.powersave_bias) {
391 if (policy->cur == policy->max) 402 if (policy->cur == policy->max)
@@ -412,15 +423,9 @@ static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info)
412 * can support the current CPU usage without triggering the up 423 * can support the current CPU usage without triggering the up
413 * policy. To be safe, we focus 10 points under the threshold. 424 * policy. To be safe, we focus 10 points under the threshold.
414 */ 425 */
415 if (load < (dbs_tuners_ins.up_threshold - 10)) { 426 if (max_load_freq < (dbs_tuners_ins.up_threshold - 10) * policy->cur) {
416 unsigned int freq_next, freq_cur; 427 unsigned int freq_next;
417 428 freq_next = max_load_freq / (dbs_tuners_ins.up_threshold - 10);
418 freq_cur = __cpufreq_driver_getavg(policy, policy->cpu);
419 if (!freq_cur)
420 freq_cur = policy->cur;
421
422 freq_next = (freq_cur * load) /
423 (dbs_tuners_ins.up_threshold - 10);
424 429
425 if (!dbs_tuners_ins.powersave_bias) { 430 if (!dbs_tuners_ins.powersave_bias) {
426 __cpufreq_driver_target(policy, freq_next, 431 __cpufreq_driver_target(policy, freq_next,