diff options
author | Arjan van de Ven <arjan@linux.intel.com> | 2006-07-26 09:40:07 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2006-07-26 10:21:40 -0400 |
commit | 153d7f3fcae7ed4e19328549aa9467acdfbced10 (patch) | |
tree | a7b15b844119663a276c7a99549ea5a06c16f19a /drivers/cpufreq/cpufreq_ondemand.c | |
parent | 44eb123126d289bac398cac0232309c228386671 (diff) |
[PATCH] Reorganize the cpufreq cpu hotplug locking to not be totally bizare
The patch below moves the cpu hotplugging higher up in the cpufreq
layering; this is needed to avoid recursive taking of the cpu hotplug
lock and to otherwise detangle the mess.
The new rules are:
1. you must do lock_cpu_hotplug() around the following functions:
__cpufreq_driver_target
__cpufreq_governor (for CPUFREQ_GOV_LIMITS operation only)
__cpufreq_set_policy
2. governer methods (.governer) must NOT take the lock_cpu_hotplug()
lock in any way; they are called with the lock taken already
3. if your governer spawns a thread that does things, like calling
__cpufreq_driver_target, your thread must honor rule #1.
4. the policy lock and other cpufreq internal locks nest within
the lock_cpu_hotplug() lock.
I'm not entirely happy about how the __cpufreq_governor rule ended up
(conditional locking rule depending on the argument) but basically all
callers pass this as a constant so it's not too horrible.
The patch also removes the cpufreq_governor() function since during the
locking audit it turned out to be entirely unused (so no need to fix it)
The patch works on my testbox, but it could use more testing
(otoh... it can't be much worse than the current code)
Signed-off-by: Arjan van de Ven <arjan@linux.intel.com>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'drivers/cpufreq/cpufreq_ondemand.c')
-rw-r--r-- | drivers/cpufreq/cpufreq_ondemand.c | 4 |
1 files changed, 2 insertions, 2 deletions
diff --git a/drivers/cpufreq/cpufreq_ondemand.c b/drivers/cpufreq/cpufreq_ondemand.c index 178f0c547eb7..52cf1f021825 100644 --- a/drivers/cpufreq/cpufreq_ondemand.c +++ b/drivers/cpufreq/cpufreq_ondemand.c | |||
@@ -309,7 +309,9 @@ static void do_dbs_timer(void *data) | |||
309 | if (!dbs_info->enable) | 309 | if (!dbs_info->enable) |
310 | return; | 310 | return; |
311 | 311 | ||
312 | lock_cpu_hotplug(); | ||
312 | dbs_check_cpu(dbs_info); | 313 | dbs_check_cpu(dbs_info); |
314 | unlock_cpu_hotplug(); | ||
313 | queue_delayed_work_on(cpu, kondemand_wq, &dbs_info->work, | 315 | queue_delayed_work_on(cpu, kondemand_wq, &dbs_info->work, |
314 | usecs_to_jiffies(dbs_tuners_ins.sampling_rate)); | 316 | usecs_to_jiffies(dbs_tuners_ins.sampling_rate)); |
315 | } | 317 | } |
@@ -412,7 +414,6 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
412 | break; | 414 | break; |
413 | 415 | ||
414 | case CPUFREQ_GOV_LIMITS: | 416 | case CPUFREQ_GOV_LIMITS: |
415 | lock_cpu_hotplug(); | ||
416 | mutex_lock(&dbs_mutex); | 417 | mutex_lock(&dbs_mutex); |
417 | if (policy->max < this_dbs_info->cur_policy->cur) | 418 | if (policy->max < this_dbs_info->cur_policy->cur) |
418 | __cpufreq_driver_target(this_dbs_info->cur_policy, | 419 | __cpufreq_driver_target(this_dbs_info->cur_policy, |
@@ -423,7 +424,6 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
423 | policy->min, | 424 | policy->min, |
424 | CPUFREQ_RELATION_L); | 425 | CPUFREQ_RELATION_L); |
425 | mutex_unlock(&dbs_mutex); | 426 | mutex_unlock(&dbs_mutex); |
426 | unlock_cpu_hotplug(); | ||
427 | break; | 427 | break; |
428 | } | 428 | } |
429 | return 0; | 429 | return 0; |