diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2009-03-26 14:04:08 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2009-03-26 14:04:08 -0400 |
commit | ada19a31a90b4f46c040c25ef4ef8ffc203c7fc6 (patch) | |
tree | 7d523d2d90dbaa973c5843d6219ec149b5949243 | |
parent | 8d80ce80e1d58ba9cd3e3972b112cccd6b4008f4 (diff) | |
parent | 36e8abf3edcd2d207193ec5741d1a2a645d470a5 (diff) |
Merge branch 'next' of git://git.kernel.org/pub/scm/linux/kernel/git/davej/cpufreq
* 'next' of git://git.kernel.org/pub/scm/linux/kernel/git/davej/cpufreq: (35 commits)
[CPUFREQ] Prevent p4-clockmod from auto-binding to the ondemand governor.
[CPUFREQ] Make cpufreq-nforce2 less obnoxious
[CPUFREQ] p4-clockmod reports wrong frequency.
[CPUFREQ] powernow-k8: Use a common exit path.
[CPUFREQ] Change link order of x86 cpufreq modules
[CPUFREQ] conservative: remove 10x from def_sampling_rate
[CPUFREQ] conservative: fixup governor to function more like ondemand logic
[CPUFREQ] conservative: fix dbs_cpufreq_notifier so freq is not locked
[CPUFREQ] conservative: amend author's email address
[CPUFREQ] Use swap() in longhaul.c
[CPUFREQ] checkpatch cleanups for acpi-cpufreq
[CPUFREQ] powernow-k8: Only print error message once, not per core.
[CPUFREQ] ondemand/conservative: sanitize sampling_rate restrictions
[CPUFREQ] ondemand/conservative: deprecate sampling_rate{min,max}
[CPUFREQ] powernow-k8: Always compile powernow-k8 driver with ACPI support
[CPUFREQ] Introduce /sys/devices/system/cpu/cpu*/cpufreq/cpuinfo_transition_latency
[CPUFREQ] checkpatch cleanups for powernow-k8
[CPUFREQ] checkpatch cleanups for ondemand governor.
[CPUFREQ] checkpatch cleanups for powernow-k7
[CPUFREQ] checkpatch cleanups for speedstep related drivers.
...
30 files changed, 1372 insertions, 995 deletions
diff --git a/Documentation/cpu-freq/governors.txt b/Documentation/cpu-freq/governors.txt index 5b0cfa67aff9..ce73f3eb5ddb 100644 --- a/Documentation/cpu-freq/governors.txt +++ b/Documentation/cpu-freq/governors.txt | |||
@@ -117,10 +117,28 @@ accessible parameters: | |||
117 | sampling_rate: measured in uS (10^-6 seconds), this is how often you | 117 | sampling_rate: measured in uS (10^-6 seconds), this is how often you |
118 | want the kernel to look at the CPU usage and to make decisions on | 118 | want the kernel to look at the CPU usage and to make decisions on |
119 | what to do about the frequency. Typically this is set to values of | 119 | what to do about the frequency. Typically this is set to values of |
120 | around '10000' or more. | 120 | around '10000' or more. It's default value is (cmp. with users-guide.txt): |
121 | 121 | transition_latency * 1000 | |
122 | show_sampling_rate_(min|max): the minimum and maximum sampling rates | 122 | The lowest value you can set is: |
123 | available that you may set 'sampling_rate' to. | 123 | transition_latency * 100 or it may get restricted to a value where it |
124 | makes not sense for the kernel anymore to poll that often which depends | ||
125 | on your HZ config variable (HZ=1000: max=20000us, HZ=250: max=5000). | ||
126 | Be aware that transition latency is in ns and sampling_rate is in us, so you | ||
127 | get the same sysfs value by default. | ||
128 | Sampling rate should always get adjusted considering the transition latency | ||
129 | To set the sampling rate 750 times as high as the transition latency | ||
130 | in the bash (as said, 1000 is default), do: | ||
131 | echo `$(($(cat cpuinfo_transition_latency) * 750 / 1000)) \ | ||
132 | >ondemand/sampling_rate | ||
133 | |||
134 | show_sampling_rate_(min|max): THIS INTERFACE IS DEPRECATED, DON'T USE IT. | ||
135 | You can use wider ranges now and the general | ||
136 | cpuinfo_transition_latency variable (cmp. with user-guide.txt) can be | ||
137 | used to obtain exactly the same info: | ||
138 | show_sampling_rate_min = transtition_latency * 500 / 1000 | ||
139 | show_sampling_rate_max = transtition_latency * 500000 / 1000 | ||
140 | (divided by 1000 is to illustrate that sampling rate is in us and | ||
141 | transition latency is exported ns). | ||
124 | 142 | ||
125 | up_threshold: defines what the average CPU usage between the samplings | 143 | up_threshold: defines what the average CPU usage between the samplings |
126 | of 'sampling_rate' needs to be for the kernel to make a decision on | 144 | of 'sampling_rate' needs to be for the kernel to make a decision on |
diff --git a/Documentation/cpu-freq/user-guide.txt b/Documentation/cpu-freq/user-guide.txt index 917918f84fc7..75f41193f3e1 100644 --- a/Documentation/cpu-freq/user-guide.txt +++ b/Documentation/cpu-freq/user-guide.txt | |||
@@ -152,6 +152,18 @@ cpuinfo_min_freq : this file shows the minimum operating | |||
152 | frequency the processor can run at(in kHz) | 152 | frequency the processor can run at(in kHz) |
153 | cpuinfo_max_freq : this file shows the maximum operating | 153 | cpuinfo_max_freq : this file shows the maximum operating |
154 | frequency the processor can run at(in kHz) | 154 | frequency the processor can run at(in kHz) |
155 | cpuinfo_transition_latency The time it takes on this CPU to | ||
156 | switch between two frequencies in nano | ||
157 | seconds. If unknown or known to be | ||
158 | that high that the driver does not | ||
159 | work with the ondemand governor, -1 | ||
160 | (CPUFREQ_ETERNAL) will be returned. | ||
161 | Using this information can be useful | ||
162 | to choose an appropriate polling | ||
163 | frequency for a kernel governor or | ||
164 | userspace daemon. Make sure to not | ||
165 | switch the frequency too often | ||
166 | resulting in performance loss. | ||
155 | scaling_driver : this file shows what cpufreq driver is | 167 | scaling_driver : this file shows what cpufreq driver is |
156 | used to set the frequency on this CPU | 168 | used to set the frequency on this CPU |
157 | 169 | ||
diff --git a/arch/x86/include/asm/timer.h b/arch/x86/include/asm/timer.h index 2bb6a835c453..4f5c24724856 100644 --- a/arch/x86/include/asm/timer.h +++ b/arch/x86/include/asm/timer.h | |||
@@ -11,8 +11,8 @@ unsigned long native_calibrate_tsc(void); | |||
11 | 11 | ||
12 | #ifdef CONFIG_X86_32 | 12 | #ifdef CONFIG_X86_32 |
13 | extern int timer_ack; | 13 | extern int timer_ack; |
14 | #endif | ||
14 | extern int recalibrate_cpu_khz(void); | 15 | extern int recalibrate_cpu_khz(void); |
15 | #endif /* CONFIG_X86_32 */ | ||
16 | 16 | ||
17 | extern int no_timer_check; | 17 | extern int no_timer_check; |
18 | 18 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/Kconfig b/arch/x86/kernel/cpu/cpufreq/Kconfig index 65792c2cc462..52c839875478 100644 --- a/arch/x86/kernel/cpu/cpufreq/Kconfig +++ b/arch/x86/kernel/cpu/cpufreq/Kconfig | |||
@@ -87,30 +87,15 @@ config X86_POWERNOW_K7_ACPI | |||
87 | config X86_POWERNOW_K8 | 87 | config X86_POWERNOW_K8 |
88 | tristate "AMD Opteron/Athlon64 PowerNow!" | 88 | tristate "AMD Opteron/Athlon64 PowerNow!" |
89 | select CPU_FREQ_TABLE | 89 | select CPU_FREQ_TABLE |
90 | depends on ACPI && ACPI_PROCESSOR | ||
90 | help | 91 | help |
91 | This adds the CPUFreq driver for mobile AMD Opteron/Athlon64 processors. | 92 | This adds the CPUFreq driver for K8/K10 Opteron/Athlon64 processors. |
92 | 93 | ||
93 | To compile this driver as a module, choose M here: the | 94 | To compile this driver as a module, choose M here: the |
94 | module will be called powernow-k8. | 95 | module will be called powernow-k8. |
95 | 96 | ||
96 | For details, take a look at <file:Documentation/cpu-freq/>. | 97 | For details, take a look at <file:Documentation/cpu-freq/>. |
97 | 98 | ||
98 | If in doubt, say N. | ||
99 | |||
100 | config X86_POWERNOW_K8_ACPI | ||
101 | bool | ||
102 | prompt "ACPI Support" if X86_32 | ||
103 | depends on ACPI && X86_POWERNOW_K8 && ACPI_PROCESSOR | ||
104 | depends on !(X86_POWERNOW_K8 = y && ACPI_PROCESSOR = m) | ||
105 | default y | ||
106 | help | ||
107 | This provides access to the K8s Processor Performance States via ACPI. | ||
108 | This driver is probably required for CPUFreq to work with multi-socket and | ||
109 | SMP systems. It is not required on at least some single-socket yet | ||
110 | multi-core systems, even if SMP is enabled. | ||
111 | |||
112 | It is safe to say Y here. | ||
113 | |||
114 | config X86_GX_SUSPMOD | 99 | config X86_GX_SUSPMOD |
115 | tristate "Cyrix MediaGX/NatSemi Geode Suspend Modulation" | 100 | tristate "Cyrix MediaGX/NatSemi Geode Suspend Modulation" |
116 | depends on X86_32 && PCI | 101 | depends on X86_32 && PCI |
diff --git a/arch/x86/kernel/cpu/cpufreq/Makefile b/arch/x86/kernel/cpu/cpufreq/Makefile index 560f7760dae5..509296df294d 100644 --- a/arch/x86/kernel/cpu/cpufreq/Makefile +++ b/arch/x86/kernel/cpu/cpufreq/Makefile | |||
@@ -1,6 +1,11 @@ | |||
1 | # Link order matters. K8 is preferred to ACPI because of firmware bugs in early | ||
2 | # K8 systems. ACPI is preferred to all other hardware-specific drivers. | ||
3 | # speedstep-* is preferred over p4-clockmod. | ||
4 | |||
5 | obj-$(CONFIG_X86_POWERNOW_K8) += powernow-k8.o | ||
6 | obj-$(CONFIG_X86_ACPI_CPUFREQ) += acpi-cpufreq.o | ||
1 | obj-$(CONFIG_X86_POWERNOW_K6) += powernow-k6.o | 7 | obj-$(CONFIG_X86_POWERNOW_K6) += powernow-k6.o |
2 | obj-$(CONFIG_X86_POWERNOW_K7) += powernow-k7.o | 8 | obj-$(CONFIG_X86_POWERNOW_K7) += powernow-k7.o |
3 | obj-$(CONFIG_X86_POWERNOW_K8) += powernow-k8.o | ||
4 | obj-$(CONFIG_X86_LONGHAUL) += longhaul.o | 9 | obj-$(CONFIG_X86_LONGHAUL) += longhaul.o |
5 | obj-$(CONFIG_X86_E_POWERSAVER) += e_powersaver.o | 10 | obj-$(CONFIG_X86_E_POWERSAVER) += e_powersaver.o |
6 | obj-$(CONFIG_ELAN_CPUFREQ) += elanfreq.o | 11 | obj-$(CONFIG_ELAN_CPUFREQ) += elanfreq.o |
@@ -10,7 +15,6 @@ obj-$(CONFIG_X86_GX_SUSPMOD) += gx-suspmod.o | |||
10 | obj-$(CONFIG_X86_SPEEDSTEP_ICH) += speedstep-ich.o | 15 | obj-$(CONFIG_X86_SPEEDSTEP_ICH) += speedstep-ich.o |
11 | obj-$(CONFIG_X86_SPEEDSTEP_LIB) += speedstep-lib.o | 16 | obj-$(CONFIG_X86_SPEEDSTEP_LIB) += speedstep-lib.o |
12 | obj-$(CONFIG_X86_SPEEDSTEP_SMI) += speedstep-smi.o | 17 | obj-$(CONFIG_X86_SPEEDSTEP_SMI) += speedstep-smi.o |
13 | obj-$(CONFIG_X86_ACPI_CPUFREQ) += acpi-cpufreq.o | ||
14 | obj-$(CONFIG_X86_SPEEDSTEP_CENTRINO) += speedstep-centrino.o | 18 | obj-$(CONFIG_X86_SPEEDSTEP_CENTRINO) += speedstep-centrino.o |
15 | obj-$(CONFIG_X86_P4_CLOCKMOD) += p4-clockmod.o | 19 | obj-$(CONFIG_X86_P4_CLOCKMOD) += p4-clockmod.o |
16 | obj-$(CONFIG_X86_CPUFREQ_NFORCE2) += cpufreq-nforce2.o | 20 | obj-$(CONFIG_X86_CPUFREQ_NFORCE2) += cpufreq-nforce2.o |
diff --git a/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c b/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c index 4b1c319d30c3..3babe1f1e912 100644 --- a/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c +++ b/arch/x86/kernel/cpu/cpufreq/acpi-cpufreq.c | |||
@@ -1,5 +1,5 @@ | |||
1 | /* | 1 | /* |
2 | * acpi-cpufreq.c - ACPI Processor P-States Driver ($Revision: 1.4 $) | 2 | * acpi-cpufreq.c - ACPI Processor P-States Driver |
3 | * | 3 | * |
4 | * Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com> | 4 | * Copyright (C) 2001, 2002 Andy Grover <andrew.grover@intel.com> |
5 | * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com> | 5 | * Copyright (C) 2001, 2002 Paul Diefenbaugh <paul.s.diefenbaugh@intel.com> |
@@ -36,16 +36,18 @@ | |||
36 | #include <linux/ftrace.h> | 36 | #include <linux/ftrace.h> |
37 | 37 | ||
38 | #include <linux/acpi.h> | 38 | #include <linux/acpi.h> |
39 | #include <linux/io.h> | ||
40 | #include <linux/delay.h> | ||
41 | #include <linux/uaccess.h> | ||
42 | |||
39 | #include <acpi/processor.h> | 43 | #include <acpi/processor.h> |
40 | 44 | ||
41 | #include <asm/io.h> | ||
42 | #include <asm/msr.h> | 45 | #include <asm/msr.h> |
43 | #include <asm/processor.h> | 46 | #include <asm/processor.h> |
44 | #include <asm/cpufeature.h> | 47 | #include <asm/cpufeature.h> |
45 | #include <asm/delay.h> | ||
46 | #include <asm/uaccess.h> | ||
47 | 48 | ||
48 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "acpi-cpufreq", msg) | 49 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
50 | "acpi-cpufreq", msg) | ||
49 | 51 | ||
50 | MODULE_AUTHOR("Paul Diefenbaugh, Dominik Brodowski"); | 52 | MODULE_AUTHOR("Paul Diefenbaugh, Dominik Brodowski"); |
51 | MODULE_DESCRIPTION("ACPI Processor P-States Driver"); | 53 | MODULE_DESCRIPTION("ACPI Processor P-States Driver"); |
@@ -95,7 +97,7 @@ static unsigned extract_io(u32 value, struct acpi_cpufreq_data *data) | |||
95 | 97 | ||
96 | perf = data->acpi_data; | 98 | perf = data->acpi_data; |
97 | 99 | ||
98 | for (i=0; i<perf->state_count; i++) { | 100 | for (i = 0; i < perf->state_count; i++) { |
99 | if (value == perf->states[i].status) | 101 | if (value == perf->states[i].status) |
100 | return data->freq_table[i].frequency; | 102 | return data->freq_table[i].frequency; |
101 | } | 103 | } |
@@ -110,7 +112,7 @@ static unsigned extract_msr(u32 msr, struct acpi_cpufreq_data *data) | |||
110 | msr &= INTEL_MSR_RANGE; | 112 | msr &= INTEL_MSR_RANGE; |
111 | perf = data->acpi_data; | 113 | perf = data->acpi_data; |
112 | 114 | ||
113 | for (i=0; data->freq_table[i].frequency != CPUFREQ_TABLE_END; i++) { | 115 | for (i = 0; data->freq_table[i].frequency != CPUFREQ_TABLE_END; i++) { |
114 | if (msr == perf->states[data->freq_table[i].index].status) | 116 | if (msr == perf->states[data->freq_table[i].index].status) |
115 | return data->freq_table[i].frequency; | 117 | return data->freq_table[i].frequency; |
116 | } | 118 | } |
@@ -138,15 +140,13 @@ struct io_addr { | |||
138 | u8 bit_width; | 140 | u8 bit_width; |
139 | }; | 141 | }; |
140 | 142 | ||
141 | typedef union { | ||
142 | struct msr_addr msr; | ||
143 | struct io_addr io; | ||
144 | } drv_addr_union; | ||
145 | |||
146 | struct drv_cmd { | 143 | struct drv_cmd { |
147 | unsigned int type; | 144 | unsigned int type; |
148 | const struct cpumask *mask; | 145 | const struct cpumask *mask; |
149 | drv_addr_union addr; | 146 | union { |
147 | struct msr_addr msr; | ||
148 | struct io_addr io; | ||
149 | } addr; | ||
150 | u32 val; | 150 | u32 val; |
151 | }; | 151 | }; |
152 | 152 | ||
@@ -369,7 +369,7 @@ static unsigned int check_freqs(const struct cpumask *mask, unsigned int freq, | |||
369 | unsigned int cur_freq; | 369 | unsigned int cur_freq; |
370 | unsigned int i; | 370 | unsigned int i; |
371 | 371 | ||
372 | for (i=0; i<100; i++) { | 372 | for (i = 0; i < 100; i++) { |
373 | cur_freq = extract_freq(get_cur_val(mask), data); | 373 | cur_freq = extract_freq(get_cur_val(mask), data); |
374 | if (cur_freq == freq) | 374 | if (cur_freq == freq) |
375 | return 1; | 375 | return 1; |
@@ -494,7 +494,7 @@ acpi_cpufreq_guess_freq(struct acpi_cpufreq_data *data, unsigned int cpu) | |||
494 | unsigned long freq; | 494 | unsigned long freq; |
495 | unsigned long freqn = perf->states[0].core_frequency * 1000; | 495 | unsigned long freqn = perf->states[0].core_frequency * 1000; |
496 | 496 | ||
497 | for (i=0; i<(perf->state_count-1); i++) { | 497 | for (i = 0; i < (perf->state_count-1); i++) { |
498 | freq = freqn; | 498 | freq = freqn; |
499 | freqn = perf->states[i+1].core_frequency * 1000; | 499 | freqn = perf->states[i+1].core_frequency * 1000; |
500 | if ((2 * cpu_khz) > (freqn + freq)) { | 500 | if ((2 * cpu_khz) > (freqn + freq)) { |
@@ -673,7 +673,7 @@ static int acpi_cpufreq_cpu_init(struct cpufreq_policy *policy) | |||
673 | 673 | ||
674 | /* detect transition latency */ | 674 | /* detect transition latency */ |
675 | policy->cpuinfo.transition_latency = 0; | 675 | policy->cpuinfo.transition_latency = 0; |
676 | for (i=0; i<perf->state_count; i++) { | 676 | for (i = 0; i < perf->state_count; i++) { |
677 | if ((perf->states[i].transition_latency * 1000) > | 677 | if ((perf->states[i].transition_latency * 1000) > |
678 | policy->cpuinfo.transition_latency) | 678 | policy->cpuinfo.transition_latency) |
679 | policy->cpuinfo.transition_latency = | 679 | policy->cpuinfo.transition_latency = |
@@ -682,8 +682,8 @@ static int acpi_cpufreq_cpu_init(struct cpufreq_policy *policy) | |||
682 | 682 | ||
683 | data->max_freq = perf->states[0].core_frequency * 1000; | 683 | data->max_freq = perf->states[0].core_frequency * 1000; |
684 | /* table init */ | 684 | /* table init */ |
685 | for (i=0; i<perf->state_count; i++) { | 685 | for (i = 0; i < perf->state_count; i++) { |
686 | if (i>0 && perf->states[i].core_frequency >= | 686 | if (i > 0 && perf->states[i].core_frequency >= |
687 | data->freq_table[valid_states-1].frequency / 1000) | 687 | data->freq_table[valid_states-1].frequency / 1000) |
688 | continue; | 688 | continue; |
689 | 689 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c b/arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c index 965ea52767ac..733093d60436 100644 --- a/arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c +++ b/arch/x86/kernel/cpu/cpufreq/cpufreq-nforce2.c | |||
@@ -32,7 +32,7 @@ | |||
32 | * nforce2_chipset: | 32 | * nforce2_chipset: |
33 | * FSB is changed using the chipset | 33 | * FSB is changed using the chipset |
34 | */ | 34 | */ |
35 | static struct pci_dev *nforce2_chipset_dev; | 35 | static struct pci_dev *nforce2_dev; |
36 | 36 | ||
37 | /* fid: | 37 | /* fid: |
38 | * multiplier * 10 | 38 | * multiplier * 10 |
@@ -56,7 +56,9 @@ MODULE_PARM_DESC(fid, "CPU multiplier to use (11.5 = 115)"); | |||
56 | MODULE_PARM_DESC(min_fsb, | 56 | MODULE_PARM_DESC(min_fsb, |
57 | "Minimum FSB to use, if not defined: current FSB - 50"); | 57 | "Minimum FSB to use, if not defined: current FSB - 50"); |
58 | 58 | ||
59 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "cpufreq-nforce2", msg) | 59 | #define PFX "cpufreq-nforce2: " |
60 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ | ||
61 | "cpufreq-nforce2", msg) | ||
60 | 62 | ||
61 | /** | 63 | /** |
62 | * nforce2_calc_fsb - calculate FSB | 64 | * nforce2_calc_fsb - calculate FSB |
@@ -118,11 +120,11 @@ static void nforce2_write_pll(int pll) | |||
118 | int temp; | 120 | int temp; |
119 | 121 | ||
120 | /* Set the pll addr. to 0x00 */ | 122 | /* Set the pll addr. to 0x00 */ |
121 | pci_write_config_dword(nforce2_chipset_dev, NFORCE2_PLLADR, 0); | 123 | pci_write_config_dword(nforce2_dev, NFORCE2_PLLADR, 0); |
122 | 124 | ||
123 | /* Now write the value in all 64 registers */ | 125 | /* Now write the value in all 64 registers */ |
124 | for (temp = 0; temp <= 0x3f; temp++) | 126 | for (temp = 0; temp <= 0x3f; temp++) |
125 | pci_write_config_dword(nforce2_chipset_dev, NFORCE2_PLLREG, pll); | 127 | pci_write_config_dword(nforce2_dev, NFORCE2_PLLREG, pll); |
126 | 128 | ||
127 | return; | 129 | return; |
128 | } | 130 | } |
@@ -139,8 +141,8 @@ static unsigned int nforce2_fsb_read(int bootfsb) | |||
139 | u32 fsb, temp = 0; | 141 | u32 fsb, temp = 0; |
140 | 142 | ||
141 | /* Get chipset boot FSB from subdevice 5 (FSB at boot-time) */ | 143 | /* Get chipset boot FSB from subdevice 5 (FSB at boot-time) */ |
142 | nforce2_sub5 = pci_get_subsys(PCI_VENDOR_ID_NVIDIA, | 144 | nforce2_sub5 = pci_get_subsys(PCI_VENDOR_ID_NVIDIA, 0x01EF, |
143 | 0x01EF, PCI_ANY_ID, PCI_ANY_ID, NULL); | 145 | PCI_ANY_ID, PCI_ANY_ID, NULL); |
144 | if (!nforce2_sub5) | 146 | if (!nforce2_sub5) |
145 | return 0; | 147 | return 0; |
146 | 148 | ||
@@ -148,13 +150,13 @@ static unsigned int nforce2_fsb_read(int bootfsb) | |||
148 | fsb /= 1000000; | 150 | fsb /= 1000000; |
149 | 151 | ||
150 | /* Check if PLL register is already set */ | 152 | /* Check if PLL register is already set */ |
151 | pci_read_config_byte(nforce2_chipset_dev, NFORCE2_PLLENABLE, (u8 *)&temp); | 153 | pci_read_config_byte(nforce2_dev, NFORCE2_PLLENABLE, (u8 *)&temp); |
152 | 154 | ||
153 | if (bootfsb || !temp) | 155 | if (bootfsb || !temp) |
154 | return fsb; | 156 | return fsb; |
155 | 157 | ||
156 | /* Use PLL register FSB value */ | 158 | /* Use PLL register FSB value */ |
157 | pci_read_config_dword(nforce2_chipset_dev, NFORCE2_PLLREG, &temp); | 159 | pci_read_config_dword(nforce2_dev, NFORCE2_PLLREG, &temp); |
158 | fsb = nforce2_calc_fsb(temp); | 160 | fsb = nforce2_calc_fsb(temp); |
159 | 161 | ||
160 | return fsb; | 162 | return fsb; |
@@ -174,18 +176,18 @@ static int nforce2_set_fsb(unsigned int fsb) | |||
174 | int pll = 0; | 176 | int pll = 0; |
175 | 177 | ||
176 | if ((fsb > max_fsb) || (fsb < NFORCE2_MIN_FSB)) { | 178 | if ((fsb > max_fsb) || (fsb < NFORCE2_MIN_FSB)) { |
177 | printk(KERN_ERR "cpufreq: FSB %d is out of range!\n", fsb); | 179 | printk(KERN_ERR PFX "FSB %d is out of range!\n", fsb); |
178 | return -EINVAL; | 180 | return -EINVAL; |
179 | } | 181 | } |
180 | 182 | ||
181 | tfsb = nforce2_fsb_read(0); | 183 | tfsb = nforce2_fsb_read(0); |
182 | if (!tfsb) { | 184 | if (!tfsb) { |
183 | printk(KERN_ERR "cpufreq: Error while reading the FSB\n"); | 185 | printk(KERN_ERR PFX "Error while reading the FSB\n"); |
184 | return -EINVAL; | 186 | return -EINVAL; |
185 | } | 187 | } |
186 | 188 | ||
187 | /* First write? Then set actual value */ | 189 | /* First write? Then set actual value */ |
188 | pci_read_config_byte(nforce2_chipset_dev, NFORCE2_PLLENABLE, (u8 *)&temp); | 190 | pci_read_config_byte(nforce2_dev, NFORCE2_PLLENABLE, (u8 *)&temp); |
189 | if (!temp) { | 191 | if (!temp) { |
190 | pll = nforce2_calc_pll(tfsb); | 192 | pll = nforce2_calc_pll(tfsb); |
191 | 193 | ||
@@ -197,7 +199,7 @@ static int nforce2_set_fsb(unsigned int fsb) | |||
197 | 199 | ||
198 | /* Enable write access */ | 200 | /* Enable write access */ |
199 | temp = 0x01; | 201 | temp = 0x01; |
200 | pci_write_config_byte(nforce2_chipset_dev, NFORCE2_PLLENABLE, (u8)temp); | 202 | pci_write_config_byte(nforce2_dev, NFORCE2_PLLENABLE, (u8)temp); |
201 | 203 | ||
202 | diff = tfsb - fsb; | 204 | diff = tfsb - fsb; |
203 | 205 | ||
@@ -222,7 +224,7 @@ static int nforce2_set_fsb(unsigned int fsb) | |||
222 | } | 224 | } |
223 | 225 | ||
224 | temp = 0x40; | 226 | temp = 0x40; |
225 | pci_write_config_byte(nforce2_chipset_dev, NFORCE2_PLLADR, (u8)temp); | 227 | pci_write_config_byte(nforce2_dev, NFORCE2_PLLADR, (u8)temp); |
226 | 228 | ||
227 | return 0; | 229 | return 0; |
228 | } | 230 | } |
@@ -244,7 +246,8 @@ static unsigned int nforce2_get(unsigned int cpu) | |||
244 | * nforce2_target - set a new CPUFreq policy | 246 | * nforce2_target - set a new CPUFreq policy |
245 | * @policy: new policy | 247 | * @policy: new policy |
246 | * @target_freq: the target frequency | 248 | * @target_freq: the target frequency |
247 | * @relation: how that frequency relates to achieved frequency (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | 249 | * @relation: how that frequency relates to achieved frequency |
250 | * (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | ||
248 | * | 251 | * |
249 | * Sets a new CPUFreq policy. | 252 | * Sets a new CPUFreq policy. |
250 | */ | 253 | */ |
@@ -276,7 +279,7 @@ static int nforce2_target(struct cpufreq_policy *policy, | |||
276 | /* local_irq_save(flags); */ | 279 | /* local_irq_save(flags); */ |
277 | 280 | ||
278 | if (nforce2_set_fsb(target_fsb) < 0) | 281 | if (nforce2_set_fsb(target_fsb) < 0) |
279 | printk(KERN_ERR "cpufreq: Changing FSB to %d failed\n", | 282 | printk(KERN_ERR PFX "Changing FSB to %d failed\n", |
280 | target_fsb); | 283 | target_fsb); |
281 | else | 284 | else |
282 | dprintk("Changed FSB successfully to %d\n", | 285 | dprintk("Changed FSB successfully to %d\n", |
@@ -327,8 +330,8 @@ static int nforce2_cpu_init(struct cpufreq_policy *policy) | |||
327 | /* FIX: Get FID from CPU */ | 330 | /* FIX: Get FID from CPU */ |
328 | if (!fid) { | 331 | if (!fid) { |
329 | if (!cpu_khz) { | 332 | if (!cpu_khz) { |
330 | printk(KERN_WARNING | 333 | printk(KERN_WARNING PFX |
331 | "cpufreq: cpu_khz not set, can't calculate multiplier!\n"); | 334 | "cpu_khz not set, can't calculate multiplier!\n"); |
332 | return -ENODEV; | 335 | return -ENODEV; |
333 | } | 336 | } |
334 | 337 | ||
@@ -343,7 +346,7 @@ static int nforce2_cpu_init(struct cpufreq_policy *policy) | |||
343 | } | 346 | } |
344 | } | 347 | } |
345 | 348 | ||
346 | printk(KERN_INFO "cpufreq: FSB currently at %i MHz, FID %d.%d\n", fsb, | 349 | printk(KERN_INFO PFX "FSB currently at %i MHz, FID %d.%d\n", fsb, |
347 | fid / 10, fid % 10); | 350 | fid / 10, fid % 10); |
348 | 351 | ||
349 | /* Set maximum FSB to FSB at boot time */ | 352 | /* Set maximum FSB to FSB at boot time */ |
@@ -392,17 +395,18 @@ static struct cpufreq_driver nforce2_driver = { | |||
392 | */ | 395 | */ |
393 | static unsigned int nforce2_detect_chipset(void) | 396 | static unsigned int nforce2_detect_chipset(void) |
394 | { | 397 | { |
395 | nforce2_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_NVIDIA, | 398 | nforce2_dev = pci_get_subsys(PCI_VENDOR_ID_NVIDIA, |
396 | PCI_DEVICE_ID_NVIDIA_NFORCE2, | 399 | PCI_DEVICE_ID_NVIDIA_NFORCE2, |
397 | PCI_ANY_ID, PCI_ANY_ID, NULL); | 400 | PCI_ANY_ID, PCI_ANY_ID, NULL); |
398 | 401 | ||
399 | if (nforce2_chipset_dev == NULL) | 402 | if (nforce2_dev == NULL) |
400 | return -ENODEV; | 403 | return -ENODEV; |
401 | 404 | ||
402 | printk(KERN_INFO "cpufreq: Detected nForce2 chipset revision %X\n", | 405 | printk(KERN_INFO PFX "Detected nForce2 chipset revision %X\n", |
403 | nforce2_chipset_dev->revision); | 406 | nforce2_dev->revision); |
404 | printk(KERN_INFO | 407 | printk(KERN_INFO PFX |
405 | "cpufreq: FSB changing is maybe unstable and can lead to crashes and data loss.\n"); | 408 | "FSB changing is maybe unstable and can lead to " |
409 | "crashes and data loss.\n"); | ||
406 | 410 | ||
407 | return 0; | 411 | return 0; |
408 | } | 412 | } |
@@ -420,7 +424,7 @@ static int __init nforce2_init(void) | |||
420 | 424 | ||
421 | /* detect chipset */ | 425 | /* detect chipset */ |
422 | if (nforce2_detect_chipset()) { | 426 | if (nforce2_detect_chipset()) { |
423 | printk(KERN_ERR "cpufreq: No nForce2 chipset.\n"); | 427 | printk(KERN_INFO PFX "No nForce2 chipset.\n"); |
424 | return -ENODEV; | 428 | return -ENODEV; |
425 | } | 429 | } |
426 | 430 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/e_powersaver.c b/arch/x86/kernel/cpu/cpufreq/e_powersaver.c index c2f930d86640..3f83ea12c47a 100644 --- a/arch/x86/kernel/cpu/cpufreq/e_powersaver.c +++ b/arch/x86/kernel/cpu/cpufreq/e_powersaver.c | |||
@@ -12,12 +12,12 @@ | |||
12 | #include <linux/cpufreq.h> | 12 | #include <linux/cpufreq.h> |
13 | #include <linux/ioport.h> | 13 | #include <linux/ioport.h> |
14 | #include <linux/slab.h> | 14 | #include <linux/slab.h> |
15 | #include <linux/timex.h> | ||
16 | #include <linux/io.h> | ||
17 | #include <linux/delay.h> | ||
15 | 18 | ||
16 | #include <asm/msr.h> | 19 | #include <asm/msr.h> |
17 | #include <asm/tsc.h> | 20 | #include <asm/tsc.h> |
18 | #include <asm/timex.h> | ||
19 | #include <asm/io.h> | ||
20 | #include <asm/delay.h> | ||
21 | 21 | ||
22 | #define EPS_BRAND_C7M 0 | 22 | #define EPS_BRAND_C7M 0 |
23 | #define EPS_BRAND_C7 1 | 23 | #define EPS_BRAND_C7 1 |
@@ -184,7 +184,7 @@ static int eps_cpu_init(struct cpufreq_policy *policy) | |||
184 | break; | 184 | break; |
185 | } | 185 | } |
186 | 186 | ||
187 | switch(brand) { | 187 | switch (brand) { |
188 | case EPS_BRAND_C7M: | 188 | case EPS_BRAND_C7M: |
189 | printk(KERN_CONT "C7-M\n"); | 189 | printk(KERN_CONT "C7-M\n"); |
190 | break; | 190 | break; |
@@ -218,17 +218,20 @@ static int eps_cpu_init(struct cpufreq_policy *policy) | |||
218 | /* Print voltage and multiplier */ | 218 | /* Print voltage and multiplier */ |
219 | rdmsr(MSR_IA32_PERF_STATUS, lo, hi); | 219 | rdmsr(MSR_IA32_PERF_STATUS, lo, hi); |
220 | current_voltage = lo & 0xff; | 220 | current_voltage = lo & 0xff; |
221 | printk(KERN_INFO "eps: Current voltage = %dmV\n", current_voltage * 16 + 700); | 221 | printk(KERN_INFO "eps: Current voltage = %dmV\n", |
222 | current_voltage * 16 + 700); | ||
222 | current_multiplier = (lo >> 8) & 0xff; | 223 | current_multiplier = (lo >> 8) & 0xff; |
223 | printk(KERN_INFO "eps: Current multiplier = %d\n", current_multiplier); | 224 | printk(KERN_INFO "eps: Current multiplier = %d\n", current_multiplier); |
224 | 225 | ||
225 | /* Print limits */ | 226 | /* Print limits */ |
226 | max_voltage = hi & 0xff; | 227 | max_voltage = hi & 0xff; |
227 | printk(KERN_INFO "eps: Highest voltage = %dmV\n", max_voltage * 16 + 700); | 228 | printk(KERN_INFO "eps: Highest voltage = %dmV\n", |
229 | max_voltage * 16 + 700); | ||
228 | max_multiplier = (hi >> 8) & 0xff; | 230 | max_multiplier = (hi >> 8) & 0xff; |
229 | printk(KERN_INFO "eps: Highest multiplier = %d\n", max_multiplier); | 231 | printk(KERN_INFO "eps: Highest multiplier = %d\n", max_multiplier); |
230 | min_voltage = (hi >> 16) & 0xff; | 232 | min_voltage = (hi >> 16) & 0xff; |
231 | printk(KERN_INFO "eps: Lowest voltage = %dmV\n", min_voltage * 16 + 700); | 233 | printk(KERN_INFO "eps: Lowest voltage = %dmV\n", |
234 | min_voltage * 16 + 700); | ||
232 | min_multiplier = (hi >> 24) & 0xff; | 235 | min_multiplier = (hi >> 24) & 0xff; |
233 | printk(KERN_INFO "eps: Lowest multiplier = %d\n", min_multiplier); | 236 | printk(KERN_INFO "eps: Lowest multiplier = %d\n", min_multiplier); |
234 | 237 | ||
@@ -318,7 +321,7 @@ static int eps_cpu_exit(struct cpufreq_policy *policy) | |||
318 | return 0; | 321 | return 0; |
319 | } | 322 | } |
320 | 323 | ||
321 | static struct freq_attr* eps_attr[] = { | 324 | static struct freq_attr *eps_attr[] = { |
322 | &cpufreq_freq_attr_scaling_available_freqs, | 325 | &cpufreq_freq_attr_scaling_available_freqs, |
323 | NULL, | 326 | NULL, |
324 | }; | 327 | }; |
@@ -356,7 +359,7 @@ static void __exit eps_exit(void) | |||
356 | cpufreq_unregister_driver(&eps_driver); | 359 | cpufreq_unregister_driver(&eps_driver); |
357 | } | 360 | } |
358 | 361 | ||
359 | MODULE_AUTHOR("Rafa³ Bilski <rafalbilski@interia.pl>"); | 362 | MODULE_AUTHOR("Rafal Bilski <rafalbilski@interia.pl>"); |
360 | MODULE_DESCRIPTION("Enhanced PowerSaver driver for VIA C7 CPU's."); | 363 | MODULE_DESCRIPTION("Enhanced PowerSaver driver for VIA C7 CPU's."); |
361 | MODULE_LICENSE("GPL"); | 364 | MODULE_LICENSE("GPL"); |
362 | 365 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/elanfreq.c b/arch/x86/kernel/cpu/cpufreq/elanfreq.c index fe613c93b366..006b278b0d5d 100644 --- a/arch/x86/kernel/cpu/cpufreq/elanfreq.c +++ b/arch/x86/kernel/cpu/cpufreq/elanfreq.c | |||
@@ -184,7 +184,8 @@ static int elanfreq_target(struct cpufreq_policy *policy, | |||
184 | { | 184 | { |
185 | unsigned int newstate = 0; | 185 | unsigned int newstate = 0; |
186 | 186 | ||
187 | if (cpufreq_frequency_table_target(policy, &elanfreq_table[0], target_freq, relation, &newstate)) | 187 | if (cpufreq_frequency_table_target(policy, &elanfreq_table[0], |
188 | target_freq, relation, &newstate)) | ||
188 | return -EINVAL; | 189 | return -EINVAL; |
189 | 190 | ||
190 | elanfreq_set_cpu_state(newstate); | 191 | elanfreq_set_cpu_state(newstate); |
@@ -301,7 +302,8 @@ static void __exit elanfreq_exit(void) | |||
301 | module_param(max_freq, int, 0444); | 302 | module_param(max_freq, int, 0444); |
302 | 303 | ||
303 | MODULE_LICENSE("GPL"); | 304 | MODULE_LICENSE("GPL"); |
304 | MODULE_AUTHOR("Robert Schwebel <r.schwebel@pengutronix.de>, Sven Geggus <sven@geggus.net>"); | 305 | MODULE_AUTHOR("Robert Schwebel <r.schwebel@pengutronix.de>, " |
306 | "Sven Geggus <sven@geggus.net>"); | ||
305 | MODULE_DESCRIPTION("cpufreq driver for AMD's Elan CPUs"); | 307 | MODULE_DESCRIPTION("cpufreq driver for AMD's Elan CPUs"); |
306 | 308 | ||
307 | module_init(elanfreq_init); | 309 | module_init(elanfreq_init); |
diff --git a/arch/x86/kernel/cpu/cpufreq/gx-suspmod.c b/arch/x86/kernel/cpu/cpufreq/gx-suspmod.c index 9d9eae82e60f..ac27ec2264d5 100644 --- a/arch/x86/kernel/cpu/cpufreq/gx-suspmod.c +++ b/arch/x86/kernel/cpu/cpufreq/gx-suspmod.c | |||
@@ -79,8 +79,9 @@ | |||
79 | #include <linux/smp.h> | 79 | #include <linux/smp.h> |
80 | #include <linux/cpufreq.h> | 80 | #include <linux/cpufreq.h> |
81 | #include <linux/pci.h> | 81 | #include <linux/pci.h> |
82 | #include <linux/errno.h> | ||
83 | |||
82 | #include <asm/processor-cyrix.h> | 84 | #include <asm/processor-cyrix.h> |
83 | #include <asm/errno.h> | ||
84 | 85 | ||
85 | /* PCI config registers, all at F0 */ | 86 | /* PCI config registers, all at F0 */ |
86 | #define PCI_PMER1 0x80 /* power management enable register 1 */ | 87 | #define PCI_PMER1 0x80 /* power management enable register 1 */ |
@@ -122,8 +123,8 @@ static struct gxfreq_params *gx_params; | |||
122 | static int stock_freq; | 123 | static int stock_freq; |
123 | 124 | ||
124 | /* PCI bus clock - defaults to 30.000 if cpu_khz is not available */ | 125 | /* PCI bus clock - defaults to 30.000 if cpu_khz is not available */ |
125 | static int pci_busclk = 0; | 126 | static int pci_busclk; |
126 | module_param (pci_busclk, int, 0444); | 127 | module_param(pci_busclk, int, 0444); |
127 | 128 | ||
128 | /* maximum duration for which the cpu may be suspended | 129 | /* maximum duration for which the cpu may be suspended |
129 | * (32us * MAX_DURATION). If no parameter is given, this defaults | 130 | * (32us * MAX_DURATION). If no parameter is given, this defaults |
@@ -132,7 +133,7 @@ module_param (pci_busclk, int, 0444); | |||
132 | * is suspended -- processing power is just 0.39% of what it used to be, | 133 | * is suspended -- processing power is just 0.39% of what it used to be, |
133 | * though. 781.25 kHz(!) for a 200 MHz processor -- wow. */ | 134 | * though. 781.25 kHz(!) for a 200 MHz processor -- wow. */ |
134 | static int max_duration = 255; | 135 | static int max_duration = 255; |
135 | module_param (max_duration, int, 0444); | 136 | module_param(max_duration, int, 0444); |
136 | 137 | ||
137 | /* For the default policy, we want at least some processing power | 138 | /* For the default policy, we want at least some processing power |
138 | * - let's say 5%. (min = maxfreq / POLICY_MIN_DIV) | 139 | * - let's say 5%. (min = maxfreq / POLICY_MIN_DIV) |
@@ -140,7 +141,8 @@ module_param (max_duration, int, 0444); | |||
140 | #define POLICY_MIN_DIV 20 | 141 | #define POLICY_MIN_DIV 20 |
141 | 142 | ||
142 | 143 | ||
143 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "gx-suspmod", msg) | 144 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
145 | "gx-suspmod", msg) | ||
144 | 146 | ||
145 | /** | 147 | /** |
146 | * we can detect a core multipiler from dir0_lsb | 148 | * we can detect a core multipiler from dir0_lsb |
@@ -166,12 +168,20 @@ static int gx_freq_mult[16] = { | |||
166 | * Low Level chipset interface * | 168 | * Low Level chipset interface * |
167 | ****************************************************************/ | 169 | ****************************************************************/ |
168 | static struct pci_device_id gx_chipset_tbl[] __initdata = { | 170 | static struct pci_device_id gx_chipset_tbl[] __initdata = { |
169 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5530_LEGACY, PCI_ANY_ID, PCI_ANY_ID }, | 171 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5530_LEGACY, |
170 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5520, PCI_ANY_ID, PCI_ANY_ID }, | 172 | PCI_ANY_ID, PCI_ANY_ID }, |
171 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5510, PCI_ANY_ID, PCI_ANY_ID }, | 173 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5520, |
174 | PCI_ANY_ID, PCI_ANY_ID }, | ||
175 | { PCI_VENDOR_ID_CYRIX, PCI_DEVICE_ID_CYRIX_5510, | ||
176 | PCI_ANY_ID, PCI_ANY_ID }, | ||
172 | { 0, }, | 177 | { 0, }, |
173 | }; | 178 | }; |
174 | 179 | ||
180 | static void gx_write_byte(int reg, int value) | ||
181 | { | ||
182 | pci_write_config_byte(gx_params->cs55x0, reg, value); | ||
183 | } | ||
184 | |||
175 | /** | 185 | /** |
176 | * gx_detect_chipset: | 186 | * gx_detect_chipset: |
177 | * | 187 | * |
@@ -200,7 +210,8 @@ static __init struct pci_dev *gx_detect_chipset(void) | |||
200 | /** | 210 | /** |
201 | * gx_get_cpuspeed: | 211 | * gx_get_cpuspeed: |
202 | * | 212 | * |
203 | * Finds out at which efficient frequency the Cyrix MediaGX/NatSemi Geode CPU runs. | 213 | * Finds out at which efficient frequency the Cyrix MediaGX/NatSemi |
214 | * Geode CPU runs. | ||
204 | */ | 215 | */ |
205 | static unsigned int gx_get_cpuspeed(unsigned int cpu) | 216 | static unsigned int gx_get_cpuspeed(unsigned int cpu) |
206 | { | 217 | { |
@@ -217,17 +228,18 @@ static unsigned int gx_get_cpuspeed(unsigned int cpu) | |||
217 | * | 228 | * |
218 | **/ | 229 | **/ |
219 | 230 | ||
220 | static unsigned int gx_validate_speed(unsigned int khz, u8 *on_duration, u8 *off_duration) | 231 | static unsigned int gx_validate_speed(unsigned int khz, u8 *on_duration, |
232 | u8 *off_duration) | ||
221 | { | 233 | { |
222 | unsigned int i; | 234 | unsigned int i; |
223 | u8 tmp_on, tmp_off; | 235 | u8 tmp_on, tmp_off; |
224 | int old_tmp_freq = stock_freq; | 236 | int old_tmp_freq = stock_freq; |
225 | int tmp_freq; | 237 | int tmp_freq; |
226 | 238 | ||
227 | *off_duration=1; | 239 | *off_duration = 1; |
228 | *on_duration=0; | 240 | *on_duration = 0; |
229 | 241 | ||
230 | for (i=max_duration; i>0; i--) { | 242 | for (i = max_duration; i > 0; i--) { |
231 | tmp_off = ((khz * i) / stock_freq) & 0xff; | 243 | tmp_off = ((khz * i) / stock_freq) & 0xff; |
232 | tmp_on = i - tmp_off; | 244 | tmp_on = i - tmp_off; |
233 | tmp_freq = (stock_freq * tmp_off) / i; | 245 | tmp_freq = (stock_freq * tmp_off) / i; |
@@ -259,26 +271,34 @@ static void gx_set_cpuspeed(unsigned int khz) | |||
259 | freqs.cpu = 0; | 271 | freqs.cpu = 0; |
260 | freqs.old = gx_get_cpuspeed(0); | 272 | freqs.old = gx_get_cpuspeed(0); |
261 | 273 | ||
262 | new_khz = gx_validate_speed(khz, &gx_params->on_duration, &gx_params->off_duration); | 274 | new_khz = gx_validate_speed(khz, &gx_params->on_duration, |
275 | &gx_params->off_duration); | ||
263 | 276 | ||
264 | freqs.new = new_khz; | 277 | freqs.new = new_khz; |
265 | 278 | ||
266 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); | 279 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); |
267 | local_irq_save(flags); | 280 | local_irq_save(flags); |
268 | 281 | ||
269 | if (new_khz != stock_freq) { /* if new khz == 100% of CPU speed, it is special case */ | 282 | |
283 | |||
284 | if (new_khz != stock_freq) { | ||
285 | /* if new khz == 100% of CPU speed, it is special case */ | ||
270 | switch (gx_params->cs55x0->device) { | 286 | switch (gx_params->cs55x0->device) { |
271 | case PCI_DEVICE_ID_CYRIX_5530_LEGACY: | 287 | case PCI_DEVICE_ID_CYRIX_5530_LEGACY: |
272 | pmer1 = gx_params->pci_pmer1 | IRQ_SPDUP | VID_SPDUP; | 288 | pmer1 = gx_params->pci_pmer1 | IRQ_SPDUP | VID_SPDUP; |
273 | /* FIXME: need to test other values -- Zwane,Miura */ | 289 | /* FIXME: need to test other values -- Zwane,Miura */ |
274 | pci_write_config_byte(gx_params->cs55x0, PCI_IRQTC, 4); /* typical 2 to 4ms */ | 290 | /* typical 2 to 4ms */ |
275 | pci_write_config_byte(gx_params->cs55x0, PCI_VIDTC, 100);/* typical 50 to 100ms */ | 291 | gx_write_byte(PCI_IRQTC, 4); |
276 | pci_write_config_byte(gx_params->cs55x0, PCI_PMER1, pmer1); | 292 | /* typical 50 to 100ms */ |
277 | 293 | gx_write_byte(PCI_VIDTC, 100); | |
278 | if (gx_params->cs55x0->revision < 0x10) { /* CS5530(rev 1.2, 1.3) */ | 294 | gx_write_byte(PCI_PMER1, pmer1); |
279 | suscfg = gx_params->pci_suscfg | SUSMOD; | 295 | |
280 | } else { /* CS5530A,B.. */ | 296 | if (gx_params->cs55x0->revision < 0x10) { |
281 | suscfg = gx_params->pci_suscfg | SUSMOD | PWRSVE; | 297 | /* CS5530(rev 1.2, 1.3) */ |
298 | suscfg = gx_params->pci_suscfg|SUSMOD; | ||
299 | } else { | ||
300 | /* CS5530A,B.. */ | ||
301 | suscfg = gx_params->pci_suscfg|SUSMOD|PWRSVE; | ||
282 | } | 302 | } |
283 | break; | 303 | break; |
284 | case PCI_DEVICE_ID_CYRIX_5520: | 304 | case PCI_DEVICE_ID_CYRIX_5520: |
@@ -294,13 +314,13 @@ static void gx_set_cpuspeed(unsigned int khz) | |||
294 | suscfg = gx_params->pci_suscfg & ~(SUSMOD); | 314 | suscfg = gx_params->pci_suscfg & ~(SUSMOD); |
295 | gx_params->off_duration = 0; | 315 | gx_params->off_duration = 0; |
296 | gx_params->on_duration = 0; | 316 | gx_params->on_duration = 0; |
297 | dprintk("suspend modulation disabled: cpu runs 100 percent speed.\n"); | 317 | dprintk("suspend modulation disabled: cpu runs 100%% speed.\n"); |
298 | } | 318 | } |
299 | 319 | ||
300 | pci_write_config_byte(gx_params->cs55x0, PCI_MODOFF, gx_params->off_duration); | 320 | gx_write_byte(PCI_MODOFF, gx_params->off_duration); |
301 | pci_write_config_byte(gx_params->cs55x0, PCI_MODON, gx_params->on_duration); | 321 | gx_write_byte(PCI_MODON, gx_params->on_duration); |
302 | 322 | ||
303 | pci_write_config_byte(gx_params->cs55x0, PCI_SUSCFG, suscfg); | 323 | gx_write_byte(PCI_SUSCFG, suscfg); |
304 | pci_read_config_byte(gx_params->cs55x0, PCI_SUSCFG, &suscfg); | 324 | pci_read_config_byte(gx_params->cs55x0, PCI_SUSCFG, &suscfg); |
305 | 325 | ||
306 | local_irq_restore(flags); | 326 | local_irq_restore(flags); |
@@ -334,7 +354,8 @@ static int cpufreq_gx_verify(struct cpufreq_policy *policy) | |||
334 | return -EINVAL; | 354 | return -EINVAL; |
335 | 355 | ||
336 | policy->cpu = 0; | 356 | policy->cpu = 0; |
337 | cpufreq_verify_within_limits(policy, (stock_freq / max_duration), stock_freq); | 357 | cpufreq_verify_within_limits(policy, (stock_freq / max_duration), |
358 | stock_freq); | ||
338 | 359 | ||
339 | /* it needs to be assured that at least one supported frequency is | 360 | /* it needs to be assured that at least one supported frequency is |
340 | * within policy->min and policy->max. If it is not, policy->max | 361 | * within policy->min and policy->max. If it is not, policy->max |
@@ -354,7 +375,8 @@ static int cpufreq_gx_verify(struct cpufreq_policy *policy) | |||
354 | policy->max = tmp_freq; | 375 | policy->max = tmp_freq; |
355 | if (policy->max < policy->min) | 376 | if (policy->max < policy->min) |
356 | policy->max = policy->min; | 377 | policy->max = policy->min; |
357 | cpufreq_verify_within_limits(policy, (stock_freq / max_duration), stock_freq); | 378 | cpufreq_verify_within_limits(policy, (stock_freq / max_duration), |
379 | stock_freq); | ||
358 | 380 | ||
359 | return 0; | 381 | return 0; |
360 | } | 382 | } |
@@ -398,18 +420,18 @@ static int cpufreq_gx_cpu_init(struct cpufreq_policy *policy) | |||
398 | return -ENODEV; | 420 | return -ENODEV; |
399 | 421 | ||
400 | /* determine maximum frequency */ | 422 | /* determine maximum frequency */ |
401 | if (pci_busclk) { | 423 | if (pci_busclk) |
402 | maxfreq = pci_busclk * gx_freq_mult[getCx86(CX86_DIR1) & 0x0f]; | 424 | maxfreq = pci_busclk * gx_freq_mult[getCx86(CX86_DIR1) & 0x0f]; |
403 | } else if (cpu_khz) { | 425 | else if (cpu_khz) |
404 | maxfreq = cpu_khz; | 426 | maxfreq = cpu_khz; |
405 | } else { | 427 | else |
406 | maxfreq = 30000 * gx_freq_mult[getCx86(CX86_DIR1) & 0x0f]; | 428 | maxfreq = 30000 * gx_freq_mult[getCx86(CX86_DIR1) & 0x0f]; |
407 | } | 429 | |
408 | stock_freq = maxfreq; | 430 | stock_freq = maxfreq; |
409 | curfreq = gx_get_cpuspeed(0); | 431 | curfreq = gx_get_cpuspeed(0); |
410 | 432 | ||
411 | dprintk("cpu max frequency is %d.\n", maxfreq); | 433 | dprintk("cpu max frequency is %d.\n", maxfreq); |
412 | dprintk("cpu current frequency is %dkHz.\n",curfreq); | 434 | dprintk("cpu current frequency is %dkHz.\n", curfreq); |
413 | 435 | ||
414 | /* setup basic struct for cpufreq API */ | 436 | /* setup basic struct for cpufreq API */ |
415 | policy->cpu = 0; | 437 | policy->cpu = 0; |
@@ -447,7 +469,8 @@ static int __init cpufreq_gx_init(void) | |||
447 | struct pci_dev *gx_pci; | 469 | struct pci_dev *gx_pci; |
448 | 470 | ||
449 | /* Test if we have the right hardware */ | 471 | /* Test if we have the right hardware */ |
450 | if ((gx_pci = gx_detect_chipset()) == NULL) | 472 | gx_pci = gx_detect_chipset(); |
473 | if (gx_pci == NULL) | ||
451 | return -ENODEV; | 474 | return -ENODEV; |
452 | 475 | ||
453 | /* check whether module parameters are sane */ | 476 | /* check whether module parameters are sane */ |
@@ -468,9 +491,11 @@ static int __init cpufreq_gx_init(void) | |||
468 | pci_read_config_byte(params->cs55x0, PCI_PMER1, &(params->pci_pmer1)); | 491 | pci_read_config_byte(params->cs55x0, PCI_PMER1, &(params->pci_pmer1)); |
469 | pci_read_config_byte(params->cs55x0, PCI_PMER2, &(params->pci_pmer2)); | 492 | pci_read_config_byte(params->cs55x0, PCI_PMER2, &(params->pci_pmer2)); |
470 | pci_read_config_byte(params->cs55x0, PCI_MODON, &(params->on_duration)); | 493 | pci_read_config_byte(params->cs55x0, PCI_MODON, &(params->on_duration)); |
471 | pci_read_config_byte(params->cs55x0, PCI_MODOFF, &(params->off_duration)); | 494 | pci_read_config_byte(params->cs55x0, PCI_MODOFF, |
495 | &(params->off_duration)); | ||
472 | 496 | ||
473 | if ((ret = cpufreq_register_driver(&gx_suspmod_driver))) { | 497 | ret = cpufreq_register_driver(&gx_suspmod_driver); |
498 | if (ret) { | ||
474 | kfree(params); | 499 | kfree(params); |
475 | return ret; /* register error! */ | 500 | return ret; /* register error! */ |
476 | } | 501 | } |
@@ -485,9 +510,9 @@ static void __exit cpufreq_gx_exit(void) | |||
485 | kfree(gx_params); | 510 | kfree(gx_params); |
486 | } | 511 | } |
487 | 512 | ||
488 | MODULE_AUTHOR ("Hiroshi Miura <miura@da-cha.org>"); | 513 | MODULE_AUTHOR("Hiroshi Miura <miura@da-cha.org>"); |
489 | MODULE_DESCRIPTION ("Cpufreq driver for Cyrix MediaGX and NatSemi Geode"); | 514 | MODULE_DESCRIPTION("Cpufreq driver for Cyrix MediaGX and NatSemi Geode"); |
490 | MODULE_LICENSE ("GPL"); | 515 | MODULE_LICENSE("GPL"); |
491 | 516 | ||
492 | module_init(cpufreq_gx_init); | 517 | module_init(cpufreq_gx_init); |
493 | module_exit(cpufreq_gx_exit); | 518 | module_exit(cpufreq_gx_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/longhaul.c b/arch/x86/kernel/cpu/cpufreq/longhaul.c index a4cff5d6e380..f1c51aea064d 100644 --- a/arch/x86/kernel/cpu/cpufreq/longhaul.c +++ b/arch/x86/kernel/cpu/cpufreq/longhaul.c | |||
@@ -30,12 +30,12 @@ | |||
30 | #include <linux/slab.h> | 30 | #include <linux/slab.h> |
31 | #include <linux/string.h> | 31 | #include <linux/string.h> |
32 | #include <linux/delay.h> | 32 | #include <linux/delay.h> |
33 | #include <linux/timex.h> | ||
34 | #include <linux/io.h> | ||
35 | #include <linux/acpi.h> | ||
36 | #include <linux/kernel.h> | ||
33 | 37 | ||
34 | #include <asm/msr.h> | 38 | #include <asm/msr.h> |
35 | #include <asm/timex.h> | ||
36 | #include <asm/io.h> | ||
37 | #include <asm/acpi.h> | ||
38 | #include <linux/acpi.h> | ||
39 | #include <acpi/processor.h> | 39 | #include <acpi/processor.h> |
40 | 40 | ||
41 | #include "longhaul.h" | 41 | #include "longhaul.h" |
@@ -58,7 +58,7 @@ | |||
58 | #define USE_NORTHBRIDGE (1 << 2) | 58 | #define USE_NORTHBRIDGE (1 << 2) |
59 | 59 | ||
60 | static int cpu_model; | 60 | static int cpu_model; |
61 | static unsigned int numscales=16; | 61 | static unsigned int numscales = 16; |
62 | static unsigned int fsb; | 62 | static unsigned int fsb; |
63 | 63 | ||
64 | static const struct mV_pos *vrm_mV_table; | 64 | static const struct mV_pos *vrm_mV_table; |
@@ -67,8 +67,8 @@ static const unsigned char *mV_vrm_table; | |||
67 | static unsigned int highest_speed, lowest_speed; /* kHz */ | 67 | static unsigned int highest_speed, lowest_speed; /* kHz */ |
68 | static unsigned int minmult, maxmult; | 68 | static unsigned int minmult, maxmult; |
69 | static int can_scale_voltage; | 69 | static int can_scale_voltage; |
70 | static struct acpi_processor *pr = NULL; | 70 | static struct acpi_processor *pr; |
71 | static struct acpi_processor_cx *cx = NULL; | 71 | static struct acpi_processor_cx *cx; |
72 | static u32 acpi_regs_addr; | 72 | static u32 acpi_regs_addr; |
73 | static u8 longhaul_flags; | 73 | static u8 longhaul_flags; |
74 | static unsigned int longhaul_index; | 74 | static unsigned int longhaul_index; |
@@ -78,12 +78,13 @@ static int scale_voltage; | |||
78 | static int disable_acpi_c3; | 78 | static int disable_acpi_c3; |
79 | static int revid_errata; | 79 | static int revid_errata; |
80 | 80 | ||
81 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "longhaul", msg) | 81 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
82 | "longhaul", msg) | ||
82 | 83 | ||
83 | 84 | ||
84 | /* Clock ratios multiplied by 10 */ | 85 | /* Clock ratios multiplied by 10 */ |
85 | static int clock_ratio[32]; | 86 | static int mults[32]; |
86 | static int eblcr_table[32]; | 87 | static int eblcr[32]; |
87 | static int longhaul_version; | 88 | static int longhaul_version; |
88 | static struct cpufreq_frequency_table *longhaul_table; | 89 | static struct cpufreq_frequency_table *longhaul_table; |
89 | 90 | ||
@@ -93,7 +94,7 @@ static char speedbuffer[8]; | |||
93 | static char *print_speed(int speed) | 94 | static char *print_speed(int speed) |
94 | { | 95 | { |
95 | if (speed < 1000) { | 96 | if (speed < 1000) { |
96 | snprintf(speedbuffer, sizeof(speedbuffer),"%dMHz", speed); | 97 | snprintf(speedbuffer, sizeof(speedbuffer), "%dMHz", speed); |
97 | return speedbuffer; | 98 | return speedbuffer; |
98 | } | 99 | } |
99 | 100 | ||
@@ -122,27 +123,28 @@ static unsigned int calc_speed(int mult) | |||
122 | 123 | ||
123 | static int longhaul_get_cpu_mult(void) | 124 | static int longhaul_get_cpu_mult(void) |
124 | { | 125 | { |
125 | unsigned long invalue=0,lo, hi; | 126 | unsigned long invalue = 0, lo, hi; |
126 | 127 | ||
127 | rdmsr (MSR_IA32_EBL_CR_POWERON, lo, hi); | 128 | rdmsr(MSR_IA32_EBL_CR_POWERON, lo, hi); |
128 | invalue = (lo & (1<<22|1<<23|1<<24|1<<25)) >>22; | 129 | invalue = (lo & (1<<22|1<<23|1<<24|1<<25))>>22; |
129 | if (longhaul_version==TYPE_LONGHAUL_V2 || longhaul_version==TYPE_POWERSAVER) { | 130 | if (longhaul_version == TYPE_LONGHAUL_V2 || |
131 | longhaul_version == TYPE_POWERSAVER) { | ||
130 | if (lo & (1<<27)) | 132 | if (lo & (1<<27)) |
131 | invalue+=16; | 133 | invalue += 16; |
132 | } | 134 | } |
133 | return eblcr_table[invalue]; | 135 | return eblcr[invalue]; |
134 | } | 136 | } |
135 | 137 | ||
136 | /* For processor with BCR2 MSR */ | 138 | /* For processor with BCR2 MSR */ |
137 | 139 | ||
138 | static void do_longhaul1(unsigned int clock_ratio_index) | 140 | static void do_longhaul1(unsigned int mults_index) |
139 | { | 141 | { |
140 | union msr_bcr2 bcr2; | 142 | union msr_bcr2 bcr2; |
141 | 143 | ||
142 | rdmsrl(MSR_VIA_BCR2, bcr2.val); | 144 | rdmsrl(MSR_VIA_BCR2, bcr2.val); |
143 | /* Enable software clock multiplier */ | 145 | /* Enable software clock multiplier */ |
144 | bcr2.bits.ESOFTBF = 1; | 146 | bcr2.bits.ESOFTBF = 1; |
145 | bcr2.bits.CLOCKMUL = clock_ratio_index & 0xff; | 147 | bcr2.bits.CLOCKMUL = mults_index & 0xff; |
146 | 148 | ||
147 | /* Sync to timer tick */ | 149 | /* Sync to timer tick */ |
148 | safe_halt(); | 150 | safe_halt(); |
@@ -161,7 +163,7 @@ static void do_longhaul1(unsigned int clock_ratio_index) | |||
161 | 163 | ||
162 | /* For processor with Longhaul MSR */ | 164 | /* For processor with Longhaul MSR */ |
163 | 165 | ||
164 | static void do_powersaver(int cx_address, unsigned int clock_ratio_index, | 166 | static void do_powersaver(int cx_address, unsigned int mults_index, |
165 | unsigned int dir) | 167 | unsigned int dir) |
166 | { | 168 | { |
167 | union msr_longhaul longhaul; | 169 | union msr_longhaul longhaul; |
@@ -173,11 +175,11 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index, | |||
173 | longhaul.bits.RevisionKey = longhaul.bits.RevisionID; | 175 | longhaul.bits.RevisionKey = longhaul.bits.RevisionID; |
174 | else | 176 | else |
175 | longhaul.bits.RevisionKey = 0; | 177 | longhaul.bits.RevisionKey = 0; |
176 | longhaul.bits.SoftBusRatio = clock_ratio_index & 0xf; | 178 | longhaul.bits.SoftBusRatio = mults_index & 0xf; |
177 | longhaul.bits.SoftBusRatio4 = (clock_ratio_index & 0x10) >> 4; | 179 | longhaul.bits.SoftBusRatio4 = (mults_index & 0x10) >> 4; |
178 | /* Setup new voltage */ | 180 | /* Setup new voltage */ |
179 | if (can_scale_voltage) | 181 | if (can_scale_voltage) |
180 | longhaul.bits.SoftVID = (clock_ratio_index >> 8) & 0x1f; | 182 | longhaul.bits.SoftVID = (mults_index >> 8) & 0x1f; |
181 | /* Sync to timer tick */ | 183 | /* Sync to timer tick */ |
182 | safe_halt(); | 184 | safe_halt(); |
183 | /* Raise voltage if necessary */ | 185 | /* Raise voltage if necessary */ |
@@ -240,14 +242,14 @@ static void do_powersaver(int cx_address, unsigned int clock_ratio_index, | |||
240 | 242 | ||
241 | /** | 243 | /** |
242 | * longhaul_set_cpu_frequency() | 244 | * longhaul_set_cpu_frequency() |
243 | * @clock_ratio_index : bitpattern of the new multiplier. | 245 | * @mults_index : bitpattern of the new multiplier. |
244 | * | 246 | * |
245 | * Sets a new clock ratio. | 247 | * Sets a new clock ratio. |
246 | */ | 248 | */ |
247 | 249 | ||
248 | static void longhaul_setstate(unsigned int table_index) | 250 | static void longhaul_setstate(unsigned int table_index) |
249 | { | 251 | { |
250 | unsigned int clock_ratio_index; | 252 | unsigned int mults_index; |
251 | int speed, mult; | 253 | int speed, mult; |
252 | struct cpufreq_freqs freqs; | 254 | struct cpufreq_freqs freqs; |
253 | unsigned long flags; | 255 | unsigned long flags; |
@@ -256,9 +258,9 @@ static void longhaul_setstate(unsigned int table_index) | |||
256 | u32 bm_timeout = 1000; | 258 | u32 bm_timeout = 1000; |
257 | unsigned int dir = 0; | 259 | unsigned int dir = 0; |
258 | 260 | ||
259 | clock_ratio_index = longhaul_table[table_index].index; | 261 | mults_index = longhaul_table[table_index].index; |
260 | /* Safety precautions */ | 262 | /* Safety precautions */ |
261 | mult = clock_ratio[clock_ratio_index & 0x1f]; | 263 | mult = mults[mults_index & 0x1f]; |
262 | if (mult == -1) | 264 | if (mult == -1) |
263 | return; | 265 | return; |
264 | speed = calc_speed(mult); | 266 | speed = calc_speed(mult); |
@@ -274,7 +276,7 @@ static void longhaul_setstate(unsigned int table_index) | |||
274 | 276 | ||
275 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); | 277 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); |
276 | 278 | ||
277 | dprintk ("Setting to FSB:%dMHz Mult:%d.%dx (%s)\n", | 279 | dprintk("Setting to FSB:%dMHz Mult:%d.%dx (%s)\n", |
278 | fsb, mult/10, mult%10, print_speed(speed/1000)); | 280 | fsb, mult/10, mult%10, print_speed(speed/1000)); |
279 | retry_loop: | 281 | retry_loop: |
280 | preempt_disable(); | 282 | preempt_disable(); |
@@ -282,8 +284,8 @@ retry_loop: | |||
282 | 284 | ||
283 | pic2_mask = inb(0xA1); | 285 | pic2_mask = inb(0xA1); |
284 | pic1_mask = inb(0x21); /* works on C3. save mask. */ | 286 | pic1_mask = inb(0x21); /* works on C3. save mask. */ |
285 | outb(0xFF,0xA1); /* Overkill */ | 287 | outb(0xFF, 0xA1); /* Overkill */ |
286 | outb(0xFE,0x21); /* TMR0 only */ | 288 | outb(0xFE, 0x21); /* TMR0 only */ |
287 | 289 | ||
288 | /* Wait while PCI bus is busy. */ | 290 | /* Wait while PCI bus is busy. */ |
289 | if (acpi_regs_addr && (longhaul_flags & USE_NORTHBRIDGE | 291 | if (acpi_regs_addr && (longhaul_flags & USE_NORTHBRIDGE |
@@ -312,7 +314,7 @@ retry_loop: | |||
312 | * Software controlled multipliers only. | 314 | * Software controlled multipliers only. |
313 | */ | 315 | */ |
314 | case TYPE_LONGHAUL_V1: | 316 | case TYPE_LONGHAUL_V1: |
315 | do_longhaul1(clock_ratio_index); | 317 | do_longhaul1(mults_index); |
316 | break; | 318 | break; |
317 | 319 | ||
318 | /* | 320 | /* |
@@ -327,9 +329,9 @@ retry_loop: | |||
327 | if (longhaul_flags & USE_ACPI_C3) { | 329 | if (longhaul_flags & USE_ACPI_C3) { |
328 | /* Don't allow wakeup */ | 330 | /* Don't allow wakeup */ |
329 | acpi_set_register(ACPI_BITREG_BUS_MASTER_RLD, 0); | 331 | acpi_set_register(ACPI_BITREG_BUS_MASTER_RLD, 0); |
330 | do_powersaver(cx->address, clock_ratio_index, dir); | 332 | do_powersaver(cx->address, mults_index, dir); |
331 | } else { | 333 | } else { |
332 | do_powersaver(0, clock_ratio_index, dir); | 334 | do_powersaver(0, mults_index, dir); |
333 | } | 335 | } |
334 | break; | 336 | break; |
335 | } | 337 | } |
@@ -341,8 +343,8 @@ retry_loop: | |||
341 | /* Enable bus master arbitration */ | 343 | /* Enable bus master arbitration */ |
342 | acpi_set_register(ACPI_BITREG_ARB_DISABLE, 0); | 344 | acpi_set_register(ACPI_BITREG_ARB_DISABLE, 0); |
343 | } | 345 | } |
344 | outb(pic2_mask,0xA1); /* restore mask */ | 346 | outb(pic2_mask, 0xA1); /* restore mask */ |
345 | outb(pic1_mask,0x21); | 347 | outb(pic1_mask, 0x21); |
346 | 348 | ||
347 | local_irq_restore(flags); | 349 | local_irq_restore(flags); |
348 | preempt_enable(); | 350 | preempt_enable(); |
@@ -392,7 +394,8 @@ retry_loop: | |||
392 | cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); | 394 | cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); |
393 | 395 | ||
394 | if (!bm_timeout) | 396 | if (!bm_timeout) |
395 | printk(KERN_INFO PFX "Warning: Timeout while waiting for idle PCI bus.\n"); | 397 | printk(KERN_INFO PFX "Warning: Timeout while waiting for " |
398 | "idle PCI bus.\n"); | ||
396 | } | 399 | } |
397 | 400 | ||
398 | /* | 401 | /* |
@@ -458,31 +461,32 @@ static int __init longhaul_get_ranges(void) | |||
458 | break; | 461 | break; |
459 | } | 462 | } |
460 | 463 | ||
461 | dprintk ("MinMult:%d.%dx MaxMult:%d.%dx\n", | 464 | dprintk("MinMult:%d.%dx MaxMult:%d.%dx\n", |
462 | minmult/10, minmult%10, maxmult/10, maxmult%10); | 465 | minmult/10, minmult%10, maxmult/10, maxmult%10); |
463 | 466 | ||
464 | highest_speed = calc_speed(maxmult); | 467 | highest_speed = calc_speed(maxmult); |
465 | lowest_speed = calc_speed(minmult); | 468 | lowest_speed = calc_speed(minmult); |
466 | dprintk ("FSB:%dMHz Lowest speed: %s Highest speed:%s\n", fsb, | 469 | dprintk("FSB:%dMHz Lowest speed: %s Highest speed:%s\n", fsb, |
467 | print_speed(lowest_speed/1000), | 470 | print_speed(lowest_speed/1000), |
468 | print_speed(highest_speed/1000)); | 471 | print_speed(highest_speed/1000)); |
469 | 472 | ||
470 | if (lowest_speed == highest_speed) { | 473 | if (lowest_speed == highest_speed) { |
471 | printk (KERN_INFO PFX "highestspeed == lowest, aborting.\n"); | 474 | printk(KERN_INFO PFX "highestspeed == lowest, aborting.\n"); |
472 | return -EINVAL; | 475 | return -EINVAL; |
473 | } | 476 | } |
474 | if (lowest_speed > highest_speed) { | 477 | if (lowest_speed > highest_speed) { |
475 | printk (KERN_INFO PFX "nonsense! lowest (%d > %d) !\n", | 478 | printk(KERN_INFO PFX "nonsense! lowest (%d > %d) !\n", |
476 | lowest_speed, highest_speed); | 479 | lowest_speed, highest_speed); |
477 | return -EINVAL; | 480 | return -EINVAL; |
478 | } | 481 | } |
479 | 482 | ||
480 | longhaul_table = kmalloc((numscales + 1) * sizeof(struct cpufreq_frequency_table), GFP_KERNEL); | 483 | longhaul_table = kmalloc((numscales + 1) * sizeof(*longhaul_table), |
481 | if(!longhaul_table) | 484 | GFP_KERNEL); |
485 | if (!longhaul_table) | ||
482 | return -ENOMEM; | 486 | return -ENOMEM; |
483 | 487 | ||
484 | for (j = 0; j < numscales; j++) { | 488 | for (j = 0; j < numscales; j++) { |
485 | ratio = clock_ratio[j]; | 489 | ratio = mults[j]; |
486 | if (ratio == -1) | 490 | if (ratio == -1) |
487 | continue; | 491 | continue; |
488 | if (ratio > maxmult || ratio < minmult) | 492 | if (ratio > maxmult || ratio < minmult) |
@@ -507,13 +511,10 @@ static int __init longhaul_get_ranges(void) | |||
507 | } | 511 | } |
508 | } | 512 | } |
509 | if (min_i != j) { | 513 | if (min_i != j) { |
510 | unsigned int temp; | 514 | swap(longhaul_table[j].frequency, |
511 | temp = longhaul_table[j].frequency; | 515 | longhaul_table[min_i].frequency); |
512 | longhaul_table[j].frequency = longhaul_table[min_i].frequency; | 516 | swap(longhaul_table[j].index, |
513 | longhaul_table[min_i].frequency = temp; | 517 | longhaul_table[min_i].index); |
514 | temp = longhaul_table[j].index; | ||
515 | longhaul_table[j].index = longhaul_table[min_i].index; | ||
516 | longhaul_table[min_i].index = temp; | ||
517 | } | 518 | } |
518 | } | 519 | } |
519 | 520 | ||
@@ -521,7 +522,7 @@ static int __init longhaul_get_ranges(void) | |||
521 | 522 | ||
522 | /* Find index we are running on */ | 523 | /* Find index we are running on */ |
523 | for (j = 0; j < k; j++) { | 524 | for (j = 0; j < k; j++) { |
524 | if (clock_ratio[longhaul_table[j].index & 0x1f] == mult) { | 525 | if (mults[longhaul_table[j].index & 0x1f] == mult) { |
525 | longhaul_index = j; | 526 | longhaul_index = j; |
526 | break; | 527 | break; |
527 | } | 528 | } |
@@ -559,20 +560,22 @@ static void __init longhaul_setup_voltagescaling(void) | |||
559 | maxvid = vrm_mV_table[longhaul.bits.MaximumVID]; | 560 | maxvid = vrm_mV_table[longhaul.bits.MaximumVID]; |
560 | 561 | ||
561 | if (minvid.mV == 0 || maxvid.mV == 0 || minvid.mV > maxvid.mV) { | 562 | if (minvid.mV == 0 || maxvid.mV == 0 || minvid.mV > maxvid.mV) { |
562 | printk (KERN_INFO PFX "Bogus values Min:%d.%03d Max:%d.%03d. " | 563 | printk(KERN_INFO PFX "Bogus values Min:%d.%03d Max:%d.%03d. " |
563 | "Voltage scaling disabled.\n", | 564 | "Voltage scaling disabled.\n", |
564 | minvid.mV/1000, minvid.mV%1000, maxvid.mV/1000, maxvid.mV%1000); | 565 | minvid.mV/1000, minvid.mV%1000, |
566 | maxvid.mV/1000, maxvid.mV%1000); | ||
565 | return; | 567 | return; |
566 | } | 568 | } |
567 | 569 | ||
568 | if (minvid.mV == maxvid.mV) { | 570 | if (minvid.mV == maxvid.mV) { |
569 | printk (KERN_INFO PFX "Claims to support voltage scaling but min & max are " | 571 | printk(KERN_INFO PFX "Claims to support voltage scaling but " |
570 | "both %d.%03d. Voltage scaling disabled\n", | 572 | "min & max are both %d.%03d. " |
573 | "Voltage scaling disabled\n", | ||
571 | maxvid.mV/1000, maxvid.mV%1000); | 574 | maxvid.mV/1000, maxvid.mV%1000); |
572 | return; | 575 | return; |
573 | } | 576 | } |
574 | 577 | ||
575 | /* How many voltage steps */ | 578 | /* How many voltage steps*/ |
576 | numvscales = maxvid.pos - minvid.pos + 1; | 579 | numvscales = maxvid.pos - minvid.pos + 1; |
577 | printk(KERN_INFO PFX | 580 | printk(KERN_INFO PFX |
578 | "Max VID=%d.%03d " | 581 | "Max VID=%d.%03d " |
@@ -586,7 +589,7 @@ static void __init longhaul_setup_voltagescaling(void) | |||
586 | j = longhaul.bits.MinMHzBR; | 589 | j = longhaul.bits.MinMHzBR; |
587 | if (longhaul.bits.MinMHzBR4) | 590 | if (longhaul.bits.MinMHzBR4) |
588 | j += 16; | 591 | j += 16; |
589 | min_vid_speed = eblcr_table[j]; | 592 | min_vid_speed = eblcr[j]; |
590 | if (min_vid_speed == -1) | 593 | if (min_vid_speed == -1) |
591 | return; | 594 | return; |
592 | switch (longhaul.bits.MinMHzFSB) { | 595 | switch (longhaul.bits.MinMHzFSB) { |
@@ -617,7 +620,8 @@ static void __init longhaul_setup_voltagescaling(void) | |||
617 | pos = minvid.pos; | 620 | pos = minvid.pos; |
618 | longhaul_table[j].index |= mV_vrm_table[pos] << 8; | 621 | longhaul_table[j].index |= mV_vrm_table[pos] << 8; |
619 | vid = vrm_mV_table[mV_vrm_table[pos]]; | 622 | vid = vrm_mV_table[mV_vrm_table[pos]]; |
620 | printk(KERN_INFO PFX "f: %d kHz, index: %d, vid: %d mV\n", speed, j, vid.mV); | 623 | printk(KERN_INFO PFX "f: %d kHz, index: %d, vid: %d mV\n", |
624 | speed, j, vid.mV); | ||
621 | j++; | 625 | j++; |
622 | } | 626 | } |
623 | 627 | ||
@@ -640,7 +644,8 @@ static int longhaul_target(struct cpufreq_policy *policy, | |||
640 | unsigned int dir = 0; | 644 | unsigned int dir = 0; |
641 | u8 vid, current_vid; | 645 | u8 vid, current_vid; |
642 | 646 | ||
643 | if (cpufreq_frequency_table_target(policy, longhaul_table, target_freq, relation, &table_index)) | 647 | if (cpufreq_frequency_table_target(policy, longhaul_table, target_freq, |
648 | relation, &table_index)) | ||
644 | return -EINVAL; | 649 | return -EINVAL; |
645 | 650 | ||
646 | /* Don't set same frequency again */ | 651 | /* Don't set same frequency again */ |
@@ -656,7 +661,8 @@ static int longhaul_target(struct cpufreq_policy *policy, | |||
656 | * this in hardware, C3 is old and we need to do this | 661 | * this in hardware, C3 is old and we need to do this |
657 | * in software. */ | 662 | * in software. */ |
658 | i = longhaul_index; | 663 | i = longhaul_index; |
659 | current_vid = (longhaul_table[longhaul_index].index >> 8) & 0x1f; | 664 | current_vid = (longhaul_table[longhaul_index].index >> 8); |
665 | current_vid &= 0x1f; | ||
660 | if (table_index > longhaul_index) | 666 | if (table_index > longhaul_index) |
661 | dir = 1; | 667 | dir = 1; |
662 | while (i != table_index) { | 668 | while (i != table_index) { |
@@ -691,9 +697,9 @@ static acpi_status longhaul_walk_callback(acpi_handle obj_handle, | |||
691 | { | 697 | { |
692 | struct acpi_device *d; | 698 | struct acpi_device *d; |
693 | 699 | ||
694 | if ( acpi_bus_get_device(obj_handle, &d) ) { | 700 | if (acpi_bus_get_device(obj_handle, &d)) |
695 | return 0; | 701 | return 0; |
696 | } | 702 | |
697 | *return_value = acpi_driver_data(d); | 703 | *return_value = acpi_driver_data(d); |
698 | return 1; | 704 | return 1; |
699 | } | 705 | } |
@@ -750,7 +756,7 @@ static int longhaul_setup_southbridge(void) | |||
750 | /* Find VT8235 southbridge */ | 756 | /* Find VT8235 southbridge */ |
751 | dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235, NULL); | 757 | dev = pci_get_device(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235, NULL); |
752 | if (dev == NULL) | 758 | if (dev == NULL) |
753 | /* Find VT8237 southbridge */ | 759 | /* Find VT8237 southbridge */ |
754 | dev = pci_get_device(PCI_VENDOR_ID_VIA, | 760 | dev = pci_get_device(PCI_VENDOR_ID_VIA, |
755 | PCI_DEVICE_ID_VIA_8237, NULL); | 761 | PCI_DEVICE_ID_VIA_8237, NULL); |
756 | if (dev != NULL) { | 762 | if (dev != NULL) { |
@@ -769,7 +775,8 @@ static int longhaul_setup_southbridge(void) | |||
769 | if (pci_cmd & 1 << 7) { | 775 | if (pci_cmd & 1 << 7) { |
770 | pci_read_config_dword(dev, 0x88, &acpi_regs_addr); | 776 | pci_read_config_dword(dev, 0x88, &acpi_regs_addr); |
771 | acpi_regs_addr &= 0xff00; | 777 | acpi_regs_addr &= 0xff00; |
772 | printk(KERN_INFO PFX "ACPI I/O at 0x%x\n", acpi_regs_addr); | 778 | printk(KERN_INFO PFX "ACPI I/O at 0x%x\n", |
779 | acpi_regs_addr); | ||
773 | } | 780 | } |
774 | 781 | ||
775 | pci_dev_put(dev); | 782 | pci_dev_put(dev); |
@@ -781,7 +788,7 @@ static int longhaul_setup_southbridge(void) | |||
781 | static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | 788 | static int __init longhaul_cpu_init(struct cpufreq_policy *policy) |
782 | { | 789 | { |
783 | struct cpuinfo_x86 *c = &cpu_data(0); | 790 | struct cpuinfo_x86 *c = &cpu_data(0); |
784 | char *cpuname=NULL; | 791 | char *cpuname = NULL; |
785 | int ret; | 792 | int ret; |
786 | u32 lo, hi; | 793 | u32 lo, hi; |
787 | 794 | ||
@@ -791,8 +798,8 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | |||
791 | cpu_model = CPU_SAMUEL; | 798 | cpu_model = CPU_SAMUEL; |
792 | cpuname = "C3 'Samuel' [C5A]"; | 799 | cpuname = "C3 'Samuel' [C5A]"; |
793 | longhaul_version = TYPE_LONGHAUL_V1; | 800 | longhaul_version = TYPE_LONGHAUL_V1; |
794 | memcpy (clock_ratio, samuel1_clock_ratio, sizeof(samuel1_clock_ratio)); | 801 | memcpy(mults, samuel1_mults, sizeof(samuel1_mults)); |
795 | memcpy (eblcr_table, samuel1_eblcr, sizeof(samuel1_eblcr)); | 802 | memcpy(eblcr, samuel1_eblcr, sizeof(samuel1_eblcr)); |
796 | break; | 803 | break; |
797 | 804 | ||
798 | case 7: | 805 | case 7: |
@@ -803,10 +810,8 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | |||
803 | cpuname = "C3 'Samuel 2' [C5B]"; | 810 | cpuname = "C3 'Samuel 2' [C5B]"; |
804 | /* Note, this is not a typo, early Samuel2's had | 811 | /* Note, this is not a typo, early Samuel2's had |
805 | * Samuel1 ratios. */ | 812 | * Samuel1 ratios. */ |
806 | memcpy(clock_ratio, samuel1_clock_ratio, | 813 | memcpy(mults, samuel1_mults, sizeof(samuel1_mults)); |
807 | sizeof(samuel1_clock_ratio)); | 814 | memcpy(eblcr, samuel2_eblcr, sizeof(samuel2_eblcr)); |
808 | memcpy(eblcr_table, samuel2_eblcr, | ||
809 | sizeof(samuel2_eblcr)); | ||
810 | break; | 815 | break; |
811 | case 1 ... 15: | 816 | case 1 ... 15: |
812 | longhaul_version = TYPE_LONGHAUL_V1; | 817 | longhaul_version = TYPE_LONGHAUL_V1; |
@@ -817,10 +822,8 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | |||
817 | cpu_model = CPU_EZRA; | 822 | cpu_model = CPU_EZRA; |
818 | cpuname = "C3 'Ezra' [C5C]"; | 823 | cpuname = "C3 'Ezra' [C5C]"; |
819 | } | 824 | } |
820 | memcpy(clock_ratio, ezra_clock_ratio, | 825 | memcpy(mults, ezra_mults, sizeof(ezra_mults)); |
821 | sizeof(ezra_clock_ratio)); | 826 | memcpy(eblcr, ezra_eblcr, sizeof(ezra_eblcr)); |
822 | memcpy(eblcr_table, ezra_eblcr, | ||
823 | sizeof(ezra_eblcr)); | ||
824 | break; | 827 | break; |
825 | } | 828 | } |
826 | break; | 829 | break; |
@@ -829,18 +832,16 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | |||
829 | cpu_model = CPU_EZRA_T; | 832 | cpu_model = CPU_EZRA_T; |
830 | cpuname = "C3 'Ezra-T' [C5M]"; | 833 | cpuname = "C3 'Ezra-T' [C5M]"; |
831 | longhaul_version = TYPE_POWERSAVER; | 834 | longhaul_version = TYPE_POWERSAVER; |
832 | numscales=32; | 835 | numscales = 32; |
833 | memcpy (clock_ratio, ezrat_clock_ratio, sizeof(ezrat_clock_ratio)); | 836 | memcpy(mults, ezrat_mults, sizeof(ezrat_mults)); |
834 | memcpy (eblcr_table, ezrat_eblcr, sizeof(ezrat_eblcr)); | 837 | memcpy(eblcr, ezrat_eblcr, sizeof(ezrat_eblcr)); |
835 | break; | 838 | break; |
836 | 839 | ||
837 | case 9: | 840 | case 9: |
838 | longhaul_version = TYPE_POWERSAVER; | 841 | longhaul_version = TYPE_POWERSAVER; |
839 | numscales = 32; | 842 | numscales = 32; |
840 | memcpy(clock_ratio, | 843 | memcpy(mults, nehemiah_mults, sizeof(nehemiah_mults)); |
841 | nehemiah_clock_ratio, | 844 | memcpy(eblcr, nehemiah_eblcr, sizeof(nehemiah_eblcr)); |
842 | sizeof(nehemiah_clock_ratio)); | ||
843 | memcpy(eblcr_table, nehemiah_eblcr, sizeof(nehemiah_eblcr)); | ||
844 | switch (c->x86_mask) { | 845 | switch (c->x86_mask) { |
845 | case 0 ... 1: | 846 | case 0 ... 1: |
846 | cpu_model = CPU_NEHEMIAH; | 847 | cpu_model = CPU_NEHEMIAH; |
@@ -869,14 +870,14 @@ static int __init longhaul_cpu_init(struct cpufreq_policy *policy) | |||
869 | longhaul_version = TYPE_LONGHAUL_V1; | 870 | longhaul_version = TYPE_LONGHAUL_V1; |
870 | } | 871 | } |
871 | 872 | ||
872 | printk (KERN_INFO PFX "VIA %s CPU detected. ", cpuname); | 873 | printk(KERN_INFO PFX "VIA %s CPU detected. ", cpuname); |
873 | switch (longhaul_version) { | 874 | switch (longhaul_version) { |
874 | case TYPE_LONGHAUL_V1: | 875 | case TYPE_LONGHAUL_V1: |
875 | case TYPE_LONGHAUL_V2: | 876 | case TYPE_LONGHAUL_V2: |
876 | printk ("Longhaul v%d supported.\n", longhaul_version); | 877 | printk(KERN_CONT "Longhaul v%d supported.\n", longhaul_version); |
877 | break; | 878 | break; |
878 | case TYPE_POWERSAVER: | 879 | case TYPE_POWERSAVER: |
879 | printk ("Powersaver supported.\n"); | 880 | printk(KERN_CONT "Powersaver supported.\n"); |
880 | break; | 881 | break; |
881 | }; | 882 | }; |
882 | 883 | ||
@@ -940,7 +941,7 @@ static int __devexit longhaul_cpu_exit(struct cpufreq_policy *policy) | |||
940 | return 0; | 941 | return 0; |
941 | } | 942 | } |
942 | 943 | ||
943 | static struct freq_attr* longhaul_attr[] = { | 944 | static struct freq_attr *longhaul_attr[] = { |
944 | &cpufreq_freq_attr_scaling_available_freqs, | 945 | &cpufreq_freq_attr_scaling_available_freqs, |
945 | NULL, | 946 | NULL, |
946 | }; | 947 | }; |
@@ -966,13 +967,15 @@ static int __init longhaul_init(void) | |||
966 | 967 | ||
967 | #ifdef CONFIG_SMP | 968 | #ifdef CONFIG_SMP |
968 | if (num_online_cpus() > 1) { | 969 | if (num_online_cpus() > 1) { |
969 | printk(KERN_ERR PFX "More than 1 CPU detected, longhaul disabled.\n"); | 970 | printk(KERN_ERR PFX "More than 1 CPU detected, " |
971 | "longhaul disabled.\n"); | ||
970 | return -ENODEV; | 972 | return -ENODEV; |
971 | } | 973 | } |
972 | #endif | 974 | #endif |
973 | #ifdef CONFIG_X86_IO_APIC | 975 | #ifdef CONFIG_X86_IO_APIC |
974 | if (cpu_has_apic) { | 976 | if (cpu_has_apic) { |
975 | printk(KERN_ERR PFX "APIC detected. Longhaul is currently broken in this configuration.\n"); | 977 | printk(KERN_ERR PFX "APIC detected. Longhaul is currently " |
978 | "broken in this configuration.\n"); | ||
976 | return -ENODEV; | 979 | return -ENODEV; |
977 | } | 980 | } |
978 | #endif | 981 | #endif |
@@ -993,8 +996,8 @@ static void __exit longhaul_exit(void) | |||
993 | { | 996 | { |
994 | int i; | 997 | int i; |
995 | 998 | ||
996 | for (i=0; i < numscales; i++) { | 999 | for (i = 0; i < numscales; i++) { |
997 | if (clock_ratio[i] == maxmult) { | 1000 | if (mults[i] == maxmult) { |
998 | longhaul_setstate(i); | 1001 | longhaul_setstate(i); |
999 | break; | 1002 | break; |
1000 | } | 1003 | } |
@@ -1007,11 +1010,11 @@ static void __exit longhaul_exit(void) | |||
1007 | /* Even if BIOS is exporting ACPI C3 state, and it is used | 1010 | /* Even if BIOS is exporting ACPI C3 state, and it is used |
1008 | * with success when CPU is idle, this state doesn't | 1011 | * with success when CPU is idle, this state doesn't |
1009 | * trigger frequency transition in some cases. */ | 1012 | * trigger frequency transition in some cases. */ |
1010 | module_param (disable_acpi_c3, int, 0644); | 1013 | module_param(disable_acpi_c3, int, 0644); |
1011 | MODULE_PARM_DESC(disable_acpi_c3, "Don't use ACPI C3 support"); | 1014 | MODULE_PARM_DESC(disable_acpi_c3, "Don't use ACPI C3 support"); |
1012 | /* Change CPU voltage with frequency. Very usefull to save | 1015 | /* Change CPU voltage with frequency. Very usefull to save |
1013 | * power, but most VIA C3 processors aren't supporting it. */ | 1016 | * power, but most VIA C3 processors aren't supporting it. */ |
1014 | module_param (scale_voltage, int, 0644); | 1017 | module_param(scale_voltage, int, 0644); |
1015 | MODULE_PARM_DESC(scale_voltage, "Scale voltage of processor"); | 1018 | MODULE_PARM_DESC(scale_voltage, "Scale voltage of processor"); |
1016 | /* Force revision key to 0 for processors which doesn't | 1019 | /* Force revision key to 0 for processors which doesn't |
1017 | * support voltage scaling, but are introducing itself as | 1020 | * support voltage scaling, but are introducing itself as |
@@ -1019,9 +1022,9 @@ MODULE_PARM_DESC(scale_voltage, "Scale voltage of processor"); | |||
1019 | module_param(revid_errata, int, 0644); | 1022 | module_param(revid_errata, int, 0644); |
1020 | MODULE_PARM_DESC(revid_errata, "Ignore CPU Revision ID"); | 1023 | MODULE_PARM_DESC(revid_errata, "Ignore CPU Revision ID"); |
1021 | 1024 | ||
1022 | MODULE_AUTHOR ("Dave Jones <davej@redhat.com>"); | 1025 | MODULE_AUTHOR("Dave Jones <davej@redhat.com>"); |
1023 | MODULE_DESCRIPTION ("Longhaul driver for VIA Cyrix processors."); | 1026 | MODULE_DESCRIPTION("Longhaul driver for VIA Cyrix processors."); |
1024 | MODULE_LICENSE ("GPL"); | 1027 | MODULE_LICENSE("GPL"); |
1025 | 1028 | ||
1026 | late_initcall(longhaul_init); | 1029 | late_initcall(longhaul_init); |
1027 | module_exit(longhaul_exit); | 1030 | module_exit(longhaul_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/longhaul.h b/arch/x86/kernel/cpu/cpufreq/longhaul.h index 4fcc320997df..e2360a469f79 100644 --- a/arch/x86/kernel/cpu/cpufreq/longhaul.h +++ b/arch/x86/kernel/cpu/cpufreq/longhaul.h | |||
@@ -49,14 +49,14 @@ union msr_longhaul { | |||
49 | 49 | ||
50 | /* | 50 | /* |
51 | * Clock ratio tables. Div/Mod by 10 to get ratio. | 51 | * Clock ratio tables. Div/Mod by 10 to get ratio. |
52 | * The eblcr ones specify the ratio read from the CPU. | 52 | * The eblcr values specify the ratio read from the CPU. |
53 | * The clock_ratio ones specify what to write to the CPU. | 53 | * The mults values specify what to write to the CPU. |
54 | */ | 54 | */ |
55 | 55 | ||
56 | /* | 56 | /* |
57 | * VIA C3 Samuel 1 & Samuel 2 (stepping 0) | 57 | * VIA C3 Samuel 1 & Samuel 2 (stepping 0) |
58 | */ | 58 | */ |
59 | static const int __initdata samuel1_clock_ratio[16] = { | 59 | static const int __initdata samuel1_mults[16] = { |
60 | -1, /* 0000 -> RESERVED */ | 60 | -1, /* 0000 -> RESERVED */ |
61 | 30, /* 0001 -> 3.0x */ | 61 | 30, /* 0001 -> 3.0x */ |
62 | 40, /* 0010 -> 4.0x */ | 62 | 40, /* 0010 -> 4.0x */ |
@@ -119,7 +119,7 @@ static const int __initdata samuel2_eblcr[16] = { | |||
119 | /* | 119 | /* |
120 | * VIA C3 Ezra | 120 | * VIA C3 Ezra |
121 | */ | 121 | */ |
122 | static const int __initdata ezra_clock_ratio[16] = { | 122 | static const int __initdata ezra_mults[16] = { |
123 | 100, /* 0000 -> 10.0x */ | 123 | 100, /* 0000 -> 10.0x */ |
124 | 30, /* 0001 -> 3.0x */ | 124 | 30, /* 0001 -> 3.0x */ |
125 | 40, /* 0010 -> 4.0x */ | 125 | 40, /* 0010 -> 4.0x */ |
@@ -160,7 +160,7 @@ static const int __initdata ezra_eblcr[16] = { | |||
160 | /* | 160 | /* |
161 | * VIA C3 (Ezra-T) [C5M]. | 161 | * VIA C3 (Ezra-T) [C5M]. |
162 | */ | 162 | */ |
163 | static const int __initdata ezrat_clock_ratio[32] = { | 163 | static const int __initdata ezrat_mults[32] = { |
164 | 100, /* 0000 -> 10.0x */ | 164 | 100, /* 0000 -> 10.0x */ |
165 | 30, /* 0001 -> 3.0x */ | 165 | 30, /* 0001 -> 3.0x */ |
166 | 40, /* 0010 -> 4.0x */ | 166 | 40, /* 0010 -> 4.0x */ |
@@ -235,7 +235,7 @@ static const int __initdata ezrat_eblcr[32] = { | |||
235 | /* | 235 | /* |
236 | * VIA C3 Nehemiah */ | 236 | * VIA C3 Nehemiah */ |
237 | 237 | ||
238 | static const int __initdata nehemiah_clock_ratio[32] = { | 238 | static const int __initdata nehemiah_mults[32] = { |
239 | 100, /* 0000 -> 10.0x */ | 239 | 100, /* 0000 -> 10.0x */ |
240 | -1, /* 0001 -> 16.0x */ | 240 | -1, /* 0001 -> 16.0x */ |
241 | 40, /* 0010 -> 4.0x */ | 241 | 40, /* 0010 -> 4.0x */ |
diff --git a/arch/x86/kernel/cpu/cpufreq/longrun.c b/arch/x86/kernel/cpu/cpufreq/longrun.c index 777a7ff075de..da5f70fcb766 100644 --- a/arch/x86/kernel/cpu/cpufreq/longrun.c +++ b/arch/x86/kernel/cpu/cpufreq/longrun.c | |||
@@ -11,12 +11,13 @@ | |||
11 | #include <linux/init.h> | 11 | #include <linux/init.h> |
12 | #include <linux/slab.h> | 12 | #include <linux/slab.h> |
13 | #include <linux/cpufreq.h> | 13 | #include <linux/cpufreq.h> |
14 | #include <linux/timex.h> | ||
14 | 15 | ||
15 | #include <asm/msr.h> | 16 | #include <asm/msr.h> |
16 | #include <asm/processor.h> | 17 | #include <asm/processor.h> |
17 | #include <asm/timex.h> | ||
18 | 18 | ||
19 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "longrun", msg) | 19 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
20 | "longrun", msg) | ||
20 | 21 | ||
21 | static struct cpufreq_driver longrun_driver; | 22 | static struct cpufreq_driver longrun_driver; |
22 | 23 | ||
@@ -51,7 +52,7 @@ static void __init longrun_get_policy(struct cpufreq_policy *policy) | |||
51 | msr_lo &= 0x0000007F; | 52 | msr_lo &= 0x0000007F; |
52 | msr_hi &= 0x0000007F; | 53 | msr_hi &= 0x0000007F; |
53 | 54 | ||
54 | if ( longrun_high_freq <= longrun_low_freq ) { | 55 | if (longrun_high_freq <= longrun_low_freq) { |
55 | /* Assume degenerate Longrun table */ | 56 | /* Assume degenerate Longrun table */ |
56 | policy->min = policy->max = longrun_high_freq; | 57 | policy->min = policy->max = longrun_high_freq; |
57 | } else { | 58 | } else { |
@@ -79,7 +80,7 @@ static int longrun_set_policy(struct cpufreq_policy *policy) | |||
79 | if (!policy) | 80 | if (!policy) |
80 | return -EINVAL; | 81 | return -EINVAL; |
81 | 82 | ||
82 | if ( longrun_high_freq <= longrun_low_freq ) { | 83 | if (longrun_high_freq <= longrun_low_freq) { |
83 | /* Assume degenerate Longrun table */ | 84 | /* Assume degenerate Longrun table */ |
84 | pctg_lo = pctg_hi = 100; | 85 | pctg_lo = pctg_hi = 100; |
85 | } else { | 86 | } else { |
@@ -152,7 +153,7 @@ static unsigned int longrun_get(unsigned int cpu) | |||
152 | cpuid(0x80860007, &eax, &ebx, &ecx, &edx); | 153 | cpuid(0x80860007, &eax, &ebx, &ecx, &edx); |
153 | dprintk("cpuid eax is %u\n", eax); | 154 | dprintk("cpuid eax is %u\n", eax); |
154 | 155 | ||
155 | return (eax * 1000); | 156 | return eax * 1000; |
156 | } | 157 | } |
157 | 158 | ||
158 | /** | 159 | /** |
@@ -196,7 +197,8 @@ static unsigned int __init longrun_determine_freqs(unsigned int *low_freq, | |||
196 | rdmsr(MSR_TMTA_LRTI_VOLT_MHZ, msr_lo, msr_hi); | 197 | rdmsr(MSR_TMTA_LRTI_VOLT_MHZ, msr_lo, msr_hi); |
197 | *high_freq = msr_lo * 1000; /* to kHz */ | 198 | *high_freq = msr_lo * 1000; /* to kHz */ |
198 | 199 | ||
199 | dprintk("longrun table interface told %u - %u kHz\n", *low_freq, *high_freq); | 200 | dprintk("longrun table interface told %u - %u kHz\n", |
201 | *low_freq, *high_freq); | ||
200 | 202 | ||
201 | if (*low_freq > *high_freq) | 203 | if (*low_freq > *high_freq) |
202 | *low_freq = *high_freq; | 204 | *low_freq = *high_freq; |
@@ -219,7 +221,7 @@ static unsigned int __init longrun_determine_freqs(unsigned int *low_freq, | |||
219 | cpuid(0x80860007, &eax, &ebx, &ecx, &edx); | 221 | cpuid(0x80860007, &eax, &ebx, &ecx, &edx); |
220 | /* try decreasing in 10% steps, some processors react only | 222 | /* try decreasing in 10% steps, some processors react only |
221 | * on some barrier values */ | 223 | * on some barrier values */ |
222 | for (try_hi = 80; try_hi > 0 && ecx > 90; try_hi -=10) { | 224 | for (try_hi = 80; try_hi > 0 && ecx > 90; try_hi -= 10) { |
223 | /* set to 0 to try_hi perf_pctg */ | 225 | /* set to 0 to try_hi perf_pctg */ |
224 | msr_lo &= 0xFFFFFF80; | 226 | msr_lo &= 0xFFFFFF80; |
225 | msr_hi &= 0xFFFFFF80; | 227 | msr_hi &= 0xFFFFFF80; |
@@ -236,7 +238,7 @@ static unsigned int __init longrun_determine_freqs(unsigned int *low_freq, | |||
236 | 238 | ||
237 | /* performance_pctg = (current_freq - low_freq)/(high_freq - low_freq) | 239 | /* performance_pctg = (current_freq - low_freq)/(high_freq - low_freq) |
238 | * eqals | 240 | * eqals |
239 | * low_freq * ( 1 - perf_pctg) = (cur_freq - high_freq * perf_pctg) | 241 | * low_freq * (1 - perf_pctg) = (cur_freq - high_freq * perf_pctg) |
240 | * | 242 | * |
241 | * high_freq * perf_pctg is stored tempoarily into "ebx". | 243 | * high_freq * perf_pctg is stored tempoarily into "ebx". |
242 | */ | 244 | */ |
@@ -317,9 +319,10 @@ static void __exit longrun_exit(void) | |||
317 | } | 319 | } |
318 | 320 | ||
319 | 321 | ||
320 | MODULE_AUTHOR ("Dominik Brodowski <linux@brodo.de>"); | 322 | MODULE_AUTHOR("Dominik Brodowski <linux@brodo.de>"); |
321 | MODULE_DESCRIPTION ("LongRun driver for Transmeta Crusoe and Efficeon processors."); | 323 | MODULE_DESCRIPTION("LongRun driver for Transmeta Crusoe and " |
322 | MODULE_LICENSE ("GPL"); | 324 | "Efficeon processors."); |
325 | MODULE_LICENSE("GPL"); | ||
323 | 326 | ||
324 | module_init(longrun_init); | 327 | module_init(longrun_init); |
325 | module_exit(longrun_exit); | 328 | module_exit(longrun_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/p4-clockmod.c b/arch/x86/kernel/cpu/cpufreq/p4-clockmod.c index 3178c3acd97e..41ed94915f97 100644 --- a/arch/x86/kernel/cpu/cpufreq/p4-clockmod.c +++ b/arch/x86/kernel/cpu/cpufreq/p4-clockmod.c | |||
@@ -27,15 +27,17 @@ | |||
27 | #include <linux/cpufreq.h> | 27 | #include <linux/cpufreq.h> |
28 | #include <linux/slab.h> | 28 | #include <linux/slab.h> |
29 | #include <linux/cpumask.h> | 29 | #include <linux/cpumask.h> |
30 | #include <linux/timex.h> | ||
30 | 31 | ||
31 | #include <asm/processor.h> | 32 | #include <asm/processor.h> |
32 | #include <asm/msr.h> | 33 | #include <asm/msr.h> |
33 | #include <asm/timex.h> | 34 | #include <asm/timer.h> |
34 | 35 | ||
35 | #include "speedstep-lib.h" | 36 | #include "speedstep-lib.h" |
36 | 37 | ||
37 | #define PFX "p4-clockmod: " | 38 | #define PFX "p4-clockmod: " |
38 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "p4-clockmod", msg) | 39 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
40 | "p4-clockmod", msg) | ||
39 | 41 | ||
40 | /* | 42 | /* |
41 | * Duty Cycle (3bits), note DC_DISABLE is not specified in | 43 | * Duty Cycle (3bits), note DC_DISABLE is not specified in |
@@ -58,7 +60,8 @@ static int cpufreq_p4_setdc(unsigned int cpu, unsigned int newstate) | |||
58 | { | 60 | { |
59 | u32 l, h; | 61 | u32 l, h; |
60 | 62 | ||
61 | if (!cpu_online(cpu) || (newstate > DC_DISABLE) || (newstate == DC_RESV)) | 63 | if (!cpu_online(cpu) || |
64 | (newstate > DC_DISABLE) || (newstate == DC_RESV)) | ||
62 | return -EINVAL; | 65 | return -EINVAL; |
63 | 66 | ||
64 | rdmsr_on_cpu(cpu, MSR_IA32_THERM_STATUS, &l, &h); | 67 | rdmsr_on_cpu(cpu, MSR_IA32_THERM_STATUS, &l, &h); |
@@ -66,7 +69,8 @@ static int cpufreq_p4_setdc(unsigned int cpu, unsigned int newstate) | |||
66 | if (l & 0x01) | 69 | if (l & 0x01) |
67 | dprintk("CPU#%d currently thermal throttled\n", cpu); | 70 | dprintk("CPU#%d currently thermal throttled\n", cpu); |
68 | 71 | ||
69 | if (has_N44_O17_errata[cpu] && (newstate == DC_25PT || newstate == DC_DFLT)) | 72 | if (has_N44_O17_errata[cpu] && |
73 | (newstate == DC_25PT || newstate == DC_DFLT)) | ||
70 | newstate = DC_38PT; | 74 | newstate = DC_38PT; |
71 | 75 | ||
72 | rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); | 76 | rdmsr_on_cpu(cpu, MSR_IA32_THERM_CONTROL, &l, &h); |
@@ -112,7 +116,8 @@ static int cpufreq_p4_target(struct cpufreq_policy *policy, | |||
112 | struct cpufreq_freqs freqs; | 116 | struct cpufreq_freqs freqs; |
113 | int i; | 117 | int i; |
114 | 118 | ||
115 | if (cpufreq_frequency_table_target(policy, &p4clockmod_table[0], target_freq, relation, &newstate)) | 119 | if (cpufreq_frequency_table_target(policy, &p4clockmod_table[0], |
120 | target_freq, relation, &newstate)) | ||
116 | return -EINVAL; | 121 | return -EINVAL; |
117 | 122 | ||
118 | freqs.old = cpufreq_p4_get(policy->cpu); | 123 | freqs.old = cpufreq_p4_get(policy->cpu); |
@@ -127,7 +132,8 @@ static int cpufreq_p4_target(struct cpufreq_policy *policy, | |||
127 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); | 132 | cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); |
128 | } | 133 | } |
129 | 134 | ||
130 | /* run on each logical CPU, see section 13.15.3 of IA32 Intel Architecture Software | 135 | /* run on each logical CPU, |
136 | * see section 13.15.3 of IA32 Intel Architecture Software | ||
131 | * Developer's Manual, Volume 3 | 137 | * Developer's Manual, Volume 3 |
132 | */ | 138 | */ |
133 | for_each_cpu(i, policy->cpus) | 139 | for_each_cpu(i, policy->cpus) |
@@ -153,28 +159,30 @@ static unsigned int cpufreq_p4_get_frequency(struct cpuinfo_x86 *c) | |||
153 | { | 159 | { |
154 | if (c->x86 == 0x06) { | 160 | if (c->x86 == 0x06) { |
155 | if (cpu_has(c, X86_FEATURE_EST)) | 161 | if (cpu_has(c, X86_FEATURE_EST)) |
156 | printk(KERN_WARNING PFX "Warning: EST-capable CPU detected. " | 162 | printk(KERN_WARNING PFX "Warning: EST-capable CPU " |
157 | "The acpi-cpufreq module offers voltage scaling" | 163 | "detected. The acpi-cpufreq module offers " |
158 | " in addition of frequency scaling. You should use " | 164 | "voltage scaling in addition of frequency " |
159 | "that instead of p4-clockmod, if possible.\n"); | 165 | "scaling. You should use that instead of " |
166 | "p4-clockmod, if possible.\n"); | ||
160 | switch (c->x86_model) { | 167 | switch (c->x86_model) { |
161 | case 0x0E: /* Core */ | 168 | case 0x0E: /* Core */ |
162 | case 0x0F: /* Core Duo */ | 169 | case 0x0F: /* Core Duo */ |
163 | case 0x16: /* Celeron Core */ | 170 | case 0x16: /* Celeron Core */ |
164 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; | 171 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; |
165 | return speedstep_get_processor_frequency(SPEEDSTEP_PROCESSOR_PCORE); | 172 | return speedstep_get_frequency(SPEEDSTEP_CPU_PCORE); |
166 | case 0x0D: /* Pentium M (Dothan) */ | 173 | case 0x0D: /* Pentium M (Dothan) */ |
167 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; | 174 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; |
168 | /* fall through */ | 175 | /* fall through */ |
169 | case 0x09: /* Pentium M (Banias) */ | 176 | case 0x09: /* Pentium M (Banias) */ |
170 | return speedstep_get_processor_frequency(SPEEDSTEP_PROCESSOR_PM); | 177 | return speedstep_get_frequency(SPEEDSTEP_CPU_PM); |
171 | } | 178 | } |
172 | } | 179 | } |
173 | 180 | ||
174 | if (c->x86 != 0xF) { | 181 | if (c->x86 != 0xF) { |
175 | if (!cpu_has(c, X86_FEATURE_EST)) | 182 | if (!cpu_has(c, X86_FEATURE_EST)) |
176 | printk(KERN_WARNING PFX "Unknown p4-clockmod-capable CPU. " | 183 | printk(KERN_WARNING PFX "Unknown CPU. " |
177 | "Please send an e-mail to <cpufreq@vger.kernel.org>\n"); | 184 | "Please send an e-mail to " |
185 | "<cpufreq@vger.kernel.org>\n"); | ||
178 | return 0; | 186 | return 0; |
179 | } | 187 | } |
180 | 188 | ||
@@ -182,16 +190,16 @@ static unsigned int cpufreq_p4_get_frequency(struct cpuinfo_x86 *c) | |||
182 | * throttling is active or not. */ | 190 | * throttling is active or not. */ |
183 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; | 191 | p4clockmod_driver.flags |= CPUFREQ_CONST_LOOPS; |
184 | 192 | ||
185 | if (speedstep_detect_processor() == SPEEDSTEP_PROCESSOR_P4M) { | 193 | if (speedstep_detect_processor() == SPEEDSTEP_CPU_P4M) { |
186 | printk(KERN_WARNING PFX "Warning: Pentium 4-M detected. " | 194 | printk(KERN_WARNING PFX "Warning: Pentium 4-M detected. " |
187 | "The speedstep-ich or acpi cpufreq modules offer " | 195 | "The speedstep-ich or acpi cpufreq modules offer " |
188 | "voltage scaling in addition of frequency scaling. " | 196 | "voltage scaling in addition of frequency scaling. " |
189 | "You should use either one instead of p4-clockmod, " | 197 | "You should use either one instead of p4-clockmod, " |
190 | "if possible.\n"); | 198 | "if possible.\n"); |
191 | return speedstep_get_processor_frequency(SPEEDSTEP_PROCESSOR_P4M); | 199 | return speedstep_get_frequency(SPEEDSTEP_CPU_P4M); |
192 | } | 200 | } |
193 | 201 | ||
194 | return speedstep_get_processor_frequency(SPEEDSTEP_PROCESSOR_P4D); | 202 | return speedstep_get_frequency(SPEEDSTEP_CPU_P4D); |
195 | } | 203 | } |
196 | 204 | ||
197 | 205 | ||
@@ -217,14 +225,20 @@ static int cpufreq_p4_cpu_init(struct cpufreq_policy *policy) | |||
217 | dprintk("has errata -- disabling low frequencies\n"); | 225 | dprintk("has errata -- disabling low frequencies\n"); |
218 | } | 226 | } |
219 | 227 | ||
228 | if (speedstep_detect_processor() == SPEEDSTEP_CPU_P4D && | ||
229 | c->x86_model < 2) { | ||
230 | /* switch to maximum frequency and measure result */ | ||
231 | cpufreq_p4_setdc(policy->cpu, DC_DISABLE); | ||
232 | recalibrate_cpu_khz(); | ||
233 | } | ||
220 | /* get max frequency */ | 234 | /* get max frequency */ |
221 | stock_freq = cpufreq_p4_get_frequency(c); | 235 | stock_freq = cpufreq_p4_get_frequency(c); |
222 | if (!stock_freq) | 236 | if (!stock_freq) |
223 | return -EINVAL; | 237 | return -EINVAL; |
224 | 238 | ||
225 | /* table init */ | 239 | /* table init */ |
226 | for (i=1; (p4clockmod_table[i].frequency != CPUFREQ_TABLE_END); i++) { | 240 | for (i = 1; (p4clockmod_table[i].frequency != CPUFREQ_TABLE_END); i++) { |
227 | if ((i<2) && (has_N44_O17_errata[policy->cpu])) | 241 | if ((i < 2) && (has_N44_O17_errata[policy->cpu])) |
228 | p4clockmod_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 242 | p4clockmod_table[i].frequency = CPUFREQ_ENTRY_INVALID; |
229 | else | 243 | else |
230 | p4clockmod_table[i].frequency = (stock_freq * i)/8; | 244 | p4clockmod_table[i].frequency = (stock_freq * i)/8; |
@@ -232,7 +246,10 @@ static int cpufreq_p4_cpu_init(struct cpufreq_policy *policy) | |||
232 | cpufreq_frequency_table_get_attr(p4clockmod_table, policy->cpu); | 246 | cpufreq_frequency_table_get_attr(p4clockmod_table, policy->cpu); |
233 | 247 | ||
234 | /* cpuinfo and default policy values */ | 248 | /* cpuinfo and default policy values */ |
235 | policy->cpuinfo.transition_latency = 1000000; /* assumed */ | 249 | |
250 | /* the transition latency is set to be 1 higher than the maximum | ||
251 | * transition latency of the ondemand governor */ | ||
252 | policy->cpuinfo.transition_latency = 10000001; | ||
236 | policy->cur = stock_freq; | 253 | policy->cur = stock_freq; |
237 | 254 | ||
238 | return cpufreq_frequency_table_cpuinfo(policy, &p4clockmod_table[0]); | 255 | return cpufreq_frequency_table_cpuinfo(policy, &p4clockmod_table[0]); |
@@ -258,12 +275,12 @@ static unsigned int cpufreq_p4_get(unsigned int cpu) | |||
258 | l = DC_DISABLE; | 275 | l = DC_DISABLE; |
259 | 276 | ||
260 | if (l != DC_DISABLE) | 277 | if (l != DC_DISABLE) |
261 | return (stock_freq * l / 8); | 278 | return stock_freq * l / 8; |
262 | 279 | ||
263 | return stock_freq; | 280 | return stock_freq; |
264 | } | 281 | } |
265 | 282 | ||
266 | static struct freq_attr* p4clockmod_attr[] = { | 283 | static struct freq_attr *p4clockmod_attr[] = { |
267 | &cpufreq_freq_attr_scaling_available_freqs, | 284 | &cpufreq_freq_attr_scaling_available_freqs, |
268 | NULL, | 285 | NULL, |
269 | }; | 286 | }; |
@@ -298,9 +315,10 @@ static int __init cpufreq_p4_init(void) | |||
298 | 315 | ||
299 | ret = cpufreq_register_driver(&p4clockmod_driver); | 316 | ret = cpufreq_register_driver(&p4clockmod_driver); |
300 | if (!ret) | 317 | if (!ret) |
301 | printk(KERN_INFO PFX "P4/Xeon(TM) CPU On-Demand Clock Modulation available\n"); | 318 | printk(KERN_INFO PFX "P4/Xeon(TM) CPU On-Demand Clock " |
319 | "Modulation available\n"); | ||
302 | 320 | ||
303 | return (ret); | 321 | return ret; |
304 | } | 322 | } |
305 | 323 | ||
306 | 324 | ||
@@ -310,9 +328,9 @@ static void __exit cpufreq_p4_exit(void) | |||
310 | } | 328 | } |
311 | 329 | ||
312 | 330 | ||
313 | MODULE_AUTHOR ("Zwane Mwaikambo <zwane@commfireservices.com>"); | 331 | MODULE_AUTHOR("Zwane Mwaikambo <zwane@commfireservices.com>"); |
314 | MODULE_DESCRIPTION ("cpufreq driver for Pentium(TM) 4/Xeon(TM)"); | 332 | MODULE_DESCRIPTION("cpufreq driver for Pentium(TM) 4/Xeon(TM)"); |
315 | MODULE_LICENSE ("GPL"); | 333 | MODULE_LICENSE("GPL"); |
316 | 334 | ||
317 | late_initcall(cpufreq_p4_init); | 335 | late_initcall(cpufreq_p4_init); |
318 | module_exit(cpufreq_p4_exit); | 336 | module_exit(cpufreq_p4_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/powernow-k6.c b/arch/x86/kernel/cpu/cpufreq/powernow-k6.c index c1ac5790c63e..f10dea409f40 100644 --- a/arch/x86/kernel/cpu/cpufreq/powernow-k6.c +++ b/arch/x86/kernel/cpu/cpufreq/powernow-k6.c | |||
@@ -1,6 +1,7 @@ | |||
1 | /* | 1 | /* |
2 | * This file was based upon code in Powertweak Linux (http://powertweak.sf.net) | 2 | * This file was based upon code in Powertweak Linux (http://powertweak.sf.net) |
3 | * (C) 2000-2003 Dave Jones, Arjan van de Ven, Janne Pänkälä, Dominik Brodowski. | 3 | * (C) 2000-2003 Dave Jones, Arjan van de Ven, Janne Pänkälä, |
4 | * Dominik Brodowski. | ||
4 | * | 5 | * |
5 | * Licensed under the terms of the GNU GPL License version 2. | 6 | * Licensed under the terms of the GNU GPL License version 2. |
6 | * | 7 | * |
@@ -13,14 +14,15 @@ | |||
13 | #include <linux/cpufreq.h> | 14 | #include <linux/cpufreq.h> |
14 | #include <linux/ioport.h> | 15 | #include <linux/ioport.h> |
15 | #include <linux/slab.h> | 16 | #include <linux/slab.h> |
16 | |||
17 | #include <asm/msr.h> | ||
18 | #include <linux/timex.h> | 17 | #include <linux/timex.h> |
19 | #include <linux/io.h> | 18 | #include <linux/io.h> |
20 | 19 | ||
20 | #include <asm/msr.h> | ||
21 | |||
21 | #define POWERNOW_IOPORT 0xfff0 /* it doesn't matter where, as long | 22 | #define POWERNOW_IOPORT 0xfff0 /* it doesn't matter where, as long |
22 | as it is unused */ | 23 | as it is unused */ |
23 | 24 | ||
25 | #define PFX "powernow-k6: " | ||
24 | static unsigned int busfreq; /* FSB, in 10 kHz */ | 26 | static unsigned int busfreq; /* FSB, in 10 kHz */ |
25 | static unsigned int max_multiplier; | 27 | static unsigned int max_multiplier; |
26 | 28 | ||
@@ -47,8 +49,8 @@ static struct cpufreq_frequency_table clock_ratio[] = { | |||
47 | */ | 49 | */ |
48 | static int powernow_k6_get_cpu_multiplier(void) | 50 | static int powernow_k6_get_cpu_multiplier(void) |
49 | { | 51 | { |
50 | u64 invalue = 0; | 52 | u64 invalue = 0; |
51 | u32 msrval; | 53 | u32 msrval; |
52 | 54 | ||
53 | msrval = POWERNOW_IOPORT + 0x1; | 55 | msrval = POWERNOW_IOPORT + 0x1; |
54 | wrmsr(MSR_K6_EPMR, msrval, 0); /* enable the PowerNow port */ | 56 | wrmsr(MSR_K6_EPMR, msrval, 0); /* enable the PowerNow port */ |
@@ -68,12 +70,12 @@ static int powernow_k6_get_cpu_multiplier(void) | |||
68 | */ | 70 | */ |
69 | static void powernow_k6_set_state(unsigned int best_i) | 71 | static void powernow_k6_set_state(unsigned int best_i) |
70 | { | 72 | { |
71 | unsigned long outvalue = 0, invalue = 0; | 73 | unsigned long outvalue = 0, invalue = 0; |
72 | unsigned long msrval; | 74 | unsigned long msrval; |
73 | struct cpufreq_freqs freqs; | 75 | struct cpufreq_freqs freqs; |
74 | 76 | ||
75 | if (clock_ratio[best_i].index > max_multiplier) { | 77 | if (clock_ratio[best_i].index > max_multiplier) { |
76 | printk(KERN_ERR "cpufreq: invalid target frequency\n"); | 78 | printk(KERN_ERR PFX "invalid target frequency\n"); |
77 | return; | 79 | return; |
78 | } | 80 | } |
79 | 81 | ||
@@ -119,7 +121,8 @@ static int powernow_k6_verify(struct cpufreq_policy *policy) | |||
119 | * powernow_k6_setpolicy - sets a new CPUFreq policy | 121 | * powernow_k6_setpolicy - sets a new CPUFreq policy |
120 | * @policy: new policy | 122 | * @policy: new policy |
121 | * @target_freq: the target frequency | 123 | * @target_freq: the target frequency |
122 | * @relation: how that frequency relates to achieved frequency (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | 124 | * @relation: how that frequency relates to achieved frequency |
125 | * (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | ||
123 | * | 126 | * |
124 | * sets a new CPUFreq policy | 127 | * sets a new CPUFreq policy |
125 | */ | 128 | */ |
@@ -127,9 +130,10 @@ static int powernow_k6_target(struct cpufreq_policy *policy, | |||
127 | unsigned int target_freq, | 130 | unsigned int target_freq, |
128 | unsigned int relation) | 131 | unsigned int relation) |
129 | { | 132 | { |
130 | unsigned int newstate = 0; | 133 | unsigned int newstate = 0; |
131 | 134 | ||
132 | if (cpufreq_frequency_table_target(policy, &clock_ratio[0], target_freq, relation, &newstate)) | 135 | if (cpufreq_frequency_table_target(policy, &clock_ratio[0], |
136 | target_freq, relation, &newstate)) | ||
133 | return -EINVAL; | 137 | return -EINVAL; |
134 | 138 | ||
135 | powernow_k6_set_state(newstate); | 139 | powernow_k6_set_state(newstate); |
@@ -140,7 +144,7 @@ static int powernow_k6_target(struct cpufreq_policy *policy, | |||
140 | 144 | ||
141 | static int powernow_k6_cpu_init(struct cpufreq_policy *policy) | 145 | static int powernow_k6_cpu_init(struct cpufreq_policy *policy) |
142 | { | 146 | { |
143 | unsigned int i; | 147 | unsigned int i, f; |
144 | int result; | 148 | int result; |
145 | 149 | ||
146 | if (policy->cpu != 0) | 150 | if (policy->cpu != 0) |
@@ -152,10 +156,11 @@ static int powernow_k6_cpu_init(struct cpufreq_policy *policy) | |||
152 | 156 | ||
153 | /* table init */ | 157 | /* table init */ |
154 | for (i = 0; (clock_ratio[i].frequency != CPUFREQ_TABLE_END); i++) { | 158 | for (i = 0; (clock_ratio[i].frequency != CPUFREQ_TABLE_END); i++) { |
155 | if (clock_ratio[i].index > max_multiplier) | 159 | f = clock_ratio[i].index; |
160 | if (f > max_multiplier) | ||
156 | clock_ratio[i].frequency = CPUFREQ_ENTRY_INVALID; | 161 | clock_ratio[i].frequency = CPUFREQ_ENTRY_INVALID; |
157 | else | 162 | else |
158 | clock_ratio[i].frequency = busfreq * clock_ratio[i].index; | 163 | clock_ratio[i].frequency = busfreq * f; |
159 | } | 164 | } |
160 | 165 | ||
161 | /* cpuinfo and default policy values */ | 166 | /* cpuinfo and default policy values */ |
@@ -185,7 +190,9 @@ static int powernow_k6_cpu_exit(struct cpufreq_policy *policy) | |||
185 | 190 | ||
186 | static unsigned int powernow_k6_get(unsigned int cpu) | 191 | static unsigned int powernow_k6_get(unsigned int cpu) |
187 | { | 192 | { |
188 | return busfreq * powernow_k6_get_cpu_multiplier(); | 193 | unsigned int ret; |
194 | ret = (busfreq * powernow_k6_get_cpu_multiplier()); | ||
195 | return ret; | ||
189 | } | 196 | } |
190 | 197 | ||
191 | static struct freq_attr *powernow_k6_attr[] = { | 198 | static struct freq_attr *powernow_k6_attr[] = { |
@@ -221,7 +228,7 @@ static int __init powernow_k6_init(void) | |||
221 | return -ENODEV; | 228 | return -ENODEV; |
222 | 229 | ||
223 | if (!request_region(POWERNOW_IOPORT, 16, "PowerNow!")) { | 230 | if (!request_region(POWERNOW_IOPORT, 16, "PowerNow!")) { |
224 | printk("cpufreq: PowerNow IOPORT region already used.\n"); | 231 | printk(KERN_INFO PFX "PowerNow IOPORT region already used.\n"); |
225 | return -EIO; | 232 | return -EIO; |
226 | } | 233 | } |
227 | 234 | ||
@@ -246,7 +253,8 @@ static void __exit powernow_k6_exit(void) | |||
246 | } | 253 | } |
247 | 254 | ||
248 | 255 | ||
249 | MODULE_AUTHOR("Arjan van de Ven, Dave Jones <davej@redhat.com>, Dominik Brodowski <linux@brodo.de>"); | 256 | MODULE_AUTHOR("Arjan van de Ven, Dave Jones <davej@redhat.com>, " |
257 | "Dominik Brodowski <linux@brodo.de>"); | ||
250 | MODULE_DESCRIPTION("PowerNow! driver for AMD K6-2+ / K6-3+ processors."); | 258 | MODULE_DESCRIPTION("PowerNow! driver for AMD K6-2+ / K6-3+ processors."); |
251 | MODULE_LICENSE("GPL"); | 259 | MODULE_LICENSE("GPL"); |
252 | 260 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/powernow-k7.c b/arch/x86/kernel/cpu/cpufreq/powernow-k7.c index 1b446d79a8fd..3c28ccd49742 100644 --- a/arch/x86/kernel/cpu/cpufreq/powernow-k7.c +++ b/arch/x86/kernel/cpu/cpufreq/powernow-k7.c | |||
@@ -6,10 +6,12 @@ | |||
6 | * Licensed under the terms of the GNU GPL License version 2. | 6 | * Licensed under the terms of the GNU GPL License version 2. |
7 | * Based upon datasheets & sample CPUs kindly provided by AMD. | 7 | * Based upon datasheets & sample CPUs kindly provided by AMD. |
8 | * | 8 | * |
9 | * Errata 5: Processor may fail to execute a FID/VID change in presence of interrupt. | 9 | * Errata 5: |
10 | * - We cli/sti on stepping A0 CPUs around the FID/VID transition. | 10 | * CPU may fail to execute a FID/VID change in presence of interrupt. |
11 | * Errata 15: Processors with half frequency multipliers may hang upon wakeup from disconnect. | 11 | * - We cli/sti on stepping A0 CPUs around the FID/VID transition. |
12 | * - We disable half multipliers if ACPI is used on A0 stepping CPUs. | 12 | * Errata 15: |
13 | * CPU with half frequency multipliers may hang upon wakeup from disconnect. | ||
14 | * - We disable half multipliers if ACPI is used on A0 stepping CPUs. | ||
13 | */ | 15 | */ |
14 | 16 | ||
15 | #include <linux/kernel.h> | 17 | #include <linux/kernel.h> |
@@ -20,11 +22,11 @@ | |||
20 | #include <linux/slab.h> | 22 | #include <linux/slab.h> |
21 | #include <linux/string.h> | 23 | #include <linux/string.h> |
22 | #include <linux/dmi.h> | 24 | #include <linux/dmi.h> |
25 | #include <linux/timex.h> | ||
26 | #include <linux/io.h> | ||
23 | 27 | ||
28 | #include <asm/timer.h> /* Needed for recalibrate_cpu_khz() */ | ||
24 | #include <asm/msr.h> | 29 | #include <asm/msr.h> |
25 | #include <asm/timer.h> | ||
26 | #include <asm/timex.h> | ||
27 | #include <asm/io.h> | ||
28 | #include <asm/system.h> | 30 | #include <asm/system.h> |
29 | 31 | ||
30 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI | 32 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI |
@@ -58,9 +60,9 @@ struct pst_s { | |||
58 | union powernow_acpi_control_t { | 60 | union powernow_acpi_control_t { |
59 | struct { | 61 | struct { |
60 | unsigned long fid:5, | 62 | unsigned long fid:5, |
61 | vid:5, | 63 | vid:5, |
62 | sgtc:20, | 64 | sgtc:20, |
63 | res1:2; | 65 | res1:2; |
64 | } bits; | 66 | } bits; |
65 | unsigned long val; | 67 | unsigned long val; |
66 | }; | 68 | }; |
@@ -94,14 +96,15 @@ static struct cpufreq_frequency_table *powernow_table; | |||
94 | 96 | ||
95 | static unsigned int can_scale_bus; | 97 | static unsigned int can_scale_bus; |
96 | static unsigned int can_scale_vid; | 98 | static unsigned int can_scale_vid; |
97 | static unsigned int minimum_speed=-1; | 99 | static unsigned int minimum_speed = -1; |
98 | static unsigned int maximum_speed; | 100 | static unsigned int maximum_speed; |
99 | static unsigned int number_scales; | 101 | static unsigned int number_scales; |
100 | static unsigned int fsb; | 102 | static unsigned int fsb; |
101 | static unsigned int latency; | 103 | static unsigned int latency; |
102 | static char have_a0; | 104 | static char have_a0; |
103 | 105 | ||
104 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "powernow-k7", msg) | 106 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
107 | "powernow-k7", msg) | ||
105 | 108 | ||
106 | static int check_fsb(unsigned int fsbspeed) | 109 | static int check_fsb(unsigned int fsbspeed) |
107 | { | 110 | { |
@@ -109,7 +112,7 @@ static int check_fsb(unsigned int fsbspeed) | |||
109 | unsigned int f = fsb / 1000; | 112 | unsigned int f = fsb / 1000; |
110 | 113 | ||
111 | delta = (fsbspeed > f) ? fsbspeed - f : f - fsbspeed; | 114 | delta = (fsbspeed > f) ? fsbspeed - f : f - fsbspeed; |
112 | return (delta < 5); | 115 | return delta < 5; |
113 | } | 116 | } |
114 | 117 | ||
115 | static int check_powernow(void) | 118 | static int check_powernow(void) |
@@ -117,24 +120,26 @@ static int check_powernow(void) | |||
117 | struct cpuinfo_x86 *c = &cpu_data(0); | 120 | struct cpuinfo_x86 *c = &cpu_data(0); |
118 | unsigned int maxei, eax, ebx, ecx, edx; | 121 | unsigned int maxei, eax, ebx, ecx, edx; |
119 | 122 | ||
120 | if ((c->x86_vendor != X86_VENDOR_AMD) || (c->x86 !=6)) { | 123 | if ((c->x86_vendor != X86_VENDOR_AMD) || (c->x86 != 6)) { |
121 | #ifdef MODULE | 124 | #ifdef MODULE |
122 | printk (KERN_INFO PFX "This module only works with AMD K7 CPUs\n"); | 125 | printk(KERN_INFO PFX "This module only works with " |
126 | "AMD K7 CPUs\n"); | ||
123 | #endif | 127 | #endif |
124 | return 0; | 128 | return 0; |
125 | } | 129 | } |
126 | 130 | ||
127 | /* Get maximum capabilities */ | 131 | /* Get maximum capabilities */ |
128 | maxei = cpuid_eax (0x80000000); | 132 | maxei = cpuid_eax(0x80000000); |
129 | if (maxei < 0x80000007) { /* Any powernow info ? */ | 133 | if (maxei < 0x80000007) { /* Any powernow info ? */ |
130 | #ifdef MODULE | 134 | #ifdef MODULE |
131 | printk (KERN_INFO PFX "No powernow capabilities detected\n"); | 135 | printk(KERN_INFO PFX "No powernow capabilities detected\n"); |
132 | #endif | 136 | #endif |
133 | return 0; | 137 | return 0; |
134 | } | 138 | } |
135 | 139 | ||
136 | if ((c->x86_model == 6) && (c->x86_mask == 0)) { | 140 | if ((c->x86_model == 6) && (c->x86_mask == 0)) { |
137 | printk (KERN_INFO PFX "K7 660[A0] core detected, enabling errata workarounds\n"); | 141 | printk(KERN_INFO PFX "K7 660[A0] core detected, " |
142 | "enabling errata workarounds\n"); | ||
138 | have_a0 = 1; | 143 | have_a0 = 1; |
139 | } | 144 | } |
140 | 145 | ||
@@ -144,37 +149,42 @@ static int check_powernow(void) | |||
144 | if (!(edx & (1 << 1 | 1 << 2))) | 149 | if (!(edx & (1 << 1 | 1 << 2))) |
145 | return 0; | 150 | return 0; |
146 | 151 | ||
147 | printk (KERN_INFO PFX "PowerNOW! Technology present. Can scale: "); | 152 | printk(KERN_INFO PFX "PowerNOW! Technology present. Can scale: "); |
148 | 153 | ||
149 | if (edx & 1 << 1) { | 154 | if (edx & 1 << 1) { |
150 | printk ("frequency"); | 155 | printk("frequency"); |
151 | can_scale_bus=1; | 156 | can_scale_bus = 1; |
152 | } | 157 | } |
153 | 158 | ||
154 | if ((edx & (1 << 1 | 1 << 2)) == 0x6) | 159 | if ((edx & (1 << 1 | 1 << 2)) == 0x6) |
155 | printk (" and "); | 160 | printk(" and "); |
156 | 161 | ||
157 | if (edx & 1 << 2) { | 162 | if (edx & 1 << 2) { |
158 | printk ("voltage"); | 163 | printk("voltage"); |
159 | can_scale_vid=1; | 164 | can_scale_vid = 1; |
160 | } | 165 | } |
161 | 166 | ||
162 | printk (".\n"); | 167 | printk(".\n"); |
163 | return 1; | 168 | return 1; |
164 | } | 169 | } |
165 | 170 | ||
171 | static void invalidate_entry(unsigned int entry) | ||
172 | { | ||
173 | powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID; | ||
174 | } | ||
166 | 175 | ||
167 | static int get_ranges (unsigned char *pst) | 176 | static int get_ranges(unsigned char *pst) |
168 | { | 177 | { |
169 | unsigned int j; | 178 | unsigned int j; |
170 | unsigned int speed; | 179 | unsigned int speed; |
171 | u8 fid, vid; | 180 | u8 fid, vid; |
172 | 181 | ||
173 | powernow_table = kzalloc((sizeof(struct cpufreq_frequency_table) * (number_scales + 1)), GFP_KERNEL); | 182 | powernow_table = kzalloc((sizeof(struct cpufreq_frequency_table) * |
183 | (number_scales + 1)), GFP_KERNEL); | ||
174 | if (!powernow_table) | 184 | if (!powernow_table) |
175 | return -ENOMEM; | 185 | return -ENOMEM; |
176 | 186 | ||
177 | for (j=0 ; j < number_scales; j++) { | 187 | for (j = 0 ; j < number_scales; j++) { |
178 | fid = *pst++; | 188 | fid = *pst++; |
179 | 189 | ||
180 | powernow_table[j].frequency = (fsb * fid_codes[fid]) / 10; | 190 | powernow_table[j].frequency = (fsb * fid_codes[fid]) / 10; |
@@ -182,10 +192,10 @@ static int get_ranges (unsigned char *pst) | |||
182 | 192 | ||
183 | speed = powernow_table[j].frequency; | 193 | speed = powernow_table[j].frequency; |
184 | 194 | ||
185 | if ((fid_codes[fid] % 10)==5) { | 195 | if ((fid_codes[fid] % 10) == 5) { |
186 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI | 196 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI |
187 | if (have_a0 == 1) | 197 | if (have_a0 == 1) |
188 | powernow_table[j].frequency = CPUFREQ_ENTRY_INVALID; | 198 | invalidate_entry(j); |
189 | #endif | 199 | #endif |
190 | } | 200 | } |
191 | 201 | ||
@@ -197,7 +207,7 @@ static int get_ranges (unsigned char *pst) | |||
197 | vid = *pst++; | 207 | vid = *pst++; |
198 | powernow_table[j].index |= (vid << 8); /* upper 8 bits */ | 208 | powernow_table[j].index |= (vid << 8); /* upper 8 bits */ |
199 | 209 | ||
200 | dprintk (" FID: 0x%x (%d.%dx [%dMHz]) " | 210 | dprintk(" FID: 0x%x (%d.%dx [%dMHz]) " |
201 | "VID: 0x%x (%d.%03dV)\n", fid, fid_codes[fid] / 10, | 211 | "VID: 0x%x (%d.%03dV)\n", fid, fid_codes[fid] / 10, |
202 | fid_codes[fid] % 10, speed/1000, vid, | 212 | fid_codes[fid] % 10, speed/1000, vid, |
203 | mobile_vid_table[vid]/1000, | 213 | mobile_vid_table[vid]/1000, |
@@ -214,13 +224,13 @@ static void change_FID(int fid) | |||
214 | { | 224 | { |
215 | union msr_fidvidctl fidvidctl; | 225 | union msr_fidvidctl fidvidctl; |
216 | 226 | ||
217 | rdmsrl (MSR_K7_FID_VID_CTL, fidvidctl.val); | 227 | rdmsrl(MSR_K7_FID_VID_CTL, fidvidctl.val); |
218 | if (fidvidctl.bits.FID != fid) { | 228 | if (fidvidctl.bits.FID != fid) { |
219 | fidvidctl.bits.SGTC = latency; | 229 | fidvidctl.bits.SGTC = latency; |
220 | fidvidctl.bits.FID = fid; | 230 | fidvidctl.bits.FID = fid; |
221 | fidvidctl.bits.VIDC = 0; | 231 | fidvidctl.bits.VIDC = 0; |
222 | fidvidctl.bits.FIDC = 1; | 232 | fidvidctl.bits.FIDC = 1; |
223 | wrmsrl (MSR_K7_FID_VID_CTL, fidvidctl.val); | 233 | wrmsrl(MSR_K7_FID_VID_CTL, fidvidctl.val); |
224 | } | 234 | } |
225 | } | 235 | } |
226 | 236 | ||
@@ -229,18 +239,18 @@ static void change_VID(int vid) | |||
229 | { | 239 | { |
230 | union msr_fidvidctl fidvidctl; | 240 | union msr_fidvidctl fidvidctl; |
231 | 241 | ||
232 | rdmsrl (MSR_K7_FID_VID_CTL, fidvidctl.val); | 242 | rdmsrl(MSR_K7_FID_VID_CTL, fidvidctl.val); |
233 | if (fidvidctl.bits.VID != vid) { | 243 | if (fidvidctl.bits.VID != vid) { |
234 | fidvidctl.bits.SGTC = latency; | 244 | fidvidctl.bits.SGTC = latency; |
235 | fidvidctl.bits.VID = vid; | 245 | fidvidctl.bits.VID = vid; |
236 | fidvidctl.bits.FIDC = 0; | 246 | fidvidctl.bits.FIDC = 0; |
237 | fidvidctl.bits.VIDC = 1; | 247 | fidvidctl.bits.VIDC = 1; |
238 | wrmsrl (MSR_K7_FID_VID_CTL, fidvidctl.val); | 248 | wrmsrl(MSR_K7_FID_VID_CTL, fidvidctl.val); |
239 | } | 249 | } |
240 | } | 250 | } |
241 | 251 | ||
242 | 252 | ||
243 | static void change_speed (unsigned int index) | 253 | static void change_speed(unsigned int index) |
244 | { | 254 | { |
245 | u8 fid, vid; | 255 | u8 fid, vid; |
246 | struct cpufreq_freqs freqs; | 256 | struct cpufreq_freqs freqs; |
@@ -257,7 +267,7 @@ static void change_speed (unsigned int index) | |||
257 | 267 | ||
258 | freqs.cpu = 0; | 268 | freqs.cpu = 0; |
259 | 269 | ||
260 | rdmsrl (MSR_K7_FID_VID_STATUS, fidvidstatus.val); | 270 | rdmsrl(MSR_K7_FID_VID_STATUS, fidvidstatus.val); |
261 | cfid = fidvidstatus.bits.CFID; | 271 | cfid = fidvidstatus.bits.CFID; |
262 | freqs.old = fsb * fid_codes[cfid] / 10; | 272 | freqs.old = fsb * fid_codes[cfid] / 10; |
263 | 273 | ||
@@ -321,12 +331,14 @@ static int powernow_acpi_init(void) | |||
321 | goto err1; | 331 | goto err1; |
322 | } | 332 | } |
323 | 333 | ||
324 | if (acpi_processor_perf->control_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) { | 334 | if (acpi_processor_perf->control_register.space_id != |
335 | ACPI_ADR_SPACE_FIXED_HARDWARE) { | ||
325 | retval = -ENODEV; | 336 | retval = -ENODEV; |
326 | goto err2; | 337 | goto err2; |
327 | } | 338 | } |
328 | 339 | ||
329 | if (acpi_processor_perf->status_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) { | 340 | if (acpi_processor_perf->status_register.space_id != |
341 | ACPI_ADR_SPACE_FIXED_HARDWARE) { | ||
330 | retval = -ENODEV; | 342 | retval = -ENODEV; |
331 | goto err2; | 343 | goto err2; |
332 | } | 344 | } |
@@ -338,7 +350,8 @@ static int powernow_acpi_init(void) | |||
338 | goto err2; | 350 | goto err2; |
339 | } | 351 | } |
340 | 352 | ||
341 | powernow_table = kzalloc((number_scales + 1) * (sizeof(struct cpufreq_frequency_table)), GFP_KERNEL); | 353 | powernow_table = kzalloc((sizeof(struct cpufreq_frequency_table) * |
354 | (number_scales + 1)), GFP_KERNEL); | ||
342 | if (!powernow_table) { | 355 | if (!powernow_table) { |
343 | retval = -ENOMEM; | 356 | retval = -ENOMEM; |
344 | goto err2; | 357 | goto err2; |
@@ -352,7 +365,7 @@ static int powernow_acpi_init(void) | |||
352 | unsigned int speed, speed_mhz; | 365 | unsigned int speed, speed_mhz; |
353 | 366 | ||
354 | pc.val = (unsigned long) state->control; | 367 | pc.val = (unsigned long) state->control; |
355 | dprintk ("acpi: P%d: %d MHz %d mW %d uS control %08x SGTC %d\n", | 368 | dprintk("acpi: P%d: %d MHz %d mW %d uS control %08x SGTC %d\n", |
356 | i, | 369 | i, |
357 | (u32) state->core_frequency, | 370 | (u32) state->core_frequency, |
358 | (u32) state->power, | 371 | (u32) state->power, |
@@ -381,12 +394,12 @@ static int powernow_acpi_init(void) | |||
381 | if (speed % 1000 > 0) | 394 | if (speed % 1000 > 0) |
382 | speed_mhz++; | 395 | speed_mhz++; |
383 | 396 | ||
384 | if ((fid_codes[fid] % 10)==5) { | 397 | if ((fid_codes[fid] % 10) == 5) { |
385 | if (have_a0 == 1) | 398 | if (have_a0 == 1) |
386 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 399 | invalidate_entry(i); |
387 | } | 400 | } |
388 | 401 | ||
389 | dprintk (" FID: 0x%x (%d.%dx [%dMHz]) " | 402 | dprintk(" FID: 0x%x (%d.%dx [%dMHz]) " |
390 | "VID: 0x%x (%d.%03dV)\n", fid, fid_codes[fid] / 10, | 403 | "VID: 0x%x (%d.%03dV)\n", fid, fid_codes[fid] / 10, |
391 | fid_codes[fid] % 10, speed_mhz, vid, | 404 | fid_codes[fid] % 10, speed_mhz, vid, |
392 | mobile_vid_table[vid]/1000, | 405 | mobile_vid_table[vid]/1000, |
@@ -422,7 +435,8 @@ err1: | |||
422 | err05: | 435 | err05: |
423 | kfree(acpi_processor_perf); | 436 | kfree(acpi_processor_perf); |
424 | err0: | 437 | err0: |
425 | printk(KERN_WARNING PFX "ACPI perflib can not be used in this platform\n"); | 438 | printk(KERN_WARNING PFX "ACPI perflib can not be used on " |
439 | "this platform\n"); | ||
426 | acpi_processor_perf = NULL; | 440 | acpi_processor_perf = NULL; |
427 | return retval; | 441 | return retval; |
428 | } | 442 | } |
@@ -435,7 +449,14 @@ static int powernow_acpi_init(void) | |||
435 | } | 449 | } |
436 | #endif | 450 | #endif |
437 | 451 | ||
438 | static int powernow_decode_bios (int maxfid, int startvid) | 452 | static void print_pst_entry(struct pst_s *pst, unsigned int j) |
453 | { | ||
454 | dprintk("PST:%d (@%p)\n", j, pst); | ||
455 | dprintk(" cpuid: 0x%x fsb: %d maxFID: 0x%x startvid: 0x%x\n", | ||
456 | pst->cpuid, pst->fsbspeed, pst->maxfid, pst->startvid); | ||
457 | } | ||
458 | |||
459 | static int powernow_decode_bios(int maxfid, int startvid) | ||
439 | { | 460 | { |
440 | struct psb_s *psb; | 461 | struct psb_s *psb; |
441 | struct pst_s *pst; | 462 | struct pst_s *pst; |
@@ -446,61 +467,67 @@ static int powernow_decode_bios (int maxfid, int startvid) | |||
446 | 467 | ||
447 | etuple = cpuid_eax(0x80000001); | 468 | etuple = cpuid_eax(0x80000001); |
448 | 469 | ||
449 | for (i=0xC0000; i < 0xffff0 ; i+=16) { | 470 | for (i = 0xC0000; i < 0xffff0 ; i += 16) { |
450 | 471 | ||
451 | p = phys_to_virt(i); | 472 | p = phys_to_virt(i); |
452 | 473 | ||
453 | if (memcmp(p, "AMDK7PNOW!", 10) == 0){ | 474 | if (memcmp(p, "AMDK7PNOW!", 10) == 0) { |
454 | dprintk ("Found PSB header at %p\n", p); | 475 | dprintk("Found PSB header at %p\n", p); |
455 | psb = (struct psb_s *) p; | 476 | psb = (struct psb_s *) p; |
456 | dprintk ("Table version: 0x%x\n", psb->tableversion); | 477 | dprintk("Table version: 0x%x\n", psb->tableversion); |
457 | if (psb->tableversion != 0x12) { | 478 | if (psb->tableversion != 0x12) { |
458 | printk (KERN_INFO PFX "Sorry, only v1.2 tables supported right now\n"); | 479 | printk(KERN_INFO PFX "Sorry, only v1.2 tables" |
480 | " supported right now\n"); | ||
459 | return -ENODEV; | 481 | return -ENODEV; |
460 | } | 482 | } |
461 | 483 | ||
462 | dprintk ("Flags: 0x%x\n", psb->flags); | 484 | dprintk("Flags: 0x%x\n", psb->flags); |
463 | if ((psb->flags & 1)==0) { | 485 | if ((psb->flags & 1) == 0) |
464 | dprintk ("Mobile voltage regulator\n"); | 486 | dprintk("Mobile voltage regulator\n"); |
465 | } else { | 487 | else |
466 | dprintk ("Desktop voltage regulator\n"); | 488 | dprintk("Desktop voltage regulator\n"); |
467 | } | ||
468 | 489 | ||
469 | latency = psb->settlingtime; | 490 | latency = psb->settlingtime; |
470 | if (latency < 100) { | 491 | if (latency < 100) { |
471 | printk(KERN_INFO PFX "BIOS set settling time to %d microseconds. " | 492 | printk(KERN_INFO PFX "BIOS set settling time " |
472 | "Should be at least 100. Correcting.\n", latency); | 493 | "to %d microseconds. " |
494 | "Should be at least 100. " | ||
495 | "Correcting.\n", latency); | ||
473 | latency = 100; | 496 | latency = 100; |
474 | } | 497 | } |
475 | dprintk ("Settling Time: %d microseconds.\n", psb->settlingtime); | 498 | dprintk("Settling Time: %d microseconds.\n", |
476 | dprintk ("Has %d PST tables. (Only dumping ones relevant to this CPU).\n", psb->numpst); | 499 | psb->settlingtime); |
500 | dprintk("Has %d PST tables. (Only dumping ones " | ||
501 | "relevant to this CPU).\n", | ||
502 | psb->numpst); | ||
477 | 503 | ||
478 | p += sizeof (struct psb_s); | 504 | p += sizeof(struct psb_s); |
479 | 505 | ||
480 | pst = (struct pst_s *) p; | 506 | pst = (struct pst_s *) p; |
481 | 507 | ||
482 | for (j=0; j<psb->numpst; j++) { | 508 | for (j = 0; j < psb->numpst; j++) { |
483 | pst = (struct pst_s *) p; | 509 | pst = (struct pst_s *) p; |
484 | number_scales = pst->numpstates; | 510 | number_scales = pst->numpstates; |
485 | 511 | ||
486 | if ((etuple == pst->cpuid) && check_fsb(pst->fsbspeed) && | 512 | if ((etuple == pst->cpuid) && |
487 | (maxfid==pst->maxfid) && (startvid==pst->startvid)) | 513 | check_fsb(pst->fsbspeed) && |
488 | { | 514 | (maxfid == pst->maxfid) && |
489 | dprintk ("PST:%d (@%p)\n", j, pst); | 515 | (startvid == pst->startvid)) { |
490 | dprintk (" cpuid: 0x%x fsb: %d maxFID: 0x%x startvid: 0x%x\n", | 516 | print_pst_entry(pst, j); |
491 | pst->cpuid, pst->fsbspeed, pst->maxfid, pst->startvid); | 517 | p = (char *)pst + sizeof(struct pst_s); |
492 | 518 | ret = get_ranges(p); | |
493 | ret = get_ranges ((char *) pst + sizeof (struct pst_s)); | ||
494 | return ret; | 519 | return ret; |
495 | } else { | 520 | } else { |
496 | unsigned int k; | 521 | unsigned int k; |
497 | p = (char *) pst + sizeof (struct pst_s); | 522 | p = (char *)pst + sizeof(struct pst_s); |
498 | for (k=0; k<number_scales; k++) | 523 | for (k = 0; k < number_scales; k++) |
499 | p+=2; | 524 | p += 2; |
500 | } | 525 | } |
501 | } | 526 | } |
502 | printk (KERN_INFO PFX "No PST tables match this cpuid (0x%x)\n", etuple); | 527 | printk(KERN_INFO PFX "No PST tables match this cpuid " |
503 | printk (KERN_INFO PFX "This is indicative of a broken BIOS.\n"); | 528 | "(0x%x)\n", etuple); |
529 | printk(KERN_INFO PFX "This is indicative of a broken " | ||
530 | "BIOS.\n"); | ||
504 | 531 | ||
505 | return -EINVAL; | 532 | return -EINVAL; |
506 | } | 533 | } |
@@ -511,13 +538,14 @@ static int powernow_decode_bios (int maxfid, int startvid) | |||
511 | } | 538 | } |
512 | 539 | ||
513 | 540 | ||
514 | static int powernow_target (struct cpufreq_policy *policy, | 541 | static int powernow_target(struct cpufreq_policy *policy, |
515 | unsigned int target_freq, | 542 | unsigned int target_freq, |
516 | unsigned int relation) | 543 | unsigned int relation) |
517 | { | 544 | { |
518 | unsigned int newstate; | 545 | unsigned int newstate; |
519 | 546 | ||
520 | if (cpufreq_frequency_table_target(policy, powernow_table, target_freq, relation, &newstate)) | 547 | if (cpufreq_frequency_table_target(policy, powernow_table, target_freq, |
548 | relation, &newstate)) | ||
521 | return -EINVAL; | 549 | return -EINVAL; |
522 | 550 | ||
523 | change_speed(newstate); | 551 | change_speed(newstate); |
@@ -526,7 +554,7 @@ static int powernow_target (struct cpufreq_policy *policy, | |||
526 | } | 554 | } |
527 | 555 | ||
528 | 556 | ||
529 | static int powernow_verify (struct cpufreq_policy *policy) | 557 | static int powernow_verify(struct cpufreq_policy *policy) |
530 | { | 558 | { |
531 | return cpufreq_frequency_table_verify(policy, powernow_table); | 559 | return cpufreq_frequency_table_verify(policy, powernow_table); |
532 | } | 560 | } |
@@ -566,18 +594,23 @@ static unsigned int powernow_get(unsigned int cpu) | |||
566 | 594 | ||
567 | if (cpu) | 595 | if (cpu) |
568 | return 0; | 596 | return 0; |
569 | rdmsrl (MSR_K7_FID_VID_STATUS, fidvidstatus.val); | 597 | rdmsrl(MSR_K7_FID_VID_STATUS, fidvidstatus.val); |
570 | cfid = fidvidstatus.bits.CFID; | 598 | cfid = fidvidstatus.bits.CFID; |
571 | 599 | ||
572 | return (fsb * fid_codes[cfid] / 10); | 600 | return fsb * fid_codes[cfid] / 10; |
573 | } | 601 | } |
574 | 602 | ||
575 | 603 | ||
576 | static int __init acer_cpufreq_pst(const struct dmi_system_id *d) | 604 | static int __init acer_cpufreq_pst(const struct dmi_system_id *d) |
577 | { | 605 | { |
578 | printk(KERN_WARNING "%s laptop with broken PST tables in BIOS detected.\n", d->ident); | 606 | printk(KERN_WARNING PFX |
579 | printk(KERN_WARNING "You need to downgrade to 3A21 (09/09/2002), or try a newer BIOS than 3A71 (01/20/2003)\n"); | 607 | "%s laptop with broken PST tables in BIOS detected.\n", |
580 | printk(KERN_WARNING "cpufreq scaling has been disabled as a result of this.\n"); | 608 | d->ident); |
609 | printk(KERN_WARNING PFX | ||
610 | "You need to downgrade to 3A21 (09/09/2002), or try a newer " | ||
611 | "BIOS than 3A71 (01/20/2003)\n"); | ||
612 | printk(KERN_WARNING PFX | ||
613 | "cpufreq scaling has been disabled as a result of this.\n"); | ||
581 | return 0; | 614 | return 0; |
582 | } | 615 | } |
583 | 616 | ||
@@ -598,7 +631,7 @@ static struct dmi_system_id __initdata powernow_dmi_table[] = { | |||
598 | { } | 631 | { } |
599 | }; | 632 | }; |
600 | 633 | ||
601 | static int __init powernow_cpu_init (struct cpufreq_policy *policy) | 634 | static int __init powernow_cpu_init(struct cpufreq_policy *policy) |
602 | { | 635 | { |
603 | union msr_fidvidstatus fidvidstatus; | 636 | union msr_fidvidstatus fidvidstatus; |
604 | int result; | 637 | int result; |
@@ -606,7 +639,7 @@ static int __init powernow_cpu_init (struct cpufreq_policy *policy) | |||
606 | if (policy->cpu != 0) | 639 | if (policy->cpu != 0) |
607 | return -ENODEV; | 640 | return -ENODEV; |
608 | 641 | ||
609 | rdmsrl (MSR_K7_FID_VID_STATUS, fidvidstatus.val); | 642 | rdmsrl(MSR_K7_FID_VID_STATUS, fidvidstatus.val); |
610 | 643 | ||
611 | recalibrate_cpu_khz(); | 644 | recalibrate_cpu_khz(); |
612 | 645 | ||
@@ -618,19 +651,21 @@ static int __init powernow_cpu_init (struct cpufreq_policy *policy) | |||
618 | dprintk("FSB: %3dMHz\n", fsb/1000); | 651 | dprintk("FSB: %3dMHz\n", fsb/1000); |
619 | 652 | ||
620 | if (dmi_check_system(powernow_dmi_table) || acpi_force) { | 653 | if (dmi_check_system(powernow_dmi_table) || acpi_force) { |
621 | printk (KERN_INFO PFX "PSB/PST known to be broken. Trying ACPI instead\n"); | 654 | printk(KERN_INFO PFX "PSB/PST known to be broken. " |
655 | "Trying ACPI instead\n"); | ||
622 | result = powernow_acpi_init(); | 656 | result = powernow_acpi_init(); |
623 | } else { | 657 | } else { |
624 | result = powernow_decode_bios(fidvidstatus.bits.MFID, fidvidstatus.bits.SVID); | 658 | result = powernow_decode_bios(fidvidstatus.bits.MFID, |
659 | fidvidstatus.bits.SVID); | ||
625 | if (result) { | 660 | if (result) { |
626 | printk (KERN_INFO PFX "Trying ACPI perflib\n"); | 661 | printk(KERN_INFO PFX "Trying ACPI perflib\n"); |
627 | maximum_speed = 0; | 662 | maximum_speed = 0; |
628 | minimum_speed = -1; | 663 | minimum_speed = -1; |
629 | latency = 0; | 664 | latency = 0; |
630 | result = powernow_acpi_init(); | 665 | result = powernow_acpi_init(); |
631 | if (result) { | 666 | if (result) { |
632 | printk (KERN_INFO PFX "ACPI and legacy methods failed\n"); | 667 | printk(KERN_INFO PFX |
633 | printk (KERN_INFO PFX "See http://www.codemonkey.org.uk/projects/cpufreq/powernow-k7.html\n"); | 668 | "ACPI and legacy methods failed\n"); |
634 | } | 669 | } |
635 | } else { | 670 | } else { |
636 | /* SGTC use the bus clock as timer */ | 671 | /* SGTC use the bus clock as timer */ |
@@ -642,10 +677,11 @@ static int __init powernow_cpu_init (struct cpufreq_policy *policy) | |||
642 | if (result) | 677 | if (result) |
643 | return result; | 678 | return result; |
644 | 679 | ||
645 | printk (KERN_INFO PFX "Minimum speed %d MHz. Maximum speed %d MHz.\n", | 680 | printk(KERN_INFO PFX "Minimum speed %d MHz. Maximum speed %d MHz.\n", |
646 | minimum_speed/1000, maximum_speed/1000); | 681 | minimum_speed/1000, maximum_speed/1000); |
647 | 682 | ||
648 | policy->cpuinfo.transition_latency = cpufreq_scale(2000000UL, fsb, latency); | 683 | policy->cpuinfo.transition_latency = |
684 | cpufreq_scale(2000000UL, fsb, latency); | ||
649 | 685 | ||
650 | policy->cur = powernow_get(0); | 686 | policy->cur = powernow_get(0); |
651 | 687 | ||
@@ -654,7 +690,8 @@ static int __init powernow_cpu_init (struct cpufreq_policy *policy) | |||
654 | return cpufreq_frequency_table_cpuinfo(policy, powernow_table); | 690 | return cpufreq_frequency_table_cpuinfo(policy, powernow_table); |
655 | } | 691 | } |
656 | 692 | ||
657 | static int powernow_cpu_exit (struct cpufreq_policy *policy) { | 693 | static int powernow_cpu_exit(struct cpufreq_policy *policy) |
694 | { | ||
658 | cpufreq_frequency_table_put_attr(policy->cpu); | 695 | cpufreq_frequency_table_put_attr(policy->cpu); |
659 | 696 | ||
660 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI | 697 | #ifdef CONFIG_X86_POWERNOW_K7_ACPI |
@@ -669,7 +706,7 @@ static int powernow_cpu_exit (struct cpufreq_policy *policy) { | |||
669 | return 0; | 706 | return 0; |
670 | } | 707 | } |
671 | 708 | ||
672 | static struct freq_attr* powernow_table_attr[] = { | 709 | static struct freq_attr *powernow_table_attr[] = { |
673 | &cpufreq_freq_attr_scaling_available_freqs, | 710 | &cpufreq_freq_attr_scaling_available_freqs, |
674 | NULL, | 711 | NULL, |
675 | }; | 712 | }; |
@@ -685,15 +722,15 @@ static struct cpufreq_driver powernow_driver = { | |||
685 | .attr = powernow_table_attr, | 722 | .attr = powernow_table_attr, |
686 | }; | 723 | }; |
687 | 724 | ||
688 | static int __init powernow_init (void) | 725 | static int __init powernow_init(void) |
689 | { | 726 | { |
690 | if (check_powernow()==0) | 727 | if (check_powernow() == 0) |
691 | return -ENODEV; | 728 | return -ENODEV; |
692 | return cpufreq_register_driver(&powernow_driver); | 729 | return cpufreq_register_driver(&powernow_driver); |
693 | } | 730 | } |
694 | 731 | ||
695 | 732 | ||
696 | static void __exit powernow_exit (void) | 733 | static void __exit powernow_exit(void) |
697 | { | 734 | { |
698 | cpufreq_unregister_driver(&powernow_driver); | 735 | cpufreq_unregister_driver(&powernow_driver); |
699 | } | 736 | } |
@@ -701,9 +738,9 @@ static void __exit powernow_exit (void) | |||
701 | module_param(acpi_force, int, 0444); | 738 | module_param(acpi_force, int, 0444); |
702 | MODULE_PARM_DESC(acpi_force, "Force ACPI to be used."); | 739 | MODULE_PARM_DESC(acpi_force, "Force ACPI to be used."); |
703 | 740 | ||
704 | MODULE_AUTHOR ("Dave Jones <davej@redhat.com>"); | 741 | MODULE_AUTHOR("Dave Jones <davej@redhat.com>"); |
705 | MODULE_DESCRIPTION ("Powernow driver for AMD K7 processors."); | 742 | MODULE_DESCRIPTION("Powernow driver for AMD K7 processors."); |
706 | MODULE_LICENSE ("GPL"); | 743 | MODULE_LICENSE("GPL"); |
707 | 744 | ||
708 | late_initcall(powernow_init); | 745 | late_initcall(powernow_init); |
709 | module_exit(powernow_exit); | 746 | module_exit(powernow_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/powernow-k8.c b/arch/x86/kernel/cpu/cpufreq/powernow-k8.c index 6428aa17b40e..a15ac94e0b9b 100644 --- a/arch/x86/kernel/cpu/cpufreq/powernow-k8.c +++ b/arch/x86/kernel/cpu/cpufreq/powernow-k8.c | |||
@@ -33,16 +33,14 @@ | |||
33 | #include <linux/string.h> | 33 | #include <linux/string.h> |
34 | #include <linux/cpumask.h> | 34 | #include <linux/cpumask.h> |
35 | #include <linux/sched.h> /* for current / set_cpus_allowed() */ | 35 | #include <linux/sched.h> /* for current / set_cpus_allowed() */ |
36 | #include <linux/io.h> | ||
37 | #include <linux/delay.h> | ||
36 | 38 | ||
37 | #include <asm/msr.h> | 39 | #include <asm/msr.h> |
38 | #include <asm/io.h> | ||
39 | #include <asm/delay.h> | ||
40 | 40 | ||
41 | #ifdef CONFIG_X86_POWERNOW_K8_ACPI | ||
42 | #include <linux/acpi.h> | 41 | #include <linux/acpi.h> |
43 | #include <linux/mutex.h> | 42 | #include <linux/mutex.h> |
44 | #include <acpi/processor.h> | 43 | #include <acpi/processor.h> |
45 | #endif | ||
46 | 44 | ||
47 | #define PFX "powernow-k8: " | 45 | #define PFX "powernow-k8: " |
48 | #define VERSION "version 2.20.00" | 46 | #define VERSION "version 2.20.00" |
@@ -71,7 +69,8 @@ static u32 find_khz_freq_from_fid(u32 fid) | |||
71 | return 1000 * find_freq_from_fid(fid); | 69 | return 1000 * find_freq_from_fid(fid); |
72 | } | 70 | } |
73 | 71 | ||
74 | static u32 find_khz_freq_from_pstate(struct cpufreq_frequency_table *data, u32 pstate) | 72 | static u32 find_khz_freq_from_pstate(struct cpufreq_frequency_table *data, |
73 | u32 pstate) | ||
75 | { | 74 | { |
76 | return data[pstate].frequency; | 75 | return data[pstate].frequency; |
77 | } | 76 | } |
@@ -186,7 +185,9 @@ static int write_new_fid(struct powernow_k8_data *data, u32 fid) | |||
186 | return 1; | 185 | return 1; |
187 | } | 186 | } |
188 | 187 | ||
189 | lo = fid | (data->currvid << MSR_C_LO_VID_SHIFT) | MSR_C_LO_INIT_FID_VID; | 188 | lo = fid; |
189 | lo |= (data->currvid << MSR_C_LO_VID_SHIFT); | ||
190 | lo |= MSR_C_LO_INIT_FID_VID; | ||
190 | 191 | ||
191 | dprintk("writing fid 0x%x, lo 0x%x, hi 0x%x\n", | 192 | dprintk("writing fid 0x%x, lo 0x%x, hi 0x%x\n", |
192 | fid, lo, data->plllock * PLL_LOCK_CONVERSION); | 193 | fid, lo, data->plllock * PLL_LOCK_CONVERSION); |
@@ -194,7 +195,9 @@ static int write_new_fid(struct powernow_k8_data *data, u32 fid) | |||
194 | do { | 195 | do { |
195 | wrmsr(MSR_FIDVID_CTL, lo, data->plllock * PLL_LOCK_CONVERSION); | 196 | wrmsr(MSR_FIDVID_CTL, lo, data->plllock * PLL_LOCK_CONVERSION); |
196 | if (i++ > 100) { | 197 | if (i++ > 100) { |
197 | printk(KERN_ERR PFX "Hardware error - pending bit very stuck - no further pstate changes possible\n"); | 198 | printk(KERN_ERR PFX |
199 | "Hardware error - pending bit very stuck - " | ||
200 | "no further pstate changes possible\n"); | ||
198 | return 1; | 201 | return 1; |
199 | } | 202 | } |
200 | } while (query_current_values_with_pending_wait(data)); | 203 | } while (query_current_values_with_pending_wait(data)); |
@@ -202,14 +205,16 @@ static int write_new_fid(struct powernow_k8_data *data, u32 fid) | |||
202 | count_off_irt(data); | 205 | count_off_irt(data); |
203 | 206 | ||
204 | if (savevid != data->currvid) { | 207 | if (savevid != data->currvid) { |
205 | printk(KERN_ERR PFX "vid change on fid trans, old 0x%x, new 0x%x\n", | 208 | printk(KERN_ERR PFX |
206 | savevid, data->currvid); | 209 | "vid change on fid trans, old 0x%x, new 0x%x\n", |
210 | savevid, data->currvid); | ||
207 | return 1; | 211 | return 1; |
208 | } | 212 | } |
209 | 213 | ||
210 | if (fid != data->currfid) { | 214 | if (fid != data->currfid) { |
211 | printk(KERN_ERR PFX "fid trans failed, fid 0x%x, curr 0x%x\n", fid, | 215 | printk(KERN_ERR PFX |
212 | data->currfid); | 216 | "fid trans failed, fid 0x%x, curr 0x%x\n", fid, |
217 | data->currfid); | ||
213 | return 1; | 218 | return 1; |
214 | } | 219 | } |
215 | 220 | ||
@@ -228,7 +233,9 @@ static int write_new_vid(struct powernow_k8_data *data, u32 vid) | |||
228 | return 1; | 233 | return 1; |
229 | } | 234 | } |
230 | 235 | ||
231 | lo = data->currfid | (vid << MSR_C_LO_VID_SHIFT) | MSR_C_LO_INIT_FID_VID; | 236 | lo = data->currfid; |
237 | lo |= (vid << MSR_C_LO_VID_SHIFT); | ||
238 | lo |= MSR_C_LO_INIT_FID_VID; | ||
232 | 239 | ||
233 | dprintk("writing vid 0x%x, lo 0x%x, hi 0x%x\n", | 240 | dprintk("writing vid 0x%x, lo 0x%x, hi 0x%x\n", |
234 | vid, lo, STOP_GRANT_5NS); | 241 | vid, lo, STOP_GRANT_5NS); |
@@ -236,20 +243,24 @@ static int write_new_vid(struct powernow_k8_data *data, u32 vid) | |||
236 | do { | 243 | do { |
237 | wrmsr(MSR_FIDVID_CTL, lo, STOP_GRANT_5NS); | 244 | wrmsr(MSR_FIDVID_CTL, lo, STOP_GRANT_5NS); |
238 | if (i++ > 100) { | 245 | if (i++ > 100) { |
239 | printk(KERN_ERR PFX "internal error - pending bit very stuck - no further pstate changes possible\n"); | 246 | printk(KERN_ERR PFX "internal error - pending bit " |
247 | "very stuck - no further pstate " | ||
248 | "changes possible\n"); | ||
240 | return 1; | 249 | return 1; |
241 | } | 250 | } |
242 | } while (query_current_values_with_pending_wait(data)); | 251 | } while (query_current_values_with_pending_wait(data)); |
243 | 252 | ||
244 | if (savefid != data->currfid) { | 253 | if (savefid != data->currfid) { |
245 | printk(KERN_ERR PFX "fid changed on vid trans, old 0x%x new 0x%x\n", | 254 | printk(KERN_ERR PFX "fid changed on vid trans, old " |
255 | "0x%x new 0x%x\n", | ||
246 | savefid, data->currfid); | 256 | savefid, data->currfid); |
247 | return 1; | 257 | return 1; |
248 | } | 258 | } |
249 | 259 | ||
250 | if (vid != data->currvid) { | 260 | if (vid != data->currvid) { |
251 | printk(KERN_ERR PFX "vid trans failed, vid 0x%x, curr 0x%x\n", vid, | 261 | printk(KERN_ERR PFX "vid trans failed, vid 0x%x, " |
252 | data->currvid); | 262 | "curr 0x%x\n", |
263 | vid, data->currvid); | ||
253 | return 1; | 264 | return 1; |
254 | } | 265 | } |
255 | 266 | ||
@@ -261,7 +272,8 @@ static int write_new_vid(struct powernow_k8_data *data, u32 vid) | |||
261 | * Decreasing vid codes represent increasing voltages: | 272 | * Decreasing vid codes represent increasing voltages: |
262 | * vid of 0 is 1.550V, vid of 0x1e is 0.800V, vid of VID_OFF is off. | 273 | * vid of 0 is 1.550V, vid of 0x1e is 0.800V, vid of VID_OFF is off. |
263 | */ | 274 | */ |
264 | static int decrease_vid_code_by_step(struct powernow_k8_data *data, u32 reqvid, u32 step) | 275 | static int decrease_vid_code_by_step(struct powernow_k8_data *data, |
276 | u32 reqvid, u32 step) | ||
265 | { | 277 | { |
266 | if ((data->currvid - reqvid) > step) | 278 | if ((data->currvid - reqvid) > step) |
267 | reqvid = data->currvid - step; | 279 | reqvid = data->currvid - step; |
@@ -283,7 +295,8 @@ static int transition_pstate(struct powernow_k8_data *data, u32 pstate) | |||
283 | } | 295 | } |
284 | 296 | ||
285 | /* Change Opteron/Athlon64 fid and vid, by the 3 phases. */ | 297 | /* Change Opteron/Athlon64 fid and vid, by the 3 phases. */ |
286 | static int transition_fid_vid(struct powernow_k8_data *data, u32 reqfid, u32 reqvid) | 298 | static int transition_fid_vid(struct powernow_k8_data *data, |
299 | u32 reqfid, u32 reqvid) | ||
287 | { | 300 | { |
288 | if (core_voltage_pre_transition(data, reqvid)) | 301 | if (core_voltage_pre_transition(data, reqvid)) |
289 | return 1; | 302 | return 1; |
@@ -298,7 +311,8 @@ static int transition_fid_vid(struct powernow_k8_data *data, u32 reqfid, u32 req | |||
298 | return 1; | 311 | return 1; |
299 | 312 | ||
300 | if ((reqfid != data->currfid) || (reqvid != data->currvid)) { | 313 | if ((reqfid != data->currfid) || (reqvid != data->currvid)) { |
301 | printk(KERN_ERR PFX "failed (cpu%d): req 0x%x 0x%x, curr 0x%x 0x%x\n", | 314 | printk(KERN_ERR PFX "failed (cpu%d): req 0x%x 0x%x, " |
315 | "curr 0x%x 0x%x\n", | ||
302 | smp_processor_id(), | 316 | smp_processor_id(), |
303 | reqfid, reqvid, data->currfid, data->currvid); | 317 | reqfid, reqvid, data->currfid, data->currvid); |
304 | return 1; | 318 | return 1; |
@@ -311,13 +325,15 @@ static int transition_fid_vid(struct powernow_k8_data *data, u32 reqfid, u32 req | |||
311 | } | 325 | } |
312 | 326 | ||
313 | /* Phase 1 - core voltage transition ... setup voltage */ | 327 | /* Phase 1 - core voltage transition ... setup voltage */ |
314 | static int core_voltage_pre_transition(struct powernow_k8_data *data, u32 reqvid) | 328 | static int core_voltage_pre_transition(struct powernow_k8_data *data, |
329 | u32 reqvid) | ||
315 | { | 330 | { |
316 | u32 rvosteps = data->rvo; | 331 | u32 rvosteps = data->rvo; |
317 | u32 savefid = data->currfid; | 332 | u32 savefid = data->currfid; |
318 | u32 maxvid, lo; | 333 | u32 maxvid, lo; |
319 | 334 | ||
320 | dprintk("ph1 (cpu%d): start, currfid 0x%x, currvid 0x%x, reqvid 0x%x, rvo 0x%x\n", | 335 | dprintk("ph1 (cpu%d): start, currfid 0x%x, currvid 0x%x, " |
336 | "reqvid 0x%x, rvo 0x%x\n", | ||
321 | smp_processor_id(), | 337 | smp_processor_id(), |
322 | data->currfid, data->currvid, reqvid, data->rvo); | 338 | data->currfid, data->currvid, reqvid, data->rvo); |
323 | 339 | ||
@@ -340,7 +356,7 @@ static int core_voltage_pre_transition(struct powernow_k8_data *data, u32 reqvid | |||
340 | } else { | 356 | } else { |
341 | dprintk("ph1: changing vid for rvo, req 0x%x\n", | 357 | dprintk("ph1: changing vid for rvo, req 0x%x\n", |
342 | data->currvid - 1); | 358 | data->currvid - 1); |
343 | if (decrease_vid_code_by_step(data, data->currvid - 1, 1)) | 359 | if (decrease_vid_code_by_step(data, data->currvid-1, 1)) |
344 | return 1; | 360 | return 1; |
345 | rvosteps--; | 361 | rvosteps--; |
346 | } | 362 | } |
@@ -350,7 +366,8 @@ static int core_voltage_pre_transition(struct powernow_k8_data *data, u32 reqvid | |||
350 | return 1; | 366 | return 1; |
351 | 367 | ||
352 | if (savefid != data->currfid) { | 368 | if (savefid != data->currfid) { |
353 | printk(KERN_ERR PFX "ph1 err, currfid changed 0x%x\n", data->currfid); | 369 | printk(KERN_ERR PFX "ph1 err, currfid changed 0x%x\n", |
370 | data->currfid); | ||
354 | return 1; | 371 | return 1; |
355 | } | 372 | } |
356 | 373 | ||
@@ -363,20 +380,24 @@ static int core_voltage_pre_transition(struct powernow_k8_data *data, u32 reqvid | |||
363 | /* Phase 2 - core frequency transition */ | 380 | /* Phase 2 - core frequency transition */ |
364 | static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid) | 381 | static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid) |
365 | { | 382 | { |
366 | u32 vcoreqfid, vcocurrfid, vcofiddiff, fid_interval, savevid = data->currvid; | 383 | u32 vcoreqfid, vcocurrfid, vcofiddiff; |
384 | u32 fid_interval, savevid = data->currvid; | ||
367 | 385 | ||
368 | if ((reqfid < HI_FID_TABLE_BOTTOM) && (data->currfid < HI_FID_TABLE_BOTTOM)) { | 386 | if ((reqfid < HI_FID_TABLE_BOTTOM) && |
369 | printk(KERN_ERR PFX "ph2: illegal lo-lo transition 0x%x 0x%x\n", | 387 | (data->currfid < HI_FID_TABLE_BOTTOM)) { |
370 | reqfid, data->currfid); | 388 | printk(KERN_ERR PFX "ph2: illegal lo-lo transition " |
389 | "0x%x 0x%x\n", reqfid, data->currfid); | ||
371 | return 1; | 390 | return 1; |
372 | } | 391 | } |
373 | 392 | ||
374 | if (data->currfid == reqfid) { | 393 | if (data->currfid == reqfid) { |
375 | printk(KERN_ERR PFX "ph2 null fid transition 0x%x\n", data->currfid); | 394 | printk(KERN_ERR PFX "ph2 null fid transition 0x%x\n", |
395 | data->currfid); | ||
376 | return 0; | 396 | return 0; |
377 | } | 397 | } |
378 | 398 | ||
379 | dprintk("ph2 (cpu%d): starting, currfid 0x%x, currvid 0x%x, reqfid 0x%x\n", | 399 | dprintk("ph2 (cpu%d): starting, currfid 0x%x, currvid 0x%x, " |
400 | "reqfid 0x%x\n", | ||
380 | smp_processor_id(), | 401 | smp_processor_id(), |
381 | data->currfid, data->currvid, reqfid); | 402 | data->currfid, data->currvid, reqfid); |
382 | 403 | ||
@@ -390,14 +411,14 @@ static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid) | |||
390 | 411 | ||
391 | if (reqfid > data->currfid) { | 412 | if (reqfid > data->currfid) { |
392 | if (data->currfid > LO_FID_TABLE_TOP) { | 413 | if (data->currfid > LO_FID_TABLE_TOP) { |
393 | if (write_new_fid(data, data->currfid + fid_interval)) { | 414 | if (write_new_fid(data, |
415 | data->currfid + fid_interval)) | ||
394 | return 1; | 416 | return 1; |
395 | } | ||
396 | } else { | 417 | } else { |
397 | if (write_new_fid | 418 | if (write_new_fid |
398 | (data, 2 + convert_fid_to_vco_fid(data->currfid))) { | 419 | (data, |
420 | 2 + convert_fid_to_vco_fid(data->currfid))) | ||
399 | return 1; | 421 | return 1; |
400 | } | ||
401 | } | 422 | } |
402 | } else { | 423 | } else { |
403 | if (write_new_fid(data, data->currfid - fid_interval)) | 424 | if (write_new_fid(data, data->currfid - fid_interval)) |
@@ -417,7 +438,8 @@ static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid) | |||
417 | 438 | ||
418 | if (data->currfid != reqfid) { | 439 | if (data->currfid != reqfid) { |
419 | printk(KERN_ERR PFX | 440 | printk(KERN_ERR PFX |
420 | "ph2: mismatch, failed fid transition, curr 0x%x, req 0x%x\n", | 441 | "ph2: mismatch, failed fid transition, " |
442 | "curr 0x%x, req 0x%x\n", | ||
421 | data->currfid, reqfid); | 443 | data->currfid, reqfid); |
422 | return 1; | 444 | return 1; |
423 | } | 445 | } |
@@ -435,7 +457,8 @@ static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid) | |||
435 | } | 457 | } |
436 | 458 | ||
437 | /* Phase 3 - core voltage transition flow ... jump to the final vid. */ | 459 | /* Phase 3 - core voltage transition flow ... jump to the final vid. */ |
438 | static int core_voltage_post_transition(struct powernow_k8_data *data, u32 reqvid) | 460 | static int core_voltage_post_transition(struct powernow_k8_data *data, |
461 | u32 reqvid) | ||
439 | { | 462 | { |
440 | u32 savefid = data->currfid; | 463 | u32 savefid = data->currfid; |
441 | u32 savereqvid = reqvid; | 464 | u32 savereqvid = reqvid; |
@@ -457,7 +480,8 @@ static int core_voltage_post_transition(struct powernow_k8_data *data, u32 reqvi | |||
457 | 480 | ||
458 | if (data->currvid != reqvid) { | 481 | if (data->currvid != reqvid) { |
459 | printk(KERN_ERR PFX | 482 | printk(KERN_ERR PFX |
460 | "ph3: failed vid transition\n, req 0x%x, curr 0x%x", | 483 | "ph3: failed vid transition\n, " |
484 | "req 0x%x, curr 0x%x", | ||
461 | reqvid, data->currvid); | 485 | reqvid, data->currvid); |
462 | return 1; | 486 | return 1; |
463 | } | 487 | } |
@@ -508,7 +532,8 @@ static int check_supported_cpu(unsigned int cpu) | |||
508 | if ((eax & CPUID_XFAM) == CPUID_XFAM_K8) { | 532 | if ((eax & CPUID_XFAM) == CPUID_XFAM_K8) { |
509 | if (((eax & CPUID_USE_XFAM_XMOD) != CPUID_USE_XFAM_XMOD) || | 533 | if (((eax & CPUID_USE_XFAM_XMOD) != CPUID_USE_XFAM_XMOD) || |
510 | ((eax & CPUID_XMOD) > CPUID_XMOD_REV_MASK)) { | 534 | ((eax & CPUID_XMOD) > CPUID_XMOD_REV_MASK)) { |
511 | printk(KERN_INFO PFX "Processor cpuid %x not supported\n", eax); | 535 | printk(KERN_INFO PFX |
536 | "Processor cpuid %x not supported\n", eax); | ||
512 | goto out; | 537 | goto out; |
513 | } | 538 | } |
514 | 539 | ||
@@ -520,8 +545,10 @@ static int check_supported_cpu(unsigned int cpu) | |||
520 | } | 545 | } |
521 | 546 | ||
522 | cpuid(CPUID_FREQ_VOLT_CAPABILITIES, &eax, &ebx, &ecx, &edx); | 547 | cpuid(CPUID_FREQ_VOLT_CAPABILITIES, &eax, &ebx, &ecx, &edx); |
523 | if ((edx & P_STATE_TRANSITION_CAPABLE) != P_STATE_TRANSITION_CAPABLE) { | 548 | if ((edx & P_STATE_TRANSITION_CAPABLE) |
524 | printk(KERN_INFO PFX "Power state transitions not supported\n"); | 549 | != P_STATE_TRANSITION_CAPABLE) { |
550 | printk(KERN_INFO PFX | ||
551 | "Power state transitions not supported\n"); | ||
525 | goto out; | 552 | goto out; |
526 | } | 553 | } |
527 | } else { /* must be a HW Pstate capable processor */ | 554 | } else { /* must be a HW Pstate capable processor */ |
@@ -539,7 +566,8 @@ out: | |||
539 | return rc; | 566 | return rc; |
540 | } | 567 | } |
541 | 568 | ||
542 | static int check_pst_table(struct powernow_k8_data *data, struct pst_s *pst, u8 maxvid) | 569 | static int check_pst_table(struct powernow_k8_data *data, struct pst_s *pst, |
570 | u8 maxvid) | ||
543 | { | 571 | { |
544 | unsigned int j; | 572 | unsigned int j; |
545 | u8 lastfid = 0xff; | 573 | u8 lastfid = 0xff; |
@@ -550,12 +578,14 @@ static int check_pst_table(struct powernow_k8_data *data, struct pst_s *pst, u8 | |||
550 | j, pst[j].vid); | 578 | j, pst[j].vid); |
551 | return -EINVAL; | 579 | return -EINVAL; |
552 | } | 580 | } |
553 | if (pst[j].vid < data->rvo) { /* vid + rvo >= 0 */ | 581 | if (pst[j].vid < data->rvo) { |
582 | /* vid + rvo >= 0 */ | ||
554 | printk(KERN_ERR FW_BUG PFX "0 vid exceeded with pstate" | 583 | printk(KERN_ERR FW_BUG PFX "0 vid exceeded with pstate" |
555 | " %d\n", j); | 584 | " %d\n", j); |
556 | return -ENODEV; | 585 | return -ENODEV; |
557 | } | 586 | } |
558 | if (pst[j].vid < maxvid + data->rvo) { /* vid + rvo >= maxvid */ | 587 | if (pst[j].vid < maxvid + data->rvo) { |
588 | /* vid + rvo >= maxvid */ | ||
559 | printk(KERN_ERR FW_BUG PFX "maxvid exceeded with pstate" | 589 | printk(KERN_ERR FW_BUG PFX "maxvid exceeded with pstate" |
560 | " %d\n", j); | 590 | " %d\n", j); |
561 | return -ENODEV; | 591 | return -ENODEV; |
@@ -579,23 +609,31 @@ static int check_pst_table(struct powernow_k8_data *data, struct pst_s *pst, u8 | |||
579 | return -EINVAL; | 609 | return -EINVAL; |
580 | } | 610 | } |
581 | if (lastfid > LO_FID_TABLE_TOP) | 611 | if (lastfid > LO_FID_TABLE_TOP) |
582 | printk(KERN_INFO FW_BUG PFX "first fid not from lo freq table\n"); | 612 | printk(KERN_INFO FW_BUG PFX |
613 | "first fid not from lo freq table\n"); | ||
583 | 614 | ||
584 | return 0; | 615 | return 0; |
585 | } | 616 | } |
586 | 617 | ||
618 | static void invalidate_entry(struct powernow_k8_data *data, unsigned int entry) | ||
619 | { | ||
620 | data->powernow_table[entry].frequency = CPUFREQ_ENTRY_INVALID; | ||
621 | } | ||
622 | |||
587 | static void print_basics(struct powernow_k8_data *data) | 623 | static void print_basics(struct powernow_k8_data *data) |
588 | { | 624 | { |
589 | int j; | 625 | int j; |
590 | for (j = 0; j < data->numps; j++) { | 626 | for (j = 0; j < data->numps; j++) { |
591 | if (data->powernow_table[j].frequency != CPUFREQ_ENTRY_INVALID) { | 627 | if (data->powernow_table[j].frequency != |
628 | CPUFREQ_ENTRY_INVALID) { | ||
592 | if (cpu_family == CPU_HW_PSTATE) { | 629 | if (cpu_family == CPU_HW_PSTATE) { |
593 | printk(KERN_INFO PFX " %d : pstate %d (%d MHz)\n", | 630 | printk(KERN_INFO PFX |
594 | j, | 631 | " %d : pstate %d (%d MHz)\n", j, |
595 | data->powernow_table[j].index, | 632 | data->powernow_table[j].index, |
596 | data->powernow_table[j].frequency/1000); | 633 | data->powernow_table[j].frequency/1000); |
597 | } else { | 634 | } else { |
598 | printk(KERN_INFO PFX " %d : fid 0x%x (%d MHz), vid 0x%x\n", | 635 | printk(KERN_INFO PFX |
636 | " %d : fid 0x%x (%d MHz), vid 0x%x\n", | ||
599 | j, | 637 | j, |
600 | data->powernow_table[j].index & 0xff, | 638 | data->powernow_table[j].index & 0xff, |
601 | data->powernow_table[j].frequency/1000, | 639 | data->powernow_table[j].frequency/1000, |
@@ -604,20 +642,25 @@ static void print_basics(struct powernow_k8_data *data) | |||
604 | } | 642 | } |
605 | } | 643 | } |
606 | if (data->batps) | 644 | if (data->batps) |
607 | printk(KERN_INFO PFX "Only %d pstates on battery\n", data->batps); | 645 | printk(KERN_INFO PFX "Only %d pstates on battery\n", |
646 | data->batps); | ||
608 | } | 647 | } |
609 | 648 | ||
610 | static int fill_powernow_table(struct powernow_k8_data *data, struct pst_s *pst, u8 maxvid) | 649 | static int fill_powernow_table(struct powernow_k8_data *data, |
650 | struct pst_s *pst, u8 maxvid) | ||
611 | { | 651 | { |
612 | struct cpufreq_frequency_table *powernow_table; | 652 | struct cpufreq_frequency_table *powernow_table; |
613 | unsigned int j; | 653 | unsigned int j; |
614 | 654 | ||
615 | if (data->batps) { /* use ACPI support to get full speed on mains power */ | 655 | if (data->batps) { |
616 | printk(KERN_WARNING PFX "Only %d pstates usable (use ACPI driver for full range\n", data->batps); | 656 | /* use ACPI support to get full speed on mains power */ |
657 | printk(KERN_WARNING PFX | ||
658 | "Only %d pstates usable (use ACPI driver for full " | ||
659 | "range\n", data->batps); | ||
617 | data->numps = data->batps; | 660 | data->numps = data->batps; |
618 | } | 661 | } |
619 | 662 | ||
620 | for ( j=1; j<data->numps; j++ ) { | 663 | for (j = 1; j < data->numps; j++) { |
621 | if (pst[j-1].fid >= pst[j].fid) { | 664 | if (pst[j-1].fid >= pst[j].fid) { |
622 | printk(KERN_ERR PFX "PST out of sequence\n"); | 665 | printk(KERN_ERR PFX "PST out of sequence\n"); |
623 | return -EINVAL; | 666 | return -EINVAL; |
@@ -640,9 +683,11 @@ static int fill_powernow_table(struct powernow_k8_data *data, struct pst_s *pst, | |||
640 | } | 683 | } |
641 | 684 | ||
642 | for (j = 0; j < data->numps; j++) { | 685 | for (j = 0; j < data->numps; j++) { |
686 | int freq; | ||
643 | powernow_table[j].index = pst[j].fid; /* lower 8 bits */ | 687 | powernow_table[j].index = pst[j].fid; /* lower 8 bits */ |
644 | powernow_table[j].index |= (pst[j].vid << 8); /* upper 8 bits */ | 688 | powernow_table[j].index |= (pst[j].vid << 8); /* upper 8 bits */ |
645 | powernow_table[j].frequency = find_khz_freq_from_fid(pst[j].fid); | 689 | freq = find_khz_freq_from_fid(pst[j].fid); |
690 | powernow_table[j].frequency = freq; | ||
646 | } | 691 | } |
647 | powernow_table[data->numps].frequency = CPUFREQ_TABLE_END; | 692 | powernow_table[data->numps].frequency = CPUFREQ_TABLE_END; |
648 | powernow_table[data->numps].index = 0; | 693 | powernow_table[data->numps].index = 0; |
@@ -658,7 +703,8 @@ static int fill_powernow_table(struct powernow_k8_data *data, struct pst_s *pst, | |||
658 | print_basics(data); | 703 | print_basics(data); |
659 | 704 | ||
660 | for (j = 0; j < data->numps; j++) | 705 | for (j = 0; j < data->numps; j++) |
661 | if ((pst[j].fid==data->currfid) && (pst[j].vid==data->currvid)) | 706 | if ((pst[j].fid == data->currfid) && |
707 | (pst[j].vid == data->currvid)) | ||
662 | return 0; | 708 | return 0; |
663 | 709 | ||
664 | dprintk("currfid/vid do not match PST, ignoring\n"); | 710 | dprintk("currfid/vid do not match PST, ignoring\n"); |
@@ -698,7 +744,8 @@ static int find_psb_table(struct powernow_k8_data *data) | |||
698 | } | 744 | } |
699 | 745 | ||
700 | data->vstable = psb->vstable; | 746 | data->vstable = psb->vstable; |
701 | dprintk("voltage stabilization time: %d(*20us)\n", data->vstable); | 747 | dprintk("voltage stabilization time: %d(*20us)\n", |
748 | data->vstable); | ||
702 | 749 | ||
703 | dprintk("flags2: 0x%x\n", psb->flags2); | 750 | dprintk("flags2: 0x%x\n", psb->flags2); |
704 | data->rvo = psb->flags2 & 3; | 751 | data->rvo = psb->flags2 & 3; |
@@ -713,11 +760,12 @@ static int find_psb_table(struct powernow_k8_data *data) | |||
713 | 760 | ||
714 | dprintk("numpst: 0x%x\n", psb->num_tables); | 761 | dprintk("numpst: 0x%x\n", psb->num_tables); |
715 | cpst = psb->num_tables; | 762 | cpst = psb->num_tables; |
716 | if ((psb->cpuid == 0x00000fc0) || (psb->cpuid == 0x00000fe0) ){ | 763 | if ((psb->cpuid == 0x00000fc0) || |
764 | (psb->cpuid == 0x00000fe0)) { | ||
717 | thiscpuid = cpuid_eax(CPUID_PROCESSOR_SIGNATURE); | 765 | thiscpuid = cpuid_eax(CPUID_PROCESSOR_SIGNATURE); |
718 | if ((thiscpuid == 0x00000fc0) || (thiscpuid == 0x00000fe0) ) { | 766 | if ((thiscpuid == 0x00000fc0) || |
767 | (thiscpuid == 0x00000fe0)) | ||
719 | cpst = 1; | 768 | cpst = 1; |
720 | } | ||
721 | } | 769 | } |
722 | if (cpst != 1) { | 770 | if (cpst != 1) { |
723 | printk(KERN_ERR FW_BUG PFX "numpst must be 1\n"); | 771 | printk(KERN_ERR FW_BUG PFX "numpst must be 1\n"); |
@@ -732,7 +780,8 @@ static int find_psb_table(struct powernow_k8_data *data) | |||
732 | 780 | ||
733 | data->numps = psb->numps; | 781 | data->numps = psb->numps; |
734 | dprintk("numpstates: 0x%x\n", data->numps); | 782 | dprintk("numpstates: 0x%x\n", data->numps); |
735 | return fill_powernow_table(data, (struct pst_s *)(psb+1), maxvid); | 783 | return fill_powernow_table(data, |
784 | (struct pst_s *)(psb+1), maxvid); | ||
736 | } | 785 | } |
737 | /* | 786 | /* |
738 | * If you see this message, complain to BIOS manufacturer. If | 787 | * If you see this message, complain to BIOS manufacturer. If |
@@ -745,28 +794,31 @@ static int find_psb_table(struct powernow_k8_data *data) | |||
745 | * BIOS and Kernel Developer's Guide, which is available on | 794 | * BIOS and Kernel Developer's Guide, which is available on |
746 | * www.amd.com | 795 | * www.amd.com |
747 | */ | 796 | */ |
748 | printk(KERN_ERR PFX "BIOS error - no PSB or ACPI _PSS objects\n"); | 797 | printk(KERN_ERR FW_BUG PFX "No PSB or ACPI _PSS objects\n"); |
749 | return -ENODEV; | 798 | return -ENODEV; |
750 | } | 799 | } |
751 | 800 | ||
752 | #ifdef CONFIG_X86_POWERNOW_K8_ACPI | 801 | static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, |
753 | static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index) | 802 | unsigned int index) |
754 | { | 803 | { |
804 | acpi_integer control; | ||
805 | |||
755 | if (!data->acpi_data.state_count || (cpu_family == CPU_HW_PSTATE)) | 806 | if (!data->acpi_data.state_count || (cpu_family == CPU_HW_PSTATE)) |
756 | return; | 807 | return; |
757 | 808 | ||
758 | data->irt = (data->acpi_data.states[index].control >> IRT_SHIFT) & IRT_MASK; | 809 | control = data->acpi_data.states[index].control; data->irt = (control |
759 | data->rvo = (data->acpi_data.states[index].control >> RVO_SHIFT) & RVO_MASK; | 810 | >> IRT_SHIFT) & IRT_MASK; data->rvo = (control >> |
760 | data->exttype = (data->acpi_data.states[index].control >> EXT_TYPE_SHIFT) & EXT_TYPE_MASK; | 811 | RVO_SHIFT) & RVO_MASK; data->exttype = (control |
761 | data->plllock = (data->acpi_data.states[index].control >> PLL_L_SHIFT) & PLL_L_MASK; | 812 | >> EXT_TYPE_SHIFT) & EXT_TYPE_MASK; |
762 | data->vidmvs = 1 << ((data->acpi_data.states[index].control >> MVS_SHIFT) & MVS_MASK); | 813 | data->plllock = (control >> PLL_L_SHIFT) & PLL_L_MASK; data->vidmvs = 1 |
763 | data->vstable = (data->acpi_data.states[index].control >> VST_SHIFT) & VST_MASK; | 814 | << ((control >> MVS_SHIFT) & MVS_MASK); data->vstable = |
764 | } | 815 | (control >> VST_SHIFT) & VST_MASK; } |
765 | 816 | ||
766 | static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) | 817 | static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) |
767 | { | 818 | { |
768 | struct cpufreq_frequency_table *powernow_table; | 819 | struct cpufreq_frequency_table *powernow_table; |
769 | int ret_val = -ENODEV; | 820 | int ret_val = -ENODEV; |
821 | acpi_integer space_id; | ||
770 | 822 | ||
771 | if (acpi_processor_register_performance(&data->acpi_data, data->cpu)) { | 823 | if (acpi_processor_register_performance(&data->acpi_data, data->cpu)) { |
772 | dprintk("register performance failed: bad ACPI data\n"); | 824 | dprintk("register performance failed: bad ACPI data\n"); |
@@ -779,11 +831,12 @@ static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) | |||
779 | goto err_out; | 831 | goto err_out; |
780 | } | 832 | } |
781 | 833 | ||
782 | if ((data->acpi_data.control_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) || | 834 | space_id = data->acpi_data.control_register.space_id; |
783 | (data->acpi_data.status_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE)) { | 835 | if ((space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) || |
836 | (space_id != ACPI_ADR_SPACE_FIXED_HARDWARE)) { | ||
784 | dprintk("Invalid control/status registers (%x - %x)\n", | 837 | dprintk("Invalid control/status registers (%x - %x)\n", |
785 | data->acpi_data.control_register.space_id, | 838 | data->acpi_data.control_register.space_id, |
786 | data->acpi_data.status_register.space_id); | 839 | space_id); |
787 | goto err_out; | 840 | goto err_out; |
788 | } | 841 | } |
789 | 842 | ||
@@ -802,7 +855,8 @@ static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) | |||
802 | if (ret_val) | 855 | if (ret_val) |
803 | goto err_out_mem; | 856 | goto err_out_mem; |
804 | 857 | ||
805 | powernow_table[data->acpi_data.state_count].frequency = CPUFREQ_TABLE_END; | 858 | powernow_table[data->acpi_data.state_count].frequency = |
859 | CPUFREQ_TABLE_END; | ||
806 | powernow_table[data->acpi_data.state_count].index = 0; | 860 | powernow_table[data->acpi_data.state_count].index = 0; |
807 | data->powernow_table = powernow_table; | 861 | data->powernow_table = powernow_table; |
808 | 862 | ||
@@ -830,13 +884,15 @@ err_out_mem: | |||
830 | err_out: | 884 | err_out: |
831 | acpi_processor_unregister_performance(&data->acpi_data, data->cpu); | 885 | acpi_processor_unregister_performance(&data->acpi_data, data->cpu); |
832 | 886 | ||
833 | /* data->acpi_data.state_count informs us at ->exit() whether ACPI was used */ | 887 | /* data->acpi_data.state_count informs us at ->exit() |
888 | * whether ACPI was used */ | ||
834 | data->acpi_data.state_count = 0; | 889 | data->acpi_data.state_count = 0; |
835 | 890 | ||
836 | return ret_val; | 891 | return ret_val; |
837 | } | 892 | } |
838 | 893 | ||
839 | static int fill_powernow_table_pstate(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table) | 894 | static int fill_powernow_table_pstate(struct powernow_k8_data *data, |
895 | struct cpufreq_frequency_table *powernow_table) | ||
840 | { | 896 | { |
841 | int i; | 897 | int i; |
842 | u32 hi = 0, lo = 0; | 898 | u32 hi = 0, lo = 0; |
@@ -848,84 +904,101 @@ static int fill_powernow_table_pstate(struct powernow_k8_data *data, struct cpuf | |||
848 | 904 | ||
849 | index = data->acpi_data.states[i].control & HW_PSTATE_MASK; | 905 | index = data->acpi_data.states[i].control & HW_PSTATE_MASK; |
850 | if (index > data->max_hw_pstate) { | 906 | if (index > data->max_hw_pstate) { |
851 | printk(KERN_ERR PFX "invalid pstate %d - bad value %d.\n", i, index); | 907 | printk(KERN_ERR PFX "invalid pstate %d - " |
852 | printk(KERN_ERR PFX "Please report to BIOS manufacturer\n"); | 908 | "bad value %d.\n", i, index); |
853 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 909 | printk(KERN_ERR PFX "Please report to BIOS " |
910 | "manufacturer\n"); | ||
911 | invalidate_entry(data, i); | ||
854 | continue; | 912 | continue; |
855 | } | 913 | } |
856 | rdmsr(MSR_PSTATE_DEF_BASE + index, lo, hi); | 914 | rdmsr(MSR_PSTATE_DEF_BASE + index, lo, hi); |
857 | if (!(hi & HW_PSTATE_VALID_MASK)) { | 915 | if (!(hi & HW_PSTATE_VALID_MASK)) { |
858 | dprintk("invalid pstate %d, ignoring\n", index); | 916 | dprintk("invalid pstate %d, ignoring\n", index); |
859 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 917 | invalidate_entry(data, i); |
860 | continue; | 918 | continue; |
861 | } | 919 | } |
862 | 920 | ||
863 | powernow_table[i].index = index; | 921 | powernow_table[i].index = index; |
864 | 922 | ||
865 | powernow_table[i].frequency = data->acpi_data.states[i].core_frequency * 1000; | 923 | powernow_table[i].frequency = |
924 | data->acpi_data.states[i].core_frequency * 1000; | ||
866 | } | 925 | } |
867 | return 0; | 926 | return 0; |
868 | } | 927 | } |
869 | 928 | ||
870 | static int fill_powernow_table_fidvid(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table) | 929 | static int fill_powernow_table_fidvid(struct powernow_k8_data *data, |
930 | struct cpufreq_frequency_table *powernow_table) | ||
871 | { | 931 | { |
872 | int i; | 932 | int i; |
873 | int cntlofreq = 0; | 933 | int cntlofreq = 0; |
934 | |||
874 | for (i = 0; i < data->acpi_data.state_count; i++) { | 935 | for (i = 0; i < data->acpi_data.state_count; i++) { |
875 | u32 fid; | 936 | u32 fid; |
876 | u32 vid; | 937 | u32 vid; |
938 | u32 freq, index; | ||
939 | acpi_integer status, control; | ||
877 | 940 | ||
878 | if (data->exttype) { | 941 | if (data->exttype) { |
879 | fid = data->acpi_data.states[i].status & EXT_FID_MASK; | 942 | status = data->acpi_data.states[i].status; |
880 | vid = (data->acpi_data.states[i].status >> VID_SHIFT) & EXT_VID_MASK; | 943 | fid = status & EXT_FID_MASK; |
944 | vid = (status >> VID_SHIFT) & EXT_VID_MASK; | ||
881 | } else { | 945 | } else { |
882 | fid = data->acpi_data.states[i].control & FID_MASK; | 946 | control = data->acpi_data.states[i].control; |
883 | vid = (data->acpi_data.states[i].control >> VID_SHIFT) & VID_MASK; | 947 | fid = control & FID_MASK; |
948 | vid = (control >> VID_SHIFT) & VID_MASK; | ||
884 | } | 949 | } |
885 | 950 | ||
886 | dprintk(" %d : fid 0x%x, vid 0x%x\n", i, fid, vid); | 951 | dprintk(" %d : fid 0x%x, vid 0x%x\n", i, fid, vid); |
887 | 952 | ||
888 | powernow_table[i].index = fid; /* lower 8 bits */ | 953 | index = fid | (vid<<8); |
889 | powernow_table[i].index |= (vid << 8); /* upper 8 bits */ | 954 | powernow_table[i].index = index; |
890 | powernow_table[i].frequency = find_khz_freq_from_fid(fid); | 955 | |
956 | freq = find_khz_freq_from_fid(fid); | ||
957 | powernow_table[i].frequency = freq; | ||
891 | 958 | ||
892 | /* verify frequency is OK */ | 959 | /* verify frequency is OK */ |
893 | if ((powernow_table[i].frequency > (MAX_FREQ * 1000)) || | 960 | if ((freq > (MAX_FREQ * 1000)) || (freq < (MIN_FREQ * 1000))) { |
894 | (powernow_table[i].frequency < (MIN_FREQ * 1000))) { | 961 | dprintk("invalid freq %u kHz, ignoring\n", freq); |
895 | dprintk("invalid freq %u kHz, ignoring\n", powernow_table[i].frequency); | 962 | invalidate_entry(data, i); |
896 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | ||
897 | continue; | 963 | continue; |
898 | } | 964 | } |
899 | 965 | ||
900 | /* verify voltage is OK - BIOSs are using "off" to indicate invalid */ | 966 | /* verify voltage is OK - |
967 | * BIOSs are using "off" to indicate invalid */ | ||
901 | if (vid == VID_OFF) { | 968 | if (vid == VID_OFF) { |
902 | dprintk("invalid vid %u, ignoring\n", vid); | 969 | dprintk("invalid vid %u, ignoring\n", vid); |
903 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 970 | invalidate_entry(data, i); |
904 | continue; | 971 | continue; |
905 | } | 972 | } |
906 | 973 | ||
907 | /* verify only 1 entry from the lo frequency table */ | 974 | /* verify only 1 entry from the lo frequency table */ |
908 | if (fid < HI_FID_TABLE_BOTTOM) { | 975 | if (fid < HI_FID_TABLE_BOTTOM) { |
909 | if (cntlofreq) { | 976 | if (cntlofreq) { |
910 | /* if both entries are the same, ignore this one ... */ | 977 | /* if both entries are the same, |
911 | if ((powernow_table[i].frequency != powernow_table[cntlofreq].frequency) || | 978 | * ignore this one ... */ |
912 | (powernow_table[i].index != powernow_table[cntlofreq].index)) { | 979 | if ((freq != powernow_table[cntlofreq].frequency) || |
913 | printk(KERN_ERR PFX "Too many lo freq table entries\n"); | 980 | (index != powernow_table[cntlofreq].index)) { |
981 | printk(KERN_ERR PFX | ||
982 | "Too many lo freq table " | ||
983 | "entries\n"); | ||
914 | return 1; | 984 | return 1; |
915 | } | 985 | } |
916 | 986 | ||
917 | dprintk("double low frequency table entry, ignoring it.\n"); | 987 | dprintk("double low frequency table entry, " |
918 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 988 | "ignoring it.\n"); |
989 | invalidate_entry(data, i); | ||
919 | continue; | 990 | continue; |
920 | } else | 991 | } else |
921 | cntlofreq = i; | 992 | cntlofreq = i; |
922 | } | 993 | } |
923 | 994 | ||
924 | if (powernow_table[i].frequency != (data->acpi_data.states[i].core_frequency * 1000)) { | 995 | if (freq != (data->acpi_data.states[i].core_frequency * 1000)) { |
925 | printk(KERN_INFO PFX "invalid freq entries %u kHz vs. %u kHz\n", | 996 | printk(KERN_INFO PFX "invalid freq entries " |
926 | powernow_table[i].frequency, | 997 | "%u kHz vs. %u kHz\n", freq, |
927 | (unsigned int) (data->acpi_data.states[i].core_frequency * 1000)); | 998 | (unsigned int) |
928 | powernow_table[i].frequency = CPUFREQ_ENTRY_INVALID; | 999 | (data->acpi_data.states[i].core_frequency |
1000 | * 1000)); | ||
1001 | invalidate_entry(data, i); | ||
929 | continue; | 1002 | continue; |
930 | } | 1003 | } |
931 | } | 1004 | } |
@@ -935,7 +1008,8 @@ static int fill_powernow_table_fidvid(struct powernow_k8_data *data, struct cpuf | |||
935 | static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data) | 1008 | static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data) |
936 | { | 1009 | { |
937 | if (data->acpi_data.state_count) | 1010 | if (data->acpi_data.state_count) |
938 | acpi_processor_unregister_performance(&data->acpi_data, data->cpu); | 1011 | acpi_processor_unregister_performance(&data->acpi_data, |
1012 | data->cpu); | ||
939 | free_cpumask_var(data->acpi_data.shared_cpu_map); | 1013 | free_cpumask_var(data->acpi_data.shared_cpu_map); |
940 | } | 1014 | } |
941 | 1015 | ||
@@ -953,15 +1027,9 @@ static int get_transition_latency(struct powernow_k8_data *data) | |||
953 | return 1000 * max_latency; | 1027 | return 1000 * max_latency; |
954 | } | 1028 | } |
955 | 1029 | ||
956 | #else | ||
957 | static int powernow_k8_cpu_init_acpi(struct powernow_k8_data *data) { return -ENODEV; } | ||
958 | static void powernow_k8_cpu_exit_acpi(struct powernow_k8_data *data) { return; } | ||
959 | static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index) { return; } | ||
960 | static int get_transition_latency(struct powernow_k8_data *data) { return 0; } | ||
961 | #endif /* CONFIG_X86_POWERNOW_K8_ACPI */ | ||
962 | |||
963 | /* Take a frequency, and issue the fid/vid transition command */ | 1030 | /* Take a frequency, and issue the fid/vid transition command */ |
964 | static int transition_frequency_fidvid(struct powernow_k8_data *data, unsigned int index) | 1031 | static int transition_frequency_fidvid(struct powernow_k8_data *data, |
1032 | unsigned int index) | ||
965 | { | 1033 | { |
966 | u32 fid = 0; | 1034 | u32 fid = 0; |
967 | u32 vid = 0; | 1035 | u32 vid = 0; |
@@ -989,7 +1057,8 @@ static int transition_frequency_fidvid(struct powernow_k8_data *data, unsigned i | |||
989 | return 0; | 1057 | return 0; |
990 | } | 1058 | } |
991 | 1059 | ||
992 | if ((fid < HI_FID_TABLE_BOTTOM) && (data->currfid < HI_FID_TABLE_BOTTOM)) { | 1060 | if ((fid < HI_FID_TABLE_BOTTOM) && |
1061 | (data->currfid < HI_FID_TABLE_BOTTOM)) { | ||
993 | printk(KERN_ERR PFX | 1062 | printk(KERN_ERR PFX |
994 | "ignoring illegal change in lo freq table-%x to 0x%x\n", | 1063 | "ignoring illegal change in lo freq table-%x to 0x%x\n", |
995 | data->currfid, fid); | 1064 | data->currfid, fid); |
@@ -1017,7 +1086,8 @@ static int transition_frequency_fidvid(struct powernow_k8_data *data, unsigned i | |||
1017 | } | 1086 | } |
1018 | 1087 | ||
1019 | /* Take a frequency, and issue the hardware pstate transition command */ | 1088 | /* Take a frequency, and issue the hardware pstate transition command */ |
1020 | static int transition_frequency_pstate(struct powernow_k8_data *data, unsigned int index) | 1089 | static int transition_frequency_pstate(struct powernow_k8_data *data, |
1090 | unsigned int index) | ||
1021 | { | 1091 | { |
1022 | u32 pstate = 0; | 1092 | u32 pstate = 0; |
1023 | int res, i; | 1093 | int res, i; |
@@ -1029,7 +1099,8 @@ static int transition_frequency_pstate(struct powernow_k8_data *data, unsigned i | |||
1029 | pstate = index & HW_PSTATE_MASK; | 1099 | pstate = index & HW_PSTATE_MASK; |
1030 | if (pstate > data->max_hw_pstate) | 1100 | if (pstate > data->max_hw_pstate) |
1031 | return 0; | 1101 | return 0; |
1032 | freqs.old = find_khz_freq_from_pstate(data->powernow_table, data->currpstate); | 1102 | freqs.old = find_khz_freq_from_pstate(data->powernow_table, |
1103 | data->currpstate); | ||
1033 | freqs.new = find_khz_freq_from_pstate(data->powernow_table, pstate); | 1104 | freqs.new = find_khz_freq_from_pstate(data->powernow_table, pstate); |
1034 | 1105 | ||
1035 | for_each_cpu_mask_nr(i, *(data->available_cores)) { | 1106 | for_each_cpu_mask_nr(i, *(data->available_cores)) { |
@@ -1048,7 +1119,8 @@ static int transition_frequency_pstate(struct powernow_k8_data *data, unsigned i | |||
1048 | } | 1119 | } |
1049 | 1120 | ||
1050 | /* Driver entry point to switch to the target frequency */ | 1121 | /* Driver entry point to switch to the target frequency */ |
1051 | static int powernowk8_target(struct cpufreq_policy *pol, unsigned targfreq, unsigned relation) | 1122 | static int powernowk8_target(struct cpufreq_policy *pol, |
1123 | unsigned targfreq, unsigned relation) | ||
1052 | { | 1124 | { |
1053 | cpumask_t oldmask; | 1125 | cpumask_t oldmask; |
1054 | struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); | 1126 | struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); |
@@ -1087,14 +1159,18 @@ static int powernowk8_target(struct cpufreq_policy *pol, unsigned targfreq, unsi | |||
1087 | dprintk("targ: curr fid 0x%x, vid 0x%x\n", | 1159 | dprintk("targ: curr fid 0x%x, vid 0x%x\n", |
1088 | data->currfid, data->currvid); | 1160 | data->currfid, data->currvid); |
1089 | 1161 | ||
1090 | if ((checkvid != data->currvid) || (checkfid != data->currfid)) { | 1162 | if ((checkvid != data->currvid) || |
1163 | (checkfid != data->currfid)) { | ||
1091 | printk(KERN_INFO PFX | 1164 | printk(KERN_INFO PFX |
1092 | "error - out of sync, fix 0x%x 0x%x, vid 0x%x 0x%x\n", | 1165 | "error - out of sync, fix 0x%x 0x%x, " |
1093 | checkfid, data->currfid, checkvid, data->currvid); | 1166 | "vid 0x%x 0x%x\n", |
1167 | checkfid, data->currfid, | ||
1168 | checkvid, data->currvid); | ||
1094 | } | 1169 | } |
1095 | } | 1170 | } |
1096 | 1171 | ||
1097 | if (cpufreq_frequency_table_target(pol, data->powernow_table, targfreq, relation, &newstate)) | 1172 | if (cpufreq_frequency_table_target(pol, data->powernow_table, |
1173 | targfreq, relation, &newstate)) | ||
1098 | goto err_out; | 1174 | goto err_out; |
1099 | 1175 | ||
1100 | mutex_lock(&fidvid_mutex); | 1176 | mutex_lock(&fidvid_mutex); |
@@ -1114,7 +1190,8 @@ static int powernowk8_target(struct cpufreq_policy *pol, unsigned targfreq, unsi | |||
1114 | mutex_unlock(&fidvid_mutex); | 1190 | mutex_unlock(&fidvid_mutex); |
1115 | 1191 | ||
1116 | if (cpu_family == CPU_HW_PSTATE) | 1192 | if (cpu_family == CPU_HW_PSTATE) |
1117 | pol->cur = find_khz_freq_from_pstate(data->powernow_table, newstate); | 1193 | pol->cur = find_khz_freq_from_pstate(data->powernow_table, |
1194 | newstate); | ||
1118 | else | 1195 | else |
1119 | pol->cur = find_khz_freq_from_fid(data->currfid); | 1196 | pol->cur = find_khz_freq_from_fid(data->currfid); |
1120 | ret = 0; | 1197 | ret = 0; |
@@ -1141,6 +1218,7 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1141 | struct powernow_k8_data *data; | 1218 | struct powernow_k8_data *data; |
1142 | cpumask_t oldmask; | 1219 | cpumask_t oldmask; |
1143 | int rc; | 1220 | int rc; |
1221 | static int print_once; | ||
1144 | 1222 | ||
1145 | if (!cpu_online(pol->cpu)) | 1223 | if (!cpu_online(pol->cpu)) |
1146 | return -ENODEV; | 1224 | return -ENODEV; |
@@ -1163,33 +1241,31 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1163 | * an UP version, and is deprecated by AMD. | 1241 | * an UP version, and is deprecated by AMD. |
1164 | */ | 1242 | */ |
1165 | if (num_online_cpus() != 1) { | 1243 | if (num_online_cpus() != 1) { |
1166 | #ifndef CONFIG_ACPI_PROCESSOR | 1244 | /* |
1167 | printk(KERN_ERR PFX "ACPI Processor support is required " | 1245 | * Replace this one with print_once as soon as such a |
1168 | "for SMP systems but is absent. Please load the " | 1246 | * thing gets introduced |
1169 | "ACPI Processor module before starting this " | 1247 | */ |
1170 | "driver.\n"); | 1248 | if (!print_once) { |
1171 | #else | 1249 | WARN_ONCE(1, KERN_ERR FW_BUG PFX "Your BIOS " |
1172 | printk(KERN_ERR FW_BUG PFX "Your BIOS does not provide" | 1250 | "does not provide ACPI _PSS objects " |
1173 | " ACPI _PSS objects in a way that Linux " | 1251 | "in a way that Linux understands. " |
1174 | "understands. Please report this to the Linux " | 1252 | "Please report this to the Linux ACPI" |
1175 | "ACPI maintainers and complain to your BIOS " | 1253 | " maintainers and complain to your " |
1176 | "vendor.\n"); | 1254 | "BIOS vendor.\n"); |
1177 | #endif | 1255 | print_once++; |
1178 | kfree(data); | 1256 | } |
1179 | return -ENODEV; | 1257 | goto err_out; |
1180 | } | 1258 | } |
1181 | if (pol->cpu != 0) { | 1259 | if (pol->cpu != 0) { |
1182 | printk(KERN_ERR FW_BUG PFX "No ACPI _PSS objects for " | 1260 | printk(KERN_ERR FW_BUG PFX "No ACPI _PSS objects for " |
1183 | "CPU other than CPU0. Complain to your BIOS " | 1261 | "CPU other than CPU0. Complain to your BIOS " |
1184 | "vendor.\n"); | 1262 | "vendor.\n"); |
1185 | kfree(data); | 1263 | goto err_out; |
1186 | return -ENODEV; | ||
1187 | } | 1264 | } |
1188 | rc = find_psb_table(data); | 1265 | rc = find_psb_table(data); |
1189 | if (rc) { | 1266 | if (rc) |
1190 | kfree(data); | 1267 | goto err_out; |
1191 | return -ENODEV; | 1268 | |
1192 | } | ||
1193 | /* Take a crude guess here. | 1269 | /* Take a crude guess here. |
1194 | * That guess was in microseconds, so multiply with 1000 */ | 1270 | * That guess was in microseconds, so multiply with 1000 */ |
1195 | pol->cpuinfo.transition_latency = ( | 1271 | pol->cpuinfo.transition_latency = ( |
@@ -1204,16 +1280,16 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1204 | 1280 | ||
1205 | if (smp_processor_id() != pol->cpu) { | 1281 | if (smp_processor_id() != pol->cpu) { |
1206 | printk(KERN_ERR PFX "limiting to cpu %u failed\n", pol->cpu); | 1282 | printk(KERN_ERR PFX "limiting to cpu %u failed\n", pol->cpu); |
1207 | goto err_out; | 1283 | goto err_out_unmask; |
1208 | } | 1284 | } |
1209 | 1285 | ||
1210 | if (pending_bit_stuck()) { | 1286 | if (pending_bit_stuck()) { |
1211 | printk(KERN_ERR PFX "failing init, change pending bit set\n"); | 1287 | printk(KERN_ERR PFX "failing init, change pending bit set\n"); |
1212 | goto err_out; | 1288 | goto err_out_unmask; |
1213 | } | 1289 | } |
1214 | 1290 | ||
1215 | if (query_current_values_with_pending_wait(data)) | 1291 | if (query_current_values_with_pending_wait(data)) |
1216 | goto err_out; | 1292 | goto err_out_unmask; |
1217 | 1293 | ||
1218 | if (cpu_family == CPU_OPTERON) | 1294 | if (cpu_family == CPU_OPTERON) |
1219 | fidvid_msr_init(); | 1295 | fidvid_msr_init(); |
@@ -1228,7 +1304,8 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1228 | data->available_cores = pol->cpus; | 1304 | data->available_cores = pol->cpus; |
1229 | 1305 | ||
1230 | if (cpu_family == CPU_HW_PSTATE) | 1306 | if (cpu_family == CPU_HW_PSTATE) |
1231 | pol->cur = find_khz_freq_from_pstate(data->powernow_table, data->currpstate); | 1307 | pol->cur = find_khz_freq_from_pstate(data->powernow_table, |
1308 | data->currpstate); | ||
1232 | else | 1309 | else |
1233 | pol->cur = find_khz_freq_from_fid(data->currfid); | 1310 | pol->cur = find_khz_freq_from_fid(data->currfid); |
1234 | dprintk("policy current frequency %d kHz\n", pol->cur); | 1311 | dprintk("policy current frequency %d kHz\n", pol->cur); |
@@ -1245,7 +1322,8 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1245 | cpufreq_frequency_table_get_attr(data->powernow_table, pol->cpu); | 1322 | cpufreq_frequency_table_get_attr(data->powernow_table, pol->cpu); |
1246 | 1323 | ||
1247 | if (cpu_family == CPU_HW_PSTATE) | 1324 | if (cpu_family == CPU_HW_PSTATE) |
1248 | dprintk("cpu_init done, current pstate 0x%x\n", data->currpstate); | 1325 | dprintk("cpu_init done, current pstate 0x%x\n", |
1326 | data->currpstate); | ||
1249 | else | 1327 | else |
1250 | dprintk("cpu_init done, current fid 0x%x, vid 0x%x\n", | 1328 | dprintk("cpu_init done, current fid 0x%x, vid 0x%x\n", |
1251 | data->currfid, data->currvid); | 1329 | data->currfid, data->currvid); |
@@ -1254,15 +1332,16 @@ static int __cpuinit powernowk8_cpu_init(struct cpufreq_policy *pol) | |||
1254 | 1332 | ||
1255 | return 0; | 1333 | return 0; |
1256 | 1334 | ||
1257 | err_out: | 1335 | err_out_unmask: |
1258 | set_cpus_allowed_ptr(current, &oldmask); | 1336 | set_cpus_allowed_ptr(current, &oldmask); |
1259 | powernow_k8_cpu_exit_acpi(data); | 1337 | powernow_k8_cpu_exit_acpi(data); |
1260 | 1338 | ||
1339 | err_out: | ||
1261 | kfree(data); | 1340 | kfree(data); |
1262 | return -ENODEV; | 1341 | return -ENODEV; |
1263 | } | 1342 | } |
1264 | 1343 | ||
1265 | static int __devexit powernowk8_cpu_exit (struct cpufreq_policy *pol) | 1344 | static int __devexit powernowk8_cpu_exit(struct cpufreq_policy *pol) |
1266 | { | 1345 | { |
1267 | struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); | 1346 | struct powernow_k8_data *data = per_cpu(powernow_data, pol->cpu); |
1268 | 1347 | ||
@@ -1279,7 +1358,7 @@ static int __devexit powernowk8_cpu_exit (struct cpufreq_policy *pol) | |||
1279 | return 0; | 1358 | return 0; |
1280 | } | 1359 | } |
1281 | 1360 | ||
1282 | static unsigned int powernowk8_get (unsigned int cpu) | 1361 | static unsigned int powernowk8_get(unsigned int cpu) |
1283 | { | 1362 | { |
1284 | struct powernow_k8_data *data; | 1363 | struct powernow_k8_data *data; |
1285 | cpumask_t oldmask = current->cpus_allowed; | 1364 | cpumask_t oldmask = current->cpus_allowed; |
@@ -1315,7 +1394,7 @@ out: | |||
1315 | return khz; | 1394 | return khz; |
1316 | } | 1395 | } |
1317 | 1396 | ||
1318 | static struct freq_attr* powernow_k8_attr[] = { | 1397 | static struct freq_attr *powernow_k8_attr[] = { |
1319 | &cpufreq_freq_attr_scaling_available_freqs, | 1398 | &cpufreq_freq_attr_scaling_available_freqs, |
1320 | NULL, | 1399 | NULL, |
1321 | }; | 1400 | }; |
@@ -1360,7 +1439,8 @@ static void __exit powernowk8_exit(void) | |||
1360 | cpufreq_unregister_driver(&cpufreq_amd64_driver); | 1439 | cpufreq_unregister_driver(&cpufreq_amd64_driver); |
1361 | } | 1440 | } |
1362 | 1441 | ||
1363 | MODULE_AUTHOR("Paul Devriendt <paul.devriendt@amd.com> and Mark Langsdorf <mark.langsdorf@amd.com>"); | 1442 | MODULE_AUTHOR("Paul Devriendt <paul.devriendt@amd.com> and " |
1443 | "Mark Langsdorf <mark.langsdorf@amd.com>"); | ||
1364 | MODULE_DESCRIPTION("AMD Athlon 64 and Opteron processor frequency driver."); | 1444 | MODULE_DESCRIPTION("AMD Athlon 64 and Opteron processor frequency driver."); |
1365 | MODULE_LICENSE("GPL"); | 1445 | MODULE_LICENSE("GPL"); |
1366 | 1446 | ||
diff --git a/arch/x86/kernel/cpu/cpufreq/powernow-k8.h b/arch/x86/kernel/cpu/cpufreq/powernow-k8.h index 8ecc75b6c7c3..6c6698feade1 100644 --- a/arch/x86/kernel/cpu/cpufreq/powernow-k8.h +++ b/arch/x86/kernel/cpu/cpufreq/powernow-k8.h | |||
@@ -45,11 +45,10 @@ struct powernow_k8_data { | |||
45 | * frequency is in kHz */ | 45 | * frequency is in kHz */ |
46 | struct cpufreq_frequency_table *powernow_table; | 46 | struct cpufreq_frequency_table *powernow_table; |
47 | 47 | ||
48 | #ifdef CONFIG_X86_POWERNOW_K8_ACPI | ||
49 | /* the acpi table needs to be kept. it's only available if ACPI was | 48 | /* the acpi table needs to be kept. it's only available if ACPI was |
50 | * used to determine valid frequency/vid/fid states */ | 49 | * used to determine valid frequency/vid/fid states */ |
51 | struct acpi_processor_performance acpi_data; | 50 | struct acpi_processor_performance acpi_data; |
52 | #endif | 51 | |
53 | /* we need to keep track of associated cores, but let cpufreq | 52 | /* we need to keep track of associated cores, but let cpufreq |
54 | * handle hotplug events - so just point at cpufreq pol->cpus | 53 | * handle hotplug events - so just point at cpufreq pol->cpus |
55 | * structure */ | 54 | * structure */ |
@@ -222,10 +221,8 @@ static int core_frequency_transition(struct powernow_k8_data *data, u32 reqfid); | |||
222 | 221 | ||
223 | static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index); | 222 | static void powernow_k8_acpi_pst_values(struct powernow_k8_data *data, unsigned int index); |
224 | 223 | ||
225 | #ifdef CONFIG_X86_POWERNOW_K8_ACPI | ||
226 | static int fill_powernow_table_pstate(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table); | 224 | static int fill_powernow_table_pstate(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table); |
227 | static int fill_powernow_table_fidvid(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table); | 225 | static int fill_powernow_table_fidvid(struct powernow_k8_data *data, struct cpufreq_frequency_table *powernow_table); |
228 | #endif | ||
229 | 226 | ||
230 | #ifdef CONFIG_SMP | 227 | #ifdef CONFIG_SMP |
231 | static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[]) | 228 | static inline void define_siblings(int cpu, cpumask_t cpu_sharedcore_mask[]) |
diff --git a/arch/x86/kernel/cpu/cpufreq/sc520_freq.c b/arch/x86/kernel/cpu/cpufreq/sc520_freq.c index 42da9bd677d6..435a996a613a 100644 --- a/arch/x86/kernel/cpu/cpufreq/sc520_freq.c +++ b/arch/x86/kernel/cpu/cpufreq/sc520_freq.c | |||
@@ -19,17 +19,19 @@ | |||
19 | 19 | ||
20 | #include <linux/delay.h> | 20 | #include <linux/delay.h> |
21 | #include <linux/cpufreq.h> | 21 | #include <linux/cpufreq.h> |
22 | #include <linux/timex.h> | ||
23 | #include <linux/io.h> | ||
22 | 24 | ||
23 | #include <asm/msr.h> | 25 | #include <asm/msr.h> |
24 | #include <asm/timex.h> | ||
25 | #include <asm/io.h> | ||
26 | 26 | ||
27 | #define MMCR_BASE 0xfffef000 /* The default base address */ | 27 | #define MMCR_BASE 0xfffef000 /* The default base address */ |
28 | #define OFFS_CPUCTL 0x2 /* CPU Control Register */ | 28 | #define OFFS_CPUCTL 0x2 /* CPU Control Register */ |
29 | 29 | ||
30 | static __u8 __iomem *cpuctl; | 30 | static __u8 __iomem *cpuctl; |
31 | 31 | ||
32 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "sc520_freq", msg) | 32 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
33 | "sc520_freq", msg) | ||
34 | #define PFX "sc520_freq: " | ||
33 | 35 | ||
34 | static struct cpufreq_frequency_table sc520_freq_table[] = { | 36 | static struct cpufreq_frequency_table sc520_freq_table[] = { |
35 | {0x01, 100000}, | 37 | {0x01, 100000}, |
@@ -43,7 +45,8 @@ static unsigned int sc520_freq_get_cpu_frequency(unsigned int cpu) | |||
43 | 45 | ||
44 | switch (clockspeed_reg & 0x03) { | 46 | switch (clockspeed_reg & 0x03) { |
45 | default: | 47 | default: |
46 | printk(KERN_ERR "sc520_freq: error: cpuctl register has unexpected value %02x\n", clockspeed_reg); | 48 | printk(KERN_ERR PFX "error: cpuctl register has unexpected " |
49 | "value %02x\n", clockspeed_reg); | ||
47 | case 0x01: | 50 | case 0x01: |
48 | return 100000; | 51 | return 100000; |
49 | case 0x02: | 52 | case 0x02: |
@@ -51,7 +54,7 @@ static unsigned int sc520_freq_get_cpu_frequency(unsigned int cpu) | |||
51 | } | 54 | } |
52 | } | 55 | } |
53 | 56 | ||
54 | static void sc520_freq_set_cpu_state (unsigned int state) | 57 | static void sc520_freq_set_cpu_state(unsigned int state) |
55 | { | 58 | { |
56 | 59 | ||
57 | struct cpufreq_freqs freqs; | 60 | struct cpufreq_freqs freqs; |
@@ -76,18 +79,19 @@ static void sc520_freq_set_cpu_state (unsigned int state) | |||
76 | cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); | 79 | cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); |
77 | }; | 80 | }; |
78 | 81 | ||
79 | static int sc520_freq_verify (struct cpufreq_policy *policy) | 82 | static int sc520_freq_verify(struct cpufreq_policy *policy) |
80 | { | 83 | { |
81 | return cpufreq_frequency_table_verify(policy, &sc520_freq_table[0]); | 84 | return cpufreq_frequency_table_verify(policy, &sc520_freq_table[0]); |
82 | } | 85 | } |
83 | 86 | ||
84 | static int sc520_freq_target (struct cpufreq_policy *policy, | 87 | static int sc520_freq_target(struct cpufreq_policy *policy, |
85 | unsigned int target_freq, | 88 | unsigned int target_freq, |
86 | unsigned int relation) | 89 | unsigned int relation) |
87 | { | 90 | { |
88 | unsigned int newstate = 0; | 91 | unsigned int newstate = 0; |
89 | 92 | ||
90 | if (cpufreq_frequency_table_target(policy, sc520_freq_table, target_freq, relation, &newstate)) | 93 | if (cpufreq_frequency_table_target(policy, sc520_freq_table, |
94 | target_freq, relation, &newstate)) | ||
91 | return -EINVAL; | 95 | return -EINVAL; |
92 | 96 | ||
93 | sc520_freq_set_cpu_state(newstate); | 97 | sc520_freq_set_cpu_state(newstate); |
@@ -116,7 +120,7 @@ static int sc520_freq_cpu_init(struct cpufreq_policy *policy) | |||
116 | 120 | ||
117 | result = cpufreq_frequency_table_cpuinfo(policy, sc520_freq_table); | 121 | result = cpufreq_frequency_table_cpuinfo(policy, sc520_freq_table); |
118 | if (result) | 122 | if (result) |
119 | return (result); | 123 | return result; |
120 | 124 | ||
121 | cpufreq_frequency_table_get_attr(sc520_freq_table, policy->cpu); | 125 | cpufreq_frequency_table_get_attr(sc520_freq_table, policy->cpu); |
122 | 126 | ||
@@ -131,7 +135,7 @@ static int sc520_freq_cpu_exit(struct cpufreq_policy *policy) | |||
131 | } | 135 | } |
132 | 136 | ||
133 | 137 | ||
134 | static struct freq_attr* sc520_freq_attr[] = { | 138 | static struct freq_attr *sc520_freq_attr[] = { |
135 | &cpufreq_freq_attr_scaling_available_freqs, | 139 | &cpufreq_freq_attr_scaling_available_freqs, |
136 | NULL, | 140 | NULL, |
137 | }; | 141 | }; |
@@ -155,13 +159,13 @@ static int __init sc520_freq_init(void) | |||
155 | int err; | 159 | int err; |
156 | 160 | ||
157 | /* Test if we have the right hardware */ | 161 | /* Test if we have the right hardware */ |
158 | if(c->x86_vendor != X86_VENDOR_AMD || | 162 | if (c->x86_vendor != X86_VENDOR_AMD || |
159 | c->x86 != 4 || c->x86_model != 9) { | 163 | c->x86 != 4 || c->x86_model != 9) { |
160 | dprintk("no Elan SC520 processor found!\n"); | 164 | dprintk("no Elan SC520 processor found!\n"); |
161 | return -ENODEV; | 165 | return -ENODEV; |
162 | } | 166 | } |
163 | cpuctl = ioremap((unsigned long)(MMCR_BASE + OFFS_CPUCTL), 1); | 167 | cpuctl = ioremap((unsigned long)(MMCR_BASE + OFFS_CPUCTL), 1); |
164 | if(!cpuctl) { | 168 | if (!cpuctl) { |
165 | printk(KERN_ERR "sc520_freq: error: failed to remap memory\n"); | 169 | printk(KERN_ERR "sc520_freq: error: failed to remap memory\n"); |
166 | return -ENOMEM; | 170 | return -ENOMEM; |
167 | } | 171 | } |
diff --git a/arch/x86/kernel/cpu/cpufreq/speedstep-ich.c b/arch/x86/kernel/cpu/cpufreq/speedstep-ich.c index dedc1e98f168..8bbb11adb315 100644 --- a/arch/x86/kernel/cpu/cpufreq/speedstep-ich.c +++ b/arch/x86/kernel/cpu/cpufreq/speedstep-ich.c | |||
@@ -39,7 +39,7 @@ static struct pci_dev *speedstep_chipset_dev; | |||
39 | 39 | ||
40 | /* speedstep_processor | 40 | /* speedstep_processor |
41 | */ | 41 | */ |
42 | static unsigned int speedstep_processor = 0; | 42 | static unsigned int speedstep_processor; |
43 | 43 | ||
44 | static u32 pmbase; | 44 | static u32 pmbase; |
45 | 45 | ||
@@ -54,7 +54,8 @@ static struct cpufreq_frequency_table speedstep_freqs[] = { | |||
54 | }; | 54 | }; |
55 | 55 | ||
56 | 56 | ||
57 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "speedstep-ich", msg) | 57 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
58 | "speedstep-ich", msg) | ||
58 | 59 | ||
59 | 60 | ||
60 | /** | 61 | /** |
@@ -62,7 +63,7 @@ static struct cpufreq_frequency_table speedstep_freqs[] = { | |||
62 | * | 63 | * |
63 | * Returns: -ENODEV if no register could be found | 64 | * Returns: -ENODEV if no register could be found |
64 | */ | 65 | */ |
65 | static int speedstep_find_register (void) | 66 | static int speedstep_find_register(void) |
66 | { | 67 | { |
67 | if (!speedstep_chipset_dev) | 68 | if (!speedstep_chipset_dev) |
68 | return -ENODEV; | 69 | return -ENODEV; |
@@ -90,7 +91,7 @@ static int speedstep_find_register (void) | |||
90 | * | 91 | * |
91 | * Tries to change the SpeedStep state. | 92 | * Tries to change the SpeedStep state. |
92 | */ | 93 | */ |
93 | static void speedstep_set_state (unsigned int state) | 94 | static void speedstep_set_state(unsigned int state) |
94 | { | 95 | { |
95 | u8 pm2_blk; | 96 | u8 pm2_blk; |
96 | u8 value; | 97 | u8 value; |
@@ -133,11 +134,11 @@ static void speedstep_set_state (unsigned int state) | |||
133 | 134 | ||
134 | dprintk("read at pmbase 0x%x + 0x50 returned 0x%x\n", pmbase, value); | 135 | dprintk("read at pmbase 0x%x + 0x50 returned 0x%x\n", pmbase, value); |
135 | 136 | ||
136 | if (state == (value & 0x1)) { | 137 | if (state == (value & 0x1)) |
137 | dprintk("change to %u MHz succeeded\n", (speedstep_get_processor_frequency(speedstep_processor) / 1000)); | 138 | dprintk("change to %u MHz succeeded\n", |
138 | } else { | 139 | speedstep_get_frequency(speedstep_processor) / 1000); |
139 | printk (KERN_ERR "cpufreq: change failed - I/O error\n"); | 140 | else |
140 | } | 141 | printk(KERN_ERR "cpufreq: change failed - I/O error\n"); |
141 | 142 | ||
142 | return; | 143 | return; |
143 | } | 144 | } |
@@ -149,7 +150,7 @@ static void speedstep_set_state (unsigned int state) | |||
149 | * Tries to activate the SpeedStep status and control registers. | 150 | * Tries to activate the SpeedStep status and control registers. |
150 | * Returns -EINVAL on an unsupported chipset, and zero on success. | 151 | * Returns -EINVAL on an unsupported chipset, and zero on success. |
151 | */ | 152 | */ |
152 | static int speedstep_activate (void) | 153 | static int speedstep_activate(void) |
153 | { | 154 | { |
154 | u16 value = 0; | 155 | u16 value = 0; |
155 | 156 | ||
@@ -175,20 +176,18 @@ static int speedstep_activate (void) | |||
175 | * functions. Returns the SPEEDSTEP_CHIPSET_-number for the detected | 176 | * functions. Returns the SPEEDSTEP_CHIPSET_-number for the detected |
176 | * chipset, or zero on failure. | 177 | * chipset, or zero on failure. |
177 | */ | 178 | */ |
178 | static unsigned int speedstep_detect_chipset (void) | 179 | static unsigned int speedstep_detect_chipset(void) |
179 | { | 180 | { |
180 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, | 181 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, |
181 | PCI_DEVICE_ID_INTEL_82801DB_12, | 182 | PCI_DEVICE_ID_INTEL_82801DB_12, |
182 | PCI_ANY_ID, | 183 | PCI_ANY_ID, PCI_ANY_ID, |
183 | PCI_ANY_ID, | ||
184 | NULL); | 184 | NULL); |
185 | if (speedstep_chipset_dev) | 185 | if (speedstep_chipset_dev) |
186 | return 4; /* 4-M */ | 186 | return 4; /* 4-M */ |
187 | 187 | ||
188 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, | 188 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, |
189 | PCI_DEVICE_ID_INTEL_82801CA_12, | 189 | PCI_DEVICE_ID_INTEL_82801CA_12, |
190 | PCI_ANY_ID, | 190 | PCI_ANY_ID, PCI_ANY_ID, |
191 | PCI_ANY_ID, | ||
192 | NULL); | 191 | NULL); |
193 | if (speedstep_chipset_dev) | 192 | if (speedstep_chipset_dev) |
194 | return 3; /* 3-M */ | 193 | return 3; /* 3-M */ |
@@ -196,8 +195,7 @@ static unsigned int speedstep_detect_chipset (void) | |||
196 | 195 | ||
197 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, | 196 | speedstep_chipset_dev = pci_get_subsys(PCI_VENDOR_ID_INTEL, |
198 | PCI_DEVICE_ID_INTEL_82801BA_10, | 197 | PCI_DEVICE_ID_INTEL_82801BA_10, |
199 | PCI_ANY_ID, | 198 | PCI_ANY_ID, PCI_ANY_ID, |
200 | PCI_ANY_ID, | ||
201 | NULL); | 199 | NULL); |
202 | if (speedstep_chipset_dev) { | 200 | if (speedstep_chipset_dev) { |
203 | /* speedstep.c causes lockups on Dell Inspirons 8000 and | 201 | /* speedstep.c causes lockups on Dell Inspirons 8000 and |
@@ -208,8 +206,7 @@ static unsigned int speedstep_detect_chipset (void) | |||
208 | 206 | ||
209 | hostbridge = pci_get_subsys(PCI_VENDOR_ID_INTEL, | 207 | hostbridge = pci_get_subsys(PCI_VENDOR_ID_INTEL, |
210 | PCI_DEVICE_ID_INTEL_82815_MC, | 208 | PCI_DEVICE_ID_INTEL_82815_MC, |
211 | PCI_ANY_ID, | 209 | PCI_ANY_ID, PCI_ANY_ID, |
212 | PCI_ANY_ID, | ||
213 | NULL); | 210 | NULL); |
214 | 211 | ||
215 | if (!hostbridge) | 212 | if (!hostbridge) |
@@ -236,7 +233,7 @@ static unsigned int _speedstep_get(const struct cpumask *cpus) | |||
236 | 233 | ||
237 | cpus_allowed = current->cpus_allowed; | 234 | cpus_allowed = current->cpus_allowed; |
238 | set_cpus_allowed_ptr(current, cpus); | 235 | set_cpus_allowed_ptr(current, cpus); |
239 | speed = speedstep_get_processor_frequency(speedstep_processor); | 236 | speed = speedstep_get_frequency(speedstep_processor); |
240 | set_cpus_allowed_ptr(current, &cpus_allowed); | 237 | set_cpus_allowed_ptr(current, &cpus_allowed); |
241 | dprintk("detected %u kHz as current frequency\n", speed); | 238 | dprintk("detected %u kHz as current frequency\n", speed); |
242 | return speed; | 239 | return speed; |
@@ -251,11 +248,12 @@ static unsigned int speedstep_get(unsigned int cpu) | |||
251 | * speedstep_target - set a new CPUFreq policy | 248 | * speedstep_target - set a new CPUFreq policy |
252 | * @policy: new policy | 249 | * @policy: new policy |
253 | * @target_freq: the target frequency | 250 | * @target_freq: the target frequency |
254 | * @relation: how that frequency relates to achieved frequency (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | 251 | * @relation: how that frequency relates to achieved frequency |
252 | * (CPUFREQ_RELATION_L or CPUFREQ_RELATION_H) | ||
255 | * | 253 | * |
256 | * Sets a new CPUFreq policy. | 254 | * Sets a new CPUFreq policy. |
257 | */ | 255 | */ |
258 | static int speedstep_target (struct cpufreq_policy *policy, | 256 | static int speedstep_target(struct cpufreq_policy *policy, |
259 | unsigned int target_freq, | 257 | unsigned int target_freq, |
260 | unsigned int relation) | 258 | unsigned int relation) |
261 | { | 259 | { |
@@ -264,7 +262,8 @@ static int speedstep_target (struct cpufreq_policy *policy, | |||
264 | cpumask_t cpus_allowed; | 262 | cpumask_t cpus_allowed; |
265 | int i; | 263 | int i; |
266 | 264 | ||
267 | if (cpufreq_frequency_table_target(policy, &speedstep_freqs[0], target_freq, relation, &newstate)) | 265 | if (cpufreq_frequency_table_target(policy, &speedstep_freqs[0], |
266 | target_freq, relation, &newstate)) | ||
268 | return -EINVAL; | 267 | return -EINVAL; |
269 | 268 | ||
270 | freqs.old = _speedstep_get(policy->cpus); | 269 | freqs.old = _speedstep_get(policy->cpus); |
@@ -308,7 +307,7 @@ static int speedstep_target (struct cpufreq_policy *policy, | |||
308 | * Limit must be within speedstep_low_freq and speedstep_high_freq, with | 307 | * Limit must be within speedstep_low_freq and speedstep_high_freq, with |
309 | * at least one border included. | 308 | * at least one border included. |
310 | */ | 309 | */ |
311 | static int speedstep_verify (struct cpufreq_policy *policy) | 310 | static int speedstep_verify(struct cpufreq_policy *policy) |
312 | { | 311 | { |
313 | return cpufreq_frequency_table_verify(policy, &speedstep_freqs[0]); | 312 | return cpufreq_frequency_table_verify(policy, &speedstep_freqs[0]); |
314 | } | 313 | } |
@@ -344,7 +343,8 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy) | |||
344 | return -EIO; | 343 | return -EIO; |
345 | 344 | ||
346 | dprintk("currently at %s speed setting - %i MHz\n", | 345 | dprintk("currently at %s speed setting - %i MHz\n", |
347 | (speed == speedstep_freqs[SPEEDSTEP_LOW].frequency) ? "low" : "high", | 346 | (speed == speedstep_freqs[SPEEDSTEP_LOW].frequency) |
347 | ? "low" : "high", | ||
348 | (speed / 1000)); | 348 | (speed / 1000)); |
349 | 349 | ||
350 | /* cpuinfo and default policy values */ | 350 | /* cpuinfo and default policy values */ |
@@ -352,9 +352,9 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy) | |||
352 | 352 | ||
353 | result = cpufreq_frequency_table_cpuinfo(policy, speedstep_freqs); | 353 | result = cpufreq_frequency_table_cpuinfo(policy, speedstep_freqs); |
354 | if (result) | 354 | if (result) |
355 | return (result); | 355 | return result; |
356 | 356 | ||
357 | cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); | 357 | cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); |
358 | 358 | ||
359 | return 0; | 359 | return 0; |
360 | } | 360 | } |
@@ -366,7 +366,7 @@ static int speedstep_cpu_exit(struct cpufreq_policy *policy) | |||
366 | return 0; | 366 | return 0; |
367 | } | 367 | } |
368 | 368 | ||
369 | static struct freq_attr* speedstep_attr[] = { | 369 | static struct freq_attr *speedstep_attr[] = { |
370 | &cpufreq_freq_attr_scaling_available_freqs, | 370 | &cpufreq_freq_attr_scaling_available_freqs, |
371 | NULL, | 371 | NULL, |
372 | }; | 372 | }; |
@@ -396,13 +396,15 @@ static int __init speedstep_init(void) | |||
396 | /* detect processor */ | 396 | /* detect processor */ |
397 | speedstep_processor = speedstep_detect_processor(); | 397 | speedstep_processor = speedstep_detect_processor(); |
398 | if (!speedstep_processor) { | 398 | if (!speedstep_processor) { |
399 | dprintk("Intel(R) SpeedStep(TM) capable processor not found\n"); | 399 | dprintk("Intel(R) SpeedStep(TM) capable processor " |
400 | "not found\n"); | ||
400 | return -ENODEV; | 401 | return -ENODEV; |
401 | } | 402 | } |
402 | 403 | ||
403 | /* detect chipset */ | 404 | /* detect chipset */ |
404 | if (!speedstep_detect_chipset()) { | 405 | if (!speedstep_detect_chipset()) { |
405 | dprintk("Intel(R) SpeedStep(TM) for this chipset not (yet) available.\n"); | 406 | dprintk("Intel(R) SpeedStep(TM) for this chipset not " |
407 | "(yet) available.\n"); | ||
406 | return -ENODEV; | 408 | return -ENODEV; |
407 | } | 409 | } |
408 | 410 | ||
@@ -431,9 +433,11 @@ static void __exit speedstep_exit(void) | |||
431 | } | 433 | } |
432 | 434 | ||
433 | 435 | ||
434 | MODULE_AUTHOR ("Dave Jones <davej@redhat.com>, Dominik Brodowski <linux@brodo.de>"); | 436 | MODULE_AUTHOR("Dave Jones <davej@redhat.com>, " |
435 | MODULE_DESCRIPTION ("Speedstep driver for Intel mobile processors on chipsets with ICH-M southbridges."); | 437 | "Dominik Brodowski <linux@brodo.de>"); |
436 | MODULE_LICENSE ("GPL"); | 438 | MODULE_DESCRIPTION("Speedstep driver for Intel mobile processors on chipsets " |
439 | "with ICH-M southbridges."); | ||
440 | MODULE_LICENSE("GPL"); | ||
437 | 441 | ||
438 | module_init(speedstep_init); | 442 | module_init(speedstep_init); |
439 | module_exit(speedstep_exit); | 443 | module_exit(speedstep_exit); |
diff --git a/arch/x86/kernel/cpu/cpufreq/speedstep-lib.c b/arch/x86/kernel/cpu/cpufreq/speedstep-lib.c index cdac7d62369b..2e3c6862657b 100644 --- a/arch/x86/kernel/cpu/cpufreq/speedstep-lib.c +++ b/arch/x86/kernel/cpu/cpufreq/speedstep-lib.c | |||
@@ -16,12 +16,16 @@ | |||
16 | #include <linux/slab.h> | 16 | #include <linux/slab.h> |
17 | 17 | ||
18 | #include <asm/msr.h> | 18 | #include <asm/msr.h> |
19 | #include <asm/tsc.h> | ||
19 | #include "speedstep-lib.h" | 20 | #include "speedstep-lib.h" |
20 | 21 | ||
21 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "speedstep-lib", msg) | 22 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
23 | "speedstep-lib", msg) | ||
24 | |||
25 | #define PFX "speedstep-lib: " | ||
22 | 26 | ||
23 | #ifdef CONFIG_X86_SPEEDSTEP_RELAXED_CAP_CHECK | 27 | #ifdef CONFIG_X86_SPEEDSTEP_RELAXED_CAP_CHECK |
24 | static int relaxed_check = 0; | 28 | static int relaxed_check; |
25 | #else | 29 | #else |
26 | #define relaxed_check 0 | 30 | #define relaxed_check 0 |
27 | #endif | 31 | #endif |
@@ -30,14 +34,14 @@ static int relaxed_check = 0; | |||
30 | * GET PROCESSOR CORE SPEED IN KHZ * | 34 | * GET PROCESSOR CORE SPEED IN KHZ * |
31 | *********************************************************************/ | 35 | *********************************************************************/ |
32 | 36 | ||
33 | static unsigned int pentium3_get_frequency (unsigned int processor) | 37 | static unsigned int pentium3_get_frequency(unsigned int processor) |
34 | { | 38 | { |
35 | /* See table 14 of p3_ds.pdf and table 22 of 29834003.pdf */ | 39 | /* See table 14 of p3_ds.pdf and table 22 of 29834003.pdf */ |
36 | struct { | 40 | struct { |
37 | unsigned int ratio; /* Frequency Multiplier (x10) */ | 41 | unsigned int ratio; /* Frequency Multiplier (x10) */ |
38 | u8 bitmap; /* power on configuration bits | 42 | u8 bitmap; /* power on configuration bits |
39 | [27, 25:22] (in MSR 0x2a) */ | 43 | [27, 25:22] (in MSR 0x2a) */ |
40 | } msr_decode_mult [] = { | 44 | } msr_decode_mult[] = { |
41 | { 30, 0x01 }, | 45 | { 30, 0x01 }, |
42 | { 35, 0x05 }, | 46 | { 35, 0x05 }, |
43 | { 40, 0x02 }, | 47 | { 40, 0x02 }, |
@@ -52,7 +56,7 @@ static unsigned int pentium3_get_frequency (unsigned int processor) | |||
52 | { 85, 0x26 }, | 56 | { 85, 0x26 }, |
53 | { 90, 0x20 }, | 57 | { 90, 0x20 }, |
54 | { 100, 0x2b }, | 58 | { 100, 0x2b }, |
55 | { 0, 0xff } /* error or unknown value */ | 59 | { 0, 0xff } /* error or unknown value */ |
56 | }; | 60 | }; |
57 | 61 | ||
58 | /* PIII(-M) FSB settings: see table b1-b of 24547206.pdf */ | 62 | /* PIII(-M) FSB settings: see table b1-b of 24547206.pdf */ |
@@ -60,7 +64,7 @@ static unsigned int pentium3_get_frequency (unsigned int processor) | |||
60 | unsigned int value; /* Front Side Bus speed in MHz */ | 64 | unsigned int value; /* Front Side Bus speed in MHz */ |
61 | u8 bitmap; /* power on configuration bits [18: 19] | 65 | u8 bitmap; /* power on configuration bits [18: 19] |
62 | (in MSR 0x2a) */ | 66 | (in MSR 0x2a) */ |
63 | } msr_decode_fsb [] = { | 67 | } msr_decode_fsb[] = { |
64 | { 66, 0x0 }, | 68 | { 66, 0x0 }, |
65 | { 100, 0x2 }, | 69 | { 100, 0x2 }, |
66 | { 133, 0x1 }, | 70 | { 133, 0x1 }, |
@@ -85,7 +89,7 @@ static unsigned int pentium3_get_frequency (unsigned int processor) | |||
85 | } | 89 | } |
86 | 90 | ||
87 | /* decode the multiplier */ | 91 | /* decode the multiplier */ |
88 | if (processor == SPEEDSTEP_PROCESSOR_PIII_C_EARLY) { | 92 | if (processor == SPEEDSTEP_CPU_PIII_C_EARLY) { |
89 | dprintk("workaround for early PIIIs\n"); | 93 | dprintk("workaround for early PIIIs\n"); |
90 | msr_lo &= 0x03c00000; | 94 | msr_lo &= 0x03c00000; |
91 | } else | 95 | } else |
@@ -97,9 +101,10 @@ static unsigned int pentium3_get_frequency (unsigned int processor) | |||
97 | j++; | 101 | j++; |
98 | } | 102 | } |
99 | 103 | ||
100 | dprintk("speed is %u\n", (msr_decode_mult[j].ratio * msr_decode_fsb[i].value * 100)); | 104 | dprintk("speed is %u\n", |
105 | (msr_decode_mult[j].ratio * msr_decode_fsb[i].value * 100)); | ||
101 | 106 | ||
102 | return (msr_decode_mult[j].ratio * msr_decode_fsb[i].value * 100); | 107 | return msr_decode_mult[j].ratio * msr_decode_fsb[i].value * 100; |
103 | } | 108 | } |
104 | 109 | ||
105 | 110 | ||
@@ -112,20 +117,23 @@ static unsigned int pentiumM_get_frequency(void) | |||
112 | 117 | ||
113 | /* see table B-2 of 24547212.pdf */ | 118 | /* see table B-2 of 24547212.pdf */ |
114 | if (msr_lo & 0x00040000) { | 119 | if (msr_lo & 0x00040000) { |
115 | printk(KERN_DEBUG "speedstep-lib: PM - invalid FSB: 0x%x 0x%x\n", msr_lo, msr_tmp); | 120 | printk(KERN_DEBUG PFX "PM - invalid FSB: 0x%x 0x%x\n", |
121 | msr_lo, msr_tmp); | ||
116 | return 0; | 122 | return 0; |
117 | } | 123 | } |
118 | 124 | ||
119 | msr_tmp = (msr_lo >> 22) & 0x1f; | 125 | msr_tmp = (msr_lo >> 22) & 0x1f; |
120 | dprintk("bits 22-26 are 0x%x, speed is %u\n", msr_tmp, (msr_tmp * 100 * 1000)); | 126 | dprintk("bits 22-26 are 0x%x, speed is %u\n", |
127 | msr_tmp, (msr_tmp * 100 * 1000)); | ||
121 | 128 | ||
122 | return (msr_tmp * 100 * 1000); | 129 | return msr_tmp * 100 * 1000; |
123 | } | 130 | } |
124 | 131 | ||
125 | static unsigned int pentium_core_get_frequency(void) | 132 | static unsigned int pentium_core_get_frequency(void) |
126 | { | 133 | { |
127 | u32 fsb = 0; | 134 | u32 fsb = 0; |
128 | u32 msr_lo, msr_tmp; | 135 | u32 msr_lo, msr_tmp; |
136 | int ret; | ||
129 | 137 | ||
130 | rdmsr(MSR_FSB_FREQ, msr_lo, msr_tmp); | 138 | rdmsr(MSR_FSB_FREQ, msr_lo, msr_tmp); |
131 | /* see table B-2 of 25366920.pdf */ | 139 | /* see table B-2 of 25366920.pdf */ |
@@ -153,12 +161,15 @@ static unsigned int pentium_core_get_frequency(void) | |||
153 | } | 161 | } |
154 | 162 | ||
155 | rdmsr(MSR_IA32_EBL_CR_POWERON, msr_lo, msr_tmp); | 163 | rdmsr(MSR_IA32_EBL_CR_POWERON, msr_lo, msr_tmp); |
156 | dprintk("PCORE - MSR_IA32_EBL_CR_POWERON: 0x%x 0x%x\n", msr_lo, msr_tmp); | 164 | dprintk("PCORE - MSR_IA32_EBL_CR_POWERON: 0x%x 0x%x\n", |
165 | msr_lo, msr_tmp); | ||
157 | 166 | ||
158 | msr_tmp = (msr_lo >> 22) & 0x1f; | 167 | msr_tmp = (msr_lo >> 22) & 0x1f; |
159 | dprintk("bits 22-26 are 0x%x, speed is %u\n", msr_tmp, (msr_tmp * fsb)); | 168 | dprintk("bits 22-26 are 0x%x, speed is %u\n", |
169 | msr_tmp, (msr_tmp * fsb)); | ||
160 | 170 | ||
161 | return (msr_tmp * fsb); | 171 | ret = (msr_tmp * fsb); |
172 | return ret; | ||
162 | } | 173 | } |
163 | 174 | ||
164 | 175 | ||
@@ -167,6 +178,16 @@ static unsigned int pentium4_get_frequency(void) | |||
167 | struct cpuinfo_x86 *c = &boot_cpu_data; | 178 | struct cpuinfo_x86 *c = &boot_cpu_data; |
168 | u32 msr_lo, msr_hi, mult; | 179 | u32 msr_lo, msr_hi, mult; |
169 | unsigned int fsb = 0; | 180 | unsigned int fsb = 0; |
181 | unsigned int ret; | ||
182 | u8 fsb_code; | ||
183 | |||
184 | /* Pentium 4 Model 0 and 1 do not have the Core Clock Frequency | ||
185 | * to System Bus Frequency Ratio Field in the Processor Frequency | ||
186 | * Configuration Register of the MSR. Therefore the current | ||
187 | * frequency cannot be calculated and has to be measured. | ||
188 | */ | ||
189 | if (c->x86_model < 2) | ||
190 | return cpu_khz; | ||
170 | 191 | ||
171 | rdmsr(0x2c, msr_lo, msr_hi); | 192 | rdmsr(0x2c, msr_lo, msr_hi); |
172 | 193 | ||
@@ -177,62 +198,61 @@ static unsigned int pentium4_get_frequency(void) | |||
177 | * revision #12 in Table B-1: MSRs in the Pentium 4 and | 198 | * revision #12 in Table B-1: MSRs in the Pentium 4 and |
178 | * Intel Xeon Processors, on page B-4 and B-5. | 199 | * Intel Xeon Processors, on page B-4 and B-5. |
179 | */ | 200 | */ |
180 | if (c->x86_model < 2) | 201 | fsb_code = (msr_lo >> 16) & 0x7; |
202 | switch (fsb_code) { | ||
203 | case 0: | ||
181 | fsb = 100 * 1000; | 204 | fsb = 100 * 1000; |
182 | else { | 205 | break; |
183 | u8 fsb_code = (msr_lo >> 16) & 0x7; | 206 | case 1: |
184 | switch (fsb_code) { | 207 | fsb = 13333 * 10; |
185 | case 0: | 208 | break; |
186 | fsb = 100 * 1000; | 209 | case 2: |
187 | break; | 210 | fsb = 200 * 1000; |
188 | case 1: | 211 | break; |
189 | fsb = 13333 * 10; | ||
190 | break; | ||
191 | case 2: | ||
192 | fsb = 200 * 1000; | ||
193 | break; | ||
194 | } | ||
195 | } | 212 | } |
196 | 213 | ||
197 | if (!fsb) | 214 | if (!fsb) |
198 | printk(KERN_DEBUG "speedstep-lib: couldn't detect FSB speed. Please send an e-mail to <linux@brodo.de>\n"); | 215 | printk(KERN_DEBUG PFX "couldn't detect FSB speed. " |
216 | "Please send an e-mail to <linux@brodo.de>\n"); | ||
199 | 217 | ||
200 | /* Multiplier. */ | 218 | /* Multiplier. */ |
201 | mult = msr_lo >> 24; | 219 | mult = msr_lo >> 24; |
202 | 220 | ||
203 | dprintk("P4 - FSB %u kHz; Multiplier %u; Speed %u kHz\n", fsb, mult, (fsb * mult)); | 221 | dprintk("P4 - FSB %u kHz; Multiplier %u; Speed %u kHz\n", |
222 | fsb, mult, (fsb * mult)); | ||
204 | 223 | ||
205 | return (fsb * mult); | 224 | ret = (fsb * mult); |
225 | return ret; | ||
206 | } | 226 | } |
207 | 227 | ||
208 | 228 | ||
209 | unsigned int speedstep_get_processor_frequency(unsigned int processor) | 229 | unsigned int speedstep_get_frequency(unsigned int processor) |
210 | { | 230 | { |
211 | switch (processor) { | 231 | switch (processor) { |
212 | case SPEEDSTEP_PROCESSOR_PCORE: | 232 | case SPEEDSTEP_CPU_PCORE: |
213 | return pentium_core_get_frequency(); | 233 | return pentium_core_get_frequency(); |
214 | case SPEEDSTEP_PROCESSOR_PM: | 234 | case SPEEDSTEP_CPU_PM: |
215 | return pentiumM_get_frequency(); | 235 | return pentiumM_get_frequency(); |
216 | case SPEEDSTEP_PROCESSOR_P4D: | 236 | case SPEEDSTEP_CPU_P4D: |
217 | case SPEEDSTEP_PROCESSOR_P4M: | 237 | case SPEEDSTEP_CPU_P4M: |
218 | return pentium4_get_frequency(); | 238 | return pentium4_get_frequency(); |
219 | case SPEEDSTEP_PROCESSOR_PIII_T: | 239 | case SPEEDSTEP_CPU_PIII_T: |
220 | case SPEEDSTEP_PROCESSOR_PIII_C: | 240 | case SPEEDSTEP_CPU_PIII_C: |
221 | case SPEEDSTEP_PROCESSOR_PIII_C_EARLY: | 241 | case SPEEDSTEP_CPU_PIII_C_EARLY: |
222 | return pentium3_get_frequency(processor); | 242 | return pentium3_get_frequency(processor); |
223 | default: | 243 | default: |
224 | return 0; | 244 | return 0; |
225 | }; | 245 | }; |
226 | return 0; | 246 | return 0; |
227 | } | 247 | } |
228 | EXPORT_SYMBOL_GPL(speedstep_get_processor_frequency); | 248 | EXPORT_SYMBOL_GPL(speedstep_get_frequency); |
229 | 249 | ||
230 | 250 | ||
231 | /********************************************************************* | 251 | /********************************************************************* |
232 | * DETECT SPEEDSTEP-CAPABLE PROCESSOR * | 252 | * DETECT SPEEDSTEP-CAPABLE PROCESSOR * |
233 | *********************************************************************/ | 253 | *********************************************************************/ |
234 | 254 | ||
235 | unsigned int speedstep_detect_processor (void) | 255 | unsigned int speedstep_detect_processor(void) |
236 | { | 256 | { |
237 | struct cpuinfo_x86 *c = &cpu_data(0); | 257 | struct cpuinfo_x86 *c = &cpu_data(0); |
238 | u32 ebx, msr_lo, msr_hi; | 258 | u32 ebx, msr_lo, msr_hi; |
@@ -261,7 +281,7 @@ unsigned int speedstep_detect_processor (void) | |||
261 | * sample has ebx = 0x0f, production has 0x0e. | 281 | * sample has ebx = 0x0f, production has 0x0e. |
262 | */ | 282 | */ |
263 | if ((ebx == 0x0e) || (ebx == 0x0f)) | 283 | if ((ebx == 0x0e) || (ebx == 0x0f)) |
264 | return SPEEDSTEP_PROCESSOR_P4M; | 284 | return SPEEDSTEP_CPU_P4M; |
265 | break; | 285 | break; |
266 | case 7: | 286 | case 7: |
267 | /* | 287 | /* |
@@ -272,7 +292,7 @@ unsigned int speedstep_detect_processor (void) | |||
272 | * samples are only of B-stepping... | 292 | * samples are only of B-stepping... |
273 | */ | 293 | */ |
274 | if (ebx == 0x0e) | 294 | if (ebx == 0x0e) |
275 | return SPEEDSTEP_PROCESSOR_P4M; | 295 | return SPEEDSTEP_CPU_P4M; |
276 | break; | 296 | break; |
277 | case 9: | 297 | case 9: |
278 | /* | 298 | /* |
@@ -288,10 +308,13 @@ unsigned int speedstep_detect_processor (void) | |||
288 | * M-P4-Ms may have either ebx=0xe or 0xf [see above] | 308 | * M-P4-Ms may have either ebx=0xe or 0xf [see above] |
289 | * M-P4/533 have either ebx=0xe or 0xf. [25317607.pdf] | 309 | * M-P4/533 have either ebx=0xe or 0xf. [25317607.pdf] |
290 | * also, M-P4M HTs have ebx=0x8, too | 310 | * also, M-P4M HTs have ebx=0x8, too |
291 | * For now, they are distinguished by the model_id string | 311 | * For now, they are distinguished by the model_id |
312 | * string | ||
292 | */ | 313 | */ |
293 | if ((ebx == 0x0e) || (strstr(c->x86_model_id,"Mobile Intel(R) Pentium(R) 4") != NULL)) | 314 | if ((ebx == 0x0e) || |
294 | return SPEEDSTEP_PROCESSOR_P4M; | 315 | (strstr(c->x86_model_id, |
316 | "Mobile Intel(R) Pentium(R) 4") != NULL)) | ||
317 | return SPEEDSTEP_CPU_P4M; | ||
295 | break; | 318 | break; |
296 | default: | 319 | default: |
297 | break; | 320 | break; |
@@ -301,7 +324,8 @@ unsigned int speedstep_detect_processor (void) | |||
301 | 324 | ||
302 | switch (c->x86_model) { | 325 | switch (c->x86_model) { |
303 | case 0x0B: /* Intel PIII [Tualatin] */ | 326 | case 0x0B: /* Intel PIII [Tualatin] */ |
304 | /* cpuid_ebx(1) is 0x04 for desktop PIII, 0x06 for mobile PIII-M */ | 327 | /* cpuid_ebx(1) is 0x04 for desktop PIII, |
328 | * 0x06 for mobile PIII-M */ | ||
305 | ebx = cpuid_ebx(0x00000001); | 329 | ebx = cpuid_ebx(0x00000001); |
306 | dprintk("ebx is %x\n", ebx); | 330 | dprintk("ebx is %x\n", ebx); |
307 | 331 | ||
@@ -313,14 +337,15 @@ unsigned int speedstep_detect_processor (void) | |||
313 | /* So far all PIII-M processors support SpeedStep. See | 337 | /* So far all PIII-M processors support SpeedStep. See |
314 | * Intel's 24540640.pdf of June 2003 | 338 | * Intel's 24540640.pdf of June 2003 |
315 | */ | 339 | */ |
316 | return SPEEDSTEP_PROCESSOR_PIII_T; | 340 | return SPEEDSTEP_CPU_PIII_T; |
317 | 341 | ||
318 | case 0x08: /* Intel PIII [Coppermine] */ | 342 | case 0x08: /* Intel PIII [Coppermine] */ |
319 | 343 | ||
320 | /* all mobile PIII Coppermines have FSB 100 MHz | 344 | /* all mobile PIII Coppermines have FSB 100 MHz |
321 | * ==> sort out a few desktop PIIIs. */ | 345 | * ==> sort out a few desktop PIIIs. */ |
322 | rdmsr(MSR_IA32_EBL_CR_POWERON, msr_lo, msr_hi); | 346 | rdmsr(MSR_IA32_EBL_CR_POWERON, msr_lo, msr_hi); |
323 | dprintk("Coppermine: MSR_IA32_EBL_CR_POWERON is 0x%x, 0x%x\n", msr_lo, msr_hi); | 347 | dprintk("Coppermine: MSR_IA32_EBL_CR_POWERON is 0x%x, 0x%x\n", |
348 | msr_lo, msr_hi); | ||
324 | msr_lo &= 0x00c0000; | 349 | msr_lo &= 0x00c0000; |
325 | if (msr_lo != 0x0080000) | 350 | if (msr_lo != 0x0080000) |
326 | return 0; | 351 | return 0; |
@@ -332,13 +357,15 @@ unsigned int speedstep_detect_processor (void) | |||
332 | * bit 56 or 57 is set | 357 | * bit 56 or 57 is set |
333 | */ | 358 | */ |
334 | rdmsr(MSR_IA32_PLATFORM_ID, msr_lo, msr_hi); | 359 | rdmsr(MSR_IA32_PLATFORM_ID, msr_lo, msr_hi); |
335 | dprintk("Coppermine: MSR_IA32_PLATFORM ID is 0x%x, 0x%x\n", msr_lo, msr_hi); | 360 | dprintk("Coppermine: MSR_IA32_PLATFORM ID is 0x%x, 0x%x\n", |
336 | if ((msr_hi & (1<<18)) && (relaxed_check ? 1 : (msr_hi & (3<<24)))) { | 361 | msr_lo, msr_hi); |
362 | if ((msr_hi & (1<<18)) && | ||
363 | (relaxed_check ? 1 : (msr_hi & (3<<24)))) { | ||
337 | if (c->x86_mask == 0x01) { | 364 | if (c->x86_mask == 0x01) { |
338 | dprintk("early PIII version\n"); | 365 | dprintk("early PIII version\n"); |
339 | return SPEEDSTEP_PROCESSOR_PIII_C_EARLY; | 366 | return SPEEDSTEP_CPU_PIII_C_EARLY; |
340 | } else | 367 | } else |
341 | return SPEEDSTEP_PROCESSOR_PIII_C; | 368 | return SPEEDSTEP_CPU_PIII_C; |
342 | } | 369 | } |
343 | 370 | ||
344 | default: | 371 | default: |
@@ -369,7 +396,7 @@ unsigned int speedstep_get_freqs(unsigned int processor, | |||
369 | dprintk("trying to determine both speeds\n"); | 396 | dprintk("trying to determine both speeds\n"); |
370 | 397 | ||
371 | /* get current speed */ | 398 | /* get current speed */ |
372 | prev_speed = speedstep_get_processor_frequency(processor); | 399 | prev_speed = speedstep_get_frequency(processor); |
373 | if (!prev_speed) | 400 | if (!prev_speed) |
374 | return -EIO; | 401 | return -EIO; |
375 | 402 | ||
@@ -379,7 +406,7 @@ unsigned int speedstep_get_freqs(unsigned int processor, | |||
379 | 406 | ||
380 | /* switch to low state */ | 407 | /* switch to low state */ |
381 | set_state(SPEEDSTEP_LOW); | 408 | set_state(SPEEDSTEP_LOW); |
382 | *low_speed = speedstep_get_processor_frequency(processor); | 409 | *low_speed = speedstep_get_frequency(processor); |
383 | if (!*low_speed) { | 410 | if (!*low_speed) { |
384 | ret = -EIO; | 411 | ret = -EIO; |
385 | goto out; | 412 | goto out; |
@@ -398,7 +425,7 @@ unsigned int speedstep_get_freqs(unsigned int processor, | |||
398 | if (transition_latency) | 425 | if (transition_latency) |
399 | do_gettimeofday(&tv2); | 426 | do_gettimeofday(&tv2); |
400 | 427 | ||
401 | *high_speed = speedstep_get_processor_frequency(processor); | 428 | *high_speed = speedstep_get_frequency(processor); |
402 | if (!*high_speed) { | 429 | if (!*high_speed) { |
403 | ret = -EIO; | 430 | ret = -EIO; |
404 | goto out; | 431 | goto out; |
@@ -426,9 +453,12 @@ unsigned int speedstep_get_freqs(unsigned int processor, | |||
426 | /* check if the latency measurement is too high or too low | 453 | /* check if the latency measurement is too high or too low |
427 | * and set it to a safe value (500uSec) in that case | 454 | * and set it to a safe value (500uSec) in that case |
428 | */ | 455 | */ |
429 | if (*transition_latency > 10000000 || *transition_latency < 50000) { | 456 | if (*transition_latency > 10000000 || |
430 | printk (KERN_WARNING "speedstep: frequency transition measured seems out of " | 457 | *transition_latency < 50000) { |
431 | "range (%u nSec), falling back to a safe one of %u nSec.\n", | 458 | printk(KERN_WARNING PFX "frequency transition " |
459 | "measured seems out of range (%u " | ||
460 | "nSec), falling back to a safe one of" | ||
461 | "%u nSec.\n", | ||
432 | *transition_latency, 500000); | 462 | *transition_latency, 500000); |
433 | *transition_latency = 500000; | 463 | *transition_latency = 500000; |
434 | } | 464 | } |
@@ -436,15 +466,16 @@ unsigned int speedstep_get_freqs(unsigned int processor, | |||
436 | 466 | ||
437 | out: | 467 | out: |
438 | local_irq_restore(flags); | 468 | local_irq_restore(flags); |
439 | return (ret); | 469 | return ret; |
440 | } | 470 | } |
441 | EXPORT_SYMBOL_GPL(speedstep_get_freqs); | 471 | EXPORT_SYMBOL_GPL(speedstep_get_freqs); |
442 | 472 | ||
443 | #ifdef CONFIG_X86_SPEEDSTEP_RELAXED_CAP_CHECK | 473 | #ifdef CONFIG_X86_SPEEDSTEP_RELAXED_CAP_CHECK |
444 | module_param(relaxed_check, int, 0444); | 474 | module_param(relaxed_check, int, 0444); |
445 | MODULE_PARM_DESC(relaxed_check, "Don't do all checks for speedstep capability."); | 475 | MODULE_PARM_DESC(relaxed_check, |
476 | "Don't do all checks for speedstep capability."); | ||
446 | #endif | 477 | #endif |
447 | 478 | ||
448 | MODULE_AUTHOR ("Dominik Brodowski <linux@brodo.de>"); | 479 | MODULE_AUTHOR("Dominik Brodowski <linux@brodo.de>"); |
449 | MODULE_DESCRIPTION ("Library for Intel SpeedStep 1 or 2 cpufreq drivers."); | 480 | MODULE_DESCRIPTION("Library for Intel SpeedStep 1 or 2 cpufreq drivers."); |
450 | MODULE_LICENSE ("GPL"); | 481 | MODULE_LICENSE("GPL"); |
diff --git a/arch/x86/kernel/cpu/cpufreq/speedstep-lib.h b/arch/x86/kernel/cpu/cpufreq/speedstep-lib.h index b11bcc608cac..2b6c04e5a304 100644 --- a/arch/x86/kernel/cpu/cpufreq/speedstep-lib.h +++ b/arch/x86/kernel/cpu/cpufreq/speedstep-lib.h | |||
@@ -12,17 +12,17 @@ | |||
12 | 12 | ||
13 | /* processors */ | 13 | /* processors */ |
14 | 14 | ||
15 | #define SPEEDSTEP_PROCESSOR_PIII_C_EARLY 0x00000001 /* Coppermine core */ | 15 | #define SPEEDSTEP_CPU_PIII_C_EARLY 0x00000001 /* Coppermine core */ |
16 | #define SPEEDSTEP_PROCESSOR_PIII_C 0x00000002 /* Coppermine core */ | 16 | #define SPEEDSTEP_CPU_PIII_C 0x00000002 /* Coppermine core */ |
17 | #define SPEEDSTEP_PROCESSOR_PIII_T 0x00000003 /* Tualatin core */ | 17 | #define SPEEDSTEP_CPU_PIII_T 0x00000003 /* Tualatin core */ |
18 | #define SPEEDSTEP_PROCESSOR_P4M 0x00000004 /* P4-M */ | 18 | #define SPEEDSTEP_CPU_P4M 0x00000004 /* P4-M */ |
19 | 19 | ||
20 | /* the following processors are not speedstep-capable and are not auto-detected | 20 | /* the following processors are not speedstep-capable and are not auto-detected |
21 | * in speedstep_detect_processor(). However, their speed can be detected using | 21 | * in speedstep_detect_processor(). However, their speed can be detected using |
22 | * the speedstep_get_processor_frequency() call. */ | 22 | * the speedstep_get_frequency() call. */ |
23 | #define SPEEDSTEP_PROCESSOR_PM 0xFFFFFF03 /* Pentium M */ | 23 | #define SPEEDSTEP_CPU_PM 0xFFFFFF03 /* Pentium M */ |
24 | #define SPEEDSTEP_PROCESSOR_P4D 0xFFFFFF04 /* desktop P4 */ | 24 | #define SPEEDSTEP_CPU_P4D 0xFFFFFF04 /* desktop P4 */ |
25 | #define SPEEDSTEP_PROCESSOR_PCORE 0xFFFFFF05 /* Core */ | 25 | #define SPEEDSTEP_CPU_PCORE 0xFFFFFF05 /* Core */ |
26 | 26 | ||
27 | /* speedstep states -- only two of them */ | 27 | /* speedstep states -- only two of them */ |
28 | 28 | ||
@@ -34,7 +34,7 @@ | |||
34 | extern unsigned int speedstep_detect_processor (void); | 34 | extern unsigned int speedstep_detect_processor (void); |
35 | 35 | ||
36 | /* detect the current speed (in khz) of the processor */ | 36 | /* detect the current speed (in khz) of the processor */ |
37 | extern unsigned int speedstep_get_processor_frequency(unsigned int processor); | 37 | extern unsigned int speedstep_get_frequency(unsigned int processor); |
38 | 38 | ||
39 | 39 | ||
40 | /* detect the low and high speeds of the processor. The callback | 40 | /* detect the low and high speeds of the processor. The callback |
diff --git a/arch/x86/kernel/cpu/cpufreq/speedstep-smi.c b/arch/x86/kernel/cpu/cpufreq/speedstep-smi.c index 8a85c93bd62a..befea088e4f5 100644 --- a/arch/x86/kernel/cpu/cpufreq/speedstep-smi.c +++ b/arch/x86/kernel/cpu/cpufreq/speedstep-smi.c | |||
@@ -19,8 +19,8 @@ | |||
19 | #include <linux/cpufreq.h> | 19 | #include <linux/cpufreq.h> |
20 | #include <linux/slab.h> | 20 | #include <linux/slab.h> |
21 | #include <linux/delay.h> | 21 | #include <linux/delay.h> |
22 | #include <linux/io.h> | ||
22 | #include <asm/ist.h> | 23 | #include <asm/ist.h> |
23 | #include <asm/io.h> | ||
24 | 24 | ||
25 | #include "speedstep-lib.h" | 25 | #include "speedstep-lib.h" |
26 | 26 | ||
@@ -30,12 +30,12 @@ | |||
30 | * If user gives it, these are used. | 30 | * If user gives it, these are used. |
31 | * | 31 | * |
32 | */ | 32 | */ |
33 | static int smi_port = 0; | 33 | static int smi_port; |
34 | static int smi_cmd = 0; | 34 | static int smi_cmd; |
35 | static unsigned int smi_sig = 0; | 35 | static unsigned int smi_sig; |
36 | 36 | ||
37 | /* info about the processor */ | 37 | /* info about the processor */ |
38 | static unsigned int speedstep_processor = 0; | 38 | static unsigned int speedstep_processor; |
39 | 39 | ||
40 | /* | 40 | /* |
41 | * There are only two frequency states for each processor. Values | 41 | * There are only two frequency states for each processor. Values |
@@ -56,12 +56,13 @@ static struct cpufreq_frequency_table speedstep_freqs[] = { | |||
56 | * of DMA activity going on? */ | 56 | * of DMA activity going on? */ |
57 | #define SMI_TRIES 5 | 57 | #define SMI_TRIES 5 |
58 | 58 | ||
59 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, "speedstep-smi", msg) | 59 | #define dprintk(msg...) cpufreq_debug_printk(CPUFREQ_DEBUG_DRIVER, \ |
60 | "speedstep-smi", msg) | ||
60 | 61 | ||
61 | /** | 62 | /** |
62 | * speedstep_smi_ownership | 63 | * speedstep_smi_ownership |
63 | */ | 64 | */ |
64 | static int speedstep_smi_ownership (void) | 65 | static int speedstep_smi_ownership(void) |
65 | { | 66 | { |
66 | u32 command, result, magic, dummy; | 67 | u32 command, result, magic, dummy; |
67 | u32 function = GET_SPEEDSTEP_OWNER; | 68 | u32 function = GET_SPEEDSTEP_OWNER; |
@@ -70,16 +71,18 @@ static int speedstep_smi_ownership (void) | |||
70 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); | 71 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); |
71 | magic = virt_to_phys(magic_data); | 72 | magic = virt_to_phys(magic_data); |
72 | 73 | ||
73 | dprintk("trying to obtain ownership with command %x at port %x\n", command, smi_port); | 74 | dprintk("trying to obtain ownership with command %x at port %x\n", |
75 | command, smi_port); | ||
74 | 76 | ||
75 | __asm__ __volatile__( | 77 | __asm__ __volatile__( |
76 | "push %%ebp\n" | 78 | "push %%ebp\n" |
77 | "out %%al, (%%dx)\n" | 79 | "out %%al, (%%dx)\n" |
78 | "pop %%ebp\n" | 80 | "pop %%ebp\n" |
79 | : "=D" (result), "=a" (dummy), "=b" (dummy), "=c" (dummy), "=d" (dummy), | 81 | : "=D" (result), |
80 | "=S" (dummy) | 82 | "=a" (dummy), "=b" (dummy), "=c" (dummy), "=d" (dummy), |
83 | "=S" (dummy) | ||
81 | : "a" (command), "b" (function), "c" (0), "d" (smi_port), | 84 | : "a" (command), "b" (function), "c" (0), "d" (smi_port), |
82 | "D" (0), "S" (magic) | 85 | "D" (0), "S" (magic) |
83 | : "memory" | 86 | : "memory" |
84 | ); | 87 | ); |
85 | 88 | ||
@@ -97,10 +100,10 @@ static int speedstep_smi_ownership (void) | |||
97 | * even hangs [cf. bugme.osdl.org # 1422] on earlier systems. Empirical testing | 100 | * even hangs [cf. bugme.osdl.org # 1422] on earlier systems. Empirical testing |
98 | * shows that the latter occurs if !(ist_info.event & 0xFFFF). | 101 | * shows that the latter occurs if !(ist_info.event & 0xFFFF). |
99 | */ | 102 | */ |
100 | static int speedstep_smi_get_freqs (unsigned int *low, unsigned int *high) | 103 | static int speedstep_smi_get_freqs(unsigned int *low, unsigned int *high) |
101 | { | 104 | { |
102 | u32 command, result = 0, edi, high_mhz, low_mhz, dummy; | 105 | u32 command, result = 0, edi, high_mhz, low_mhz, dummy; |
103 | u32 state=0; | 106 | u32 state = 0; |
104 | u32 function = GET_SPEEDSTEP_FREQS; | 107 | u32 function = GET_SPEEDSTEP_FREQS; |
105 | 108 | ||
106 | if (!(ist_info.event & 0xFFFF)) { | 109 | if (!(ist_info.event & 0xFFFF)) { |
@@ -110,17 +113,25 @@ static int speedstep_smi_get_freqs (unsigned int *low, unsigned int *high) | |||
110 | 113 | ||
111 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); | 114 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); |
112 | 115 | ||
113 | dprintk("trying to determine frequencies with command %x at port %x\n", command, smi_port); | 116 | dprintk("trying to determine frequencies with command %x at port %x\n", |
117 | command, smi_port); | ||
114 | 118 | ||
115 | __asm__ __volatile__( | 119 | __asm__ __volatile__( |
116 | "push %%ebp\n" | 120 | "push %%ebp\n" |
117 | "out %%al, (%%dx)\n" | 121 | "out %%al, (%%dx)\n" |
118 | "pop %%ebp" | 122 | "pop %%ebp" |
119 | : "=a" (result), "=b" (high_mhz), "=c" (low_mhz), "=d" (state), "=D" (edi), "=S" (dummy) | 123 | : "=a" (result), |
120 | : "a" (command), "b" (function), "c" (state), "d" (smi_port), "S" (0), "D" (0) | 124 | "=b" (high_mhz), |
125 | "=c" (low_mhz), | ||
126 | "=d" (state), "=D" (edi), "=S" (dummy) | ||
127 | : "a" (command), | ||
128 | "b" (function), | ||
129 | "c" (state), | ||
130 | "d" (smi_port), "S" (0), "D" (0) | ||
121 | ); | 131 | ); |
122 | 132 | ||
123 | dprintk("result %x, low_freq %u, high_freq %u\n", result, low_mhz, high_mhz); | 133 | dprintk("result %x, low_freq %u, high_freq %u\n", |
134 | result, low_mhz, high_mhz); | ||
124 | 135 | ||
125 | /* abort if results are obviously incorrect... */ | 136 | /* abort if results are obviously incorrect... */ |
126 | if ((high_mhz + low_mhz) < 600) | 137 | if ((high_mhz + low_mhz) < 600) |
@@ -137,26 +148,30 @@ static int speedstep_smi_get_freqs (unsigned int *low, unsigned int *high) | |||
137 | * @state: processor frequency state (SPEEDSTEP_LOW or SPEEDSTEP_HIGH) | 148 | * @state: processor frequency state (SPEEDSTEP_LOW or SPEEDSTEP_HIGH) |
138 | * | 149 | * |
139 | */ | 150 | */ |
140 | static int speedstep_get_state (void) | 151 | static int speedstep_get_state(void) |
141 | { | 152 | { |
142 | u32 function=GET_SPEEDSTEP_STATE; | 153 | u32 function = GET_SPEEDSTEP_STATE; |
143 | u32 result, state, edi, command, dummy; | 154 | u32 result, state, edi, command, dummy; |
144 | 155 | ||
145 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); | 156 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); |
146 | 157 | ||
147 | dprintk("trying to determine current setting with command %x at port %x\n", command, smi_port); | 158 | dprintk("trying to determine current setting with command %x " |
159 | "at port %x\n", command, smi_port); | ||
148 | 160 | ||
149 | __asm__ __volatile__( | 161 | __asm__ __volatile__( |
150 | "push %%ebp\n" | 162 | "push %%ebp\n" |
151 | "out %%al, (%%dx)\n" | 163 | "out %%al, (%%dx)\n" |
152 | "pop %%ebp\n" | 164 | "pop %%ebp\n" |
153 | : "=a" (result), "=b" (state), "=D" (edi), "=c" (dummy), "=d" (dummy), "=S" (dummy) | 165 | : "=a" (result), |
154 | : "a" (command), "b" (function), "c" (0), "d" (smi_port), "S" (0), "D" (0) | 166 | "=b" (state), "=D" (edi), |
167 | "=c" (dummy), "=d" (dummy), "=S" (dummy) | ||
168 | : "a" (command), "b" (function), "c" (0), | ||
169 | "d" (smi_port), "S" (0), "D" (0) | ||
155 | ); | 170 | ); |
156 | 171 | ||
157 | dprintk("state is %x, result is %x\n", state, result); | 172 | dprintk("state is %x, result is %x\n", state, result); |
158 | 173 | ||
159 | return (state & 1); | 174 | return state & 1; |
160 | } | 175 | } |
161 | 176 | ||
162 | 177 | ||
@@ -165,11 +180,11 @@ static int speedstep_get_state (void) | |||
165 | * @state: new processor frequency state (SPEEDSTEP_LOW or SPEEDSTEP_HIGH) | 180 | * @state: new processor frequency state (SPEEDSTEP_LOW or SPEEDSTEP_HIGH) |
166 | * | 181 | * |
167 | */ | 182 | */ |
168 | static void speedstep_set_state (unsigned int state) | 183 | static void speedstep_set_state(unsigned int state) |
169 | { | 184 | { |
170 | unsigned int result = 0, command, new_state, dummy; | 185 | unsigned int result = 0, command, new_state, dummy; |
171 | unsigned long flags; | 186 | unsigned long flags; |
172 | unsigned int function=SET_SPEEDSTEP_STATE; | 187 | unsigned int function = SET_SPEEDSTEP_STATE; |
173 | unsigned int retry = 0; | 188 | unsigned int retry = 0; |
174 | 189 | ||
175 | if (state > 0x1) | 190 | if (state > 0x1) |
@@ -180,11 +195,14 @@ static void speedstep_set_state (unsigned int state) | |||
180 | 195 | ||
181 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); | 196 | command = (smi_sig & 0xffffff00) | (smi_cmd & 0xff); |
182 | 197 | ||
183 | dprintk("trying to set frequency to state %u with command %x at port %x\n", state, command, smi_port); | 198 | dprintk("trying to set frequency to state %u " |
199 | "with command %x at port %x\n", | ||
200 | state, command, smi_port); | ||
184 | 201 | ||
185 | do { | 202 | do { |
186 | if (retry) { | 203 | if (retry) { |
187 | dprintk("retry %u, previous result %u, waiting...\n", retry, result); | 204 | dprintk("retry %u, previous result %u, waiting...\n", |
205 | retry, result); | ||
188 | mdelay(retry * 50); | 206 | mdelay(retry * 50); |
189 | } | 207 | } |
190 | retry++; | 208 | retry++; |
@@ -192,20 +210,26 @@ static void speedstep_set_state (unsigned int state) | |||
192 | "push %%ebp\n" | 210 | "push %%ebp\n" |
193 | "out %%al, (%%dx)\n" | 211 | "out %%al, (%%dx)\n" |
194 | "pop %%ebp" | 212 | "pop %%ebp" |
195 | : "=b" (new_state), "=D" (result), "=c" (dummy), "=a" (dummy), | 213 | : "=b" (new_state), "=D" (result), |
196 | "=d" (dummy), "=S" (dummy) | 214 | "=c" (dummy), "=a" (dummy), |
197 | : "a" (command), "b" (function), "c" (state), "d" (smi_port), "S" (0), "D" (0) | 215 | "=d" (dummy), "=S" (dummy) |
216 | : "a" (command), "b" (function), "c" (state), | ||
217 | "d" (smi_port), "S" (0), "D" (0) | ||
198 | ); | 218 | ); |
199 | } while ((new_state != state) && (retry <= SMI_TRIES)); | 219 | } while ((new_state != state) && (retry <= SMI_TRIES)); |
200 | 220 | ||
201 | /* enable IRQs */ | 221 | /* enable IRQs */ |
202 | local_irq_restore(flags); | 222 | local_irq_restore(flags); |
203 | 223 | ||
204 | if (new_state == state) { | 224 | if (new_state == state) |
205 | dprintk("change to %u MHz succeeded after %u tries with result %u\n", (speedstep_freqs[new_state].frequency / 1000), retry, result); | 225 | dprintk("change to %u MHz succeeded after %u tries " |
206 | } else { | 226 | "with result %u\n", |
207 | printk(KERN_ERR "cpufreq: change to state %u failed with new_state %u and result %u\n", state, new_state, result); | 227 | (speedstep_freqs[new_state].frequency / 1000), |
208 | } | 228 | retry, result); |
229 | else | ||
230 | printk(KERN_ERR "cpufreq: change to state %u " | ||
231 | "failed with new_state %u and result %u\n", | ||
232 | state, new_state, result); | ||
209 | 233 | ||
210 | return; | 234 | return; |
211 | } | 235 | } |
@@ -219,13 +243,14 @@ static void speedstep_set_state (unsigned int state) | |||
219 | * | 243 | * |
220 | * Sets a new CPUFreq policy/freq. | 244 | * Sets a new CPUFreq policy/freq. |
221 | */ | 245 | */ |
222 | static int speedstep_target (struct cpufreq_policy *policy, | 246 | static int speedstep_target(struct cpufreq_policy *policy, |
223 | unsigned int target_freq, unsigned int relation) | 247 | unsigned int target_freq, unsigned int relation) |
224 | { | 248 | { |
225 | unsigned int newstate = 0; | 249 | unsigned int newstate = 0; |
226 | struct cpufreq_freqs freqs; | 250 | struct cpufreq_freqs freqs; |
227 | 251 | ||
228 | if (cpufreq_frequency_table_target(policy, &speedstep_freqs[0], target_freq, relation, &newstate)) | 252 | if (cpufreq_frequency_table_target(policy, &speedstep_freqs[0], |
253 | target_freq, relation, &newstate)) | ||
229 | return -EINVAL; | 254 | return -EINVAL; |
230 | 255 | ||
231 | freqs.old = speedstep_freqs[speedstep_get_state()].frequency; | 256 | freqs.old = speedstep_freqs[speedstep_get_state()].frequency; |
@@ -250,7 +275,7 @@ static int speedstep_target (struct cpufreq_policy *policy, | |||
250 | * Limit must be within speedstep_low_freq and speedstep_high_freq, with | 275 | * Limit must be within speedstep_low_freq and speedstep_high_freq, with |
251 | * at least one border included. | 276 | * at least one border included. |
252 | */ | 277 | */ |
253 | static int speedstep_verify (struct cpufreq_policy *policy) | 278 | static int speedstep_verify(struct cpufreq_policy *policy) |
254 | { | 279 | { |
255 | return cpufreq_frequency_table_verify(policy, &speedstep_freqs[0]); | 280 | return cpufreq_frequency_table_verify(policy, &speedstep_freqs[0]); |
256 | } | 281 | } |
@@ -259,7 +284,8 @@ static int speedstep_verify (struct cpufreq_policy *policy) | |||
259 | static int speedstep_cpu_init(struct cpufreq_policy *policy) | 284 | static int speedstep_cpu_init(struct cpufreq_policy *policy) |
260 | { | 285 | { |
261 | int result; | 286 | int result; |
262 | unsigned int speed,state; | 287 | unsigned int speed, state; |
288 | unsigned int *low, *high; | ||
263 | 289 | ||
264 | /* capability check */ | 290 | /* capability check */ |
265 | if (policy->cpu != 0) | 291 | if (policy->cpu != 0) |
@@ -272,19 +298,23 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy) | |||
272 | } | 298 | } |
273 | 299 | ||
274 | /* detect low and high frequency */ | 300 | /* detect low and high frequency */ |
275 | result = speedstep_smi_get_freqs(&speedstep_freqs[SPEEDSTEP_LOW].frequency, | 301 | low = &speedstep_freqs[SPEEDSTEP_LOW].frequency; |
276 | &speedstep_freqs[SPEEDSTEP_HIGH].frequency); | 302 | high = &speedstep_freqs[SPEEDSTEP_HIGH].frequency; |
303 | |||
304 | result = speedstep_smi_get_freqs(low, high); | ||
277 | if (result) { | 305 | if (result) { |
278 | /* fall back to speedstep_lib.c dection mechanism: try both states out */ | 306 | /* fall back to speedstep_lib.c dection mechanism: |
279 | dprintk("could not detect low and high frequencies by SMI call.\n"); | 307 | * try both states out */ |
308 | dprintk("could not detect low and high frequencies " | ||
309 | "by SMI call.\n"); | ||
280 | result = speedstep_get_freqs(speedstep_processor, | 310 | result = speedstep_get_freqs(speedstep_processor, |
281 | &speedstep_freqs[SPEEDSTEP_LOW].frequency, | 311 | low, high, |
282 | &speedstep_freqs[SPEEDSTEP_HIGH].frequency, | ||
283 | NULL, | 312 | NULL, |
284 | &speedstep_set_state); | 313 | &speedstep_set_state); |
285 | 314 | ||
286 | if (result) { | 315 | if (result) { |
287 | dprintk("could not detect two different speeds -- aborting.\n"); | 316 | dprintk("could not detect two different speeds" |
317 | " -- aborting.\n"); | ||
288 | return result; | 318 | return result; |
289 | } else | 319 | } else |
290 | dprintk("workaround worked.\n"); | 320 | dprintk("workaround worked.\n"); |
@@ -295,7 +325,8 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy) | |||
295 | speed = speedstep_freqs[state].frequency; | 325 | speed = speedstep_freqs[state].frequency; |
296 | 326 | ||
297 | dprintk("currently at %s speed setting - %i MHz\n", | 327 | dprintk("currently at %s speed setting - %i MHz\n", |
298 | (speed == speedstep_freqs[SPEEDSTEP_LOW].frequency) ? "low" : "high", | 328 | (speed == speedstep_freqs[SPEEDSTEP_LOW].frequency) |
329 | ? "low" : "high", | ||
299 | (speed / 1000)); | 330 | (speed / 1000)); |
300 | 331 | ||
301 | /* cpuinfo and default policy values */ | 332 | /* cpuinfo and default policy values */ |
@@ -304,7 +335,7 @@ static int speedstep_cpu_init(struct cpufreq_policy *policy) | |||
304 | 335 | ||
305 | result = cpufreq_frequency_table_cpuinfo(policy, speedstep_freqs); | 336 | result = cpufreq_frequency_table_cpuinfo(policy, speedstep_freqs); |
306 | if (result) | 337 | if (result) |
307 | return (result); | 338 | return result; |
308 | 339 | ||
309 | cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); | 340 | cpufreq_frequency_table_get_attr(speedstep_freqs, policy->cpu); |
310 | 341 | ||
@@ -321,7 +352,7 @@ static unsigned int speedstep_get(unsigned int cpu) | |||
321 | { | 352 | { |
322 | if (cpu) | 353 | if (cpu) |
323 | return -ENODEV; | 354 | return -ENODEV; |
324 | return speedstep_get_processor_frequency(speedstep_processor); | 355 | return speedstep_get_frequency(speedstep_processor); |
325 | } | 356 | } |
326 | 357 | ||
327 | 358 | ||
@@ -335,7 +366,7 @@ static int speedstep_resume(struct cpufreq_policy *policy) | |||
335 | return result; | 366 | return result; |
336 | } | 367 | } |
337 | 368 | ||
338 | static struct freq_attr* speedstep_attr[] = { | 369 | static struct freq_attr *speedstep_attr[] = { |
339 | &cpufreq_freq_attr_scaling_available_freqs, | 370 | &cpufreq_freq_attr_scaling_available_freqs, |
340 | NULL, | 371 | NULL, |
341 | }; | 372 | }; |
@@ -364,21 +395,23 @@ static int __init speedstep_init(void) | |||
364 | speedstep_processor = speedstep_detect_processor(); | 395 | speedstep_processor = speedstep_detect_processor(); |
365 | 396 | ||
366 | switch (speedstep_processor) { | 397 | switch (speedstep_processor) { |
367 | case SPEEDSTEP_PROCESSOR_PIII_T: | 398 | case SPEEDSTEP_CPU_PIII_T: |
368 | case SPEEDSTEP_PROCESSOR_PIII_C: | 399 | case SPEEDSTEP_CPU_PIII_C: |
369 | case SPEEDSTEP_PROCESSOR_PIII_C_EARLY: | 400 | case SPEEDSTEP_CPU_PIII_C_EARLY: |
370 | break; | 401 | break; |
371 | default: | 402 | default: |
372 | speedstep_processor = 0; | 403 | speedstep_processor = 0; |
373 | } | 404 | } |
374 | 405 | ||
375 | if (!speedstep_processor) { | 406 | if (!speedstep_processor) { |
376 | dprintk ("No supported Intel CPU detected.\n"); | 407 | dprintk("No supported Intel CPU detected.\n"); |
377 | return -ENODEV; | 408 | return -ENODEV; |
378 | } | 409 | } |
379 | 410 | ||
380 | dprintk("signature:0x%.8lx, command:0x%.8lx, event:0x%.8lx, perf_level:0x%.8lx.\n", | 411 | dprintk("signature:0x%.8lx, command:0x%.8lx, " |
381 | ist_info.signature, ist_info.command, ist_info.event, ist_info.perf_level); | 412 | "event:0x%.8lx, perf_level:0x%.8lx.\n", |
413 | ist_info.signature, ist_info.command, | ||
414 | ist_info.event, ist_info.perf_level); | ||
382 | 415 | ||
383 | /* Error if no IST-SMI BIOS or no PARM | 416 | /* Error if no IST-SMI BIOS or no PARM |
384 | sig= 'ISGE' aka 'Intel Speedstep Gate E' */ | 417 | sig= 'ISGE' aka 'Intel Speedstep Gate E' */ |
@@ -416,17 +449,20 @@ static void __exit speedstep_exit(void) | |||
416 | cpufreq_unregister_driver(&speedstep_driver); | 449 | cpufreq_unregister_driver(&speedstep_driver); |
417 | } | 450 | } |
418 | 451 | ||
419 | module_param(smi_port, int, 0444); | 452 | module_param(smi_port, int, 0444); |
420 | module_param(smi_cmd, int, 0444); | 453 | module_param(smi_cmd, int, 0444); |
421 | module_param(smi_sig, uint, 0444); | 454 | module_param(smi_sig, uint, 0444); |
422 | 455 | ||
423 | MODULE_PARM_DESC(smi_port, "Override the BIOS-given IST port with this value -- Intel's default setting is 0xb2"); | 456 | MODULE_PARM_DESC(smi_port, "Override the BIOS-given IST port with this value " |
424 | MODULE_PARM_DESC(smi_cmd, "Override the BIOS-given IST command with this value -- Intel's default setting is 0x82"); | 457 | "-- Intel's default setting is 0xb2"); |
425 | MODULE_PARM_DESC(smi_sig, "Set to 1 to fake the IST signature when using the SMI interface."); | 458 | MODULE_PARM_DESC(smi_cmd, "Override the BIOS-given IST command with this value " |
459 | "-- Intel's default setting is 0x82"); | ||
460 | MODULE_PARM_DESC(smi_sig, "Set to 1 to fake the IST signature when using the " | ||
461 | "SMI interface."); | ||
426 | 462 | ||
427 | MODULE_AUTHOR ("Hiroshi Miura"); | 463 | MODULE_AUTHOR("Hiroshi Miura"); |
428 | MODULE_DESCRIPTION ("Speedstep driver for IST applet SMI interface."); | 464 | MODULE_DESCRIPTION("Speedstep driver for IST applet SMI interface."); |
429 | MODULE_LICENSE ("GPL"); | 465 | MODULE_LICENSE("GPL"); |
430 | 466 | ||
431 | module_init(speedstep_init); | 467 | module_init(speedstep_init); |
432 | module_exit(speedstep_exit); | 468 | module_exit(speedstep_exit); |
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c index d5cebb52d45b..b8e7aaf7ef75 100644 --- a/arch/x86/kernel/tsc.c +++ b/arch/x86/kernel/tsc.c | |||
@@ -543,8 +543,6 @@ unsigned long native_calibrate_tsc(void) | |||
543 | return tsc_pit_min; | 543 | return tsc_pit_min; |
544 | } | 544 | } |
545 | 545 | ||
546 | #ifdef CONFIG_X86_32 | ||
547 | /* Only called from the Powernow K7 cpu freq driver */ | ||
548 | int recalibrate_cpu_khz(void) | 546 | int recalibrate_cpu_khz(void) |
549 | { | 547 | { |
550 | #ifndef CONFIG_SMP | 548 | #ifndef CONFIG_SMP |
@@ -566,7 +564,6 @@ int recalibrate_cpu_khz(void) | |||
566 | 564 | ||
567 | EXPORT_SYMBOL(recalibrate_cpu_khz); | 565 | EXPORT_SYMBOL(recalibrate_cpu_khz); |
568 | 566 | ||
569 | #endif /* CONFIG_X86_32 */ | ||
570 | 567 | ||
571 | /* Accelerators for sched_clock() | 568 | /* Accelerators for sched_clock() |
572 | * convert from cycles(64bits) => nanoseconds (64bits) | 569 | * convert from cycles(64bits) => nanoseconds (64bits) |
diff --git a/drivers/cpufreq/cpufreq.c b/drivers/cpufreq/cpufreq.c index d6daf3c507d3..d270e8eb3e67 100644 --- a/drivers/cpufreq/cpufreq.c +++ b/drivers/cpufreq/cpufreq.c | |||
@@ -104,7 +104,8 @@ EXPORT_SYMBOL_GPL(unlock_policy_rwsem_write); | |||
104 | 104 | ||
105 | 105 | ||
106 | /* internal prototypes */ | 106 | /* internal prototypes */ |
107 | static int __cpufreq_governor(struct cpufreq_policy *policy, unsigned int event); | 107 | static int __cpufreq_governor(struct cpufreq_policy *policy, |
108 | unsigned int event); | ||
108 | static unsigned int __cpufreq_get(unsigned int cpu); | 109 | static unsigned int __cpufreq_get(unsigned int cpu); |
109 | static void handle_update(struct work_struct *work); | 110 | static void handle_update(struct work_struct *work); |
110 | 111 | ||
@@ -128,7 +129,7 @@ static int __init init_cpufreq_transition_notifier_list(void) | |||
128 | pure_initcall(init_cpufreq_transition_notifier_list); | 129 | pure_initcall(init_cpufreq_transition_notifier_list); |
129 | 130 | ||
130 | static LIST_HEAD(cpufreq_governor_list); | 131 | static LIST_HEAD(cpufreq_governor_list); |
131 | static DEFINE_MUTEX (cpufreq_governor_mutex); | 132 | static DEFINE_MUTEX(cpufreq_governor_mutex); |
132 | 133 | ||
133 | struct cpufreq_policy *cpufreq_cpu_get(unsigned int cpu) | 134 | struct cpufreq_policy *cpufreq_cpu_get(unsigned int cpu) |
134 | { | 135 | { |
@@ -371,7 +372,7 @@ static struct cpufreq_governor *__find_governor(const char *str_governor) | |||
371 | struct cpufreq_governor *t; | 372 | struct cpufreq_governor *t; |
372 | 373 | ||
373 | list_for_each_entry(t, &cpufreq_governor_list, governor_list) | 374 | list_for_each_entry(t, &cpufreq_governor_list, governor_list) |
374 | if (!strnicmp(str_governor,t->name,CPUFREQ_NAME_LEN)) | 375 | if (!strnicmp(str_governor, t->name, CPUFREQ_NAME_LEN)) |
375 | return t; | 376 | return t; |
376 | 377 | ||
377 | return NULL; | 378 | return NULL; |
@@ -429,15 +430,11 @@ static int cpufreq_parse_governor(char *str_governor, unsigned int *policy, | |||
429 | 430 | ||
430 | mutex_unlock(&cpufreq_governor_mutex); | 431 | mutex_unlock(&cpufreq_governor_mutex); |
431 | } | 432 | } |
432 | out: | 433 | out: |
433 | return err; | 434 | return err; |
434 | } | 435 | } |
435 | 436 | ||
436 | 437 | ||
437 | /* drivers/base/cpu.c */ | ||
438 | extern struct sysdev_class cpu_sysdev_class; | ||
439 | |||
440 | |||
441 | /** | 438 | /** |
442 | * cpufreq_per_cpu_attr_read() / show_##file_name() - | 439 | * cpufreq_per_cpu_attr_read() / show_##file_name() - |
443 | * print out cpufreq information | 440 | * print out cpufreq information |
@@ -450,11 +447,12 @@ extern struct sysdev_class cpu_sysdev_class; | |||
450 | static ssize_t show_##file_name \ | 447 | static ssize_t show_##file_name \ |
451 | (struct cpufreq_policy *policy, char *buf) \ | 448 | (struct cpufreq_policy *policy, char *buf) \ |
452 | { \ | 449 | { \ |
453 | return sprintf (buf, "%u\n", policy->object); \ | 450 | return sprintf(buf, "%u\n", policy->object); \ |
454 | } | 451 | } |
455 | 452 | ||
456 | show_one(cpuinfo_min_freq, cpuinfo.min_freq); | 453 | show_one(cpuinfo_min_freq, cpuinfo.min_freq); |
457 | show_one(cpuinfo_max_freq, cpuinfo.max_freq); | 454 | show_one(cpuinfo_max_freq, cpuinfo.max_freq); |
455 | show_one(cpuinfo_transition_latency, cpuinfo.transition_latency); | ||
458 | show_one(scaling_min_freq, min); | 456 | show_one(scaling_min_freq, min); |
459 | show_one(scaling_max_freq, max); | 457 | show_one(scaling_max_freq, max); |
460 | show_one(scaling_cur_freq, cur); | 458 | show_one(scaling_cur_freq, cur); |
@@ -476,7 +474,7 @@ static ssize_t store_##file_name \ | |||
476 | if (ret) \ | 474 | if (ret) \ |
477 | return -EINVAL; \ | 475 | return -EINVAL; \ |
478 | \ | 476 | \ |
479 | ret = sscanf (buf, "%u", &new_policy.object); \ | 477 | ret = sscanf(buf, "%u", &new_policy.object); \ |
480 | if (ret != 1) \ | 478 | if (ret != 1) \ |
481 | return -EINVAL; \ | 479 | return -EINVAL; \ |
482 | \ | 480 | \ |
@@ -486,8 +484,8 @@ static ssize_t store_##file_name \ | |||
486 | return ret ? ret : count; \ | 484 | return ret ? ret : count; \ |
487 | } | 485 | } |
488 | 486 | ||
489 | store_one(scaling_min_freq,min); | 487 | store_one(scaling_min_freq, min); |
490 | store_one(scaling_max_freq,max); | 488 | store_one(scaling_max_freq, max); |
491 | 489 | ||
492 | /** | 490 | /** |
493 | * show_cpuinfo_cur_freq - current CPU frequency as detected by hardware | 491 | * show_cpuinfo_cur_freq - current CPU frequency as detected by hardware |
@@ -507,12 +505,13 @@ static ssize_t show_cpuinfo_cur_freq(struct cpufreq_policy *policy, | |||
507 | */ | 505 | */ |
508 | static ssize_t show_scaling_governor(struct cpufreq_policy *policy, char *buf) | 506 | static ssize_t show_scaling_governor(struct cpufreq_policy *policy, char *buf) |
509 | { | 507 | { |
510 | if(policy->policy == CPUFREQ_POLICY_POWERSAVE) | 508 | if (policy->policy == CPUFREQ_POLICY_POWERSAVE) |
511 | return sprintf(buf, "powersave\n"); | 509 | return sprintf(buf, "powersave\n"); |
512 | else if (policy->policy == CPUFREQ_POLICY_PERFORMANCE) | 510 | else if (policy->policy == CPUFREQ_POLICY_PERFORMANCE) |
513 | return sprintf(buf, "performance\n"); | 511 | return sprintf(buf, "performance\n"); |
514 | else if (policy->governor) | 512 | else if (policy->governor) |
515 | return scnprintf(buf, CPUFREQ_NAME_LEN, "%s\n", policy->governor->name); | 513 | return scnprintf(buf, CPUFREQ_NAME_LEN, "%s\n", |
514 | policy->governor->name); | ||
516 | return -EINVAL; | 515 | return -EINVAL; |
517 | } | 516 | } |
518 | 517 | ||
@@ -531,7 +530,7 @@ static ssize_t store_scaling_governor(struct cpufreq_policy *policy, | |||
531 | if (ret) | 530 | if (ret) |
532 | return ret; | 531 | return ret; |
533 | 532 | ||
534 | ret = sscanf (buf, "%15s", str_governor); | 533 | ret = sscanf(buf, "%15s", str_governor); |
535 | if (ret != 1) | 534 | if (ret != 1) |
536 | return -EINVAL; | 535 | return -EINVAL; |
537 | 536 | ||
@@ -575,7 +574,8 @@ static ssize_t show_scaling_available_governors(struct cpufreq_policy *policy, | |||
575 | } | 574 | } |
576 | 575 | ||
577 | list_for_each_entry(t, &cpufreq_governor_list, governor_list) { | 576 | list_for_each_entry(t, &cpufreq_governor_list, governor_list) { |
578 | if (i >= (ssize_t) ((PAGE_SIZE / sizeof(char)) - (CPUFREQ_NAME_LEN + 2))) | 577 | if (i >= (ssize_t) ((PAGE_SIZE / sizeof(char)) |
578 | - (CPUFREQ_NAME_LEN + 2))) | ||
579 | goto out; | 579 | goto out; |
580 | i += scnprintf(&buf[i], CPUFREQ_NAME_LEN, "%s ", t->name); | 580 | i += scnprintf(&buf[i], CPUFREQ_NAME_LEN, "%s ", t->name); |
581 | } | 581 | } |
@@ -594,7 +594,7 @@ static ssize_t show_cpus(const struct cpumask *mask, char *buf) | |||
594 | i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), " "); | 594 | i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), " "); |
595 | i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), "%u", cpu); | 595 | i += scnprintf(&buf[i], (PAGE_SIZE - i - 2), "%u", cpu); |
596 | if (i >= (PAGE_SIZE - 5)) | 596 | if (i >= (PAGE_SIZE - 5)) |
597 | break; | 597 | break; |
598 | } | 598 | } |
599 | i += sprintf(&buf[i], "\n"); | 599 | i += sprintf(&buf[i], "\n"); |
600 | return i; | 600 | return i; |
@@ -660,6 +660,7 @@ __ATTR(_name, 0644, show_##_name, store_##_name) | |||
660 | define_one_ro0400(cpuinfo_cur_freq); | 660 | define_one_ro0400(cpuinfo_cur_freq); |
661 | define_one_ro(cpuinfo_min_freq); | 661 | define_one_ro(cpuinfo_min_freq); |
662 | define_one_ro(cpuinfo_max_freq); | 662 | define_one_ro(cpuinfo_max_freq); |
663 | define_one_ro(cpuinfo_transition_latency); | ||
663 | define_one_ro(scaling_available_governors); | 664 | define_one_ro(scaling_available_governors); |
664 | define_one_ro(scaling_driver); | 665 | define_one_ro(scaling_driver); |
665 | define_one_ro(scaling_cur_freq); | 666 | define_one_ro(scaling_cur_freq); |
@@ -673,6 +674,7 @@ define_one_rw(scaling_setspeed); | |||
673 | static struct attribute *default_attrs[] = { | 674 | static struct attribute *default_attrs[] = { |
674 | &cpuinfo_min_freq.attr, | 675 | &cpuinfo_min_freq.attr, |
675 | &cpuinfo_max_freq.attr, | 676 | &cpuinfo_max_freq.attr, |
677 | &cpuinfo_transition_latency.attr, | ||
676 | &scaling_min_freq.attr, | 678 | &scaling_min_freq.attr, |
677 | &scaling_max_freq.attr, | 679 | &scaling_max_freq.attr, |
678 | &affected_cpus.attr, | 680 | &affected_cpus.attr, |
@@ -684,10 +686,10 @@ static struct attribute *default_attrs[] = { | |||
684 | NULL | 686 | NULL |
685 | }; | 687 | }; |
686 | 688 | ||
687 | #define to_policy(k) container_of(k,struct cpufreq_policy,kobj) | 689 | #define to_policy(k) container_of(k, struct cpufreq_policy, kobj) |
688 | #define to_attr(a) container_of(a,struct freq_attr,attr) | 690 | #define to_attr(a) container_of(a, struct freq_attr, attr) |
689 | 691 | ||
690 | static ssize_t show(struct kobject *kobj, struct attribute *attr ,char *buf) | 692 | static ssize_t show(struct kobject *kobj, struct attribute *attr, char *buf) |
691 | { | 693 | { |
692 | struct cpufreq_policy *policy = to_policy(kobj); | 694 | struct cpufreq_policy *policy = to_policy(kobj); |
693 | struct freq_attr *fattr = to_attr(attr); | 695 | struct freq_attr *fattr = to_attr(attr); |
@@ -853,10 +855,10 @@ static int cpufreq_add_dev(struct sys_device *sys_dev) | |||
853 | if (cpu == j) | 855 | if (cpu == j) |
854 | continue; | 856 | continue; |
855 | 857 | ||
856 | /* check for existing affected CPUs. They may not be aware | 858 | /* Check for existing affected CPUs. |
857 | * of it due to CPU Hotplug. | 859 | * They may not be aware of it due to CPU Hotplug. |
858 | */ | 860 | */ |
859 | managed_policy = cpufreq_cpu_get(j); // FIXME: Where is this released? What about error paths? | 861 | managed_policy = cpufreq_cpu_get(j); /* FIXME: Where is this released? What about error paths? */ |
860 | if (unlikely(managed_policy)) { | 862 | if (unlikely(managed_policy)) { |
861 | 863 | ||
862 | /* Set proper policy_cpu */ | 864 | /* Set proper policy_cpu */ |
@@ -1127,8 +1129,8 @@ static void handle_update(struct work_struct *work) | |||
1127 | * @old_freq: CPU frequency the kernel thinks the CPU runs at | 1129 | * @old_freq: CPU frequency the kernel thinks the CPU runs at |
1128 | * @new_freq: CPU frequency the CPU actually runs at | 1130 | * @new_freq: CPU frequency the CPU actually runs at |
1129 | * | 1131 | * |
1130 | * We adjust to current frequency first, and need to clean up later. So either call | 1132 | * We adjust to current frequency first, and need to clean up later. |
1131 | * to cpufreq_update_policy() or schedule handle_update()). | 1133 | * So either call to cpufreq_update_policy() or schedule handle_update()). |
1132 | */ | 1134 | */ |
1133 | static void cpufreq_out_of_sync(unsigned int cpu, unsigned int old_freq, | 1135 | static void cpufreq_out_of_sync(unsigned int cpu, unsigned int old_freq, |
1134 | unsigned int new_freq) | 1136 | unsigned int new_freq) |
@@ -1610,7 +1612,8 @@ EXPORT_SYMBOL_GPL(cpufreq_unregister_governor); | |||
1610 | 1612 | ||
1611 | /** | 1613 | /** |
1612 | * cpufreq_get_policy - get the current cpufreq_policy | 1614 | * cpufreq_get_policy - get the current cpufreq_policy |
1613 | * @policy: struct cpufreq_policy into which the current cpufreq_policy is written | 1615 | * @policy: struct cpufreq_policy into which the current cpufreq_policy |
1616 | * is written | ||
1614 | * | 1617 | * |
1615 | * Reads the current cpufreq policy. | 1618 | * Reads the current cpufreq policy. |
1616 | */ | 1619 | */ |
diff --git a/drivers/cpufreq/cpufreq_conservative.c b/drivers/cpufreq/cpufreq_conservative.c index 0320962c4ec5..2ecd95e4ab1a 100644 --- a/drivers/cpufreq/cpufreq_conservative.c +++ b/drivers/cpufreq/cpufreq_conservative.c | |||
@@ -4,7 +4,7 @@ | |||
4 | * Copyright (C) 2001 Russell King | 4 | * Copyright (C) 2001 Russell King |
5 | * (C) 2003 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>. | 5 | * (C) 2003 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>. |
6 | * Jun Nakajima <jun.nakajima@intel.com> | 6 | * Jun Nakajima <jun.nakajima@intel.com> |
7 | * (C) 2004 Alexander Clouter <alex-kernel@digriz.org.uk> | 7 | * (C) 2009 Alexander Clouter <alex@digriz.org.uk> |
8 | * | 8 | * |
9 | * This program is free software; you can redistribute it and/or modify | 9 | * This program is free software; you can redistribute it and/or modify |
10 | * it under the terms of the GNU General Public License version 2 as | 10 | * it under the terms of the GNU General Public License version 2 as |
@@ -13,22 +13,17 @@ | |||
13 | 13 | ||
14 | #include <linux/kernel.h> | 14 | #include <linux/kernel.h> |
15 | #include <linux/module.h> | 15 | #include <linux/module.h> |
16 | #include <linux/smp.h> | ||
17 | #include <linux/init.h> | 16 | #include <linux/init.h> |
18 | #include <linux/interrupt.h> | ||
19 | #include <linux/ctype.h> | ||
20 | #include <linux/cpufreq.h> | 17 | #include <linux/cpufreq.h> |
21 | #include <linux/sysctl.h> | ||
22 | #include <linux/types.h> | ||
23 | #include <linux/fs.h> | ||
24 | #include <linux/sysfs.h> | ||
25 | #include <linux/cpu.h> | 18 | #include <linux/cpu.h> |
26 | #include <linux/kmod.h> | ||
27 | #include <linux/workqueue.h> | ||
28 | #include <linux/jiffies.h> | 19 | #include <linux/jiffies.h> |
29 | #include <linux/kernel_stat.h> | 20 | #include <linux/kernel_stat.h> |
30 | #include <linux/percpu.h> | ||
31 | #include <linux/mutex.h> | 21 | #include <linux/mutex.h> |
22 | #include <linux/hrtimer.h> | ||
23 | #include <linux/tick.h> | ||
24 | #include <linux/ktime.h> | ||
25 | #include <linux/sched.h> | ||
26 | |||
32 | /* | 27 | /* |
33 | * dbs is used in this file as a shortform for demandbased switching | 28 | * dbs is used in this file as a shortform for demandbased switching |
34 | * It helps to keep variable names smaller, simpler | 29 | * It helps to keep variable names smaller, simpler |
@@ -43,19 +38,31 @@ | |||
43 | * latency of the processor. The governor will work on any processor with | 38 | * latency of the processor. The governor will work on any processor with |
44 | * transition latency <= 10mS, using appropriate sampling | 39 | * transition latency <= 10mS, using appropriate sampling |
45 | * rate. | 40 | * rate. |
46 | * For CPUs with transition latency > 10mS (mostly drivers | 41 | * For CPUs with transition latency > 10mS (mostly drivers with CPUFREQ_ETERNAL) |
47 | * with CPUFREQ_ETERNAL), this governor will not work. | 42 | * this governor will not work. |
48 | * All times here are in uS. | 43 | * All times here are in uS. |
49 | */ | 44 | */ |
50 | static unsigned int def_sampling_rate; | 45 | static unsigned int def_sampling_rate; |
51 | #define MIN_SAMPLING_RATE_RATIO (2) | 46 | #define MIN_SAMPLING_RATE_RATIO (2) |
52 | /* for correct statistics, we need at least 10 ticks between each measure */ | 47 | /* for correct statistics, we need at least 10 ticks between each measure */ |
53 | #define MIN_STAT_SAMPLING_RATE \ | 48 | #define MIN_STAT_SAMPLING_RATE \ |
54 | (MIN_SAMPLING_RATE_RATIO * jiffies_to_usecs(10)) | 49 | (MIN_SAMPLING_RATE_RATIO * jiffies_to_usecs(10)) |
55 | #define MIN_SAMPLING_RATE \ | 50 | #define MIN_SAMPLING_RATE \ |
56 | (def_sampling_rate / MIN_SAMPLING_RATE_RATIO) | 51 | (def_sampling_rate / MIN_SAMPLING_RATE_RATIO) |
52 | /* Above MIN_SAMPLING_RATE will vanish with its sysfs file soon | ||
53 | * Define the minimal settable sampling rate to the greater of: | ||
54 | * - "HW transition latency" * 100 (same as default sampling / 10) | ||
55 | * - MIN_STAT_SAMPLING_RATE | ||
56 | * To avoid that userspace shoots itself. | ||
57 | */ | ||
58 | static unsigned int minimum_sampling_rate(void) | ||
59 | { | ||
60 | return max(def_sampling_rate / 10, MIN_STAT_SAMPLING_RATE); | ||
61 | } | ||
62 | |||
63 | /* This will also vanish soon with removing sampling_rate_max */ | ||
57 | #define MAX_SAMPLING_RATE (500 * def_sampling_rate) | 64 | #define MAX_SAMPLING_RATE (500 * def_sampling_rate) |
58 | #define DEF_SAMPLING_RATE_LATENCY_MULTIPLIER (1000) | 65 | #define LATENCY_MULTIPLIER (1000) |
59 | #define DEF_SAMPLING_DOWN_FACTOR (1) | 66 | #define DEF_SAMPLING_DOWN_FACTOR (1) |
60 | #define MAX_SAMPLING_DOWN_FACTOR (10) | 67 | #define MAX_SAMPLING_DOWN_FACTOR (10) |
61 | #define TRANSITION_LATENCY_LIMIT (10 * 1000 * 1000) | 68 | #define TRANSITION_LATENCY_LIMIT (10 * 1000 * 1000) |
@@ -63,12 +70,15 @@ static unsigned int def_sampling_rate; | |||
63 | static void do_dbs_timer(struct work_struct *work); | 70 | static void do_dbs_timer(struct work_struct *work); |
64 | 71 | ||
65 | struct cpu_dbs_info_s { | 72 | struct cpu_dbs_info_s { |
73 | cputime64_t prev_cpu_idle; | ||
74 | cputime64_t prev_cpu_wall; | ||
75 | cputime64_t prev_cpu_nice; | ||
66 | struct cpufreq_policy *cur_policy; | 76 | struct cpufreq_policy *cur_policy; |
67 | unsigned int prev_cpu_idle_up; | 77 | struct delayed_work work; |
68 | unsigned int prev_cpu_idle_down; | ||
69 | unsigned int enable; | ||
70 | unsigned int down_skip; | 78 | unsigned int down_skip; |
71 | unsigned int requested_freq; | 79 | unsigned int requested_freq; |
80 | int cpu; | ||
81 | unsigned int enable:1; | ||
72 | }; | 82 | }; |
73 | static DEFINE_PER_CPU(struct cpu_dbs_info_s, cpu_dbs_info); | 83 | static DEFINE_PER_CPU(struct cpu_dbs_info_s, cpu_dbs_info); |
74 | 84 | ||
@@ -82,19 +92,18 @@ static unsigned int dbs_enable; /* number of CPUs using this policy */ | |||
82 | * cpu_hotplug lock should be taken before that. Note that cpu_hotplug lock | 92 | * cpu_hotplug lock should be taken before that. Note that cpu_hotplug lock |
83 | * is recursive for the same process. -Venki | 93 | * is recursive for the same process. -Venki |
84 | */ | 94 | */ |
85 | static DEFINE_MUTEX (dbs_mutex); | 95 | static DEFINE_MUTEX(dbs_mutex); |
86 | static DECLARE_DELAYED_WORK(dbs_work, do_dbs_timer); | ||
87 | 96 | ||
88 | struct dbs_tuners { | 97 | static struct workqueue_struct *kconservative_wq; |
98 | |||
99 | static struct dbs_tuners { | ||
89 | unsigned int sampling_rate; | 100 | unsigned int sampling_rate; |
90 | unsigned int sampling_down_factor; | 101 | unsigned int sampling_down_factor; |
91 | unsigned int up_threshold; | 102 | unsigned int up_threshold; |
92 | unsigned int down_threshold; | 103 | unsigned int down_threshold; |
93 | unsigned int ignore_nice; | 104 | unsigned int ignore_nice; |
94 | unsigned int freq_step; | 105 | unsigned int freq_step; |
95 | }; | 106 | } dbs_tuners_ins = { |
96 | |||
97 | static struct dbs_tuners dbs_tuners_ins = { | ||
98 | .up_threshold = DEF_FREQUENCY_UP_THRESHOLD, | 107 | .up_threshold = DEF_FREQUENCY_UP_THRESHOLD, |
99 | .down_threshold = DEF_FREQUENCY_DOWN_THRESHOLD, | 108 | .down_threshold = DEF_FREQUENCY_DOWN_THRESHOLD, |
100 | .sampling_down_factor = DEF_SAMPLING_DOWN_FACTOR, | 109 | .sampling_down_factor = DEF_SAMPLING_DOWN_FACTOR, |
@@ -102,18 +111,37 @@ static struct dbs_tuners dbs_tuners_ins = { | |||
102 | .freq_step = 5, | 111 | .freq_step = 5, |
103 | }; | 112 | }; |
104 | 113 | ||
105 | static inline unsigned int get_cpu_idle_time(unsigned int cpu) | 114 | static inline cputime64_t get_cpu_idle_time_jiffy(unsigned int cpu, |
115 | cputime64_t *wall) | ||
106 | { | 116 | { |
107 | unsigned int add_nice = 0, ret; | 117 | cputime64_t idle_time; |
118 | cputime64_t cur_wall_time; | ||
119 | cputime64_t busy_time; | ||
108 | 120 | ||
109 | if (dbs_tuners_ins.ignore_nice) | 121 | cur_wall_time = jiffies64_to_cputime64(get_jiffies_64()); |
110 | add_nice = kstat_cpu(cpu).cpustat.nice; | 122 | busy_time = cputime64_add(kstat_cpu(cpu).cpustat.user, |
123 | kstat_cpu(cpu).cpustat.system); | ||
124 | |||
125 | busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.irq); | ||
126 | busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.softirq); | ||
127 | busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.steal); | ||
128 | busy_time = cputime64_add(busy_time, kstat_cpu(cpu).cpustat.nice); | ||
129 | |||
130 | idle_time = cputime64_sub(cur_wall_time, busy_time); | ||
131 | if (wall) | ||
132 | *wall = cur_wall_time; | ||
133 | |||
134 | return idle_time; | ||
135 | } | ||
136 | |||
137 | static inline cputime64_t get_cpu_idle_time(unsigned int cpu, cputime64_t *wall) | ||
138 | { | ||
139 | u64 idle_time = get_cpu_idle_time_us(cpu, wall); | ||
111 | 140 | ||
112 | ret = kstat_cpu(cpu).cpustat.idle + | 141 | if (idle_time == -1ULL) |
113 | kstat_cpu(cpu).cpustat.iowait + | 142 | return get_cpu_idle_time_jiffy(cpu, wall); |
114 | add_nice; | ||
115 | 143 | ||
116 | return ret; | 144 | return idle_time; |
117 | } | 145 | } |
118 | 146 | ||
119 | /* keep track of frequency transitions */ | 147 | /* keep track of frequency transitions */ |
@@ -125,10 +153,21 @@ dbs_cpufreq_notifier(struct notifier_block *nb, unsigned long val, | |||
125 | struct cpu_dbs_info_s *this_dbs_info = &per_cpu(cpu_dbs_info, | 153 | struct cpu_dbs_info_s *this_dbs_info = &per_cpu(cpu_dbs_info, |
126 | freq->cpu); | 154 | freq->cpu); |
127 | 155 | ||
156 | struct cpufreq_policy *policy; | ||
157 | |||
128 | if (!this_dbs_info->enable) | 158 | if (!this_dbs_info->enable) |
129 | return 0; | 159 | return 0; |
130 | 160 | ||
131 | this_dbs_info->requested_freq = freq->new; | 161 | policy = this_dbs_info->cur_policy; |
162 | |||
163 | /* | ||
164 | * we only care if our internally tracked freq moves outside | ||
165 | * the 'valid' ranges of freqency available to us otherwise | ||
166 | * we do not change it | ||
167 | */ | ||
168 | if (this_dbs_info->requested_freq > policy->max | ||
169 | || this_dbs_info->requested_freq < policy->min) | ||
170 | this_dbs_info->requested_freq = freq->new; | ||
132 | 171 | ||
133 | return 0; | 172 | return 0; |
134 | } | 173 | } |
@@ -140,16 +179,31 @@ static struct notifier_block dbs_cpufreq_notifier_block = { | |||
140 | /************************** sysfs interface ************************/ | 179 | /************************** sysfs interface ************************/ |
141 | static ssize_t show_sampling_rate_max(struct cpufreq_policy *policy, char *buf) | 180 | static ssize_t show_sampling_rate_max(struct cpufreq_policy *policy, char *buf) |
142 | { | 181 | { |
143 | return sprintf (buf, "%u\n", MAX_SAMPLING_RATE); | 182 | static int print_once; |
183 | |||
184 | if (!print_once) { | ||
185 | printk(KERN_INFO "CPUFREQ: conservative sampling_rate_max " | ||
186 | "sysfs file is deprecated - used by: %s\n", | ||
187 | current->comm); | ||
188 | print_once = 1; | ||
189 | } | ||
190 | return sprintf(buf, "%u\n", MAX_SAMPLING_RATE); | ||
144 | } | 191 | } |
145 | 192 | ||
146 | static ssize_t show_sampling_rate_min(struct cpufreq_policy *policy, char *buf) | 193 | static ssize_t show_sampling_rate_min(struct cpufreq_policy *policy, char *buf) |
147 | { | 194 | { |
148 | return sprintf (buf, "%u\n", MIN_SAMPLING_RATE); | 195 | static int print_once; |
196 | |||
197 | if (!print_once) { | ||
198 | printk(KERN_INFO "CPUFREQ: conservative sampling_rate_max " | ||
199 | "sysfs file is deprecated - used by: %s\n", current->comm); | ||
200 | print_once = 1; | ||
201 | } | ||
202 | return sprintf(buf, "%u\n", MIN_SAMPLING_RATE); | ||
149 | } | 203 | } |
150 | 204 | ||
151 | #define define_one_ro(_name) \ | 205 | #define define_one_ro(_name) \ |
152 | static struct freq_attr _name = \ | 206 | static struct freq_attr _name = \ |
153 | __ATTR(_name, 0444, show_##_name, NULL) | 207 | __ATTR(_name, 0444, show_##_name, NULL) |
154 | 208 | ||
155 | define_one_ro(sampling_rate_max); | 209 | define_one_ro(sampling_rate_max); |
@@ -174,7 +228,8 @@ static ssize_t store_sampling_down_factor(struct cpufreq_policy *unused, | |||
174 | { | 228 | { |
175 | unsigned int input; | 229 | unsigned int input; |
176 | int ret; | 230 | int ret; |
177 | ret = sscanf (buf, "%u", &input); | 231 | ret = sscanf(buf, "%u", &input); |
232 | |||
178 | if (ret != 1 || input > MAX_SAMPLING_DOWN_FACTOR || input < 1) | 233 | if (ret != 1 || input > MAX_SAMPLING_DOWN_FACTOR || input < 1) |
179 | return -EINVAL; | 234 | return -EINVAL; |
180 | 235 | ||
@@ -190,15 +245,13 @@ static ssize_t store_sampling_rate(struct cpufreq_policy *unused, | |||
190 | { | 245 | { |
191 | unsigned int input; | 246 | unsigned int input; |
192 | int ret; | 247 | int ret; |
193 | ret = sscanf (buf, "%u", &input); | 248 | ret = sscanf(buf, "%u", &input); |
194 | 249 | ||
195 | mutex_lock(&dbs_mutex); | 250 | if (ret != 1) |
196 | if (ret != 1 || input > MAX_SAMPLING_RATE || input < MIN_SAMPLING_RATE) { | ||
197 | mutex_unlock(&dbs_mutex); | ||
198 | return -EINVAL; | 251 | return -EINVAL; |
199 | } | ||
200 | 252 | ||
201 | dbs_tuners_ins.sampling_rate = input; | 253 | mutex_lock(&dbs_mutex); |
254 | dbs_tuners_ins.sampling_rate = max(input, minimum_sampling_rate()); | ||
202 | mutex_unlock(&dbs_mutex); | 255 | mutex_unlock(&dbs_mutex); |
203 | 256 | ||
204 | return count; | 257 | return count; |
@@ -209,10 +262,11 @@ static ssize_t store_up_threshold(struct cpufreq_policy *unused, | |||
209 | { | 262 | { |
210 | unsigned int input; | 263 | unsigned int input; |
211 | int ret; | 264 | int ret; |
212 | ret = sscanf (buf, "%u", &input); | 265 | ret = sscanf(buf, "%u", &input); |
213 | 266 | ||
214 | mutex_lock(&dbs_mutex); | 267 | mutex_lock(&dbs_mutex); |
215 | if (ret != 1 || input > 100 || input <= dbs_tuners_ins.down_threshold) { | 268 | if (ret != 1 || input > 100 || |
269 | input <= dbs_tuners_ins.down_threshold) { | ||
216 | mutex_unlock(&dbs_mutex); | 270 | mutex_unlock(&dbs_mutex); |
217 | return -EINVAL; | 271 | return -EINVAL; |
218 | } | 272 | } |
@@ -228,10 +282,12 @@ static ssize_t store_down_threshold(struct cpufreq_policy *unused, | |||
228 | { | 282 | { |
229 | unsigned int input; | 283 | unsigned int input; |
230 | int ret; | 284 | int ret; |
231 | ret = sscanf (buf, "%u", &input); | 285 | ret = sscanf(buf, "%u", &input); |
232 | 286 | ||
233 | mutex_lock(&dbs_mutex); | 287 | mutex_lock(&dbs_mutex); |
234 | if (ret != 1 || input > 100 || input >= dbs_tuners_ins.up_threshold) { | 288 | /* cannot be lower than 11 otherwise freq will not fall */ |
289 | if (ret != 1 || input < 11 || input > 100 || | ||
290 | input >= dbs_tuners_ins.up_threshold) { | ||
235 | mutex_unlock(&dbs_mutex); | 291 | mutex_unlock(&dbs_mutex); |
236 | return -EINVAL; | 292 | return -EINVAL; |
237 | } | 293 | } |
@@ -264,12 +320,14 @@ static ssize_t store_ignore_nice_load(struct cpufreq_policy *policy, | |||
264 | } | 320 | } |
265 | dbs_tuners_ins.ignore_nice = input; | 321 | dbs_tuners_ins.ignore_nice = input; |
266 | 322 | ||
267 | /* we need to re-evaluate prev_cpu_idle_up and prev_cpu_idle_down */ | 323 | /* we need to re-evaluate prev_cpu_idle */ |
268 | for_each_online_cpu(j) { | 324 | for_each_online_cpu(j) { |
269 | struct cpu_dbs_info_s *j_dbs_info; | 325 | struct cpu_dbs_info_s *dbs_info; |
270 | j_dbs_info = &per_cpu(cpu_dbs_info, j); | 326 | dbs_info = &per_cpu(cpu_dbs_info, j); |
271 | j_dbs_info->prev_cpu_idle_up = get_cpu_idle_time(j); | 327 | dbs_info->prev_cpu_idle = get_cpu_idle_time(j, |
272 | j_dbs_info->prev_cpu_idle_down = j_dbs_info->prev_cpu_idle_up; | 328 | &dbs_info->prev_cpu_wall); |
329 | if (dbs_tuners_ins.ignore_nice) | ||
330 | dbs_info->prev_cpu_nice = kstat_cpu(j).cpustat.nice; | ||
273 | } | 331 | } |
274 | mutex_unlock(&dbs_mutex); | 332 | mutex_unlock(&dbs_mutex); |
275 | 333 | ||
@@ -281,7 +339,6 @@ static ssize_t store_freq_step(struct cpufreq_policy *policy, | |||
281 | { | 339 | { |
282 | unsigned int input; | 340 | unsigned int input; |
283 | int ret; | 341 | int ret; |
284 | |||
285 | ret = sscanf(buf, "%u", &input); | 342 | ret = sscanf(buf, "%u", &input); |
286 | 343 | ||
287 | if (ret != 1) | 344 | if (ret != 1) |
@@ -310,7 +367,7 @@ define_one_rw(down_threshold); | |||
310 | define_one_rw(ignore_nice_load); | 367 | define_one_rw(ignore_nice_load); |
311 | define_one_rw(freq_step); | 368 | define_one_rw(freq_step); |
312 | 369 | ||
313 | static struct attribute * dbs_attributes[] = { | 370 | static struct attribute *dbs_attributes[] = { |
314 | &sampling_rate_max.attr, | 371 | &sampling_rate_max.attr, |
315 | &sampling_rate_min.attr, | 372 | &sampling_rate_min.attr, |
316 | &sampling_rate.attr, | 373 | &sampling_rate.attr, |
@@ -329,55 +386,78 @@ static struct attribute_group dbs_attr_group = { | |||
329 | 386 | ||
330 | /************************** sysfs end ************************/ | 387 | /************************** sysfs end ************************/ |
331 | 388 | ||
332 | static void dbs_check_cpu(int cpu) | 389 | static void dbs_check_cpu(struct cpu_dbs_info_s *this_dbs_info) |
333 | { | 390 | { |
334 | unsigned int idle_ticks, up_idle_ticks, down_idle_ticks; | 391 | unsigned int load = 0; |
335 | unsigned int tmp_idle_ticks, total_idle_ticks; | ||
336 | unsigned int freq_target; | 392 | unsigned int freq_target; |
337 | unsigned int freq_down_sampling_rate; | ||
338 | struct cpu_dbs_info_s *this_dbs_info = &per_cpu(cpu_dbs_info, cpu); | ||
339 | struct cpufreq_policy *policy; | ||
340 | 393 | ||
341 | if (!this_dbs_info->enable) | 394 | struct cpufreq_policy *policy; |
342 | return; | 395 | unsigned int j; |
343 | 396 | ||
344 | policy = this_dbs_info->cur_policy; | 397 | policy = this_dbs_info->cur_policy; |
345 | 398 | ||
346 | /* | 399 | /* |
347 | * The default safe range is 20% to 80% | 400 | * Every sampling_rate, we check, if current idle time is less |
348 | * Every sampling_rate, we check | 401 | * than 20% (default), then we try to increase frequency |
349 | * - If current idle time is less than 20%, then we try to | 402 | * Every sampling_rate*sampling_down_factor, we check, if current |
350 | * increase frequency | 403 | * idle time is more than 80%, then we try to decrease frequency |
351 | * Every sampling_rate*sampling_down_factor, we check | ||
352 | * - If current idle time is more than 80%, then we try to | ||
353 | * decrease frequency | ||
354 | * | 404 | * |
355 | * Any frequency increase takes it to the maximum frequency. | 405 | * Any frequency increase takes it to the maximum frequency. |
356 | * Frequency reduction happens at minimum steps of | 406 | * Frequency reduction happens at minimum steps of |
357 | * 5% (default) of max_frequency | 407 | * 5% (default) of maximum frequency |
358 | */ | 408 | */ |
359 | 409 | ||
360 | /* Check for frequency increase */ | 410 | /* Get Absolute Load */ |
361 | idle_ticks = UINT_MAX; | 411 | for_each_cpu(j, policy->cpus) { |
412 | struct cpu_dbs_info_s *j_dbs_info; | ||
413 | cputime64_t cur_wall_time, cur_idle_time; | ||
414 | unsigned int idle_time, wall_time; | ||
362 | 415 | ||
363 | /* Check for frequency increase */ | 416 | j_dbs_info = &per_cpu(cpu_dbs_info, j); |
364 | total_idle_ticks = get_cpu_idle_time(cpu); | 417 | |
365 | tmp_idle_ticks = total_idle_ticks - | 418 | cur_idle_time = get_cpu_idle_time(j, &cur_wall_time); |
366 | this_dbs_info->prev_cpu_idle_up; | 419 | |
367 | this_dbs_info->prev_cpu_idle_up = total_idle_ticks; | 420 | wall_time = (unsigned int) cputime64_sub(cur_wall_time, |
421 | j_dbs_info->prev_cpu_wall); | ||
422 | j_dbs_info->prev_cpu_wall = cur_wall_time; | ||
423 | |||
424 | idle_time = (unsigned int) cputime64_sub(cur_idle_time, | ||
425 | j_dbs_info->prev_cpu_idle); | ||
426 | j_dbs_info->prev_cpu_idle = cur_idle_time; | ||
427 | |||
428 | if (dbs_tuners_ins.ignore_nice) { | ||
429 | cputime64_t cur_nice; | ||
430 | unsigned long cur_nice_jiffies; | ||
431 | |||
432 | cur_nice = cputime64_sub(kstat_cpu(j).cpustat.nice, | ||
433 | j_dbs_info->prev_cpu_nice); | ||
434 | /* | ||
435 | * Assumption: nice time between sampling periods will | ||
436 | * be less than 2^32 jiffies for 32 bit sys | ||
437 | */ | ||
438 | cur_nice_jiffies = (unsigned long) | ||
439 | cputime64_to_jiffies64(cur_nice); | ||
368 | 440 | ||
369 | if (tmp_idle_ticks < idle_ticks) | 441 | j_dbs_info->prev_cpu_nice = kstat_cpu(j).cpustat.nice; |
370 | idle_ticks = tmp_idle_ticks; | 442 | idle_time += jiffies_to_usecs(cur_nice_jiffies); |
443 | } | ||
444 | |||
445 | if (unlikely(!wall_time || wall_time < idle_time)) | ||
446 | continue; | ||
447 | |||
448 | load = 100 * (wall_time - idle_time) / wall_time; | ||
449 | } | ||
371 | 450 | ||
372 | /* Scale idle ticks by 100 and compare with up and down ticks */ | 451 | /* |
373 | idle_ticks *= 100; | 452 | * break out if we 'cannot' reduce the speed as the user might |
374 | up_idle_ticks = (100 - dbs_tuners_ins.up_threshold) * | 453 | * want freq_step to be zero |
375 | usecs_to_jiffies(dbs_tuners_ins.sampling_rate); | 454 | */ |
455 | if (dbs_tuners_ins.freq_step == 0) | ||
456 | return; | ||
376 | 457 | ||
377 | if (idle_ticks < up_idle_ticks) { | 458 | /* Check for frequency increase */ |
459 | if (load > dbs_tuners_ins.up_threshold) { | ||
378 | this_dbs_info->down_skip = 0; | 460 | this_dbs_info->down_skip = 0; |
379 | this_dbs_info->prev_cpu_idle_down = | ||
380 | this_dbs_info->prev_cpu_idle_up; | ||
381 | 461 | ||
382 | /* if we are already at full speed then break out early */ | 462 | /* if we are already at full speed then break out early */ |
383 | if (this_dbs_info->requested_freq == policy->max) | 463 | if (this_dbs_info->requested_freq == policy->max) |
@@ -398,49 +478,24 @@ static void dbs_check_cpu(int cpu) | |||
398 | return; | 478 | return; |
399 | } | 479 | } |
400 | 480 | ||
401 | /* Check for frequency decrease */ | 481 | /* |
402 | this_dbs_info->down_skip++; | 482 | * The optimal frequency is the frequency that is the lowest that |
403 | if (this_dbs_info->down_skip < dbs_tuners_ins.sampling_down_factor) | 483 | * can support the current CPU usage without triggering the up |
404 | return; | 484 | * policy. To be safe, we focus 10 points under the threshold. |
405 | 485 | */ | |
406 | /* Check for frequency decrease */ | 486 | if (load < (dbs_tuners_ins.down_threshold - 10)) { |
407 | total_idle_ticks = this_dbs_info->prev_cpu_idle_up; | ||
408 | tmp_idle_ticks = total_idle_ticks - | ||
409 | this_dbs_info->prev_cpu_idle_down; | ||
410 | this_dbs_info->prev_cpu_idle_down = total_idle_ticks; | ||
411 | |||
412 | if (tmp_idle_ticks < idle_ticks) | ||
413 | idle_ticks = tmp_idle_ticks; | ||
414 | |||
415 | /* Scale idle ticks by 100 and compare with up and down ticks */ | ||
416 | idle_ticks *= 100; | ||
417 | this_dbs_info->down_skip = 0; | ||
418 | |||
419 | freq_down_sampling_rate = dbs_tuners_ins.sampling_rate * | ||
420 | dbs_tuners_ins.sampling_down_factor; | ||
421 | down_idle_ticks = (100 - dbs_tuners_ins.down_threshold) * | ||
422 | usecs_to_jiffies(freq_down_sampling_rate); | ||
423 | |||
424 | if (idle_ticks > down_idle_ticks) { | ||
425 | /* | ||
426 | * if we are already at the lowest speed then break out early | ||
427 | * or if we 'cannot' reduce the speed as the user might want | ||
428 | * freq_target to be zero | ||
429 | */ | ||
430 | if (this_dbs_info->requested_freq == policy->min | ||
431 | || dbs_tuners_ins.freq_step == 0) | ||
432 | return; | ||
433 | |||
434 | freq_target = (dbs_tuners_ins.freq_step * policy->max) / 100; | 487 | freq_target = (dbs_tuners_ins.freq_step * policy->max) / 100; |
435 | 488 | ||
436 | /* max freq cannot be less than 100. But who knows.... */ | ||
437 | if (unlikely(freq_target == 0)) | ||
438 | freq_target = 5; | ||
439 | |||
440 | this_dbs_info->requested_freq -= freq_target; | 489 | this_dbs_info->requested_freq -= freq_target; |
441 | if (this_dbs_info->requested_freq < policy->min) | 490 | if (this_dbs_info->requested_freq < policy->min) |
442 | this_dbs_info->requested_freq = policy->min; | 491 | this_dbs_info->requested_freq = policy->min; |
443 | 492 | ||
493 | /* | ||
494 | * if we cannot reduce the frequency anymore, break out early | ||
495 | */ | ||
496 | if (policy->cur == policy->min) | ||
497 | return; | ||
498 | |||
444 | __cpufreq_driver_target(policy, this_dbs_info->requested_freq, | 499 | __cpufreq_driver_target(policy, this_dbs_info->requested_freq, |
445 | CPUFREQ_RELATION_H); | 500 | CPUFREQ_RELATION_H); |
446 | return; | 501 | return; |
@@ -449,27 +504,45 @@ static void dbs_check_cpu(int cpu) | |||
449 | 504 | ||
450 | static void do_dbs_timer(struct work_struct *work) | 505 | static void do_dbs_timer(struct work_struct *work) |
451 | { | 506 | { |
452 | int i; | 507 | struct cpu_dbs_info_s *dbs_info = |
453 | mutex_lock(&dbs_mutex); | 508 | container_of(work, struct cpu_dbs_info_s, work.work); |
454 | for_each_online_cpu(i) | 509 | unsigned int cpu = dbs_info->cpu; |
455 | dbs_check_cpu(i); | 510 | |
456 | schedule_delayed_work(&dbs_work, | 511 | /* We want all CPUs to do sampling nearly on same jiffy */ |
457 | usecs_to_jiffies(dbs_tuners_ins.sampling_rate)); | 512 | int delay = usecs_to_jiffies(dbs_tuners_ins.sampling_rate); |
458 | mutex_unlock(&dbs_mutex); | 513 | |
514 | delay -= jiffies % delay; | ||
515 | |||
516 | if (lock_policy_rwsem_write(cpu) < 0) | ||
517 | return; | ||
518 | |||
519 | if (!dbs_info->enable) { | ||
520 | unlock_policy_rwsem_write(cpu); | ||
521 | return; | ||
522 | } | ||
523 | |||
524 | dbs_check_cpu(dbs_info); | ||
525 | |||
526 | queue_delayed_work_on(cpu, kconservative_wq, &dbs_info->work, delay); | ||
527 | unlock_policy_rwsem_write(cpu); | ||
459 | } | 528 | } |
460 | 529 | ||
461 | static inline void dbs_timer_init(void) | 530 | static inline void dbs_timer_init(struct cpu_dbs_info_s *dbs_info) |
462 | { | 531 | { |
463 | init_timer_deferrable(&dbs_work.timer); | 532 | /* We want all CPUs to do sampling nearly on same jiffy */ |
464 | schedule_delayed_work(&dbs_work, | 533 | int delay = usecs_to_jiffies(dbs_tuners_ins.sampling_rate); |
465 | usecs_to_jiffies(dbs_tuners_ins.sampling_rate)); | 534 | delay -= jiffies % delay; |
466 | return; | 535 | |
536 | dbs_info->enable = 1; | ||
537 | INIT_DELAYED_WORK_DEFERRABLE(&dbs_info->work, do_dbs_timer); | ||
538 | queue_delayed_work_on(dbs_info->cpu, kconservative_wq, &dbs_info->work, | ||
539 | delay); | ||
467 | } | 540 | } |
468 | 541 | ||
469 | static inline void dbs_timer_exit(void) | 542 | static inline void dbs_timer_exit(struct cpu_dbs_info_s *dbs_info) |
470 | { | 543 | { |
471 | cancel_delayed_work(&dbs_work); | 544 | dbs_info->enable = 0; |
472 | return; | 545 | cancel_delayed_work(&dbs_info->work); |
473 | } | 546 | } |
474 | 547 | ||
475 | static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | 548 | static int cpufreq_governor_dbs(struct cpufreq_policy *policy, |
@@ -503,11 +576,13 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
503 | j_dbs_info = &per_cpu(cpu_dbs_info, j); | 576 | j_dbs_info = &per_cpu(cpu_dbs_info, j); |
504 | j_dbs_info->cur_policy = policy; | 577 | j_dbs_info->cur_policy = policy; |
505 | 578 | ||
506 | j_dbs_info->prev_cpu_idle_up = get_cpu_idle_time(cpu); | 579 | j_dbs_info->prev_cpu_idle = get_cpu_idle_time(j, |
507 | j_dbs_info->prev_cpu_idle_down | 580 | &j_dbs_info->prev_cpu_wall); |
508 | = j_dbs_info->prev_cpu_idle_up; | 581 | if (dbs_tuners_ins.ignore_nice) { |
582 | j_dbs_info->prev_cpu_nice = | ||
583 | kstat_cpu(j).cpustat.nice; | ||
584 | } | ||
509 | } | 585 | } |
510 | this_dbs_info->enable = 1; | ||
511 | this_dbs_info->down_skip = 0; | 586 | this_dbs_info->down_skip = 0; |
512 | this_dbs_info->requested_freq = policy->cur; | 587 | this_dbs_info->requested_freq = policy->cur; |
513 | 588 | ||
@@ -523,38 +598,36 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
523 | if (latency == 0) | 598 | if (latency == 0) |
524 | latency = 1; | 599 | latency = 1; |
525 | 600 | ||
526 | def_sampling_rate = 10 * latency * | 601 | def_sampling_rate = |
527 | DEF_SAMPLING_RATE_LATENCY_MULTIPLIER; | 602 | max(latency * LATENCY_MULTIPLIER, |
528 | 603 | MIN_STAT_SAMPLING_RATE); | |
529 | if (def_sampling_rate < MIN_STAT_SAMPLING_RATE) | ||
530 | def_sampling_rate = MIN_STAT_SAMPLING_RATE; | ||
531 | 604 | ||
532 | dbs_tuners_ins.sampling_rate = def_sampling_rate; | 605 | dbs_tuners_ins.sampling_rate = def_sampling_rate; |
533 | 606 | ||
534 | dbs_timer_init(); | ||
535 | cpufreq_register_notifier( | 607 | cpufreq_register_notifier( |
536 | &dbs_cpufreq_notifier_block, | 608 | &dbs_cpufreq_notifier_block, |
537 | CPUFREQ_TRANSITION_NOTIFIER); | 609 | CPUFREQ_TRANSITION_NOTIFIER); |
538 | } | 610 | } |
611 | dbs_timer_init(this_dbs_info); | ||
539 | 612 | ||
540 | mutex_unlock(&dbs_mutex); | 613 | mutex_unlock(&dbs_mutex); |
614 | |||
541 | break; | 615 | break; |
542 | 616 | ||
543 | case CPUFREQ_GOV_STOP: | 617 | case CPUFREQ_GOV_STOP: |
544 | mutex_lock(&dbs_mutex); | 618 | mutex_lock(&dbs_mutex); |
545 | this_dbs_info->enable = 0; | 619 | dbs_timer_exit(this_dbs_info); |
546 | sysfs_remove_group(&policy->kobj, &dbs_attr_group); | 620 | sysfs_remove_group(&policy->kobj, &dbs_attr_group); |
547 | dbs_enable--; | 621 | dbs_enable--; |
622 | |||
548 | /* | 623 | /* |
549 | * Stop the timerschedule work, when this governor | 624 | * Stop the timerschedule work, when this governor |
550 | * is used for first time | 625 | * is used for first time |
551 | */ | 626 | */ |
552 | if (dbs_enable == 0) { | 627 | if (dbs_enable == 0) |
553 | dbs_timer_exit(); | ||
554 | cpufreq_unregister_notifier( | 628 | cpufreq_unregister_notifier( |
555 | &dbs_cpufreq_notifier_block, | 629 | &dbs_cpufreq_notifier_block, |
556 | CPUFREQ_TRANSITION_NOTIFIER); | 630 | CPUFREQ_TRANSITION_NOTIFIER); |
557 | } | ||
558 | 631 | ||
559 | mutex_unlock(&dbs_mutex); | 632 | mutex_unlock(&dbs_mutex); |
560 | 633 | ||
@@ -571,6 +644,7 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
571 | this_dbs_info->cur_policy, | 644 | this_dbs_info->cur_policy, |
572 | policy->min, CPUFREQ_RELATION_L); | 645 | policy->min, CPUFREQ_RELATION_L); |
573 | mutex_unlock(&dbs_mutex); | 646 | mutex_unlock(&dbs_mutex); |
647 | |||
574 | break; | 648 | break; |
575 | } | 649 | } |
576 | return 0; | 650 | return 0; |
@@ -588,23 +662,33 @@ struct cpufreq_governor cpufreq_gov_conservative = { | |||
588 | 662 | ||
589 | static int __init cpufreq_gov_dbs_init(void) | 663 | static int __init cpufreq_gov_dbs_init(void) |
590 | { | 664 | { |
591 | return cpufreq_register_governor(&cpufreq_gov_conservative); | 665 | int err; |
666 | |||
667 | kconservative_wq = create_workqueue("kconservative"); | ||
668 | if (!kconservative_wq) { | ||
669 | printk(KERN_ERR "Creation of kconservative failed\n"); | ||
670 | return -EFAULT; | ||
671 | } | ||
672 | |||
673 | err = cpufreq_register_governor(&cpufreq_gov_conservative); | ||
674 | if (err) | ||
675 | destroy_workqueue(kconservative_wq); | ||
676 | |||
677 | return err; | ||
592 | } | 678 | } |
593 | 679 | ||
594 | static void __exit cpufreq_gov_dbs_exit(void) | 680 | static void __exit cpufreq_gov_dbs_exit(void) |
595 | { | 681 | { |
596 | /* Make sure that the scheduled work is indeed not running */ | ||
597 | flush_scheduled_work(); | ||
598 | |||
599 | cpufreq_unregister_governor(&cpufreq_gov_conservative); | 682 | cpufreq_unregister_governor(&cpufreq_gov_conservative); |
683 | destroy_workqueue(kconservative_wq); | ||
600 | } | 684 | } |
601 | 685 | ||
602 | 686 | ||
603 | MODULE_AUTHOR ("Alexander Clouter <alex-kernel@digriz.org.uk>"); | 687 | MODULE_AUTHOR("Alexander Clouter <alex@digriz.org.uk>"); |
604 | MODULE_DESCRIPTION ("'cpufreq_conservative' - A dynamic cpufreq governor for " | 688 | MODULE_DESCRIPTION("'cpufreq_conservative' - A dynamic cpufreq governor for " |
605 | "Low Latency Frequency Transition capable processors " | 689 | "Low Latency Frequency Transition capable processors " |
606 | "optimised for use in a battery environment"); | 690 | "optimised for use in a battery environment"); |
607 | MODULE_LICENSE ("GPL"); | 691 | MODULE_LICENSE("GPL"); |
608 | 692 | ||
609 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_CONSERVATIVE | 693 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_CONSERVATIVE |
610 | fs_initcall(cpufreq_gov_dbs_init); | 694 | fs_initcall(cpufreq_gov_dbs_init); |
diff --git a/drivers/cpufreq/cpufreq_ondemand.c b/drivers/cpufreq/cpufreq_ondemand.c index 6f45b1658a67..338f428a15b7 100644 --- a/drivers/cpufreq/cpufreq_ondemand.c +++ b/drivers/cpufreq/cpufreq_ondemand.c | |||
@@ -21,6 +21,7 @@ | |||
21 | #include <linux/hrtimer.h> | 21 | #include <linux/hrtimer.h> |
22 | #include <linux/tick.h> | 22 | #include <linux/tick.h> |
23 | #include <linux/ktime.h> | 23 | #include <linux/ktime.h> |
24 | #include <linux/sched.h> | ||
24 | 25 | ||
25 | /* | 26 | /* |
26 | * dbs is used in this file as a shortform for demandbased switching | 27 | * dbs is used in this file as a shortform for demandbased switching |
@@ -51,8 +52,20 @@ static unsigned int def_sampling_rate; | |||
51 | (MIN_SAMPLING_RATE_RATIO * jiffies_to_usecs(10)) | 52 | (MIN_SAMPLING_RATE_RATIO * jiffies_to_usecs(10)) |
52 | #define MIN_SAMPLING_RATE \ | 53 | #define MIN_SAMPLING_RATE \ |
53 | (def_sampling_rate / MIN_SAMPLING_RATE_RATIO) | 54 | (def_sampling_rate / MIN_SAMPLING_RATE_RATIO) |
55 | /* Above MIN_SAMPLING_RATE will vanish with its sysfs file soon | ||
56 | * Define the minimal settable sampling rate to the greater of: | ||
57 | * - "HW transition latency" * 100 (same as default sampling / 10) | ||
58 | * - MIN_STAT_SAMPLING_RATE | ||
59 | * To avoid that userspace shoots itself. | ||
60 | */ | ||
61 | static unsigned int minimum_sampling_rate(void) | ||
62 | { | ||
63 | return max(def_sampling_rate / 10, MIN_STAT_SAMPLING_RATE); | ||
64 | } | ||
65 | |||
66 | /* This will also vanish soon with removing sampling_rate_max */ | ||
54 | #define MAX_SAMPLING_RATE (500 * def_sampling_rate) | 67 | #define MAX_SAMPLING_RATE (500 * def_sampling_rate) |
55 | #define DEF_SAMPLING_RATE_LATENCY_MULTIPLIER (1000) | 68 | #define LATENCY_MULTIPLIER (1000) |
56 | #define TRANSITION_LATENCY_LIMIT (10 * 1000 * 1000) | 69 | #define TRANSITION_LATENCY_LIMIT (10 * 1000 * 1000) |
57 | 70 | ||
58 | static void do_dbs_timer(struct work_struct *work); | 71 | static void do_dbs_timer(struct work_struct *work); |
@@ -65,14 +78,14 @@ struct cpu_dbs_info_s { | |||
65 | cputime64_t prev_cpu_wall; | 78 | cputime64_t prev_cpu_wall; |
66 | cputime64_t prev_cpu_nice; | 79 | cputime64_t prev_cpu_nice; |
67 | struct cpufreq_policy *cur_policy; | 80 | struct cpufreq_policy *cur_policy; |
68 | struct delayed_work work; | 81 | struct delayed_work work; |
69 | struct cpufreq_frequency_table *freq_table; | 82 | struct cpufreq_frequency_table *freq_table; |
70 | unsigned int freq_lo; | 83 | unsigned int freq_lo; |
71 | unsigned int freq_lo_jiffies; | 84 | unsigned int freq_lo_jiffies; |
72 | unsigned int freq_hi_jiffies; | 85 | unsigned int freq_hi_jiffies; |
73 | int cpu; | 86 | int cpu; |
74 | unsigned int enable:1, | 87 | unsigned int enable:1, |
75 | sample_type:1; | 88 | sample_type:1; |
76 | }; | 89 | }; |
77 | static DEFINE_PER_CPU(struct cpu_dbs_info_s, cpu_dbs_info); | 90 | static DEFINE_PER_CPU(struct cpu_dbs_info_s, cpu_dbs_info); |
78 | 91 | ||
@@ -203,12 +216,28 @@ static void ondemand_powersave_bias_init(void) | |||
203 | /************************** sysfs interface ************************/ | 216 | /************************** sysfs interface ************************/ |
204 | static ssize_t show_sampling_rate_max(struct cpufreq_policy *policy, char *buf) | 217 | static ssize_t show_sampling_rate_max(struct cpufreq_policy *policy, char *buf) |
205 | { | 218 | { |
206 | return sprintf (buf, "%u\n", MAX_SAMPLING_RATE); | 219 | static int print_once; |
220 | |||
221 | if (!print_once) { | ||
222 | printk(KERN_INFO "CPUFREQ: ondemand sampling_rate_max " | ||
223 | "sysfs file is deprecated - used by: %s\n", | ||
224 | current->comm); | ||
225 | print_once = 1; | ||
226 | } | ||
227 | return sprintf(buf, "%u\n", MAX_SAMPLING_RATE); | ||
207 | } | 228 | } |
208 | 229 | ||
209 | static ssize_t show_sampling_rate_min(struct cpufreq_policy *policy, char *buf) | 230 | static ssize_t show_sampling_rate_min(struct cpufreq_policy *policy, char *buf) |
210 | { | 231 | { |
211 | return sprintf (buf, "%u\n", MIN_SAMPLING_RATE); | 232 | static int print_once; |
233 | |||
234 | if (!print_once) { | ||
235 | printk(KERN_INFO "CPUFREQ: ondemand sampling_rate_min " | ||
236 | "sysfs file is deprecated - used by: %s\n", | ||
237 | current->comm); | ||
238 | print_once = 1; | ||
239 | } | ||
240 | return sprintf(buf, "%u\n", MIN_SAMPLING_RATE); | ||
212 | } | 241 | } |
213 | 242 | ||
214 | #define define_one_ro(_name) \ | 243 | #define define_one_ro(_name) \ |
@@ -238,13 +267,11 @@ static ssize_t store_sampling_rate(struct cpufreq_policy *unused, | |||
238 | ret = sscanf(buf, "%u", &input); | 267 | ret = sscanf(buf, "%u", &input); |
239 | 268 | ||
240 | mutex_lock(&dbs_mutex); | 269 | mutex_lock(&dbs_mutex); |
241 | if (ret != 1 || input > MAX_SAMPLING_RATE | 270 | if (ret != 1) { |
242 | || input < MIN_SAMPLING_RATE) { | ||
243 | mutex_unlock(&dbs_mutex); | 271 | mutex_unlock(&dbs_mutex); |
244 | return -EINVAL; | 272 | return -EINVAL; |
245 | } | 273 | } |
246 | 274 | dbs_tuners_ins.sampling_rate = max(input, minimum_sampling_rate()); | |
247 | dbs_tuners_ins.sampling_rate = input; | ||
248 | mutex_unlock(&dbs_mutex); | 275 | mutex_unlock(&dbs_mutex); |
249 | 276 | ||
250 | return count; | 277 | return count; |
@@ -279,14 +306,14 @@ static ssize_t store_ignore_nice_load(struct cpufreq_policy *policy, | |||
279 | unsigned int j; | 306 | unsigned int j; |
280 | 307 | ||
281 | ret = sscanf(buf, "%u", &input); | 308 | ret = sscanf(buf, "%u", &input); |
282 | if ( ret != 1 ) | 309 | if (ret != 1) |
283 | return -EINVAL; | 310 | return -EINVAL; |
284 | 311 | ||
285 | if ( input > 1 ) | 312 | if (input > 1) |
286 | input = 1; | 313 | input = 1; |
287 | 314 | ||
288 | mutex_lock(&dbs_mutex); | 315 | mutex_lock(&dbs_mutex); |
289 | if ( input == dbs_tuners_ins.ignore_nice ) { /* nothing to do */ | 316 | if (input == dbs_tuners_ins.ignore_nice) { /* nothing to do */ |
290 | mutex_unlock(&dbs_mutex); | 317 | mutex_unlock(&dbs_mutex); |
291 | return count; | 318 | return count; |
292 | } | 319 | } |
@@ -337,7 +364,7 @@ define_one_rw(up_threshold); | |||
337 | define_one_rw(ignore_nice_load); | 364 | define_one_rw(ignore_nice_load); |
338 | define_one_rw(powersave_bias); | 365 | define_one_rw(powersave_bias); |
339 | 366 | ||
340 | static struct attribute * dbs_attributes[] = { | 367 | static struct attribute *dbs_attributes[] = { |
341 | &sampling_rate_max.attr, | 368 | &sampling_rate_max.attr, |
342 | &sampling_rate_min.attr, | 369 | &sampling_rate_min.attr, |
343 | &sampling_rate.attr, | 370 | &sampling_rate.attr, |
@@ -512,8 +539,7 @@ static void do_dbs_timer(struct work_struct *work) | |||
512 | } | 539 | } |
513 | } else { | 540 | } else { |
514 | __cpufreq_driver_target(dbs_info->cur_policy, | 541 | __cpufreq_driver_target(dbs_info->cur_policy, |
515 | dbs_info->freq_lo, | 542 | dbs_info->freq_lo, CPUFREQ_RELATION_H); |
516 | CPUFREQ_RELATION_H); | ||
517 | } | 543 | } |
518 | queue_delayed_work_on(cpu, kondemand_wq, &dbs_info->work, delay); | 544 | queue_delayed_work_on(cpu, kondemand_wq, &dbs_info->work, delay); |
519 | unlock_policy_rwsem_write(cpu); | 545 | unlock_policy_rwsem_write(cpu); |
@@ -530,7 +556,7 @@ static inline void dbs_timer_init(struct cpu_dbs_info_s *dbs_info) | |||
530 | dbs_info->sample_type = DBS_NORMAL_SAMPLE; | 556 | dbs_info->sample_type = DBS_NORMAL_SAMPLE; |
531 | INIT_DELAYED_WORK_DEFERRABLE(&dbs_info->work, do_dbs_timer); | 557 | INIT_DELAYED_WORK_DEFERRABLE(&dbs_info->work, do_dbs_timer); |
532 | queue_delayed_work_on(dbs_info->cpu, kondemand_wq, &dbs_info->work, | 558 | queue_delayed_work_on(dbs_info->cpu, kondemand_wq, &dbs_info->work, |
533 | delay); | 559 | delay); |
534 | } | 560 | } |
535 | 561 | ||
536 | static inline void dbs_timer_exit(struct cpu_dbs_info_s *dbs_info) | 562 | static inline void dbs_timer_exit(struct cpu_dbs_info_s *dbs_info) |
@@ -591,11 +617,9 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
591 | if (latency == 0) | 617 | if (latency == 0) |
592 | latency = 1; | 618 | latency = 1; |
593 | 619 | ||
594 | def_sampling_rate = latency * | 620 | def_sampling_rate = |
595 | DEF_SAMPLING_RATE_LATENCY_MULTIPLIER; | 621 | max(latency * LATENCY_MULTIPLIER, |
596 | 622 | MIN_STAT_SAMPLING_RATE); | |
597 | if (def_sampling_rate < MIN_STAT_SAMPLING_RATE) | ||
598 | def_sampling_rate = MIN_STAT_SAMPLING_RATE; | ||
599 | 623 | ||
600 | dbs_tuners_ins.sampling_rate = def_sampling_rate; | 624 | dbs_tuners_ins.sampling_rate = def_sampling_rate; |
601 | } | 625 | } |
@@ -617,12 +641,10 @@ static int cpufreq_governor_dbs(struct cpufreq_policy *policy, | |||
617 | mutex_lock(&dbs_mutex); | 641 | mutex_lock(&dbs_mutex); |
618 | if (policy->max < this_dbs_info->cur_policy->cur) | 642 | if (policy->max < this_dbs_info->cur_policy->cur) |
619 | __cpufreq_driver_target(this_dbs_info->cur_policy, | 643 | __cpufreq_driver_target(this_dbs_info->cur_policy, |
620 | policy->max, | 644 | policy->max, CPUFREQ_RELATION_H); |
621 | CPUFREQ_RELATION_H); | ||
622 | else if (policy->min > this_dbs_info->cur_policy->cur) | 645 | else if (policy->min > this_dbs_info->cur_policy->cur) |
623 | __cpufreq_driver_target(this_dbs_info->cur_policy, | 646 | __cpufreq_driver_target(this_dbs_info->cur_policy, |
624 | policy->min, | 647 | policy->min, CPUFREQ_RELATION_L); |
625 | CPUFREQ_RELATION_L); | ||
626 | mutex_unlock(&dbs_mutex); | 648 | mutex_unlock(&dbs_mutex); |
627 | break; | 649 | break; |
628 | } | 650 | } |
@@ -677,7 +699,7 @@ static void __exit cpufreq_gov_dbs_exit(void) | |||
677 | MODULE_AUTHOR("Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>"); | 699 | MODULE_AUTHOR("Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>"); |
678 | MODULE_AUTHOR("Alexey Starikovskiy <alexey.y.starikovskiy@intel.com>"); | 700 | MODULE_AUTHOR("Alexey Starikovskiy <alexey.y.starikovskiy@intel.com>"); |
679 | MODULE_DESCRIPTION("'cpufreq_ondemand' - A dynamic cpufreq governor for " | 701 | MODULE_DESCRIPTION("'cpufreq_ondemand' - A dynamic cpufreq governor for " |
680 | "Low Latency Frequency Transition capable processors"); | 702 | "Low Latency Frequency Transition capable processors"); |
681 | MODULE_LICENSE("GPL"); | 703 | MODULE_LICENSE("GPL"); |
682 | 704 | ||
683 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND | 705 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_ONDEMAND |
diff --git a/drivers/cpufreq/cpufreq_stats.c b/drivers/cpufreq/cpufreq_stats.c index c0ff97d375d7..5a62d678dd19 100644 --- a/drivers/cpufreq/cpufreq_stats.c +++ b/drivers/cpufreq/cpufreq_stats.c | |||
@@ -2,7 +2,7 @@ | |||
2 | * drivers/cpufreq/cpufreq_stats.c | 2 | * drivers/cpufreq/cpufreq_stats.c |
3 | * | 3 | * |
4 | * Copyright (C) 2003-2004 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>. | 4 | * Copyright (C) 2003-2004 Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>. |
5 | * (C) 2004 Zou Nan hai <nanhai.zou@intel.com>. | 5 | * (C) 2004 Zou Nan hai <nanhai.zou@intel.com>. |
6 | * | 6 | * |
7 | * This program is free software; you can redistribute it and/or modify | 7 | * This program is free software; you can redistribute it and/or modify |
8 | * it under the terms of the GNU General Public License version 2 as | 8 | * it under the terms of the GNU General Public License version 2 as |
@@ -23,7 +23,7 @@ | |||
23 | 23 | ||
24 | static spinlock_t cpufreq_stats_lock; | 24 | static spinlock_t cpufreq_stats_lock; |
25 | 25 | ||
26 | #define CPUFREQ_STATDEVICE_ATTR(_name,_mode,_show) \ | 26 | #define CPUFREQ_STATDEVICE_ATTR(_name, _mode, _show) \ |
27 | static struct freq_attr _attr_##_name = {\ | 27 | static struct freq_attr _attr_##_name = {\ |
28 | .attr = {.name = __stringify(_name), .mode = _mode, }, \ | 28 | .attr = {.name = __stringify(_name), .mode = _mode, }, \ |
29 | .show = _show,\ | 29 | .show = _show,\ |
@@ -50,8 +50,7 @@ struct cpufreq_stats_attribute { | |||
50 | ssize_t(*show) (struct cpufreq_stats *, char *); | 50 | ssize_t(*show) (struct cpufreq_stats *, char *); |
51 | }; | 51 | }; |
52 | 52 | ||
53 | static int | 53 | static int cpufreq_stats_update(unsigned int cpu) |
54 | cpufreq_stats_update (unsigned int cpu) | ||
55 | { | 54 | { |
56 | struct cpufreq_stats *stat; | 55 | struct cpufreq_stats *stat; |
57 | unsigned long long cur_time; | 56 | unsigned long long cur_time; |
@@ -68,8 +67,7 @@ cpufreq_stats_update (unsigned int cpu) | |||
68 | return 0; | 67 | return 0; |
69 | } | 68 | } |
70 | 69 | ||
71 | static ssize_t | 70 | static ssize_t show_total_trans(struct cpufreq_policy *policy, char *buf) |
72 | show_total_trans(struct cpufreq_policy *policy, char *buf) | ||
73 | { | 71 | { |
74 | struct cpufreq_stats *stat = per_cpu(cpufreq_stats_table, policy->cpu); | 72 | struct cpufreq_stats *stat = per_cpu(cpufreq_stats_table, policy->cpu); |
75 | if (!stat) | 73 | if (!stat) |
@@ -78,8 +76,7 @@ show_total_trans(struct cpufreq_policy *policy, char *buf) | |||
78 | per_cpu(cpufreq_stats_table, stat->cpu)->total_trans); | 76 | per_cpu(cpufreq_stats_table, stat->cpu)->total_trans); |
79 | } | 77 | } |
80 | 78 | ||
81 | static ssize_t | 79 | static ssize_t show_time_in_state(struct cpufreq_policy *policy, char *buf) |
82 | show_time_in_state(struct cpufreq_policy *policy, char *buf) | ||
83 | { | 80 | { |
84 | ssize_t len = 0; | 81 | ssize_t len = 0; |
85 | int i; | 82 | int i; |
@@ -89,14 +86,14 @@ show_time_in_state(struct cpufreq_policy *policy, char *buf) | |||
89 | cpufreq_stats_update(stat->cpu); | 86 | cpufreq_stats_update(stat->cpu); |
90 | for (i = 0; i < stat->state_num; i++) { | 87 | for (i = 0; i < stat->state_num; i++) { |
91 | len += sprintf(buf + len, "%u %llu\n", stat->freq_table[i], | 88 | len += sprintf(buf + len, "%u %llu\n", stat->freq_table[i], |
92 | (unsigned long long)cputime64_to_clock_t(stat->time_in_state[i])); | 89 | (unsigned long long) |
90 | cputime64_to_clock_t(stat->time_in_state[i])); | ||
93 | } | 91 | } |
94 | return len; | 92 | return len; |
95 | } | 93 | } |
96 | 94 | ||
97 | #ifdef CONFIG_CPU_FREQ_STAT_DETAILS | 95 | #ifdef CONFIG_CPU_FREQ_STAT_DETAILS |
98 | static ssize_t | 96 | static ssize_t show_trans_table(struct cpufreq_policy *policy, char *buf) |
99 | show_trans_table(struct cpufreq_policy *policy, char *buf) | ||
100 | { | 97 | { |
101 | ssize_t len = 0; | 98 | ssize_t len = 0; |
102 | int i, j; | 99 | int i, j; |
@@ -139,11 +136,11 @@ show_trans_table(struct cpufreq_policy *policy, char *buf) | |||
139 | return PAGE_SIZE; | 136 | return PAGE_SIZE; |
140 | return len; | 137 | return len; |
141 | } | 138 | } |
142 | CPUFREQ_STATDEVICE_ATTR(trans_table,0444,show_trans_table); | 139 | CPUFREQ_STATDEVICE_ATTR(trans_table, 0444, show_trans_table); |
143 | #endif | 140 | #endif |
144 | 141 | ||
145 | CPUFREQ_STATDEVICE_ATTR(total_trans,0444,show_total_trans); | 142 | CPUFREQ_STATDEVICE_ATTR(total_trans, 0444, show_total_trans); |
146 | CPUFREQ_STATDEVICE_ATTR(time_in_state,0444,show_time_in_state); | 143 | CPUFREQ_STATDEVICE_ATTR(time_in_state, 0444, show_time_in_state); |
147 | 144 | ||
148 | static struct attribute *default_attrs[] = { | 145 | static struct attribute *default_attrs[] = { |
149 | &_attr_total_trans.attr, | 146 | &_attr_total_trans.attr, |
@@ -158,8 +155,7 @@ static struct attribute_group stats_attr_group = { | |||
158 | .name = "stats" | 155 | .name = "stats" |
159 | }; | 156 | }; |
160 | 157 | ||
161 | static int | 158 | static int freq_table_get_index(struct cpufreq_stats *stat, unsigned int freq) |
162 | freq_table_get_index(struct cpufreq_stats *stat, unsigned int freq) | ||
163 | { | 159 | { |
164 | int index; | 160 | int index; |
165 | for (index = 0; index < stat->max_state; index++) | 161 | for (index = 0; index < stat->max_state; index++) |
@@ -183,8 +179,7 @@ static void cpufreq_stats_free_table(unsigned int cpu) | |||
183 | cpufreq_cpu_put(policy); | 179 | cpufreq_cpu_put(policy); |
184 | } | 180 | } |
185 | 181 | ||
186 | static int | 182 | static int cpufreq_stats_create_table(struct cpufreq_policy *policy, |
187 | cpufreq_stats_create_table (struct cpufreq_policy *policy, | ||
188 | struct cpufreq_frequency_table *table) | 183 | struct cpufreq_frequency_table *table) |
189 | { | 184 | { |
190 | unsigned int i, j, count = 0, ret = 0; | 185 | unsigned int i, j, count = 0, ret = 0; |
@@ -194,7 +189,8 @@ cpufreq_stats_create_table (struct cpufreq_policy *policy, | |||
194 | unsigned int cpu = policy->cpu; | 189 | unsigned int cpu = policy->cpu; |
195 | if (per_cpu(cpufreq_stats_table, cpu)) | 190 | if (per_cpu(cpufreq_stats_table, cpu)) |
196 | return -EBUSY; | 191 | return -EBUSY; |
197 | if ((stat = kzalloc(sizeof(struct cpufreq_stats), GFP_KERNEL)) == NULL) | 192 | stat = kzalloc(sizeof(struct cpufreq_stats), GFP_KERNEL); |
193 | if ((stat) == NULL) | ||
198 | return -ENOMEM; | 194 | return -ENOMEM; |
199 | 195 | ||
200 | data = cpufreq_cpu_get(cpu); | 196 | data = cpufreq_cpu_get(cpu); |
@@ -203,13 +199,14 @@ cpufreq_stats_create_table (struct cpufreq_policy *policy, | |||
203 | goto error_get_fail; | 199 | goto error_get_fail; |
204 | } | 200 | } |
205 | 201 | ||
206 | if ((ret = sysfs_create_group(&data->kobj, &stats_attr_group))) | 202 | ret = sysfs_create_group(&data->kobj, &stats_attr_group); |
203 | if (ret) | ||
207 | goto error_out; | 204 | goto error_out; |
208 | 205 | ||
209 | stat->cpu = cpu; | 206 | stat->cpu = cpu; |
210 | per_cpu(cpufreq_stats_table, cpu) = stat; | 207 | per_cpu(cpufreq_stats_table, cpu) = stat; |
211 | 208 | ||
212 | for (i=0; table[i].frequency != CPUFREQ_TABLE_END; i++) { | 209 | for (i = 0; table[i].frequency != CPUFREQ_TABLE_END; i++) { |
213 | unsigned int freq = table[i].frequency; | 210 | unsigned int freq = table[i].frequency; |
214 | if (freq == CPUFREQ_ENTRY_INVALID) | 211 | if (freq == CPUFREQ_ENTRY_INVALID) |
215 | continue; | 212 | continue; |
@@ -255,9 +252,8 @@ error_get_fail: | |||
255 | return ret; | 252 | return ret; |
256 | } | 253 | } |
257 | 254 | ||
258 | static int | 255 | static int cpufreq_stat_notifier_policy(struct notifier_block *nb, |
259 | cpufreq_stat_notifier_policy (struct notifier_block *nb, unsigned long val, | 256 | unsigned long val, void *data) |
260 | void *data) | ||
261 | { | 257 | { |
262 | int ret; | 258 | int ret; |
263 | struct cpufreq_policy *policy = data; | 259 | struct cpufreq_policy *policy = data; |
@@ -268,14 +264,14 @@ cpufreq_stat_notifier_policy (struct notifier_block *nb, unsigned long val, | |||
268 | table = cpufreq_frequency_get_table(cpu); | 264 | table = cpufreq_frequency_get_table(cpu); |
269 | if (!table) | 265 | if (!table) |
270 | return 0; | 266 | return 0; |
271 | if ((ret = cpufreq_stats_create_table(policy, table))) | 267 | ret = cpufreq_stats_create_table(policy, table); |
268 | if (ret) | ||
272 | return ret; | 269 | return ret; |
273 | return 0; | 270 | return 0; |
274 | } | 271 | } |
275 | 272 | ||
276 | static int | 273 | static int cpufreq_stat_notifier_trans(struct notifier_block *nb, |
277 | cpufreq_stat_notifier_trans (struct notifier_block *nb, unsigned long val, | 274 | unsigned long val, void *data) |
278 | void *data) | ||
279 | { | 275 | { |
280 | struct cpufreq_freqs *freq = data; | 276 | struct cpufreq_freqs *freq = data; |
281 | struct cpufreq_stats *stat; | 277 | struct cpufreq_stats *stat; |
@@ -340,19 +336,20 @@ static struct notifier_block notifier_trans_block = { | |||
340 | .notifier_call = cpufreq_stat_notifier_trans | 336 | .notifier_call = cpufreq_stat_notifier_trans |
341 | }; | 337 | }; |
342 | 338 | ||
343 | static int | 339 | static int __init cpufreq_stats_init(void) |
344 | __init cpufreq_stats_init(void) | ||
345 | { | 340 | { |
346 | int ret; | 341 | int ret; |
347 | unsigned int cpu; | 342 | unsigned int cpu; |
348 | 343 | ||
349 | spin_lock_init(&cpufreq_stats_lock); | 344 | spin_lock_init(&cpufreq_stats_lock); |
350 | if ((ret = cpufreq_register_notifier(¬ifier_policy_block, | 345 | ret = cpufreq_register_notifier(¬ifier_policy_block, |
351 | CPUFREQ_POLICY_NOTIFIER))) | 346 | CPUFREQ_POLICY_NOTIFIER); |
347 | if (ret) | ||
352 | return ret; | 348 | return ret; |
353 | 349 | ||
354 | if ((ret = cpufreq_register_notifier(¬ifier_trans_block, | 350 | ret = cpufreq_register_notifier(¬ifier_trans_block, |
355 | CPUFREQ_TRANSITION_NOTIFIER))) { | 351 | CPUFREQ_TRANSITION_NOTIFIER); |
352 | if (ret) { | ||
356 | cpufreq_unregister_notifier(¬ifier_policy_block, | 353 | cpufreq_unregister_notifier(¬ifier_policy_block, |
357 | CPUFREQ_POLICY_NOTIFIER); | 354 | CPUFREQ_POLICY_NOTIFIER); |
358 | return ret; | 355 | return ret; |
@@ -364,8 +361,7 @@ __init cpufreq_stats_init(void) | |||
364 | } | 361 | } |
365 | return 0; | 362 | return 0; |
366 | } | 363 | } |
367 | static void | 364 | static void __exit cpufreq_stats_exit(void) |
368 | __exit cpufreq_stats_exit(void) | ||
369 | { | 365 | { |
370 | unsigned int cpu; | 366 | unsigned int cpu; |
371 | 367 | ||
@@ -379,10 +375,10 @@ __exit cpufreq_stats_exit(void) | |||
379 | } | 375 | } |
380 | } | 376 | } |
381 | 377 | ||
382 | MODULE_AUTHOR ("Zou Nan hai <nanhai.zou@intel.com>"); | 378 | MODULE_AUTHOR("Zou Nan hai <nanhai.zou@intel.com>"); |
383 | MODULE_DESCRIPTION ("'cpufreq_stats' - A driver to export cpufreq stats " | 379 | MODULE_DESCRIPTION("'cpufreq_stats' - A driver to export cpufreq stats " |
384 | "through sysfs filesystem"); | 380 | "through sysfs filesystem"); |
385 | MODULE_LICENSE ("GPL"); | 381 | MODULE_LICENSE("GPL"); |
386 | 382 | ||
387 | module_init(cpufreq_stats_init); | 383 | module_init(cpufreq_stats_init); |
388 | module_exit(cpufreq_stats_exit); | 384 | module_exit(cpufreq_stats_exit); |
diff --git a/drivers/cpufreq/cpufreq_userspace.c b/drivers/cpufreq/cpufreq_userspace.c index 1442bbada053..66d2d1d6c80f 100644 --- a/drivers/cpufreq/cpufreq_userspace.c +++ b/drivers/cpufreq/cpufreq_userspace.c | |||
@@ -24,9 +24,6 @@ | |||
24 | #include <linux/sysfs.h> | 24 | #include <linux/sysfs.h> |
25 | #include <linux/mutex.h> | 25 | #include <linux/mutex.h> |
26 | 26 | ||
27 | #include <asm/uaccess.h> | ||
28 | |||
29 | |||
30 | /** | 27 | /** |
31 | * A few values needed by the userspace governor | 28 | * A few values needed by the userspace governor |
32 | */ | 29 | */ |
@@ -37,7 +34,7 @@ static DEFINE_PER_CPU(unsigned int, cpu_set_freq); /* CPU freq desired by | |||
37 | userspace */ | 34 | userspace */ |
38 | static DEFINE_PER_CPU(unsigned int, cpu_is_managed); | 35 | static DEFINE_PER_CPU(unsigned int, cpu_is_managed); |
39 | 36 | ||
40 | static DEFINE_MUTEX (userspace_mutex); | 37 | static DEFINE_MUTEX(userspace_mutex); |
41 | static int cpus_using_userspace_governor; | 38 | static int cpus_using_userspace_governor; |
42 | 39 | ||
43 | #define dprintk(msg...) \ | 40 | #define dprintk(msg...) \ |
@@ -46,9 +43,9 @@ static int cpus_using_userspace_governor; | |||
46 | /* keep track of frequency transitions */ | 43 | /* keep track of frequency transitions */ |
47 | static int | 44 | static int |
48 | userspace_cpufreq_notifier(struct notifier_block *nb, unsigned long val, | 45 | userspace_cpufreq_notifier(struct notifier_block *nb, unsigned long val, |
49 | void *data) | 46 | void *data) |
50 | { | 47 | { |
51 | struct cpufreq_freqs *freq = data; | 48 | struct cpufreq_freqs *freq = data; |
52 | 49 | ||
53 | if (!per_cpu(cpu_is_managed, freq->cpu)) | 50 | if (!per_cpu(cpu_is_managed, freq->cpu)) |
54 | return 0; | 51 | return 0; |
@@ -57,11 +54,11 @@ userspace_cpufreq_notifier(struct notifier_block *nb, unsigned long val, | |||
57 | freq->cpu, freq->new); | 54 | freq->cpu, freq->new); |
58 | per_cpu(cpu_cur_freq, freq->cpu) = freq->new; | 55 | per_cpu(cpu_cur_freq, freq->cpu) = freq->new; |
59 | 56 | ||
60 | return 0; | 57 | return 0; |
61 | } | 58 | } |
62 | 59 | ||
63 | static struct notifier_block userspace_cpufreq_notifier_block = { | 60 | static struct notifier_block userspace_cpufreq_notifier_block = { |
64 | .notifier_call = userspace_cpufreq_notifier | 61 | .notifier_call = userspace_cpufreq_notifier |
65 | }; | 62 | }; |
66 | 63 | ||
67 | 64 | ||
@@ -93,8 +90,11 @@ static int cpufreq_set(struct cpufreq_policy *policy, unsigned int freq) | |||
93 | * We're safe from concurrent calls to ->target() here | 90 | * We're safe from concurrent calls to ->target() here |
94 | * as we hold the userspace_mutex lock. If we were calling | 91 | * as we hold the userspace_mutex lock. If we were calling |
95 | * cpufreq_driver_target, a deadlock situation might occur: | 92 | * cpufreq_driver_target, a deadlock situation might occur: |
96 | * A: cpufreq_set (lock userspace_mutex) -> cpufreq_driver_target(lock policy->lock) | 93 | * A: cpufreq_set (lock userspace_mutex) -> |
97 | * B: cpufreq_set_policy(lock policy->lock) -> __cpufreq_governor -> cpufreq_governor_userspace (lock userspace_mutex) | 94 | * cpufreq_driver_target(lock policy->lock) |
95 | * B: cpufreq_set_policy(lock policy->lock) -> | ||
96 | * __cpufreq_governor -> | ||
97 | * cpufreq_governor_userspace (lock userspace_mutex) | ||
98 | */ | 98 | */ |
99 | ret = __cpufreq_driver_target(policy, freq, CPUFREQ_RELATION_L); | 99 | ret = __cpufreq_driver_target(policy, freq, CPUFREQ_RELATION_L); |
100 | 100 | ||
@@ -210,9 +210,10 @@ static void __exit cpufreq_gov_userspace_exit(void) | |||
210 | } | 210 | } |
211 | 211 | ||
212 | 212 | ||
213 | MODULE_AUTHOR ("Dominik Brodowski <linux@brodo.de>, Russell King <rmk@arm.linux.org.uk>"); | 213 | MODULE_AUTHOR("Dominik Brodowski <linux@brodo.de>, " |
214 | MODULE_DESCRIPTION ("CPUfreq policy governor 'userspace'"); | 214 | "Russell King <rmk@arm.linux.org.uk>"); |
215 | MODULE_LICENSE ("GPL"); | 215 | MODULE_DESCRIPTION("CPUfreq policy governor 'userspace'"); |
216 | MODULE_LICENSE("GPL"); | ||
216 | 217 | ||
217 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_USERSPACE | 218 | #ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_USERSPACE |
218 | fs_initcall(cpufreq_gov_userspace_init); | 219 | fs_initcall(cpufreq_gov_userspace_init); |
diff --git a/drivers/cpufreq/freq_table.c b/drivers/cpufreq/freq_table.c index 9071d80fbba2..a9bd3a05a684 100644 --- a/drivers/cpufreq/freq_table.c +++ b/drivers/cpufreq/freq_table.c | |||
@@ -28,7 +28,7 @@ int cpufreq_frequency_table_cpuinfo(struct cpufreq_policy *policy, | |||
28 | unsigned int max_freq = 0; | 28 | unsigned int max_freq = 0; |
29 | unsigned int i; | 29 | unsigned int i; |
30 | 30 | ||
31 | for (i=0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { | 31 | for (i = 0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { |
32 | unsigned int freq = table[i].frequency; | 32 | unsigned int freq = table[i].frequency; |
33 | if (freq == CPUFREQ_ENTRY_INVALID) { | 33 | if (freq == CPUFREQ_ENTRY_INVALID) { |
34 | dprintk("table entry %u is invalid, skipping\n", i); | 34 | dprintk("table entry %u is invalid, skipping\n", i); |
@@ -70,7 +70,7 @@ int cpufreq_frequency_table_verify(struct cpufreq_policy *policy, | |||
70 | cpufreq_verify_within_limits(policy, policy->cpuinfo.min_freq, | 70 | cpufreq_verify_within_limits(policy, policy->cpuinfo.min_freq, |
71 | policy->cpuinfo.max_freq); | 71 | policy->cpuinfo.max_freq); |
72 | 72 | ||
73 | for (i=0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { | 73 | for (i = 0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { |
74 | unsigned int freq = table[i].frequency; | 74 | unsigned int freq = table[i].frequency; |
75 | if (freq == CPUFREQ_ENTRY_INVALID) | 75 | if (freq == CPUFREQ_ENTRY_INVALID) |
76 | continue; | 76 | continue; |
@@ -125,13 +125,13 @@ int cpufreq_frequency_table_target(struct cpufreq_policy *policy, | |||
125 | if (!cpu_online(policy->cpu)) | 125 | if (!cpu_online(policy->cpu)) |
126 | return -EINVAL; | 126 | return -EINVAL; |
127 | 127 | ||
128 | for (i=0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { | 128 | for (i = 0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { |
129 | unsigned int freq = table[i].frequency; | 129 | unsigned int freq = table[i].frequency; |
130 | if (freq == CPUFREQ_ENTRY_INVALID) | 130 | if (freq == CPUFREQ_ENTRY_INVALID) |
131 | continue; | 131 | continue; |
132 | if ((freq < policy->min) || (freq > policy->max)) | 132 | if ((freq < policy->min) || (freq > policy->max)) |
133 | continue; | 133 | continue; |
134 | switch(relation) { | 134 | switch (relation) { |
135 | case CPUFREQ_RELATION_H: | 135 | case CPUFREQ_RELATION_H: |
136 | if (freq <= target_freq) { | 136 | if (freq <= target_freq) { |
137 | if (freq >= optimal.frequency) { | 137 | if (freq >= optimal.frequency) { |
@@ -178,7 +178,7 @@ static DEFINE_PER_CPU(struct cpufreq_frequency_table *, show_table); | |||
178 | /** | 178 | /** |
179 | * show_available_freqs - show available frequencies for the specified CPU | 179 | * show_available_freqs - show available frequencies for the specified CPU |
180 | */ | 180 | */ |
181 | static ssize_t show_available_freqs (struct cpufreq_policy *policy, char *buf) | 181 | static ssize_t show_available_freqs(struct cpufreq_policy *policy, char *buf) |
182 | { | 182 | { |
183 | unsigned int i = 0; | 183 | unsigned int i = 0; |
184 | unsigned int cpu = policy->cpu; | 184 | unsigned int cpu = policy->cpu; |
@@ -190,7 +190,7 @@ static ssize_t show_available_freqs (struct cpufreq_policy *policy, char *buf) | |||
190 | 190 | ||
191 | table = per_cpu(show_table, cpu); | 191 | table = per_cpu(show_table, cpu); |
192 | 192 | ||
193 | for (i=0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { | 193 | for (i = 0; (table[i].frequency != CPUFREQ_TABLE_END); i++) { |
194 | if (table[i].frequency == CPUFREQ_ENTRY_INVALID) | 194 | if (table[i].frequency == CPUFREQ_ENTRY_INVALID) |
195 | continue; | 195 | continue; |
196 | count += sprintf(&buf[count], "%d ", table[i].frequency); | 196 | count += sprintf(&buf[count], "%d ", table[i].frequency); |
@@ -234,6 +234,6 @@ struct cpufreq_frequency_table *cpufreq_frequency_get_table(unsigned int cpu) | |||
234 | } | 234 | } |
235 | EXPORT_SYMBOL_GPL(cpufreq_frequency_get_table); | 235 | EXPORT_SYMBOL_GPL(cpufreq_frequency_get_table); |
236 | 236 | ||
237 | MODULE_AUTHOR ("Dominik Brodowski <linux@brodo.de>"); | 237 | MODULE_AUTHOR("Dominik Brodowski <linux@brodo.de>"); |
238 | MODULE_DESCRIPTION ("CPUfreq frequency table helpers"); | 238 | MODULE_DESCRIPTION("CPUfreq frequency table helpers"); |
239 | MODULE_LICENSE ("GPL"); | 239 | MODULE_LICENSE("GPL"); |