diff options
Diffstat (limited to 'arch/x86/kernel/tsc_64.c')
-rw-r--r-- | arch/x86/kernel/tsc_64.c | 100 |
1 files changed, 72 insertions, 28 deletions
diff --git a/arch/x86/kernel/tsc_64.c b/arch/x86/kernel/tsc_64.c index 9c70af45b42..947554ddabb 100644 --- a/arch/x86/kernel/tsc_64.c +++ b/arch/x86/kernel/tsc_64.c | |||
@@ -10,6 +10,7 @@ | |||
10 | 10 | ||
11 | #include <asm/hpet.h> | 11 | #include <asm/hpet.h> |
12 | #include <asm/timex.h> | 12 | #include <asm/timex.h> |
13 | #include <asm/timer.h> | ||
13 | 14 | ||
14 | static int notsc __initdata = 0; | 15 | static int notsc __initdata = 0; |
15 | 16 | ||
@@ -18,19 +19,51 @@ EXPORT_SYMBOL(cpu_khz); | |||
18 | unsigned int tsc_khz; | 19 | unsigned int tsc_khz; |
19 | EXPORT_SYMBOL(tsc_khz); | 20 | EXPORT_SYMBOL(tsc_khz); |
20 | 21 | ||
21 | static unsigned int cyc2ns_scale __read_mostly; | 22 | /* Accelerators for sched_clock() |
23 | * convert from cycles(64bits) => nanoseconds (64bits) | ||
24 | * basic equation: | ||
25 | * ns = cycles / (freq / ns_per_sec) | ||
26 | * ns = cycles * (ns_per_sec / freq) | ||
27 | * ns = cycles * (10^9 / (cpu_khz * 10^3)) | ||
28 | * ns = cycles * (10^6 / cpu_khz) | ||
29 | * | ||
30 | * Then we use scaling math (suggested by george@mvista.com) to get: | ||
31 | * ns = cycles * (10^6 * SC / cpu_khz) / SC | ||
32 | * ns = cycles * cyc2ns_scale / SC | ||
33 | * | ||
34 | * And since SC is a constant power of two, we can convert the div | ||
35 | * into a shift. | ||
36 | * | ||
37 | * We can use khz divisor instead of mhz to keep a better precision, since | ||
38 | * cyc2ns_scale is limited to 10^6 * 2^10, which fits in 32 bits. | ||
39 | * (mathieu.desnoyers@polymtl.ca) | ||
40 | * | ||
41 | * -johnstul@us.ibm.com "math is hard, lets go shopping!" | ||
42 | */ | ||
43 | DEFINE_PER_CPU(unsigned long, cyc2ns); | ||
22 | 44 | ||
23 | static inline void set_cyc2ns_scale(unsigned long khz) | 45 | static void set_cyc2ns_scale(unsigned long cpu_khz, int cpu) |
24 | { | 46 | { |
25 | cyc2ns_scale = (NSEC_PER_MSEC << NS_SCALE) / khz; | 47 | unsigned long flags, prev_scale, *scale; |
26 | } | 48 | unsigned long long tsc_now, ns_now; |
27 | 49 | ||
28 | static unsigned long long cycles_2_ns(unsigned long long cyc) | 50 | local_irq_save(flags); |
29 | { | 51 | sched_clock_idle_sleep_event(); |
30 | return (cyc * cyc2ns_scale) >> NS_SCALE; | 52 | |
53 | scale = &per_cpu(cyc2ns, cpu); | ||
54 | |||
55 | rdtscll(tsc_now); | ||
56 | ns_now = __cycles_2_ns(tsc_now); | ||
57 | |||
58 | prev_scale = *scale; | ||
59 | if (cpu_khz) | ||
60 | *scale = (NSEC_PER_MSEC << CYC2NS_SCALE_FACTOR)/cpu_khz; | ||
61 | |||
62 | sched_clock_idle_wakeup_event(0); | ||
63 | local_irq_restore(flags); | ||
31 | } | 64 | } |
32 | 65 | ||
33 | unsigned long long sched_clock(void) | 66 | unsigned long long native_sched_clock(void) |
34 | { | 67 | { |
35 | unsigned long a = 0; | 68 | unsigned long a = 0; |
36 | 69 | ||
@@ -44,12 +77,27 @@ unsigned long long sched_clock(void) | |||
44 | return cycles_2_ns(a); | 77 | return cycles_2_ns(a); |
45 | } | 78 | } |
46 | 79 | ||
80 | /* We need to define a real function for sched_clock, to override the | ||
81 | weak default version */ | ||
82 | #ifdef CONFIG_PARAVIRT | ||
83 | unsigned long long sched_clock(void) | ||
84 | { | ||
85 | return paravirt_sched_clock(); | ||
86 | } | ||
87 | #else | ||
88 | unsigned long long | ||
89 | sched_clock(void) __attribute__((alias("native_sched_clock"))); | ||
90 | #endif | ||
91 | |||
92 | |||
47 | static int tsc_unstable; | 93 | static int tsc_unstable; |
48 | 94 | ||
49 | inline int check_tsc_unstable(void) | 95 | int check_tsc_unstable(void) |
50 | { | 96 | { |
51 | return tsc_unstable; | 97 | return tsc_unstable; |
52 | } | 98 | } |
99 | EXPORT_SYMBOL_GPL(check_tsc_unstable); | ||
100 | |||
53 | #ifdef CONFIG_CPU_FREQ | 101 | #ifdef CONFIG_CPU_FREQ |
54 | 102 | ||
55 | /* Frequency scaling support. Adjust the TSC based timer when the cpu frequency | 103 | /* Frequency scaling support. Adjust the TSC based timer when the cpu frequency |
@@ -100,7 +148,9 @@ static int time_cpufreq_notifier(struct notifier_block *nb, unsigned long val, | |||
100 | mark_tsc_unstable("cpufreq changes"); | 148 | mark_tsc_unstable("cpufreq changes"); |
101 | } | 149 | } |
102 | 150 | ||
103 | set_cyc2ns_scale(tsc_khz_ref); | 151 | preempt_disable(); |
152 | set_cyc2ns_scale(tsc_khz_ref, smp_processor_id()); | ||
153 | preempt_enable(); | ||
104 | 154 | ||
105 | return 0; | 155 | return 0; |
106 | } | 156 | } |
@@ -133,12 +183,12 @@ static unsigned long __init tsc_read_refs(unsigned long *pm, | |||
133 | int i; | 183 | int i; |
134 | 184 | ||
135 | for (i = 0; i < MAX_RETRIES; i++) { | 185 | for (i = 0; i < MAX_RETRIES; i++) { |
136 | t1 = get_cycles_sync(); | 186 | t1 = get_cycles(); |
137 | if (hpet) | 187 | if (hpet) |
138 | *hpet = hpet_readl(HPET_COUNTER) & 0xFFFFFFFF; | 188 | *hpet = hpet_readl(HPET_COUNTER) & 0xFFFFFFFF; |
139 | else | 189 | else |
140 | *pm = acpi_pm_read_early(); | 190 | *pm = acpi_pm_read_early(); |
141 | t2 = get_cycles_sync(); | 191 | t2 = get_cycles(); |
142 | if ((t2 - t1) < SMI_TRESHOLD) | 192 | if ((t2 - t1) < SMI_TRESHOLD) |
143 | return t2; | 193 | return t2; |
144 | } | 194 | } |
@@ -151,7 +201,7 @@ static unsigned long __init tsc_read_refs(unsigned long *pm, | |||
151 | void __init tsc_calibrate(void) | 201 | void __init tsc_calibrate(void) |
152 | { | 202 | { |
153 | unsigned long flags, tsc1, tsc2, tr1, tr2, pm1, pm2, hpet1, hpet2; | 203 | unsigned long flags, tsc1, tsc2, tr1, tr2, pm1, pm2, hpet1, hpet2; |
154 | int hpet = is_hpet_enabled(); | 204 | int hpet = is_hpet_enabled(), cpu; |
155 | 205 | ||
156 | local_irq_save(flags); | 206 | local_irq_save(flags); |
157 | 207 | ||
@@ -162,9 +212,9 @@ void __init tsc_calibrate(void) | |||
162 | outb(0xb0, 0x43); | 212 | outb(0xb0, 0x43); |
163 | outb((CLOCK_TICK_RATE / (1000 / 50)) & 0xff, 0x42); | 213 | outb((CLOCK_TICK_RATE / (1000 / 50)) & 0xff, 0x42); |
164 | outb((CLOCK_TICK_RATE / (1000 / 50)) >> 8, 0x42); | 214 | outb((CLOCK_TICK_RATE / (1000 / 50)) >> 8, 0x42); |
165 | tr1 = get_cycles_sync(); | 215 | tr1 = get_cycles(); |
166 | while ((inb(0x61) & 0x20) == 0); | 216 | while ((inb(0x61) & 0x20) == 0); |
167 | tr2 = get_cycles_sync(); | 217 | tr2 = get_cycles(); |
168 | 218 | ||
169 | tsc2 = tsc_read_refs(&pm2, hpet ? &hpet2 : NULL); | 219 | tsc2 = tsc_read_refs(&pm2, hpet ? &hpet2 : NULL); |
170 | 220 | ||
@@ -206,7 +256,9 @@ void __init tsc_calibrate(void) | |||
206 | } | 256 | } |
207 | 257 | ||
208 | tsc_khz = tsc2 / tsc1; | 258 | tsc_khz = tsc2 / tsc1; |
209 | set_cyc2ns_scale(tsc_khz); | 259 | |
260 | for_each_possible_cpu(cpu) | ||
261 | set_cyc2ns_scale(tsc_khz, cpu); | ||
210 | } | 262 | } |
211 | 263 | ||
212 | /* | 264 | /* |
@@ -222,17 +274,9 @@ __cpuinit int unsynchronized_tsc(void) | |||
222 | if (apic_is_clustered_box()) | 274 | if (apic_is_clustered_box()) |
223 | return 1; | 275 | return 1; |
224 | #endif | 276 | #endif |
225 | /* Most intel systems have synchronized TSCs except for | 277 | |
226 | multi node systems */ | 278 | if (boot_cpu_has(X86_FEATURE_CONSTANT_TSC)) |
227 | if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) { | ||
228 | #ifdef CONFIG_ACPI | ||
229 | /* But TSC doesn't tick in C3 so don't use it there */ | ||
230 | if (acpi_gbl_FADT.header.length > 0 && | ||
231 | acpi_gbl_FADT.C3latency < 1000) | ||
232 | return 1; | ||
233 | #endif | ||
234 | return 0; | 279 | return 0; |
235 | } | ||
236 | 280 | ||
237 | /* Assume multi socket systems are not synchronized */ | 281 | /* Assume multi socket systems are not synchronized */ |
238 | return num_present_cpus() > 1; | 282 | return num_present_cpus() > 1; |
@@ -250,13 +294,13 @@ __setup("notsc", notsc_setup); | |||
250 | /* clock source code: */ | 294 | /* clock source code: */ |
251 | static cycle_t read_tsc(void) | 295 | static cycle_t read_tsc(void) |
252 | { | 296 | { |
253 | cycle_t ret = (cycle_t)get_cycles_sync(); | 297 | cycle_t ret = (cycle_t)get_cycles(); |
254 | return ret; | 298 | return ret; |
255 | } | 299 | } |
256 | 300 | ||
257 | static cycle_t __vsyscall_fn vread_tsc(void) | 301 | static cycle_t __vsyscall_fn vread_tsc(void) |
258 | { | 302 | { |
259 | cycle_t ret = (cycle_t)get_cycles_sync(); | 303 | cycle_t ret = (cycle_t)vget_cycles(); |
260 | return ret; | 304 | return ret; |
261 | } | 305 | } |
262 | 306 | ||