diff options
Diffstat (limited to 'include/asm-i386/msr.h')
-rw-r--r-- | include/asm-i386/msr.h | 400 |
1 files changed, 117 insertions, 283 deletions
diff --git a/include/asm-i386/msr.h b/include/asm-i386/msr.h index 2ad3f30b1a68..9559894c7658 100644 --- a/include/asm-i386/msr.h +++ b/include/asm-i386/msr.h | |||
@@ -1,6 +1,79 @@ | |||
1 | #ifndef __ASM_MSR_H | 1 | #ifndef __ASM_MSR_H |
2 | #define __ASM_MSR_H | 2 | #define __ASM_MSR_H |
3 | 3 | ||
4 | #include <asm/msr-index.h> | ||
5 | |||
6 | #ifdef __KERNEL__ | ||
7 | #ifndef __ASSEMBLY__ | ||
8 | |||
9 | #include <asm/errno.h> | ||
10 | |||
11 | static inline unsigned long long native_read_msr(unsigned int msr) | ||
12 | { | ||
13 | unsigned long long val; | ||
14 | |||
15 | asm volatile("rdmsr" : "=A" (val) : "c" (msr)); | ||
16 | return val; | ||
17 | } | ||
18 | |||
19 | static inline unsigned long long native_read_msr_safe(unsigned int msr, | ||
20 | int *err) | ||
21 | { | ||
22 | unsigned long long val; | ||
23 | |||
24 | asm volatile("2: rdmsr ; xorl %0,%0\n" | ||
25 | "1:\n\t" | ||
26 | ".section .fixup,\"ax\"\n\t" | ||
27 | "3: movl %3,%0 ; jmp 1b\n\t" | ||
28 | ".previous\n\t" | ||
29 | ".section __ex_table,\"a\"\n" | ||
30 | " .align 4\n\t" | ||
31 | " .long 2b,3b\n\t" | ||
32 | ".previous" | ||
33 | : "=r" (*err), "=A" (val) | ||
34 | : "c" (msr), "i" (-EFAULT)); | ||
35 | |||
36 | return val; | ||
37 | } | ||
38 | |||
39 | static inline void native_write_msr(unsigned int msr, unsigned long long val) | ||
40 | { | ||
41 | asm volatile("wrmsr" : : "c" (msr), "A"(val)); | ||
42 | } | ||
43 | |||
44 | static inline int native_write_msr_safe(unsigned int msr, | ||
45 | unsigned long long val) | ||
46 | { | ||
47 | int err; | ||
48 | asm volatile("2: wrmsr ; xorl %0,%0\n" | ||
49 | "1:\n\t" | ||
50 | ".section .fixup,\"ax\"\n\t" | ||
51 | "3: movl %4,%0 ; jmp 1b\n\t" | ||
52 | ".previous\n\t" | ||
53 | ".section __ex_table,\"a\"\n" | ||
54 | " .align 4\n\t" | ||
55 | " .long 2b,3b\n\t" | ||
56 | ".previous" | ||
57 | : "=a" (err) | ||
58 | : "c" (msr), "0" ((u32)val), "d" ((u32)(val>>32)), | ||
59 | "i" (-EFAULT)); | ||
60 | return err; | ||
61 | } | ||
62 | |||
63 | static inline unsigned long long native_read_tsc(void) | ||
64 | { | ||
65 | unsigned long long val; | ||
66 | asm volatile("rdtsc" : "=A" (val)); | ||
67 | return val; | ||
68 | } | ||
69 | |||
70 | static inline unsigned long long native_read_pmc(void) | ||
71 | { | ||
72 | unsigned long long val; | ||
73 | asm volatile("rdpmc" : "=A" (val)); | ||
74 | return val; | ||
75 | } | ||
76 | |||
4 | #ifdef CONFIG_PARAVIRT | 77 | #ifdef CONFIG_PARAVIRT |
5 | #include <asm/paravirt.h> | 78 | #include <asm/paravirt.h> |
6 | #else | 79 | #else |
@@ -11,22 +84,20 @@ | |||
11 | * pointer indirection), this allows gcc to optimize better | 84 | * pointer indirection), this allows gcc to optimize better |
12 | */ | 85 | */ |
13 | 86 | ||
14 | #define rdmsr(msr,val1,val2) \ | 87 | #define rdmsr(msr,val1,val2) \ |
15 | __asm__ __volatile__("rdmsr" \ | 88 | do { \ |
16 | : "=a" (val1), "=d" (val2) \ | 89 | unsigned long long __val = native_read_msr(msr); \ |
17 | : "c" (msr)) | 90 | val1 = __val; \ |
91 | val2 = __val >> 32; \ | ||
92 | } while(0) | ||
18 | 93 | ||
19 | #define wrmsr(msr,val1,val2) \ | 94 | #define wrmsr(msr,val1,val2) \ |
20 | __asm__ __volatile__("wrmsr" \ | 95 | native_write_msr(msr, ((unsigned long long)val2 << 32) | val1) |
21 | : /* no outputs */ \ | ||
22 | : "c" (msr), "a" (val1), "d" (val2)) | ||
23 | 96 | ||
24 | #define rdmsrl(msr,val) do { \ | 97 | #define rdmsrl(msr,val) \ |
25 | unsigned long l__,h__; \ | 98 | do { \ |
26 | rdmsr (msr, l__, h__); \ | 99 | (val) = native_read_msr(msr); \ |
27 | val = l__; \ | 100 | } while(0) |
28 | val |= ((u64)h__<<32); \ | ||
29 | } while(0) | ||
30 | 101 | ||
31 | static inline void wrmsrl (unsigned long msr, unsigned long long val) | 102 | static inline void wrmsrl (unsigned long msr, unsigned long long val) |
32 | { | 103 | { |
@@ -37,50 +108,41 @@ static inline void wrmsrl (unsigned long msr, unsigned long long val) | |||
37 | } | 108 | } |
38 | 109 | ||
39 | /* wrmsr with exception handling */ | 110 | /* wrmsr with exception handling */ |
40 | #define wrmsr_safe(msr,a,b) ({ int ret__; \ | 111 | #define wrmsr_safe(msr,val1,val2) \ |
41 | asm volatile("2: wrmsr ; xorl %0,%0\n" \ | 112 | (native_write_msr_safe(msr, ((unsigned long long)val2 << 32) | val1)) |
42 | "1:\n\t" \ | ||
43 | ".section .fixup,\"ax\"\n\t" \ | ||
44 | "3: movl %4,%0 ; jmp 1b\n\t" \ | ||
45 | ".previous\n\t" \ | ||
46 | ".section __ex_table,\"a\"\n" \ | ||
47 | " .align 4\n\t" \ | ||
48 | " .long 2b,3b\n\t" \ | ||
49 | ".previous" \ | ||
50 | : "=a" (ret__) \ | ||
51 | : "c" (msr), "0" (a), "d" (b), "i" (-EFAULT));\ | ||
52 | ret__; }) | ||
53 | 113 | ||
54 | /* rdmsr with exception handling */ | 114 | /* rdmsr with exception handling */ |
55 | #define rdmsr_safe(msr,a,b) ({ int ret__; \ | 115 | #define rdmsr_safe(msr,p1,p2) \ |
56 | asm volatile("2: rdmsr ; xorl %0,%0\n" \ | 116 | ({ \ |
57 | "1:\n\t" \ | 117 | int __err; \ |
58 | ".section .fixup,\"ax\"\n\t" \ | 118 | unsigned long long __val = native_read_msr_safe(msr, &__err);\ |
59 | "3: movl %4,%0 ; jmp 1b\n\t" \ | 119 | (*p1) = __val; \ |
60 | ".previous\n\t" \ | 120 | (*p2) = __val >> 32; \ |
61 | ".section __ex_table,\"a\"\n" \ | 121 | __err; \ |
62 | " .align 4\n\t" \ | 122 | }) |
63 | " .long 2b,3b\n\t" \ | 123 | |
64 | ".previous" \ | 124 | #define rdtsc(low,high) \ |
65 | : "=r" (ret__), "=a" (*(a)), "=d" (*(b)) \ | 125 | do { \ |
66 | : "c" (msr), "i" (-EFAULT));\ | 126 | u64 _l = native_read_tsc(); \ |
67 | ret__; }) | 127 | (low) = (u32)_l; \ |
68 | 128 | (high) = _l >> 32; \ | |
69 | #define rdtsc(low,high) \ | 129 | } while(0) |
70 | __asm__ __volatile__("rdtsc" : "=a" (low), "=d" (high)) | 130 | |
71 | 131 | #define rdtscl(low) \ | |
72 | #define rdtscl(low) \ | 132 | do { \ |
73 | __asm__ __volatile__("rdtsc" : "=a" (low) : : "edx") | 133 | (low) = native_read_tsc(); \ |
74 | 134 | } while(0) | |
75 | #define rdtscll(val) \ | 135 | |
76 | __asm__ __volatile__("rdtsc" : "=A" (val)) | 136 | #define rdtscll(val) ((val) = native_read_tsc()) |
77 | 137 | ||
78 | #define write_tsc(val1,val2) wrmsr(0x10, val1, val2) | 138 | #define write_tsc(val1,val2) wrmsr(0x10, val1, val2) |
79 | 139 | ||
80 | #define rdpmc(counter,low,high) \ | 140 | #define rdpmc(counter,low,high) \ |
81 | __asm__ __volatile__("rdpmc" \ | 141 | do { \ |
82 | : "=a" (low), "=d" (high) \ | 142 | u64 _l = native_read_pmc(); \ |
83 | : "c" (counter)) | 143 | low = (u32)_l; \ |
144 | high = _l >> 32; \ | ||
145 | } while(0) | ||
84 | #endif /* !CONFIG_PARAVIRT */ | 146 | #endif /* !CONFIG_PARAVIRT */ |
85 | 147 | ||
86 | #ifdef CONFIG_SMP | 148 | #ifdef CONFIG_SMP |
@@ -96,234 +158,6 @@ static inline void wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) | |||
96 | wrmsr(msr_no, l, h); | 158 | wrmsr(msr_no, l, h); |
97 | } | 159 | } |
98 | #endif /* CONFIG_SMP */ | 160 | #endif /* CONFIG_SMP */ |
99 | 161 | #endif | |
100 | /* symbolic names for some interesting MSRs */ | 162 | #endif |
101 | /* Intel defined MSRs. */ | ||
102 | #define MSR_IA32_P5_MC_ADDR 0 | ||
103 | #define MSR_IA32_P5_MC_TYPE 1 | ||
104 | #define MSR_IA32_PLATFORM_ID 0x17 | ||
105 | #define MSR_IA32_EBL_CR_POWERON 0x2a | ||
106 | |||
107 | #define MSR_IA32_APICBASE 0x1b | ||
108 | #define MSR_IA32_APICBASE_BSP (1<<8) | ||
109 | #define MSR_IA32_APICBASE_ENABLE (1<<11) | ||
110 | #define MSR_IA32_APICBASE_BASE (0xfffff<<12) | ||
111 | |||
112 | #define MSR_IA32_UCODE_WRITE 0x79 | ||
113 | #define MSR_IA32_UCODE_REV 0x8b | ||
114 | |||
115 | #define MSR_P6_PERFCTR0 0xc1 | ||
116 | #define MSR_P6_PERFCTR1 0xc2 | ||
117 | #define MSR_FSB_FREQ 0xcd | ||
118 | |||
119 | |||
120 | #define MSR_IA32_BBL_CR_CTL 0x119 | ||
121 | |||
122 | #define MSR_IA32_SYSENTER_CS 0x174 | ||
123 | #define MSR_IA32_SYSENTER_ESP 0x175 | ||
124 | #define MSR_IA32_SYSENTER_EIP 0x176 | ||
125 | |||
126 | #define MSR_IA32_MCG_CAP 0x179 | ||
127 | #define MSR_IA32_MCG_STATUS 0x17a | ||
128 | #define MSR_IA32_MCG_CTL 0x17b | ||
129 | |||
130 | /* P4/Xeon+ specific */ | ||
131 | #define MSR_IA32_MCG_EAX 0x180 | ||
132 | #define MSR_IA32_MCG_EBX 0x181 | ||
133 | #define MSR_IA32_MCG_ECX 0x182 | ||
134 | #define MSR_IA32_MCG_EDX 0x183 | ||
135 | #define MSR_IA32_MCG_ESI 0x184 | ||
136 | #define MSR_IA32_MCG_EDI 0x185 | ||
137 | #define MSR_IA32_MCG_EBP 0x186 | ||
138 | #define MSR_IA32_MCG_ESP 0x187 | ||
139 | #define MSR_IA32_MCG_EFLAGS 0x188 | ||
140 | #define MSR_IA32_MCG_EIP 0x189 | ||
141 | #define MSR_IA32_MCG_RESERVED 0x18A | ||
142 | |||
143 | #define MSR_P6_EVNTSEL0 0x186 | ||
144 | #define MSR_P6_EVNTSEL1 0x187 | ||
145 | |||
146 | #define MSR_IA32_PERF_STATUS 0x198 | ||
147 | #define MSR_IA32_PERF_CTL 0x199 | ||
148 | |||
149 | #define MSR_IA32_MPERF 0xE7 | ||
150 | #define MSR_IA32_APERF 0xE8 | ||
151 | |||
152 | #define MSR_IA32_THERM_CONTROL 0x19a | ||
153 | #define MSR_IA32_THERM_INTERRUPT 0x19b | ||
154 | #define MSR_IA32_THERM_STATUS 0x19c | ||
155 | #define MSR_IA32_MISC_ENABLE 0x1a0 | ||
156 | |||
157 | #define MSR_IA32_DEBUGCTLMSR 0x1d9 | ||
158 | #define MSR_IA32_LASTBRANCHFROMIP 0x1db | ||
159 | #define MSR_IA32_LASTBRANCHTOIP 0x1dc | ||
160 | #define MSR_IA32_LASTINTFROMIP 0x1dd | ||
161 | #define MSR_IA32_LASTINTTOIP 0x1de | ||
162 | |||
163 | #define MSR_IA32_MC0_CTL 0x400 | ||
164 | #define MSR_IA32_MC0_STATUS 0x401 | ||
165 | #define MSR_IA32_MC0_ADDR 0x402 | ||
166 | #define MSR_IA32_MC0_MISC 0x403 | ||
167 | |||
168 | #define MSR_IA32_PEBS_ENABLE 0x3f1 | ||
169 | #define MSR_IA32_DS_AREA 0x600 | ||
170 | #define MSR_IA32_PERF_CAPABILITIES 0x345 | ||
171 | |||
172 | /* Pentium IV performance counter MSRs */ | ||
173 | #define MSR_P4_BPU_PERFCTR0 0x300 | ||
174 | #define MSR_P4_BPU_PERFCTR1 0x301 | ||
175 | #define MSR_P4_BPU_PERFCTR2 0x302 | ||
176 | #define MSR_P4_BPU_PERFCTR3 0x303 | ||
177 | #define MSR_P4_MS_PERFCTR0 0x304 | ||
178 | #define MSR_P4_MS_PERFCTR1 0x305 | ||
179 | #define MSR_P4_MS_PERFCTR2 0x306 | ||
180 | #define MSR_P4_MS_PERFCTR3 0x307 | ||
181 | #define MSR_P4_FLAME_PERFCTR0 0x308 | ||
182 | #define MSR_P4_FLAME_PERFCTR1 0x309 | ||
183 | #define MSR_P4_FLAME_PERFCTR2 0x30a | ||
184 | #define MSR_P4_FLAME_PERFCTR3 0x30b | ||
185 | #define MSR_P4_IQ_PERFCTR0 0x30c | ||
186 | #define MSR_P4_IQ_PERFCTR1 0x30d | ||
187 | #define MSR_P4_IQ_PERFCTR2 0x30e | ||
188 | #define MSR_P4_IQ_PERFCTR3 0x30f | ||
189 | #define MSR_P4_IQ_PERFCTR4 0x310 | ||
190 | #define MSR_P4_IQ_PERFCTR5 0x311 | ||
191 | #define MSR_P4_BPU_CCCR0 0x360 | ||
192 | #define MSR_P4_BPU_CCCR1 0x361 | ||
193 | #define MSR_P4_BPU_CCCR2 0x362 | ||
194 | #define MSR_P4_BPU_CCCR3 0x363 | ||
195 | #define MSR_P4_MS_CCCR0 0x364 | ||
196 | #define MSR_P4_MS_CCCR1 0x365 | ||
197 | #define MSR_P4_MS_CCCR2 0x366 | ||
198 | #define MSR_P4_MS_CCCR3 0x367 | ||
199 | #define MSR_P4_FLAME_CCCR0 0x368 | ||
200 | #define MSR_P4_FLAME_CCCR1 0x369 | ||
201 | #define MSR_P4_FLAME_CCCR2 0x36a | ||
202 | #define MSR_P4_FLAME_CCCR3 0x36b | ||
203 | #define MSR_P4_IQ_CCCR0 0x36c | ||
204 | #define MSR_P4_IQ_CCCR1 0x36d | ||
205 | #define MSR_P4_IQ_CCCR2 0x36e | ||
206 | #define MSR_P4_IQ_CCCR3 0x36f | ||
207 | #define MSR_P4_IQ_CCCR4 0x370 | ||
208 | #define MSR_P4_IQ_CCCR5 0x371 | ||
209 | #define MSR_P4_ALF_ESCR0 0x3ca | ||
210 | #define MSR_P4_ALF_ESCR1 0x3cb | ||
211 | #define MSR_P4_BPU_ESCR0 0x3b2 | ||
212 | #define MSR_P4_BPU_ESCR1 0x3b3 | ||
213 | #define MSR_P4_BSU_ESCR0 0x3a0 | ||
214 | #define MSR_P4_BSU_ESCR1 0x3a1 | ||
215 | #define MSR_P4_CRU_ESCR0 0x3b8 | ||
216 | #define MSR_P4_CRU_ESCR1 0x3b9 | ||
217 | #define MSR_P4_CRU_ESCR2 0x3cc | ||
218 | #define MSR_P4_CRU_ESCR3 0x3cd | ||
219 | #define MSR_P4_CRU_ESCR4 0x3e0 | ||
220 | #define MSR_P4_CRU_ESCR5 0x3e1 | ||
221 | #define MSR_P4_DAC_ESCR0 0x3a8 | ||
222 | #define MSR_P4_DAC_ESCR1 0x3a9 | ||
223 | #define MSR_P4_FIRM_ESCR0 0x3a4 | ||
224 | #define MSR_P4_FIRM_ESCR1 0x3a5 | ||
225 | #define MSR_P4_FLAME_ESCR0 0x3a6 | ||
226 | #define MSR_P4_FLAME_ESCR1 0x3a7 | ||
227 | #define MSR_P4_FSB_ESCR0 0x3a2 | ||
228 | #define MSR_P4_FSB_ESCR1 0x3a3 | ||
229 | #define MSR_P4_IQ_ESCR0 0x3ba | ||
230 | #define MSR_P4_IQ_ESCR1 0x3bb | ||
231 | #define MSR_P4_IS_ESCR0 0x3b4 | ||
232 | #define MSR_P4_IS_ESCR1 0x3b5 | ||
233 | #define MSR_P4_ITLB_ESCR0 0x3b6 | ||
234 | #define MSR_P4_ITLB_ESCR1 0x3b7 | ||
235 | #define MSR_P4_IX_ESCR0 0x3c8 | ||
236 | #define MSR_P4_IX_ESCR1 0x3c9 | ||
237 | #define MSR_P4_MOB_ESCR0 0x3aa | ||
238 | #define MSR_P4_MOB_ESCR1 0x3ab | ||
239 | #define MSR_P4_MS_ESCR0 0x3c0 | ||
240 | #define MSR_P4_MS_ESCR1 0x3c1 | ||
241 | #define MSR_P4_PMH_ESCR0 0x3ac | ||
242 | #define MSR_P4_PMH_ESCR1 0x3ad | ||
243 | #define MSR_P4_RAT_ESCR0 0x3bc | ||
244 | #define MSR_P4_RAT_ESCR1 0x3bd | ||
245 | #define MSR_P4_SAAT_ESCR0 0x3ae | ||
246 | #define MSR_P4_SAAT_ESCR1 0x3af | ||
247 | #define MSR_P4_SSU_ESCR0 0x3be | ||
248 | #define MSR_P4_SSU_ESCR1 0x3bf /* guess: not defined in manual */ | ||
249 | #define MSR_P4_TBPU_ESCR0 0x3c2 | ||
250 | #define MSR_P4_TBPU_ESCR1 0x3c3 | ||
251 | #define MSR_P4_TC_ESCR0 0x3c4 | ||
252 | #define MSR_P4_TC_ESCR1 0x3c5 | ||
253 | #define MSR_P4_U2L_ESCR0 0x3b0 | ||
254 | #define MSR_P4_U2L_ESCR1 0x3b1 | ||
255 | |||
256 | /* AMD Defined MSRs */ | ||
257 | #define MSR_K6_EFER 0xC0000080 | ||
258 | #define MSR_K6_STAR 0xC0000081 | ||
259 | #define MSR_K6_WHCR 0xC0000082 | ||
260 | #define MSR_K6_UWCCR 0xC0000085 | ||
261 | #define MSR_K6_EPMR 0xC0000086 | ||
262 | #define MSR_K6_PSOR 0xC0000087 | ||
263 | #define MSR_K6_PFIR 0xC0000088 | ||
264 | |||
265 | #define MSR_K7_EVNTSEL0 0xC0010000 | ||
266 | #define MSR_K7_EVNTSEL1 0xC0010001 | ||
267 | #define MSR_K7_EVNTSEL2 0xC0010002 | ||
268 | #define MSR_K7_EVNTSEL3 0xC0010003 | ||
269 | #define MSR_K7_PERFCTR0 0xC0010004 | ||
270 | #define MSR_K7_PERFCTR1 0xC0010005 | ||
271 | #define MSR_K7_PERFCTR2 0xC0010006 | ||
272 | #define MSR_K7_PERFCTR3 0xC0010007 | ||
273 | #define MSR_K7_HWCR 0xC0010015 | ||
274 | #define MSR_K7_CLK_CTL 0xC001001b | ||
275 | #define MSR_K7_FID_VID_CTL 0xC0010041 | ||
276 | #define MSR_K7_FID_VID_STATUS 0xC0010042 | ||
277 | |||
278 | #define MSR_K8_ENABLE_C1E 0xC0010055 | ||
279 | |||
280 | /* extended feature register */ | ||
281 | #define MSR_EFER 0xc0000080 | ||
282 | |||
283 | /* EFER bits: */ | ||
284 | |||
285 | /* Execute Disable enable */ | ||
286 | #define _EFER_NX 11 | ||
287 | #define EFER_NX (1<<_EFER_NX) | ||
288 | |||
289 | /* Centaur-Hauls/IDT defined MSRs. */ | ||
290 | #define MSR_IDT_FCR1 0x107 | ||
291 | #define MSR_IDT_FCR2 0x108 | ||
292 | #define MSR_IDT_FCR3 0x109 | ||
293 | #define MSR_IDT_FCR4 0x10a | ||
294 | |||
295 | #define MSR_IDT_MCR0 0x110 | ||
296 | #define MSR_IDT_MCR1 0x111 | ||
297 | #define MSR_IDT_MCR2 0x112 | ||
298 | #define MSR_IDT_MCR3 0x113 | ||
299 | #define MSR_IDT_MCR4 0x114 | ||
300 | #define MSR_IDT_MCR5 0x115 | ||
301 | #define MSR_IDT_MCR6 0x116 | ||
302 | #define MSR_IDT_MCR7 0x117 | ||
303 | #define MSR_IDT_MCR_CTRL 0x120 | ||
304 | |||
305 | /* VIA Cyrix defined MSRs*/ | ||
306 | #define MSR_VIA_FCR 0x1107 | ||
307 | #define MSR_VIA_LONGHAUL 0x110a | ||
308 | #define MSR_VIA_RNG 0x110b | ||
309 | #define MSR_VIA_BCR2 0x1147 | ||
310 | |||
311 | /* Transmeta defined MSRs */ | ||
312 | #define MSR_TMTA_LONGRUN_CTRL 0x80868010 | ||
313 | #define MSR_TMTA_LONGRUN_FLAGS 0x80868011 | ||
314 | #define MSR_TMTA_LRTI_READOUT 0x80868018 | ||
315 | #define MSR_TMTA_LRTI_VOLT_MHZ 0x8086801a | ||
316 | |||
317 | /* Intel Core-based CPU performance counters */ | ||
318 | #define MSR_CORE_PERF_FIXED_CTR0 0x309 | ||
319 | #define MSR_CORE_PERF_FIXED_CTR1 0x30a | ||
320 | #define MSR_CORE_PERF_FIXED_CTR2 0x30b | ||
321 | #define MSR_CORE_PERF_FIXED_CTR_CTRL 0x38d | ||
322 | #define MSR_CORE_PERF_GLOBAL_STATUS 0x38e | ||
323 | #define MSR_CORE_PERF_GLOBAL_CTRL 0x38f | ||
324 | #define MSR_CORE_PERF_GLOBAL_OVF_CTRL 0x390 | ||
325 | |||
326 | /* Geode defined MSRs */ | ||
327 | #define MSR_GEODE_BUSCONT_CONF0 0x1900 | ||
328 | |||
329 | #endif /* __ASM_MSR_H */ | 163 | #endif /* __ASM_MSR_H */ |