diff options
author | Robert Richter <robert.richter@amd.com> | 2009-07-16 07:04:43 -0400 |
---|---|---|
committer | Robert Richter <robert.richter@amd.com> | 2009-07-20 10:43:19 -0400 |
commit | d8471ad3ab613a1ba7abd3aad46659de39a2871c (patch) | |
tree | 34aff4195cd8a00342439ed63a8d0ab041132dd6 | |
parent | a5659d17adb815fb35e11745e2f39c3f0bfd579f (diff) |
oprofile: Introduce op_x86_phys_to_virt()
This new function translates physical to virtual counter numbers.
Signed-off-by: Robert Richter <robert.richter@amd.com>
-rw-r--r-- | arch/x86/oprofile/nmi_int.c | 43 | ||||
-rw-r--r-- | arch/x86/oprofile/op_model_amd.c | 80 | ||||
-rw-r--r-- | arch/x86/oprofile/op_x86_model.h | 1 |
3 files changed, 55 insertions, 69 deletions
diff --git a/arch/x86/oprofile/nmi_int.c b/arch/x86/oprofile/nmi_int.c index b211d335e075..02b57b8d0e61 100644 --- a/arch/x86/oprofile/nmi_int.c +++ b/arch/x86/oprofile/nmi_int.c | |||
@@ -27,12 +27,6 @@ | |||
27 | #include "op_counter.h" | 27 | #include "op_counter.h" |
28 | #include "op_x86_model.h" | 28 | #include "op_x86_model.h" |
29 | 29 | ||
30 | |||
31 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | ||
32 | DEFINE_PER_CPU(int, switch_index); | ||
33 | #endif | ||
34 | |||
35 | |||
36 | static struct op_x86_model_spec const *model; | 30 | static struct op_x86_model_spec const *model; |
37 | static DEFINE_PER_CPU(struct op_msrs, cpu_msrs); | 31 | static DEFINE_PER_CPU(struct op_msrs, cpu_msrs); |
38 | static DEFINE_PER_CPU(unsigned long, saved_lvtpc); | 32 | static DEFINE_PER_CPU(unsigned long, saved_lvtpc); |
@@ -103,6 +97,21 @@ static void nmi_cpu_save_registers(struct op_msrs *msrs) | |||
103 | } | 97 | } |
104 | } | 98 | } |
105 | 99 | ||
100 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | ||
101 | |||
102 | static DEFINE_PER_CPU(int, switch_index); | ||
103 | |||
104 | inline int op_x86_phys_to_virt(int phys) | ||
105 | { | ||
106 | return __get_cpu_var(switch_index) + phys; | ||
107 | } | ||
108 | |||
109 | #else | ||
110 | |||
111 | inline int op_x86_phys_to_virt(int phys) { return phys; } | ||
112 | |||
113 | #endif | ||
114 | |||
106 | static void free_msrs(void) | 115 | static void free_msrs(void) |
107 | { | 116 | { |
108 | int i; | 117 | int i; |
@@ -248,31 +257,25 @@ static int nmi_setup(void) | |||
248 | 257 | ||
249 | static void nmi_cpu_save_mpx_registers(struct op_msrs *msrs) | 258 | static void nmi_cpu_save_mpx_registers(struct op_msrs *msrs) |
250 | { | 259 | { |
251 | unsigned int si = __get_cpu_var(switch_index); | ||
252 | struct op_msr *multiplex = msrs->multiplex; | 260 | struct op_msr *multiplex = msrs->multiplex; |
253 | unsigned int i; | 261 | int i; |
254 | 262 | ||
255 | for (i = 0; i < model->num_counters; ++i) { | 263 | for (i = 0; i < model->num_counters; ++i) { |
256 | int offset = i + si; | 264 | int virt = op_x86_phys_to_virt(i); |
257 | if (multiplex[offset].addr) { | 265 | if (multiplex[virt].addr) |
258 | rdmsrl(multiplex[offset].addr, | 266 | rdmsrl(multiplex[virt].addr, multiplex[virt].saved); |
259 | multiplex[offset].saved); | ||
260 | } | ||
261 | } | 267 | } |
262 | } | 268 | } |
263 | 269 | ||
264 | static void nmi_cpu_restore_mpx_registers(struct op_msrs *msrs) | 270 | static void nmi_cpu_restore_mpx_registers(struct op_msrs *msrs) |
265 | { | 271 | { |
266 | unsigned int si = __get_cpu_var(switch_index); | ||
267 | struct op_msr *multiplex = msrs->multiplex; | 272 | struct op_msr *multiplex = msrs->multiplex; |
268 | unsigned int i; | 273 | int i; |
269 | 274 | ||
270 | for (i = 0; i < model->num_counters; ++i) { | 275 | for (i = 0; i < model->num_counters; ++i) { |
271 | int offset = i + si; | 276 | int virt = op_x86_phys_to_virt(i); |
272 | if (multiplex[offset].addr) { | 277 | if (multiplex[virt].addr) |
273 | wrmsrl(multiplex[offset].addr, | 278 | wrmsrl(multiplex[virt].addr, multiplex[virt].saved); |
274 | multiplex[offset].saved); | ||
275 | } | ||
276 | } | 279 | } |
277 | } | 280 | } |
278 | 281 | ||
diff --git a/arch/x86/oprofile/op_model_amd.c b/arch/x86/oprofile/op_model_amd.c index dcfd4505cacc..67f830d12e0e 100644 --- a/arch/x86/oprofile/op_model_amd.c +++ b/arch/x86/oprofile/op_model_amd.c | |||
@@ -42,9 +42,6 @@ | |||
42 | #define MSR_AMD_EVENTSEL_RESERVED ((0xFFFFFCF0ULL<<32)|(1ULL<<21)) | 42 | #define MSR_AMD_EVENTSEL_RESERVED ((0xFFFFFCF0ULL<<32)|(1ULL<<21)) |
43 | 43 | ||
44 | static unsigned long reset_value[NUM_VIRT_COUNTERS]; | 44 | static unsigned long reset_value[NUM_VIRT_COUNTERS]; |
45 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | ||
46 | DECLARE_PER_CPU(int, switch_index); | ||
47 | #endif | ||
48 | 45 | ||
49 | #ifdef CONFIG_OPROFILE_IBS | 46 | #ifdef CONFIG_OPROFILE_IBS |
50 | 47 | ||
@@ -141,21 +138,20 @@ static void op_amd_setup_ctrs(struct op_x86_model_spec const *model, | |||
141 | 138 | ||
142 | /* enable active counters */ | 139 | /* enable active counters */ |
143 | for (i = 0; i < NUM_COUNTERS; ++i) { | 140 | for (i = 0; i < NUM_COUNTERS; ++i) { |
144 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | 141 | int virt = op_x86_phys_to_virt(i); |
145 | int offset = i + __get_cpu_var(switch_index); | 142 | if (!counter_config[virt].enabled) |
146 | #else | 143 | continue; |
147 | int offset = i; | 144 | if (!msrs->counters[i].addr) |
148 | #endif | 145 | continue; |
149 | if (counter_config[offset].enabled && msrs->counters[i].addr) { | 146 | |
150 | /* setup counter registers */ | 147 | /* setup counter registers */ |
151 | wrmsrl(msrs->counters[i].addr, -(u64)reset_value[offset]); | 148 | wrmsrl(msrs->counters[i].addr, -(u64)reset_value[virt]); |
152 | 149 | ||
153 | /* setup control registers */ | 150 | /* setup control registers */ |
154 | rdmsrl(msrs->controls[i].addr, val); | 151 | rdmsrl(msrs->controls[i].addr, val); |
155 | val &= model->reserved; | 152 | val &= model->reserved; |
156 | val |= op_x86_get_ctrl(model, &counter_config[offset]); | 153 | val |= op_x86_get_ctrl(model, &counter_config[virt]); |
157 | wrmsrl(msrs->controls[i].addr, val); | 154 | wrmsrl(msrs->controls[i].addr, val); |
158 | } | ||
159 | } | 155 | } |
160 | } | 156 | } |
161 | 157 | ||
@@ -170,14 +166,13 @@ static void op_amd_switch_ctrl(struct op_x86_model_spec const *model, | |||
170 | 166 | ||
171 | /* enable active counters */ | 167 | /* enable active counters */ |
172 | for (i = 0; i < NUM_COUNTERS; ++i) { | 168 | for (i = 0; i < NUM_COUNTERS; ++i) { |
173 | int offset = i + __get_cpu_var(switch_index); | 169 | int virt = op_x86_phys_to_virt(i); |
174 | if (counter_config[offset].enabled) { | 170 | if (!counter_config[virt].enabled) |
175 | /* setup control registers */ | 171 | continue; |
176 | rdmsrl(msrs->controls[i].addr, val); | 172 | rdmsrl(msrs->controls[i].addr, val); |
177 | val &= model->reserved; | 173 | val &= model->reserved; |
178 | val |= op_x86_get_ctrl(model, &counter_config[offset]); | 174 | val |= op_x86_get_ctrl(model, &counter_config[virt]); |
179 | wrmsrl(msrs->controls[i].addr, val); | 175 | wrmsrl(msrs->controls[i].addr, val); |
180 | } | ||
181 | } | 176 | } |
182 | } | 177 | } |
183 | 178 | ||
@@ -292,19 +287,15 @@ static int op_amd_check_ctrs(struct pt_regs * const regs, | |||
292 | int i; | 287 | int i; |
293 | 288 | ||
294 | for (i = 0; i < NUM_COUNTERS; ++i) { | 289 | for (i = 0; i < NUM_COUNTERS; ++i) { |
295 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | 290 | int virt = op_x86_phys_to_virt(i); |
296 | int offset = i + __get_cpu_var(switch_index); | 291 | if (!reset_value[virt]) |
297 | #else | ||
298 | int offset = i; | ||
299 | #endif | ||
300 | if (!reset_value[offset]) | ||
301 | continue; | 292 | continue; |
302 | rdmsrl(msrs->counters[i].addr, val); | 293 | rdmsrl(msrs->counters[i].addr, val); |
303 | /* bit is clear if overflowed: */ | 294 | /* bit is clear if overflowed: */ |
304 | if (val & OP_CTR_OVERFLOW) | 295 | if (val & OP_CTR_OVERFLOW) |
305 | continue; | 296 | continue; |
306 | oprofile_add_sample(regs, offset); | 297 | oprofile_add_sample(regs, virt); |
307 | wrmsrl(msrs->counters[i].addr, -(u64)reset_value[offset]); | 298 | wrmsrl(msrs->counters[i].addr, -(u64)reset_value[virt]); |
308 | } | 299 | } |
309 | 300 | ||
310 | op_amd_handle_ibs(regs, msrs); | 301 | op_amd_handle_ibs(regs, msrs); |
@@ -319,16 +310,11 @@ static void op_amd_start(struct op_msrs const * const msrs) | |||
319 | int i; | 310 | int i; |
320 | 311 | ||
321 | for (i = 0; i < NUM_COUNTERS; ++i) { | 312 | for (i = 0; i < NUM_COUNTERS; ++i) { |
322 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | 313 | if (!reset_value[op_x86_phys_to_virt(i)]) |
323 | int offset = i + __get_cpu_var(switch_index); | 314 | continue; |
324 | #else | 315 | rdmsrl(msrs->controls[i].addr, val); |
325 | int offset = i; | 316 | val |= ARCH_PERFMON_EVENTSEL0_ENABLE; |
326 | #endif | 317 | wrmsrl(msrs->controls[i].addr, val); |
327 | if (reset_value[offset]) { | ||
328 | rdmsrl(msrs->controls[i].addr, val); | ||
329 | val |= ARCH_PERFMON_EVENTSEL0_ENABLE; | ||
330 | wrmsrl(msrs->controls[i].addr, val); | ||
331 | } | ||
332 | } | 318 | } |
333 | 319 | ||
334 | op_amd_start_ibs(); | 320 | op_amd_start_ibs(); |
@@ -344,11 +330,7 @@ static void op_amd_stop(struct op_msrs const * const msrs) | |||
344 | * pm callback | 330 | * pm callback |
345 | */ | 331 | */ |
346 | for (i = 0; i < NUM_COUNTERS; ++i) { | 332 | for (i = 0; i < NUM_COUNTERS; ++i) { |
347 | #ifdef CONFIG_OPROFILE_EVENT_MULTIPLEX | 333 | if (!reset_value[op_x86_phys_to_virt(i)]) |
348 | if (!reset_value[i + per_cpu(switch_index, smp_processor_id())]) | ||
349 | #else | ||
350 | if (!reset_value[i]) | ||
351 | #endif | ||
352 | continue; | 334 | continue; |
353 | rdmsrl(msrs->controls[i].addr, val); | 335 | rdmsrl(msrs->controls[i].addr, val); |
354 | val &= ~ARCH_PERFMON_EVENTSEL0_ENABLE; | 336 | val &= ~ARCH_PERFMON_EVENTSEL0_ENABLE; |
diff --git a/arch/x86/oprofile/op_x86_model.h b/arch/x86/oprofile/op_x86_model.h index 0d07d23cb062..e874dc3565ae 100644 --- a/arch/x86/oprofile/op_x86_model.h +++ b/arch/x86/oprofile/op_x86_model.h | |||
@@ -60,6 +60,7 @@ struct op_counter_config; | |||
60 | 60 | ||
61 | extern u64 op_x86_get_ctrl(struct op_x86_model_spec const *model, | 61 | extern u64 op_x86_get_ctrl(struct op_x86_model_spec const *model, |
62 | struct op_counter_config *counter_config); | 62 | struct op_counter_config *counter_config); |
63 | extern int op_x86_phys_to_virt(int phys); | ||
63 | 64 | ||
64 | extern struct op_x86_model_spec const op_ppro_spec; | 65 | extern struct op_x86_model_spec const op_ppro_spec; |
65 | extern struct op_x86_model_spec const op_p4_spec; | 66 | extern struct op_x86_model_spec const op_p4_spec; |