diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2011-03-18 13:38:34 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2011-03-18 13:38:34 -0400 |
commit | 619297855aa16646246ea4b1f6e05f1b2455c808 (patch) | |
tree | a07fe1b24c372f5eabf244555db41fdf574c1205 /arch | |
parent | e16b396ce314b2bcdfe6c173fe075bf8e3432368 (diff) | |
parent | 1ef1d1c2353967e2d61ecaddf76edfd058a778b4 (diff) |
Merge branch 'perf-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip
* 'perf-fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/linux-2.6-tip: (30 commits)
trace, filters: Initialize the match variable in process_ops() properly
trace, documentation: Fix branch profiling location in debugfs
oprofile, s390: Cleanups
oprofile, s390: Remove hwsampler_files.c and merge it into init.c
perf: Fix tear-down of inherited group events
perf: Reorder & optimize perf_event_context to remove alignment padding on 64 bit builds
perf: Handle stopped state with tracepoints
perf: Fix the software events state check
perf, powerpc: Handle events that raise an exception without overflowing
perf, x86: Use INTEL_*_CONSTRAINT() for all PEBS event constraints
perf, x86: Clean up SandyBridge PEBS events
perf lock: Fix sorting by wait_min
perf tools: Version incorrect with some versions of grep
perf evlist: New command to list the names of events present in a perf.data file
perf script: Add support for H/W and S/W events
perf script: Add support for dumping symbols
perf script: Support custom field selection for output
perf script: Move printing of 'common' data from print_event and rename
perf tracing: Remove print_graph_cpu and print_graph_proc from trace-event-parse
perf script: Change process_event prototype
...
Diffstat (limited to 'arch')
-rw-r--r-- | arch/powerpc/include/asm/reg.h | 1 | ||||
-rw-r--r-- | arch/powerpc/kernel/perf_event.c | 24 | ||||
-rw-r--r-- | arch/s390/oprofile/Makefile | 2 | ||||
-rw-r--r-- | arch/s390/oprofile/hwsampler.c | 1256 | ||||
-rw-r--r-- | arch/s390/oprofile/hwsampler.h | 113 | ||||
-rw-r--r-- | arch/s390/oprofile/init.c | 165 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/perf_event.c | 2 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/perf_event_intel_ds.c | 112 |
8 files changed, 1603 insertions, 72 deletions
diff --git a/arch/powerpc/include/asm/reg.h b/arch/powerpc/include/asm/reg.h index 1bc6a12f3725..7e4abebe76c0 100644 --- a/arch/powerpc/include/asm/reg.h +++ b/arch/powerpc/include/asm/reg.h | |||
@@ -892,6 +892,7 @@ | |||
892 | #define PV_970 0x0039 | 892 | #define PV_970 0x0039 |
893 | #define PV_POWER5 0x003A | 893 | #define PV_POWER5 0x003A |
894 | #define PV_POWER5p 0x003B | 894 | #define PV_POWER5p 0x003B |
895 | #define PV_POWER7 0x003F | ||
895 | #define PV_970FX 0x003C | 896 | #define PV_970FX 0x003C |
896 | #define PV_630 0x0040 | 897 | #define PV_630 0x0040 |
897 | #define PV_630p 0x0041 | 898 | #define PV_630p 0x0041 |
diff --git a/arch/powerpc/kernel/perf_event.c b/arch/powerpc/kernel/perf_event.c index ab6f6beadb57..97e0ae414940 100644 --- a/arch/powerpc/kernel/perf_event.c +++ b/arch/powerpc/kernel/perf_event.c | |||
@@ -1269,6 +1269,28 @@ unsigned long perf_instruction_pointer(struct pt_regs *regs) | |||
1269 | return ip; | 1269 | return ip; |
1270 | } | 1270 | } |
1271 | 1271 | ||
1272 | static bool pmc_overflow(unsigned long val) | ||
1273 | { | ||
1274 | if ((int)val < 0) | ||
1275 | return true; | ||
1276 | |||
1277 | /* | ||
1278 | * Events on POWER7 can roll back if a speculative event doesn't | ||
1279 | * eventually complete. Unfortunately in some rare cases they will | ||
1280 | * raise a performance monitor exception. We need to catch this to | ||
1281 | * ensure we reset the PMC. In all cases the PMC will be 256 or less | ||
1282 | * cycles from overflow. | ||
1283 | * | ||
1284 | * We only do this if the first pass fails to find any overflowing | ||
1285 | * PMCs because a user might set a period of less than 256 and we | ||
1286 | * don't want to mistakenly reset them. | ||
1287 | */ | ||
1288 | if (__is_processor(PV_POWER7) && ((0x80000000 - val) <= 256)) | ||
1289 | return true; | ||
1290 | |||
1291 | return false; | ||
1292 | } | ||
1293 | |||
1272 | /* | 1294 | /* |
1273 | * Performance monitor interrupt stuff | 1295 | * Performance monitor interrupt stuff |
1274 | */ | 1296 | */ |
@@ -1316,7 +1338,7 @@ static void perf_event_interrupt(struct pt_regs *regs) | |||
1316 | if (is_limited_pmc(i + 1)) | 1338 | if (is_limited_pmc(i + 1)) |
1317 | continue; | 1339 | continue; |
1318 | val = read_pmc(i + 1); | 1340 | val = read_pmc(i + 1); |
1319 | if ((int)val < 0) | 1341 | if (pmc_overflow(val)) |
1320 | write_pmc(i + 1, 0); | 1342 | write_pmc(i + 1, 0); |
1321 | } | 1343 | } |
1322 | } | 1344 | } |
diff --git a/arch/s390/oprofile/Makefile b/arch/s390/oprofile/Makefile index 537b2d840e69..d698cddcfbdd 100644 --- a/arch/s390/oprofile/Makefile +++ b/arch/s390/oprofile/Makefile | |||
@@ -6,4 +6,4 @@ DRIVER_OBJS = $(addprefix ../../../drivers/oprofile/, \ | |||
6 | oprofilefs.o oprofile_stats.o \ | 6 | oprofilefs.o oprofile_stats.o \ |
7 | timer_int.o ) | 7 | timer_int.o ) |
8 | 8 | ||
9 | oprofile-y := $(DRIVER_OBJS) init.o backtrace.o | 9 | oprofile-y := $(DRIVER_OBJS) init.o backtrace.o hwsampler.o |
diff --git a/arch/s390/oprofile/hwsampler.c b/arch/s390/oprofile/hwsampler.c new file mode 100644 index 000000000000..3d48f4db246d --- /dev/null +++ b/arch/s390/oprofile/hwsampler.c | |||
@@ -0,0 +1,1256 @@ | |||
1 | /** | ||
2 | * arch/s390/oprofile/hwsampler.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2010 | ||
5 | * Author: Heinz Graalfs <graalfs@de.ibm.com> | ||
6 | */ | ||
7 | |||
8 | #include <linux/kernel.h> | ||
9 | #include <linux/module.h> | ||
10 | #include <linux/smp.h> | ||
11 | #include <linux/errno.h> | ||
12 | #include <linux/workqueue.h> | ||
13 | #include <linux/interrupt.h> | ||
14 | #include <linux/notifier.h> | ||
15 | #include <linux/cpu.h> | ||
16 | #include <linux/semaphore.h> | ||
17 | #include <linux/oom.h> | ||
18 | #include <linux/oprofile.h> | ||
19 | |||
20 | #include <asm/lowcore.h> | ||
21 | #include <asm/s390_ext.h> | ||
22 | |||
23 | #include "hwsampler.h" | ||
24 | |||
25 | #define MAX_NUM_SDB 511 | ||
26 | #define MIN_NUM_SDB 1 | ||
27 | |||
28 | #define ALERT_REQ_MASK 0x4000000000000000ul | ||
29 | #define BUFFER_FULL_MASK 0x8000000000000000ul | ||
30 | |||
31 | #define EI_IEA (1 << 31) /* invalid entry address */ | ||
32 | #define EI_ISE (1 << 30) /* incorrect SDBT entry */ | ||
33 | #define EI_PRA (1 << 29) /* program request alert */ | ||
34 | #define EI_SACA (1 << 23) /* sampler authorization change alert */ | ||
35 | #define EI_LSDA (1 << 22) /* loss of sample data alert */ | ||
36 | |||
37 | DECLARE_PER_CPU(struct hws_cpu_buffer, sampler_cpu_buffer); | ||
38 | |||
39 | struct hws_execute_parms { | ||
40 | void *buffer; | ||
41 | signed int rc; | ||
42 | }; | ||
43 | |||
44 | DEFINE_PER_CPU(struct hws_cpu_buffer, sampler_cpu_buffer); | ||
45 | EXPORT_PER_CPU_SYMBOL(sampler_cpu_buffer); | ||
46 | |||
47 | static DEFINE_MUTEX(hws_sem); | ||
48 | static DEFINE_MUTEX(hws_sem_oom); | ||
49 | |||
50 | static unsigned char hws_flush_all; | ||
51 | static unsigned int hws_oom; | ||
52 | static struct workqueue_struct *hws_wq; | ||
53 | |||
54 | static unsigned int hws_state; | ||
55 | enum { | ||
56 | HWS_INIT = 1, | ||
57 | HWS_DEALLOCATED, | ||
58 | HWS_STOPPED, | ||
59 | HWS_STARTED, | ||
60 | HWS_STOPPING }; | ||
61 | |||
62 | /* set to 1 if called by kernel during memory allocation */ | ||
63 | static unsigned char oom_killer_was_active; | ||
64 | /* size of SDBT and SDB as of allocate API */ | ||
65 | static unsigned long num_sdbt = 100; | ||
66 | static unsigned long num_sdb = 511; | ||
67 | /* sampling interval (machine cycles) */ | ||
68 | static unsigned long interval; | ||
69 | |||
70 | static unsigned long min_sampler_rate; | ||
71 | static unsigned long max_sampler_rate; | ||
72 | |||
73 | static int ssctl(void *buffer) | ||
74 | { | ||
75 | int cc; | ||
76 | |||
77 | /* set in order to detect a program check */ | ||
78 | cc = 1; | ||
79 | |||
80 | asm volatile( | ||
81 | "0: .insn s,0xB2870000,0(%1)\n" | ||
82 | "1: ipm %0\n" | ||
83 | " srl %0,28\n" | ||
84 | "2:\n" | ||
85 | EX_TABLE(0b, 2b) EX_TABLE(1b, 2b) | ||
86 | : "+d" (cc), "+a" (buffer) | ||
87 | : "m" (*((struct hws_ssctl_request_block *)buffer)) | ||
88 | : "cc", "memory"); | ||
89 | |||
90 | return cc ? -EINVAL : 0 ; | ||
91 | } | ||
92 | |||
93 | static int qsi(void *buffer) | ||
94 | { | ||
95 | int cc; | ||
96 | cc = 1; | ||
97 | |||
98 | asm volatile( | ||
99 | "0: .insn s,0xB2860000,0(%1)\n" | ||
100 | "1: lhi %0,0\n" | ||
101 | "2:\n" | ||
102 | EX_TABLE(0b, 2b) EX_TABLE(1b, 2b) | ||
103 | : "=d" (cc), "+a" (buffer) | ||
104 | : "m" (*((struct hws_qsi_info_block *)buffer)) | ||
105 | : "cc", "memory"); | ||
106 | |||
107 | return cc ? -EINVAL : 0; | ||
108 | } | ||
109 | |||
110 | static void execute_qsi(void *parms) | ||
111 | { | ||
112 | struct hws_execute_parms *ep = parms; | ||
113 | |||
114 | ep->rc = qsi(ep->buffer); | ||
115 | } | ||
116 | |||
117 | static void execute_ssctl(void *parms) | ||
118 | { | ||
119 | struct hws_execute_parms *ep = parms; | ||
120 | |||
121 | ep->rc = ssctl(ep->buffer); | ||
122 | } | ||
123 | |||
124 | static int smp_ctl_ssctl_stop(int cpu) | ||
125 | { | ||
126 | int rc; | ||
127 | struct hws_execute_parms ep; | ||
128 | struct hws_cpu_buffer *cb; | ||
129 | |||
130 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
131 | |||
132 | cb->ssctl.es = 0; | ||
133 | cb->ssctl.cs = 0; | ||
134 | |||
135 | ep.buffer = &cb->ssctl; | ||
136 | smp_call_function_single(cpu, execute_ssctl, &ep, 1); | ||
137 | rc = ep.rc; | ||
138 | if (rc) { | ||
139 | printk(KERN_ERR "hwsampler: CPU %d CPUMF SSCTL failed.\n", cpu); | ||
140 | dump_stack(); | ||
141 | } | ||
142 | |||
143 | ep.buffer = &cb->qsi; | ||
144 | smp_call_function_single(cpu, execute_qsi, &ep, 1); | ||
145 | |||
146 | if (cb->qsi.es || cb->qsi.cs) { | ||
147 | printk(KERN_EMERG "CPUMF sampling did not stop properly.\n"); | ||
148 | dump_stack(); | ||
149 | } | ||
150 | |||
151 | return rc; | ||
152 | } | ||
153 | |||
154 | static int smp_ctl_ssctl_deactivate(int cpu) | ||
155 | { | ||
156 | int rc; | ||
157 | struct hws_execute_parms ep; | ||
158 | struct hws_cpu_buffer *cb; | ||
159 | |||
160 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
161 | |||
162 | cb->ssctl.es = 1; | ||
163 | cb->ssctl.cs = 0; | ||
164 | |||
165 | ep.buffer = &cb->ssctl; | ||
166 | smp_call_function_single(cpu, execute_ssctl, &ep, 1); | ||
167 | rc = ep.rc; | ||
168 | if (rc) | ||
169 | printk(KERN_ERR "hwsampler: CPU %d CPUMF SSCTL failed.\n", cpu); | ||
170 | |||
171 | ep.buffer = &cb->qsi; | ||
172 | smp_call_function_single(cpu, execute_qsi, &ep, 1); | ||
173 | |||
174 | if (cb->qsi.cs) | ||
175 | printk(KERN_EMERG "CPUMF sampling was not set inactive.\n"); | ||
176 | |||
177 | return rc; | ||
178 | } | ||
179 | |||
180 | static int smp_ctl_ssctl_enable_activate(int cpu, unsigned long interval) | ||
181 | { | ||
182 | int rc; | ||
183 | struct hws_execute_parms ep; | ||
184 | struct hws_cpu_buffer *cb; | ||
185 | |||
186 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
187 | |||
188 | cb->ssctl.h = 1; | ||
189 | cb->ssctl.tear = cb->first_sdbt; | ||
190 | cb->ssctl.dear = *(unsigned long *) cb->first_sdbt; | ||
191 | cb->ssctl.interval = interval; | ||
192 | cb->ssctl.es = 1; | ||
193 | cb->ssctl.cs = 1; | ||
194 | |||
195 | ep.buffer = &cb->ssctl; | ||
196 | smp_call_function_single(cpu, execute_ssctl, &ep, 1); | ||
197 | rc = ep.rc; | ||
198 | if (rc) | ||
199 | printk(KERN_ERR "hwsampler: CPU %d CPUMF SSCTL failed.\n", cpu); | ||
200 | |||
201 | ep.buffer = &cb->qsi; | ||
202 | smp_call_function_single(cpu, execute_qsi, &ep, 1); | ||
203 | if (ep.rc) | ||
204 | printk(KERN_ERR "hwsampler: CPU %d CPUMF QSI failed.\n", cpu); | ||
205 | |||
206 | return rc; | ||
207 | } | ||
208 | |||
209 | static int smp_ctl_qsi(int cpu) | ||
210 | { | ||
211 | struct hws_execute_parms ep; | ||
212 | struct hws_cpu_buffer *cb; | ||
213 | |||
214 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
215 | |||
216 | ep.buffer = &cb->qsi; | ||
217 | smp_call_function_single(cpu, execute_qsi, &ep, 1); | ||
218 | |||
219 | return ep.rc; | ||
220 | } | ||
221 | |||
222 | static inline unsigned long *trailer_entry_ptr(unsigned long v) | ||
223 | { | ||
224 | void *ret; | ||
225 | |||
226 | ret = (void *)v; | ||
227 | ret += PAGE_SIZE; | ||
228 | ret -= sizeof(struct hws_trailer_entry); | ||
229 | |||
230 | return (unsigned long *) ret; | ||
231 | } | ||
232 | |||
233 | /* prototypes for external interrupt handler and worker */ | ||
234 | static void hws_ext_handler(unsigned int ext_int_code, | ||
235 | unsigned int param32, unsigned long param64); | ||
236 | |||
237 | static void worker(struct work_struct *work); | ||
238 | |||
239 | static void add_samples_to_oprofile(unsigned cpu, unsigned long *, | ||
240 | unsigned long *dear); | ||
241 | |||
242 | static void init_all_cpu_buffers(void) | ||
243 | { | ||
244 | int cpu; | ||
245 | struct hws_cpu_buffer *cb; | ||
246 | |||
247 | for_each_online_cpu(cpu) { | ||
248 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
249 | memset(cb, 0, sizeof(struct hws_cpu_buffer)); | ||
250 | } | ||
251 | } | ||
252 | |||
253 | static int is_link_entry(unsigned long *s) | ||
254 | { | ||
255 | return *s & 0x1ul ? 1 : 0; | ||
256 | } | ||
257 | |||
258 | static unsigned long *get_next_sdbt(unsigned long *s) | ||
259 | { | ||
260 | return (unsigned long *) (*s & ~0x1ul); | ||
261 | } | ||
262 | |||
263 | static int prepare_cpu_buffers(void) | ||
264 | { | ||
265 | int cpu; | ||
266 | int rc; | ||
267 | struct hws_cpu_buffer *cb; | ||
268 | |||
269 | rc = 0; | ||
270 | for_each_online_cpu(cpu) { | ||
271 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
272 | atomic_set(&cb->ext_params, 0); | ||
273 | cb->worker_entry = 0; | ||
274 | cb->sample_overflow = 0; | ||
275 | cb->req_alert = 0; | ||
276 | cb->incorrect_sdbt_entry = 0; | ||
277 | cb->invalid_entry_address = 0; | ||
278 | cb->loss_of_sample_data = 0; | ||
279 | cb->sample_auth_change_alert = 0; | ||
280 | cb->finish = 0; | ||
281 | cb->oom = 0; | ||
282 | cb->stop_mode = 0; | ||
283 | } | ||
284 | |||
285 | return rc; | ||
286 | } | ||
287 | |||
288 | /* | ||
289 | * allocate_sdbt() - allocate sampler memory | ||
290 | * @cpu: the cpu for which sampler memory is allocated | ||
291 | * | ||
292 | * A 4K page is allocated for each requested SDBT. | ||
293 | * A maximum of 511 4K pages are allocated for the SDBs in each of the SDBTs. | ||
294 | * Set ALERT_REQ mask in each SDBs trailer. | ||
295 | * Returns zero if successful, <0 otherwise. | ||
296 | */ | ||
297 | static int allocate_sdbt(int cpu) | ||
298 | { | ||
299 | int j, k, rc; | ||
300 | unsigned long *sdbt; | ||
301 | unsigned long sdb; | ||
302 | unsigned long *tail; | ||
303 | unsigned long *trailer; | ||
304 | struct hws_cpu_buffer *cb; | ||
305 | |||
306 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
307 | |||
308 | if (cb->first_sdbt) | ||
309 | return -EINVAL; | ||
310 | |||
311 | sdbt = NULL; | ||
312 | tail = sdbt; | ||
313 | |||
314 | for (j = 0; j < num_sdbt; j++) { | ||
315 | sdbt = (unsigned long *)get_zeroed_page(GFP_KERNEL); | ||
316 | |||
317 | mutex_lock(&hws_sem_oom); | ||
318 | /* OOM killer might have been activated */ | ||
319 | barrier(); | ||
320 | if (oom_killer_was_active || !sdbt) { | ||
321 | if (sdbt) | ||
322 | free_page((unsigned long)sdbt); | ||
323 | |||
324 | goto allocate_sdbt_error; | ||
325 | } | ||
326 | if (cb->first_sdbt == 0) | ||
327 | cb->first_sdbt = (unsigned long)sdbt; | ||
328 | |||
329 | /* link current page to tail of chain */ | ||
330 | if (tail) | ||
331 | *tail = (unsigned long)(void *)sdbt + 1; | ||
332 | |||
333 | mutex_unlock(&hws_sem_oom); | ||
334 | |||
335 | for (k = 0; k < num_sdb; k++) { | ||
336 | /* get and set SDB page */ | ||
337 | sdb = get_zeroed_page(GFP_KERNEL); | ||
338 | |||
339 | mutex_lock(&hws_sem_oom); | ||
340 | /* OOM killer might have been activated */ | ||
341 | barrier(); | ||
342 | if (oom_killer_was_active || !sdb) { | ||
343 | if (sdb) | ||
344 | free_page(sdb); | ||
345 | |||
346 | goto allocate_sdbt_error; | ||
347 | } | ||
348 | *sdbt = sdb; | ||
349 | trailer = trailer_entry_ptr(*sdbt); | ||
350 | *trailer = ALERT_REQ_MASK; | ||
351 | sdbt++; | ||
352 | mutex_unlock(&hws_sem_oom); | ||
353 | } | ||
354 | tail = sdbt; | ||
355 | } | ||
356 | mutex_lock(&hws_sem_oom); | ||
357 | if (oom_killer_was_active) | ||
358 | goto allocate_sdbt_error; | ||
359 | |||
360 | rc = 0; | ||
361 | if (tail) | ||
362 | *tail = (unsigned long) | ||
363 | ((void *)cb->first_sdbt) + 1; | ||
364 | |||
365 | allocate_sdbt_exit: | ||
366 | mutex_unlock(&hws_sem_oom); | ||
367 | return rc; | ||
368 | |||
369 | allocate_sdbt_error: | ||
370 | rc = -ENOMEM; | ||
371 | goto allocate_sdbt_exit; | ||
372 | } | ||
373 | |||
374 | /* | ||
375 | * deallocate_sdbt() - deallocate all sampler memory | ||
376 | * | ||
377 | * For each online CPU all SDBT trees are deallocated. | ||
378 | * Returns the number of freed pages. | ||
379 | */ | ||
380 | static int deallocate_sdbt(void) | ||
381 | { | ||
382 | int cpu; | ||
383 | int counter; | ||
384 | |||
385 | counter = 0; | ||
386 | |||
387 | for_each_online_cpu(cpu) { | ||
388 | unsigned long start; | ||
389 | unsigned long sdbt; | ||
390 | unsigned long *curr; | ||
391 | struct hws_cpu_buffer *cb; | ||
392 | |||
393 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
394 | |||
395 | if (!cb->first_sdbt) | ||
396 | continue; | ||
397 | |||
398 | sdbt = cb->first_sdbt; | ||
399 | curr = (unsigned long *) sdbt; | ||
400 | start = sdbt; | ||
401 | |||
402 | /* we'll free the SDBT after all SDBs are processed... */ | ||
403 | while (1) { | ||
404 | if (!*curr || !sdbt) | ||
405 | break; | ||
406 | |||
407 | /* watch for link entry reset if found */ | ||
408 | if (is_link_entry(curr)) { | ||
409 | curr = get_next_sdbt(curr); | ||
410 | if (sdbt) | ||
411 | free_page(sdbt); | ||
412 | |||
413 | /* we are done if we reach the start */ | ||
414 | if ((unsigned long) curr == start) | ||
415 | break; | ||
416 | else | ||
417 | sdbt = (unsigned long) curr; | ||
418 | } else { | ||
419 | /* process SDB pointer */ | ||
420 | if (*curr) { | ||
421 | free_page(*curr); | ||
422 | curr++; | ||
423 | } | ||
424 | } | ||
425 | counter++; | ||
426 | } | ||
427 | cb->first_sdbt = 0; | ||
428 | } | ||
429 | return counter; | ||
430 | } | ||
431 | |||
432 | static int start_sampling(int cpu) | ||
433 | { | ||
434 | int rc; | ||
435 | struct hws_cpu_buffer *cb; | ||
436 | |||
437 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
438 | rc = smp_ctl_ssctl_enable_activate(cpu, interval); | ||
439 | if (rc) { | ||
440 | printk(KERN_INFO "hwsampler: CPU %d ssctl failed.\n", cpu); | ||
441 | goto start_exit; | ||
442 | } | ||
443 | |||
444 | rc = -EINVAL; | ||
445 | if (!cb->qsi.es) { | ||
446 | printk(KERN_INFO "hwsampler: CPU %d ssctl not enabled.\n", cpu); | ||
447 | goto start_exit; | ||
448 | } | ||
449 | |||
450 | if (!cb->qsi.cs) { | ||
451 | printk(KERN_INFO "hwsampler: CPU %d ssctl not active.\n", cpu); | ||
452 | goto start_exit; | ||
453 | } | ||
454 | |||
455 | printk(KERN_INFO | ||
456 | "hwsampler: CPU %d, CPUMF Sampling started, interval %lu.\n", | ||
457 | cpu, interval); | ||
458 | |||
459 | rc = 0; | ||
460 | |||
461 | start_exit: | ||
462 | return rc; | ||
463 | } | ||
464 | |||
465 | static int stop_sampling(int cpu) | ||
466 | { | ||
467 | unsigned long v; | ||
468 | int rc; | ||
469 | struct hws_cpu_buffer *cb; | ||
470 | |||
471 | rc = smp_ctl_qsi(cpu); | ||
472 | WARN_ON(rc); | ||
473 | |||
474 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
475 | if (!rc && !cb->qsi.es) | ||
476 | printk(KERN_INFO "hwsampler: CPU %d, already stopped.\n", cpu); | ||
477 | |||
478 | rc = smp_ctl_ssctl_stop(cpu); | ||
479 | if (rc) { | ||
480 | printk(KERN_INFO "hwsampler: CPU %d, ssctl stop error %d.\n", | ||
481 | cpu, rc); | ||
482 | goto stop_exit; | ||
483 | } | ||
484 | |||
485 | printk(KERN_INFO "hwsampler: CPU %d, CPUMF Sampling stopped.\n", cpu); | ||
486 | |||
487 | stop_exit: | ||
488 | v = cb->req_alert; | ||
489 | if (v) | ||
490 | printk(KERN_ERR "hwsampler: CPU %d CPUMF Request alert," | ||
491 | " count=%lu.\n", cpu, v); | ||
492 | |||
493 | v = cb->loss_of_sample_data; | ||
494 | if (v) | ||
495 | printk(KERN_ERR "hwsampler: CPU %d CPUMF Loss of sample data," | ||
496 | " count=%lu.\n", cpu, v); | ||
497 | |||
498 | v = cb->invalid_entry_address; | ||
499 | if (v) | ||
500 | printk(KERN_ERR "hwsampler: CPU %d CPUMF Invalid entry address," | ||
501 | " count=%lu.\n", cpu, v); | ||
502 | |||
503 | v = cb->incorrect_sdbt_entry; | ||
504 | if (v) | ||
505 | printk(KERN_ERR | ||
506 | "hwsampler: CPU %d CPUMF Incorrect SDBT address," | ||
507 | " count=%lu.\n", cpu, v); | ||
508 | |||
509 | v = cb->sample_auth_change_alert; | ||
510 | if (v) | ||
511 | printk(KERN_ERR | ||
512 | "hwsampler: CPU %d CPUMF Sample authorization change," | ||
513 | " count=%lu.\n", cpu, v); | ||
514 | |||
515 | return rc; | ||
516 | } | ||
517 | |||
518 | static int check_hardware_prerequisites(void) | ||
519 | { | ||
520 | unsigned long long facility_bits[2]; | ||
521 | |||
522 | memcpy(facility_bits, S390_lowcore.stfle_fac_list, 32); | ||
523 | if (!(facility_bits[1] & (1ULL << 59))) | ||
524 | return -EOPNOTSUPP; | ||
525 | |||
526 | return 0; | ||
527 | } | ||
528 | /* | ||
529 | * hws_oom_callback() - the OOM callback function | ||
530 | * | ||
531 | * In case the callback is invoked during memory allocation for the | ||
532 | * hw sampler, all obtained memory is deallocated and a flag is set | ||
533 | * so main sampler memory allocation can exit with a failure code. | ||
534 | * In case the callback is invoked during sampling the hw sampler | ||
535 | * is deactivated for all CPUs. | ||
536 | */ | ||
537 | static int hws_oom_callback(struct notifier_block *nfb, | ||
538 | unsigned long dummy, void *parm) | ||
539 | { | ||
540 | unsigned long *freed; | ||
541 | int cpu; | ||
542 | struct hws_cpu_buffer *cb; | ||
543 | |||
544 | freed = parm; | ||
545 | |||
546 | mutex_lock(&hws_sem_oom); | ||
547 | |||
548 | if (hws_state == HWS_DEALLOCATED) { | ||
549 | /* during memory allocation */ | ||
550 | if (oom_killer_was_active == 0) { | ||
551 | oom_killer_was_active = 1; | ||
552 | *freed += deallocate_sdbt(); | ||
553 | } | ||
554 | } else { | ||
555 | int i; | ||
556 | cpu = get_cpu(); | ||
557 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
558 | |||
559 | if (!cb->oom) { | ||
560 | for_each_online_cpu(i) { | ||
561 | smp_ctl_ssctl_deactivate(i); | ||
562 | cb->oom = 1; | ||
563 | } | ||
564 | cb->finish = 1; | ||
565 | |||
566 | printk(KERN_INFO | ||
567 | "hwsampler: CPU %d, OOM notify during CPUMF Sampling.\n", | ||
568 | cpu); | ||
569 | } | ||
570 | } | ||
571 | |||
572 | mutex_unlock(&hws_sem_oom); | ||
573 | |||
574 | return NOTIFY_OK; | ||
575 | } | ||
576 | |||
577 | static struct notifier_block hws_oom_notifier = { | ||
578 | .notifier_call = hws_oom_callback | ||
579 | }; | ||
580 | |||
581 | static int hws_cpu_callback(struct notifier_block *nfb, | ||
582 | unsigned long action, void *hcpu) | ||
583 | { | ||
584 | /* We do not have sampler space available for all possible CPUs. | ||
585 | All CPUs should be online when hw sampling is activated. */ | ||
586 | return NOTIFY_BAD; | ||
587 | } | ||
588 | |||
589 | static struct notifier_block hws_cpu_notifier = { | ||
590 | .notifier_call = hws_cpu_callback | ||
591 | }; | ||
592 | |||
593 | /** | ||
594 | * hwsampler_deactivate() - set hardware sampling temporarily inactive | ||
595 | * @cpu: specifies the CPU to be set inactive. | ||
596 | * | ||
597 | * Returns 0 on success, !0 on failure. | ||
598 | */ | ||
599 | int hwsampler_deactivate(unsigned int cpu) | ||
600 | { | ||
601 | /* | ||
602 | * Deactivate hw sampling temporarily and flush the buffer | ||
603 | * by pushing all the pending samples to oprofile buffer. | ||
604 | * | ||
605 | * This function can be called under one of the following conditions: | ||
606 | * Memory unmap, task is exiting. | ||
607 | */ | ||
608 | int rc; | ||
609 | struct hws_cpu_buffer *cb; | ||
610 | |||
611 | rc = 0; | ||
612 | mutex_lock(&hws_sem); | ||
613 | |||
614 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
615 | if (hws_state == HWS_STARTED) { | ||
616 | rc = smp_ctl_qsi(cpu); | ||
617 | WARN_ON(rc); | ||
618 | if (cb->qsi.cs) { | ||
619 | rc = smp_ctl_ssctl_deactivate(cpu); | ||
620 | if (rc) { | ||
621 | printk(KERN_INFO | ||
622 | "hwsampler: CPU %d, CPUMF Deactivation failed.\n", cpu); | ||
623 | cb->finish = 1; | ||
624 | hws_state = HWS_STOPPING; | ||
625 | } else { | ||
626 | hws_flush_all = 1; | ||
627 | /* Add work to queue to read pending samples.*/ | ||
628 | queue_work_on(cpu, hws_wq, &cb->worker); | ||
629 | } | ||
630 | } | ||
631 | } | ||
632 | mutex_unlock(&hws_sem); | ||
633 | |||
634 | if (hws_wq) | ||
635 | flush_workqueue(hws_wq); | ||
636 | |||
637 | return rc; | ||
638 | } | ||
639 | |||
640 | /** | ||
641 | * hwsampler_activate() - activate/resume hardware sampling which was deactivated | ||
642 | * @cpu: specifies the CPU to be set active. | ||
643 | * | ||
644 | * Returns 0 on success, !0 on failure. | ||
645 | */ | ||
646 | int hwsampler_activate(unsigned int cpu) | ||
647 | { | ||
648 | /* | ||
649 | * Re-activate hw sampling. This should be called in pair with | ||
650 | * hwsampler_deactivate(). | ||
651 | */ | ||
652 | int rc; | ||
653 | struct hws_cpu_buffer *cb; | ||
654 | |||
655 | rc = 0; | ||
656 | mutex_lock(&hws_sem); | ||
657 | |||
658 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
659 | if (hws_state == HWS_STARTED) { | ||
660 | rc = smp_ctl_qsi(cpu); | ||
661 | WARN_ON(rc); | ||
662 | if (!cb->qsi.cs) { | ||
663 | hws_flush_all = 0; | ||
664 | rc = smp_ctl_ssctl_enable_activate(cpu, interval); | ||
665 | if (rc) { | ||
666 | printk(KERN_ERR | ||
667 | "CPU %d, CPUMF activate sampling failed.\n", | ||
668 | cpu); | ||
669 | } | ||
670 | } | ||
671 | } | ||
672 | |||
673 | mutex_unlock(&hws_sem); | ||
674 | |||
675 | return rc; | ||
676 | } | ||
677 | |||
678 | static void hws_ext_handler(unsigned int ext_int_code, | ||
679 | unsigned int param32, unsigned long param64) | ||
680 | { | ||
681 | int cpu; | ||
682 | struct hws_cpu_buffer *cb; | ||
683 | |||
684 | cpu = smp_processor_id(); | ||
685 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
686 | |||
687 | atomic_xchg( | ||
688 | &cb->ext_params, | ||
689 | atomic_read(&cb->ext_params) | ||
690 | | S390_lowcore.ext_params); | ||
691 | |||
692 | if (hws_wq) | ||
693 | queue_work(hws_wq, &cb->worker); | ||
694 | } | ||
695 | |||
696 | static int check_qsi_on_setup(void) | ||
697 | { | ||
698 | int rc; | ||
699 | unsigned int cpu; | ||
700 | struct hws_cpu_buffer *cb; | ||
701 | |||
702 | for_each_online_cpu(cpu) { | ||
703 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
704 | rc = smp_ctl_qsi(cpu); | ||
705 | WARN_ON(rc); | ||
706 | if (rc) | ||
707 | return -EOPNOTSUPP; | ||
708 | |||
709 | if (!cb->qsi.as) { | ||
710 | printk(KERN_INFO "hwsampler: CPUMF sampling is not authorized.\n"); | ||
711 | return -EINVAL; | ||
712 | } | ||
713 | |||
714 | if (cb->qsi.es) { | ||
715 | printk(KERN_WARNING "hwsampler: CPUMF is still enabled.\n"); | ||
716 | rc = smp_ctl_ssctl_stop(cpu); | ||
717 | if (rc) | ||
718 | return -EINVAL; | ||
719 | |||
720 | printk(KERN_INFO | ||
721 | "CPU %d, CPUMF Sampling stopped now.\n", cpu); | ||
722 | } | ||
723 | } | ||
724 | return 0; | ||
725 | } | ||
726 | |||
727 | static int check_qsi_on_start(void) | ||
728 | { | ||
729 | unsigned int cpu; | ||
730 | int rc; | ||
731 | struct hws_cpu_buffer *cb; | ||
732 | |||
733 | for_each_online_cpu(cpu) { | ||
734 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
735 | rc = smp_ctl_qsi(cpu); | ||
736 | WARN_ON(rc); | ||
737 | |||
738 | if (!cb->qsi.as) | ||
739 | return -EINVAL; | ||
740 | |||
741 | if (cb->qsi.es) | ||
742 | return -EINVAL; | ||
743 | |||
744 | if (cb->qsi.cs) | ||
745 | return -EINVAL; | ||
746 | } | ||
747 | return 0; | ||
748 | } | ||
749 | |||
750 | static void worker_on_start(unsigned int cpu) | ||
751 | { | ||
752 | struct hws_cpu_buffer *cb; | ||
753 | |||
754 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
755 | cb->worker_entry = cb->first_sdbt; | ||
756 | } | ||
757 | |||
758 | static int worker_check_error(unsigned int cpu, int ext_params) | ||
759 | { | ||
760 | int rc; | ||
761 | unsigned long *sdbt; | ||
762 | struct hws_cpu_buffer *cb; | ||
763 | |||
764 | rc = 0; | ||
765 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
766 | sdbt = (unsigned long *) cb->worker_entry; | ||
767 | |||
768 | if (!sdbt || !*sdbt) | ||
769 | return -EINVAL; | ||
770 | |||
771 | if (ext_params & EI_IEA) | ||
772 | cb->req_alert++; | ||
773 | |||
774 | if (ext_params & EI_LSDA) | ||
775 | cb->loss_of_sample_data++; | ||
776 | |||
777 | if (ext_params & EI_IEA) { | ||
778 | cb->invalid_entry_address++; | ||
779 | rc = -EINVAL; | ||
780 | } | ||
781 | |||
782 | if (ext_params & EI_ISE) { | ||
783 | cb->incorrect_sdbt_entry++; | ||
784 | rc = -EINVAL; | ||
785 | } | ||
786 | |||
787 | if (ext_params & EI_SACA) { | ||
788 | cb->sample_auth_change_alert++; | ||
789 | rc = -EINVAL; | ||
790 | } | ||
791 | |||
792 | return rc; | ||
793 | } | ||
794 | |||
795 | static void worker_on_finish(unsigned int cpu) | ||
796 | { | ||
797 | int rc, i; | ||
798 | struct hws_cpu_buffer *cb; | ||
799 | |||
800 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
801 | |||
802 | if (cb->finish) { | ||
803 | rc = smp_ctl_qsi(cpu); | ||
804 | WARN_ON(rc); | ||
805 | if (cb->qsi.es) { | ||
806 | printk(KERN_INFO | ||
807 | "hwsampler: CPU %d, CPUMF Stop/Deactivate sampling.\n", | ||
808 | cpu); | ||
809 | rc = smp_ctl_ssctl_stop(cpu); | ||
810 | if (rc) | ||
811 | printk(KERN_INFO | ||
812 | "hwsampler: CPU %d, CPUMF Deactivation failed.\n", | ||
813 | cpu); | ||
814 | |||
815 | for_each_online_cpu(i) { | ||
816 | if (i == cpu) | ||
817 | continue; | ||
818 | if (!cb->finish) { | ||
819 | cb->finish = 1; | ||
820 | queue_work_on(i, hws_wq, | ||
821 | &cb->worker); | ||
822 | } | ||
823 | } | ||
824 | } | ||
825 | } | ||
826 | } | ||
827 | |||
828 | static void worker_on_interrupt(unsigned int cpu) | ||
829 | { | ||
830 | unsigned long *sdbt; | ||
831 | unsigned char done; | ||
832 | struct hws_cpu_buffer *cb; | ||
833 | |||
834 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
835 | |||
836 | sdbt = (unsigned long *) cb->worker_entry; | ||
837 | |||
838 | done = 0; | ||
839 | /* do not proceed if stop was entered, | ||
840 | * forget the buffers not yet processed */ | ||
841 | while (!done && !cb->stop_mode) { | ||
842 | unsigned long *trailer; | ||
843 | struct hws_trailer_entry *te; | ||
844 | unsigned long *dear = 0; | ||
845 | |||
846 | trailer = trailer_entry_ptr(*sdbt); | ||
847 | /* leave loop if no more work to do */ | ||
848 | if (!(*trailer & BUFFER_FULL_MASK)) { | ||
849 | done = 1; | ||
850 | if (!hws_flush_all) | ||
851 | continue; | ||
852 | } | ||
853 | |||
854 | te = (struct hws_trailer_entry *)trailer; | ||
855 | cb->sample_overflow += te->overflow; | ||
856 | |||
857 | add_samples_to_oprofile(cpu, sdbt, dear); | ||
858 | |||
859 | /* reset trailer */ | ||
860 | xchg((unsigned char *) te, 0x40); | ||
861 | |||
862 | /* advance to next sdb slot in current sdbt */ | ||
863 | sdbt++; | ||
864 | /* in case link bit is set use address w/o link bit */ | ||
865 | if (is_link_entry(sdbt)) | ||
866 | sdbt = get_next_sdbt(sdbt); | ||
867 | |||
868 | cb->worker_entry = (unsigned long)sdbt; | ||
869 | } | ||
870 | } | ||
871 | |||
872 | static void add_samples_to_oprofile(unsigned int cpu, unsigned long *sdbt, | ||
873 | unsigned long *dear) | ||
874 | { | ||
875 | struct hws_data_entry *sample_data_ptr; | ||
876 | unsigned long *trailer; | ||
877 | |||
878 | trailer = trailer_entry_ptr(*sdbt); | ||
879 | if (dear) { | ||
880 | if (dear > trailer) | ||
881 | return; | ||
882 | trailer = dear; | ||
883 | } | ||
884 | |||
885 | sample_data_ptr = (struct hws_data_entry *)(*sdbt); | ||
886 | |||
887 | while ((unsigned long *)sample_data_ptr < trailer) { | ||
888 | struct pt_regs *regs = NULL; | ||
889 | struct task_struct *tsk = NULL; | ||
890 | |||
891 | /* | ||
892 | * Check sampling mode, 1 indicates basic (=customer) sampling | ||
893 | * mode. | ||
894 | */ | ||
895 | if (sample_data_ptr->def != 1) { | ||
896 | /* sample slot is not yet written */ | ||
897 | break; | ||
898 | } else { | ||
899 | /* make sure we don't use it twice, | ||
900 | * the next time the sampler will set it again */ | ||
901 | sample_data_ptr->def = 0; | ||
902 | } | ||
903 | |||
904 | /* Get pt_regs. */ | ||
905 | if (sample_data_ptr->P == 1) { | ||
906 | /* userspace sample */ | ||
907 | unsigned int pid = sample_data_ptr->prim_asn; | ||
908 | rcu_read_lock(); | ||
909 | tsk = pid_task(find_vpid(pid), PIDTYPE_PID); | ||
910 | if (tsk) | ||
911 | regs = task_pt_regs(tsk); | ||
912 | rcu_read_unlock(); | ||
913 | } else { | ||
914 | /* kernelspace sample */ | ||
915 | regs = task_pt_regs(current); | ||
916 | } | ||
917 | |||
918 | mutex_lock(&hws_sem); | ||
919 | oprofile_add_ext_hw_sample(sample_data_ptr->ia, regs, 0, | ||
920 | !sample_data_ptr->P, tsk); | ||
921 | mutex_unlock(&hws_sem); | ||
922 | |||
923 | sample_data_ptr++; | ||
924 | } | ||
925 | } | ||
926 | |||
927 | static void worker(struct work_struct *work) | ||
928 | { | ||
929 | unsigned int cpu; | ||
930 | int ext_params; | ||
931 | struct hws_cpu_buffer *cb; | ||
932 | |||
933 | cb = container_of(work, struct hws_cpu_buffer, worker); | ||
934 | cpu = smp_processor_id(); | ||
935 | ext_params = atomic_xchg(&cb->ext_params, 0); | ||
936 | |||
937 | if (!cb->worker_entry) | ||
938 | worker_on_start(cpu); | ||
939 | |||
940 | if (worker_check_error(cpu, ext_params)) | ||
941 | return; | ||
942 | |||
943 | if (!cb->finish) | ||
944 | worker_on_interrupt(cpu); | ||
945 | |||
946 | if (cb->finish) | ||
947 | worker_on_finish(cpu); | ||
948 | } | ||
949 | |||
950 | /** | ||
951 | * hwsampler_allocate() - allocate memory for the hardware sampler | ||
952 | * @sdbt: number of SDBTs per online CPU (must be > 0) | ||
953 | * @sdb: number of SDBs per SDBT (minimum 1, maximum 511) | ||
954 | * | ||
955 | * Returns 0 on success, !0 on failure. | ||
956 | */ | ||
957 | int hwsampler_allocate(unsigned long sdbt, unsigned long sdb) | ||
958 | { | ||
959 | int cpu, rc; | ||
960 | mutex_lock(&hws_sem); | ||
961 | |||
962 | rc = -EINVAL; | ||
963 | if (hws_state != HWS_DEALLOCATED) | ||
964 | goto allocate_exit; | ||
965 | |||
966 | if (sdbt < 1) | ||
967 | goto allocate_exit; | ||
968 | |||
969 | if (sdb > MAX_NUM_SDB || sdb < MIN_NUM_SDB) | ||
970 | goto allocate_exit; | ||
971 | |||
972 | num_sdbt = sdbt; | ||
973 | num_sdb = sdb; | ||
974 | |||
975 | oom_killer_was_active = 0; | ||
976 | register_oom_notifier(&hws_oom_notifier); | ||
977 | |||
978 | for_each_online_cpu(cpu) { | ||
979 | if (allocate_sdbt(cpu)) { | ||
980 | unregister_oom_notifier(&hws_oom_notifier); | ||
981 | goto allocate_error; | ||
982 | } | ||
983 | } | ||
984 | unregister_oom_notifier(&hws_oom_notifier); | ||
985 | if (oom_killer_was_active) | ||
986 | goto allocate_error; | ||
987 | |||
988 | hws_state = HWS_STOPPED; | ||
989 | rc = 0; | ||
990 | |||
991 | allocate_exit: | ||
992 | mutex_unlock(&hws_sem); | ||
993 | return rc; | ||
994 | |||
995 | allocate_error: | ||
996 | rc = -ENOMEM; | ||
997 | printk(KERN_ERR "hwsampler: CPUMF Memory allocation failed.\n"); | ||
998 | goto allocate_exit; | ||
999 | } | ||
1000 | |||
1001 | /** | ||
1002 | * hwsampler_deallocate() - deallocate hardware sampler memory | ||
1003 | * | ||
1004 | * Returns 0 on success, !0 on failure. | ||
1005 | */ | ||
1006 | int hwsampler_deallocate() | ||
1007 | { | ||
1008 | int rc; | ||
1009 | |||
1010 | mutex_lock(&hws_sem); | ||
1011 | |||
1012 | rc = -EINVAL; | ||
1013 | if (hws_state != HWS_STOPPED) | ||
1014 | goto deallocate_exit; | ||
1015 | |||
1016 | smp_ctl_clear_bit(0, 5); /* set bit 58 CR0 off */ | ||
1017 | deallocate_sdbt(); | ||
1018 | |||
1019 | hws_state = HWS_DEALLOCATED; | ||
1020 | rc = 0; | ||
1021 | |||
1022 | deallocate_exit: | ||
1023 | mutex_unlock(&hws_sem); | ||
1024 | |||
1025 | return rc; | ||
1026 | } | ||
1027 | |||
1028 | long hwsampler_query_min_interval(void) | ||
1029 | { | ||
1030 | if (min_sampler_rate) | ||
1031 | return min_sampler_rate; | ||
1032 | else | ||
1033 | return -EINVAL; | ||
1034 | } | ||
1035 | |||
1036 | long hwsampler_query_max_interval(void) | ||
1037 | { | ||
1038 | if (max_sampler_rate) | ||
1039 | return max_sampler_rate; | ||
1040 | else | ||
1041 | return -EINVAL; | ||
1042 | } | ||
1043 | |||
1044 | unsigned long hwsampler_get_sample_overflow_count(unsigned int cpu) | ||
1045 | { | ||
1046 | struct hws_cpu_buffer *cb; | ||
1047 | |||
1048 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
1049 | |||
1050 | return cb->sample_overflow; | ||
1051 | } | ||
1052 | |||
1053 | int hwsampler_setup() | ||
1054 | { | ||
1055 | int rc; | ||
1056 | int cpu; | ||
1057 | struct hws_cpu_buffer *cb; | ||
1058 | |||
1059 | mutex_lock(&hws_sem); | ||
1060 | |||
1061 | rc = -EINVAL; | ||
1062 | if (hws_state) | ||
1063 | goto setup_exit; | ||
1064 | |||
1065 | hws_state = HWS_INIT; | ||
1066 | |||
1067 | init_all_cpu_buffers(); | ||
1068 | |||
1069 | rc = check_hardware_prerequisites(); | ||
1070 | if (rc) | ||
1071 | goto setup_exit; | ||
1072 | |||
1073 | rc = check_qsi_on_setup(); | ||
1074 | if (rc) | ||
1075 | goto setup_exit; | ||
1076 | |||
1077 | rc = -EINVAL; | ||
1078 | hws_wq = create_workqueue("hwsampler"); | ||
1079 | if (!hws_wq) | ||
1080 | goto setup_exit; | ||
1081 | |||
1082 | register_cpu_notifier(&hws_cpu_notifier); | ||
1083 | |||
1084 | for_each_online_cpu(cpu) { | ||
1085 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
1086 | INIT_WORK(&cb->worker, worker); | ||
1087 | rc = smp_ctl_qsi(cpu); | ||
1088 | WARN_ON(rc); | ||
1089 | if (min_sampler_rate != cb->qsi.min_sampl_rate) { | ||
1090 | if (min_sampler_rate) { | ||
1091 | printk(KERN_WARNING | ||
1092 | "hwsampler: different min sampler rate values.\n"); | ||
1093 | if (min_sampler_rate < cb->qsi.min_sampl_rate) | ||
1094 | min_sampler_rate = | ||
1095 | cb->qsi.min_sampl_rate; | ||
1096 | } else | ||
1097 | min_sampler_rate = cb->qsi.min_sampl_rate; | ||
1098 | } | ||
1099 | if (max_sampler_rate != cb->qsi.max_sampl_rate) { | ||
1100 | if (max_sampler_rate) { | ||
1101 | printk(KERN_WARNING | ||
1102 | "hwsampler: different max sampler rate values.\n"); | ||
1103 | if (max_sampler_rate > cb->qsi.max_sampl_rate) | ||
1104 | max_sampler_rate = | ||
1105 | cb->qsi.max_sampl_rate; | ||
1106 | } else | ||
1107 | max_sampler_rate = cb->qsi.max_sampl_rate; | ||
1108 | } | ||
1109 | } | ||
1110 | register_external_interrupt(0x1407, hws_ext_handler); | ||
1111 | |||
1112 | hws_state = HWS_DEALLOCATED; | ||
1113 | rc = 0; | ||
1114 | |||
1115 | setup_exit: | ||
1116 | mutex_unlock(&hws_sem); | ||
1117 | return rc; | ||
1118 | } | ||
1119 | |||
1120 | int hwsampler_shutdown() | ||
1121 | { | ||
1122 | int rc; | ||
1123 | |||
1124 | mutex_lock(&hws_sem); | ||
1125 | |||
1126 | rc = -EINVAL; | ||
1127 | if (hws_state == HWS_DEALLOCATED || hws_state == HWS_STOPPED) { | ||
1128 | mutex_unlock(&hws_sem); | ||
1129 | |||
1130 | if (hws_wq) | ||
1131 | flush_workqueue(hws_wq); | ||
1132 | |||
1133 | mutex_lock(&hws_sem); | ||
1134 | |||
1135 | if (hws_state == HWS_STOPPED) { | ||
1136 | smp_ctl_clear_bit(0, 5); /* set bit 58 CR0 off */ | ||
1137 | deallocate_sdbt(); | ||
1138 | } | ||
1139 | if (hws_wq) { | ||
1140 | destroy_workqueue(hws_wq); | ||
1141 | hws_wq = NULL; | ||
1142 | } | ||
1143 | |||
1144 | unregister_external_interrupt(0x1407, hws_ext_handler); | ||
1145 | hws_state = HWS_INIT; | ||
1146 | rc = 0; | ||
1147 | } | ||
1148 | mutex_unlock(&hws_sem); | ||
1149 | |||
1150 | unregister_cpu_notifier(&hws_cpu_notifier); | ||
1151 | |||
1152 | return rc; | ||
1153 | } | ||
1154 | |||
1155 | /** | ||
1156 | * hwsampler_start_all() - start hardware sampling on all online CPUs | ||
1157 | * @rate: specifies the used interval when samples are taken | ||
1158 | * | ||
1159 | * Returns 0 on success, !0 on failure. | ||
1160 | */ | ||
1161 | int hwsampler_start_all(unsigned long rate) | ||
1162 | { | ||
1163 | int rc, cpu; | ||
1164 | |||
1165 | mutex_lock(&hws_sem); | ||
1166 | |||
1167 | hws_oom = 0; | ||
1168 | |||
1169 | rc = -EINVAL; | ||
1170 | if (hws_state != HWS_STOPPED) | ||
1171 | goto start_all_exit; | ||
1172 | |||
1173 | interval = rate; | ||
1174 | |||
1175 | /* fail if rate is not valid */ | ||
1176 | if (interval < min_sampler_rate || interval > max_sampler_rate) | ||
1177 | goto start_all_exit; | ||
1178 | |||
1179 | rc = check_qsi_on_start(); | ||
1180 | if (rc) | ||
1181 | goto start_all_exit; | ||
1182 | |||
1183 | rc = prepare_cpu_buffers(); | ||
1184 | if (rc) | ||
1185 | goto start_all_exit; | ||
1186 | |||
1187 | for_each_online_cpu(cpu) { | ||
1188 | rc = start_sampling(cpu); | ||
1189 | if (rc) | ||
1190 | break; | ||
1191 | } | ||
1192 | if (rc) { | ||
1193 | for_each_online_cpu(cpu) { | ||
1194 | stop_sampling(cpu); | ||
1195 | } | ||
1196 | goto start_all_exit; | ||
1197 | } | ||
1198 | hws_state = HWS_STARTED; | ||
1199 | rc = 0; | ||
1200 | |||
1201 | start_all_exit: | ||
1202 | mutex_unlock(&hws_sem); | ||
1203 | |||
1204 | if (rc) | ||
1205 | return rc; | ||
1206 | |||
1207 | register_oom_notifier(&hws_oom_notifier); | ||
1208 | hws_oom = 1; | ||
1209 | hws_flush_all = 0; | ||
1210 | /* now let them in, 1407 CPUMF external interrupts */ | ||
1211 | smp_ctl_set_bit(0, 5); /* set CR0 bit 58 */ | ||
1212 | |||
1213 | return 0; | ||
1214 | } | ||
1215 | |||
1216 | /** | ||
1217 | * hwsampler_stop_all() - stop hardware sampling on all online CPUs | ||
1218 | * | ||
1219 | * Returns 0 on success, !0 on failure. | ||
1220 | */ | ||
1221 | int hwsampler_stop_all() | ||
1222 | { | ||
1223 | int tmp_rc, rc, cpu; | ||
1224 | struct hws_cpu_buffer *cb; | ||
1225 | |||
1226 | mutex_lock(&hws_sem); | ||
1227 | |||
1228 | rc = 0; | ||
1229 | if (hws_state == HWS_INIT) { | ||
1230 | mutex_unlock(&hws_sem); | ||
1231 | return rc; | ||
1232 | } | ||
1233 | hws_state = HWS_STOPPING; | ||
1234 | mutex_unlock(&hws_sem); | ||
1235 | |||
1236 | for_each_online_cpu(cpu) { | ||
1237 | cb = &per_cpu(sampler_cpu_buffer, cpu); | ||
1238 | cb->stop_mode = 1; | ||
1239 | tmp_rc = stop_sampling(cpu); | ||
1240 | if (tmp_rc) | ||
1241 | rc = tmp_rc; | ||
1242 | } | ||
1243 | |||
1244 | if (hws_wq) | ||
1245 | flush_workqueue(hws_wq); | ||
1246 | |||
1247 | mutex_lock(&hws_sem); | ||
1248 | if (hws_oom) { | ||
1249 | unregister_oom_notifier(&hws_oom_notifier); | ||
1250 | hws_oom = 0; | ||
1251 | } | ||
1252 | hws_state = HWS_STOPPED; | ||
1253 | mutex_unlock(&hws_sem); | ||
1254 | |||
1255 | return rc; | ||
1256 | } | ||
diff --git a/arch/s390/oprofile/hwsampler.h b/arch/s390/oprofile/hwsampler.h new file mode 100644 index 000000000000..8c72b59316b5 --- /dev/null +++ b/arch/s390/oprofile/hwsampler.h | |||
@@ -0,0 +1,113 @@ | |||
1 | /* | ||
2 | * CPUMF HW sampler functions and internal structures | ||
3 | * | ||
4 | * Copyright IBM Corp. 2010 | ||
5 | * Author(s): Heinz Graalfs <graalfs@de.ibm.com> | ||
6 | */ | ||
7 | |||
8 | #ifndef HWSAMPLER_H_ | ||
9 | #define HWSAMPLER_H_ | ||
10 | |||
11 | #include <linux/workqueue.h> | ||
12 | |||
13 | struct hws_qsi_info_block /* QUERY SAMPLING information block */ | ||
14 | { /* Bit(s) */ | ||
15 | unsigned int b0_13:14; /* 0-13: zeros */ | ||
16 | unsigned int as:1; /* 14: sampling authorisation control*/ | ||
17 | unsigned int b15_21:7; /* 15-21: zeros */ | ||
18 | unsigned int es:1; /* 22: sampling enable control */ | ||
19 | unsigned int b23_29:7; /* 23-29: zeros */ | ||
20 | unsigned int cs:1; /* 30: sampling activation control */ | ||
21 | unsigned int:1; /* 31: reserved */ | ||
22 | unsigned int bsdes:16; /* 4-5: size of sampling entry */ | ||
23 | unsigned int:16; /* 6-7: reserved */ | ||
24 | unsigned long min_sampl_rate; /* 8-15: minimum sampling interval */ | ||
25 | unsigned long max_sampl_rate; /* 16-23: maximum sampling interval*/ | ||
26 | unsigned long tear; /* 24-31: TEAR contents */ | ||
27 | unsigned long dear; /* 32-39: DEAR contents */ | ||
28 | unsigned int rsvrd0; /* 40-43: reserved */ | ||
29 | unsigned int cpu_speed; /* 44-47: CPU speed */ | ||
30 | unsigned long long rsvrd1; /* 48-55: reserved */ | ||
31 | unsigned long long rsvrd2; /* 56-63: reserved */ | ||
32 | }; | ||
33 | |||
34 | struct hws_ssctl_request_block /* SET SAMPLING CONTROLS req block */ | ||
35 | { /* bytes 0 - 7 Bit(s) */ | ||
36 | unsigned int s:1; /* 0: maximum buffer indicator */ | ||
37 | unsigned int h:1; /* 1: part. level reserved for VM use*/ | ||
38 | unsigned long b2_53:52; /* 2-53: zeros */ | ||
39 | unsigned int es:1; /* 54: sampling enable control */ | ||
40 | unsigned int b55_61:7; /* 55-61: - zeros */ | ||
41 | unsigned int cs:1; /* 62: sampling activation control */ | ||
42 | unsigned int b63:1; /* 63: zero */ | ||
43 | unsigned long interval; /* 8-15: sampling interval */ | ||
44 | unsigned long tear; /* 16-23: TEAR contents */ | ||
45 | unsigned long dear; /* 24-31: DEAR contents */ | ||
46 | /* 32-63: */ | ||
47 | unsigned long rsvrd1; /* reserved */ | ||
48 | unsigned long rsvrd2; /* reserved */ | ||
49 | unsigned long rsvrd3; /* reserved */ | ||
50 | unsigned long rsvrd4; /* reserved */ | ||
51 | }; | ||
52 | |||
53 | struct hws_cpu_buffer { | ||
54 | unsigned long first_sdbt; /* @ of 1st SDB-Table for this CP*/ | ||
55 | unsigned long worker_entry; | ||
56 | unsigned long sample_overflow; /* taken from SDB ... */ | ||
57 | struct hws_qsi_info_block qsi; | ||
58 | struct hws_ssctl_request_block ssctl; | ||
59 | struct work_struct worker; | ||
60 | atomic_t ext_params; | ||
61 | unsigned long req_alert; | ||
62 | unsigned long loss_of_sample_data; | ||
63 | unsigned long invalid_entry_address; | ||
64 | unsigned long incorrect_sdbt_entry; | ||
65 | unsigned long sample_auth_change_alert; | ||
66 | unsigned int finish:1; | ||
67 | unsigned int oom:1; | ||
68 | unsigned int stop_mode:1; | ||
69 | }; | ||
70 | |||
71 | struct hws_data_entry { | ||
72 | unsigned int def:16; /* 0-15 Data Entry Format */ | ||
73 | unsigned int R:4; /* 16-19 reserved */ | ||
74 | unsigned int U:4; /* 20-23 Number of unique instruct. */ | ||
75 | unsigned int z:2; /* zeros */ | ||
76 | unsigned int T:1; /* 26 PSW DAT mode */ | ||
77 | unsigned int W:1; /* 27 PSW wait state */ | ||
78 | unsigned int P:1; /* 28 PSW Problem state */ | ||
79 | unsigned int AS:2; /* 29-30 PSW address-space control */ | ||
80 | unsigned int I:1; /* 31 entry valid or invalid */ | ||
81 | unsigned int:16; | ||
82 | unsigned int prim_asn:16; /* primary ASN */ | ||
83 | unsigned long long ia; /* Instruction Address */ | ||
84 | unsigned long long lpp; /* Logical-Partition Program Param. */ | ||
85 | unsigned long long vpp; /* Virtual-Machine Program Param. */ | ||
86 | }; | ||
87 | |||
88 | struct hws_trailer_entry { | ||
89 | unsigned int f:1; /* 0 - Block Full Indicator */ | ||
90 | unsigned int a:1; /* 1 - Alert request control */ | ||
91 | unsigned long:62; /* 2 - 63: Reserved */ | ||
92 | unsigned long overflow; /* 64 - sample Overflow count */ | ||
93 | unsigned long timestamp; /* 16 - time-stamp */ | ||
94 | unsigned long timestamp1; /* */ | ||
95 | unsigned long reserved1; /* 32 -Reserved */ | ||
96 | unsigned long reserved2; /* */ | ||
97 | unsigned long progusage1; /* 48 - reserved for programming use */ | ||
98 | unsigned long progusage2; /* */ | ||
99 | }; | ||
100 | |||
101 | int hwsampler_setup(void); | ||
102 | int hwsampler_shutdown(void); | ||
103 | int hwsampler_allocate(unsigned long sdbt, unsigned long sdb); | ||
104 | int hwsampler_deallocate(void); | ||
105 | long hwsampler_query_min_interval(void); | ||
106 | long hwsampler_query_max_interval(void); | ||
107 | int hwsampler_start_all(unsigned long interval); | ||
108 | int hwsampler_stop_all(void); | ||
109 | int hwsampler_deactivate(unsigned int cpu); | ||
110 | int hwsampler_activate(unsigned int cpu); | ||
111 | unsigned long hwsampler_get_sample_overflow_count(unsigned int cpu); | ||
112 | |||
113 | #endif /*HWSAMPLER_H_*/ | ||
diff --git a/arch/s390/oprofile/init.c b/arch/s390/oprofile/init.c index 7a995113b918..16c76def4a9d 100644 --- a/arch/s390/oprofile/init.c +++ b/arch/s390/oprofile/init.c | |||
@@ -4,23 +4,182 @@ | |||
4 | * S390 Version | 4 | * S390 Version |
5 | * Copyright (C) 2003 IBM Deutschland Entwicklung GmbH, IBM Corporation | 5 | * Copyright (C) 2003 IBM Deutschland Entwicklung GmbH, IBM Corporation |
6 | * Author(s): Thomas Spatzier (tspat@de.ibm.com) | 6 | * Author(s): Thomas Spatzier (tspat@de.ibm.com) |
7 | * Author(s): Mahesh Salgaonkar (mahesh@linux.vnet.ibm.com) | ||
8 | * Author(s): Heinz Graalfs (graalfs@linux.vnet.ibm.com) | ||
7 | * | 9 | * |
8 | * @remark Copyright 2002 OProfile authors | 10 | * @remark Copyright 2002-2011 OProfile authors |
9 | */ | 11 | */ |
10 | 12 | ||
11 | #include <linux/oprofile.h> | 13 | #include <linux/oprofile.h> |
12 | #include <linux/init.h> | 14 | #include <linux/init.h> |
13 | #include <linux/errno.h> | 15 | #include <linux/errno.h> |
16 | #include <linux/oprofile.h> | ||
17 | #include <linux/errno.h> | ||
18 | #include <linux/fs.h> | ||
19 | |||
20 | #include "../../../drivers/oprofile/oprof.h" | ||
21 | #include "hwsampler.h" | ||
22 | |||
23 | #define DEFAULT_INTERVAL 4096 | ||
24 | |||
25 | #define DEFAULT_SDBT_BLOCKS 1 | ||
26 | #define DEFAULT_SDB_BLOCKS 511 | ||
27 | |||
28 | static unsigned long oprofile_hw_interval = DEFAULT_INTERVAL; | ||
29 | static unsigned long oprofile_min_interval; | ||
30 | static unsigned long oprofile_max_interval; | ||
31 | |||
32 | static unsigned long oprofile_sdbt_blocks = DEFAULT_SDBT_BLOCKS; | ||
33 | static unsigned long oprofile_sdb_blocks = DEFAULT_SDB_BLOCKS; | ||
14 | 34 | ||
35 | static int hwsampler_file; | ||
36 | static int hwsampler_running; /* start_mutex must be held to change */ | ||
37 | |||
38 | static struct oprofile_operations timer_ops; | ||
15 | 39 | ||
16 | extern void s390_backtrace(struct pt_regs * const regs, unsigned int depth); | 40 | extern void s390_backtrace(struct pt_regs * const regs, unsigned int depth); |
17 | 41 | ||
18 | int __init oprofile_arch_init(struct oprofile_operations* ops) | 42 | static int oprofile_hwsampler_start(void) |
43 | { | ||
44 | int retval; | ||
45 | |||
46 | hwsampler_running = hwsampler_file; | ||
47 | |||
48 | if (!hwsampler_running) | ||
49 | return timer_ops.start(); | ||
50 | |||
51 | retval = hwsampler_allocate(oprofile_sdbt_blocks, oprofile_sdb_blocks); | ||
52 | if (retval) | ||
53 | return retval; | ||
54 | |||
55 | retval = hwsampler_start_all(oprofile_hw_interval); | ||
56 | if (retval) | ||
57 | hwsampler_deallocate(); | ||
58 | |||
59 | return retval; | ||
60 | } | ||
61 | |||
62 | static void oprofile_hwsampler_stop(void) | ||
63 | { | ||
64 | if (!hwsampler_running) { | ||
65 | timer_ops.stop(); | ||
66 | return; | ||
67 | } | ||
68 | |||
69 | hwsampler_stop_all(); | ||
70 | hwsampler_deallocate(); | ||
71 | return; | ||
72 | } | ||
73 | |||
74 | static ssize_t hwsampler_read(struct file *file, char __user *buf, | ||
75 | size_t count, loff_t *offset) | ||
76 | { | ||
77 | return oprofilefs_ulong_to_user(hwsampler_file, buf, count, offset); | ||
78 | } | ||
79 | |||
80 | static ssize_t hwsampler_write(struct file *file, char const __user *buf, | ||
81 | size_t count, loff_t *offset) | ||
82 | { | ||
83 | unsigned long val; | ||
84 | int retval; | ||
85 | |||
86 | if (*offset) | ||
87 | return -EINVAL; | ||
88 | |||
89 | retval = oprofilefs_ulong_from_user(&val, buf, count); | ||
90 | if (retval) | ||
91 | return retval; | ||
92 | |||
93 | if (oprofile_started) | ||
94 | /* | ||
95 | * save to do without locking as we set | ||
96 | * hwsampler_running in start() when start_mutex is | ||
97 | * held | ||
98 | */ | ||
99 | return -EBUSY; | ||
100 | |||
101 | hwsampler_file = val; | ||
102 | |||
103 | return count; | ||
104 | } | ||
105 | |||
106 | static const struct file_operations hwsampler_fops = { | ||
107 | .read = hwsampler_read, | ||
108 | .write = hwsampler_write, | ||
109 | }; | ||
110 | |||
111 | static int oprofile_create_hwsampling_files(struct super_block *sb, | ||
112 | struct dentry *root) | ||
113 | { | ||
114 | struct dentry *hw_dir; | ||
115 | |||
116 | /* reinitialize default values */ | ||
117 | hwsampler_file = 1; | ||
118 | |||
119 | hw_dir = oprofilefs_mkdir(sb, root, "hwsampling"); | ||
120 | if (!hw_dir) | ||
121 | return -EINVAL; | ||
122 | |||
123 | oprofilefs_create_file(sb, hw_dir, "hwsampler", &hwsampler_fops); | ||
124 | oprofilefs_create_ulong(sb, hw_dir, "hw_interval", | ||
125 | &oprofile_hw_interval); | ||
126 | oprofilefs_create_ro_ulong(sb, hw_dir, "hw_min_interval", | ||
127 | &oprofile_min_interval); | ||
128 | oprofilefs_create_ro_ulong(sb, hw_dir, "hw_max_interval", | ||
129 | &oprofile_max_interval); | ||
130 | oprofilefs_create_ulong(sb, hw_dir, "hw_sdbt_blocks", | ||
131 | &oprofile_sdbt_blocks); | ||
132 | |||
133 | return 0; | ||
134 | } | ||
135 | |||
136 | static int oprofile_hwsampler_init(struct oprofile_operations *ops) | ||
137 | { | ||
138 | if (hwsampler_setup()) | ||
139 | return -ENODEV; | ||
140 | |||
141 | /* | ||
142 | * create hwsampler files only if hwsampler_setup() succeeds. | ||
143 | */ | ||
144 | oprofile_min_interval = hwsampler_query_min_interval(); | ||
145 | if (oprofile_min_interval < 0) { | ||
146 | oprofile_min_interval = 0; | ||
147 | return -ENODEV; | ||
148 | } | ||
149 | oprofile_max_interval = hwsampler_query_max_interval(); | ||
150 | if (oprofile_max_interval < 0) { | ||
151 | oprofile_max_interval = 0; | ||
152 | return -ENODEV; | ||
153 | } | ||
154 | |||
155 | if (oprofile_timer_init(ops)) | ||
156 | return -ENODEV; | ||
157 | |||
158 | printk(KERN_INFO "oprofile: using hardware sampling\n"); | ||
159 | |||
160 | memcpy(&timer_ops, ops, sizeof(timer_ops)); | ||
161 | |||
162 | ops->start = oprofile_hwsampler_start; | ||
163 | ops->stop = oprofile_hwsampler_stop; | ||
164 | ops->create_files = oprofile_create_hwsampling_files; | ||
165 | |||
166 | return 0; | ||
167 | } | ||
168 | |||
169 | static void oprofile_hwsampler_exit(void) | ||
170 | { | ||
171 | oprofile_timer_exit(); | ||
172 | hwsampler_shutdown(); | ||
173 | } | ||
174 | |||
175 | int __init oprofile_arch_init(struct oprofile_operations *ops) | ||
19 | { | 176 | { |
20 | ops->backtrace = s390_backtrace; | 177 | ops->backtrace = s390_backtrace; |
21 | return -ENODEV; | 178 | |
179 | return oprofile_hwsampler_init(ops); | ||
22 | } | 180 | } |
23 | 181 | ||
24 | void oprofile_arch_exit(void) | 182 | void oprofile_arch_exit(void) |
25 | { | 183 | { |
184 | oprofile_hwsampler_exit(); | ||
26 | } | 185 | } |
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c index 26604188aa49..e8dbe179587f 100644 --- a/arch/x86/kernel/cpu/perf_event.c +++ b/arch/x86/kernel/cpu/perf_event.c | |||
@@ -178,8 +178,6 @@ struct cpu_hw_events { | |||
178 | */ | 178 | */ |
179 | #define INTEL_UEVENT_CONSTRAINT(c, n) \ | 179 | #define INTEL_UEVENT_CONSTRAINT(c, n) \ |
180 | EVENT_CONSTRAINT(c, n, INTEL_ARCH_EVENT_MASK) | 180 | EVENT_CONSTRAINT(c, n, INTEL_ARCH_EVENT_MASK) |
181 | #define PEBS_EVENT_CONSTRAINT(c, n) \ | ||
182 | INTEL_UEVENT_CONSTRAINT(c, n) | ||
183 | 181 | ||
184 | #define EVENT_CONSTRAINT_END \ | 182 | #define EVENT_CONSTRAINT_END \ |
185 | EVENT_CONSTRAINT(0, 0, 0) | 183 | EVENT_CONSTRAINT(0, 0, 0) |
diff --git a/arch/x86/kernel/cpu/perf_event_intel_ds.c b/arch/x86/kernel/cpu/perf_event_intel_ds.c index b95c66ae4a2a..bab491b8ee25 100644 --- a/arch/x86/kernel/cpu/perf_event_intel_ds.c +++ b/arch/x86/kernel/cpu/perf_event_intel_ds.c | |||
@@ -362,87 +362,69 @@ static int intel_pmu_drain_bts_buffer(void) | |||
362 | * PEBS | 362 | * PEBS |
363 | */ | 363 | */ |
364 | static struct event_constraint intel_core2_pebs_event_constraints[] = { | 364 | static struct event_constraint intel_core2_pebs_event_constraints[] = { |
365 | PEBS_EVENT_CONSTRAINT(0x00c0, 0x1), /* INST_RETIRED.ANY */ | 365 | INTEL_UEVENT_CONSTRAINT(0x00c0, 0x1), /* INST_RETIRED.ANY */ |
366 | PEBS_EVENT_CONSTRAINT(0xfec1, 0x1), /* X87_OPS_RETIRED.ANY */ | 366 | INTEL_UEVENT_CONSTRAINT(0xfec1, 0x1), /* X87_OPS_RETIRED.ANY */ |
367 | PEBS_EVENT_CONSTRAINT(0x00c5, 0x1), /* BR_INST_RETIRED.MISPRED */ | 367 | INTEL_UEVENT_CONSTRAINT(0x00c5, 0x1), /* BR_INST_RETIRED.MISPRED */ |
368 | PEBS_EVENT_CONSTRAINT(0x1fc7, 0x1), /* SIMD_INST_RETURED.ANY */ | 368 | INTEL_UEVENT_CONSTRAINT(0x1fc7, 0x1), /* SIMD_INST_RETURED.ANY */ |
369 | INTEL_EVENT_CONSTRAINT(0xcb, 0x1), /* MEM_LOAD_RETIRED.* */ | 369 | INTEL_EVENT_CONSTRAINT(0xcb, 0x1), /* MEM_LOAD_RETIRED.* */ |
370 | EVENT_CONSTRAINT_END | 370 | EVENT_CONSTRAINT_END |
371 | }; | 371 | }; |
372 | 372 | ||
373 | static struct event_constraint intel_atom_pebs_event_constraints[] = { | 373 | static struct event_constraint intel_atom_pebs_event_constraints[] = { |
374 | PEBS_EVENT_CONSTRAINT(0x00c0, 0x1), /* INST_RETIRED.ANY */ | 374 | INTEL_UEVENT_CONSTRAINT(0x00c0, 0x1), /* INST_RETIRED.ANY */ |
375 | PEBS_EVENT_CONSTRAINT(0x00c5, 0x1), /* MISPREDICTED_BRANCH_RETIRED */ | 375 | INTEL_UEVENT_CONSTRAINT(0x00c5, 0x1), /* MISPREDICTED_BRANCH_RETIRED */ |
376 | INTEL_EVENT_CONSTRAINT(0xcb, 0x1), /* MEM_LOAD_RETIRED.* */ | 376 | INTEL_EVENT_CONSTRAINT(0xcb, 0x1), /* MEM_LOAD_RETIRED.* */ |
377 | EVENT_CONSTRAINT_END | 377 | EVENT_CONSTRAINT_END |
378 | }; | 378 | }; |
379 | 379 | ||
380 | static struct event_constraint intel_nehalem_pebs_event_constraints[] = { | 380 | static struct event_constraint intel_nehalem_pebs_event_constraints[] = { |
381 | INTEL_EVENT_CONSTRAINT(0x0b, 0xf), /* MEM_INST_RETIRED.* */ | 381 | INTEL_EVENT_CONSTRAINT(0x0b, 0xf), /* MEM_INST_RETIRED.* */ |
382 | INTEL_EVENT_CONSTRAINT(0x0f, 0xf), /* MEM_UNCORE_RETIRED.* */ | 382 | INTEL_EVENT_CONSTRAINT(0x0f, 0xf), /* MEM_UNCORE_RETIRED.* */ |
383 | PEBS_EVENT_CONSTRAINT(0x010c, 0xf), /* MEM_STORE_RETIRED.DTLB_MISS */ | 383 | INTEL_UEVENT_CONSTRAINT(0x010c, 0xf), /* MEM_STORE_RETIRED.DTLB_MISS */ |
384 | INTEL_EVENT_CONSTRAINT(0xc0, 0xf), /* INST_RETIRED.ANY */ | 384 | INTEL_EVENT_CONSTRAINT(0xc0, 0xf), /* INST_RETIRED.ANY */ |
385 | INTEL_EVENT_CONSTRAINT(0xc2, 0xf), /* UOPS_RETIRED.* */ | 385 | INTEL_EVENT_CONSTRAINT(0xc2, 0xf), /* UOPS_RETIRED.* */ |
386 | INTEL_EVENT_CONSTRAINT(0xc4, 0xf), /* BR_INST_RETIRED.* */ | 386 | INTEL_EVENT_CONSTRAINT(0xc4, 0xf), /* BR_INST_RETIRED.* */ |
387 | PEBS_EVENT_CONSTRAINT(0x02c5, 0xf), /* BR_MISP_RETIRED.NEAR_CALL */ | 387 | INTEL_UEVENT_CONSTRAINT(0x02c5, 0xf), /* BR_MISP_RETIRED.NEAR_CALL */ |
388 | INTEL_EVENT_CONSTRAINT(0xc7, 0xf), /* SSEX_UOPS_RETIRED.* */ | 388 | INTEL_EVENT_CONSTRAINT(0xc7, 0xf), /* SSEX_UOPS_RETIRED.* */ |
389 | PEBS_EVENT_CONSTRAINT(0x20c8, 0xf), /* ITLB_MISS_RETIRED */ | 389 | INTEL_UEVENT_CONSTRAINT(0x20c8, 0xf), /* ITLB_MISS_RETIRED */ |
390 | INTEL_EVENT_CONSTRAINT(0xcb, 0xf), /* MEM_LOAD_RETIRED.* */ | 390 | INTEL_EVENT_CONSTRAINT(0xcb, 0xf), /* MEM_LOAD_RETIRED.* */ |
391 | INTEL_EVENT_CONSTRAINT(0xf7, 0xf), /* FP_ASSIST.* */ | 391 | INTEL_EVENT_CONSTRAINT(0xf7, 0xf), /* FP_ASSIST.* */ |
392 | EVENT_CONSTRAINT_END | 392 | EVENT_CONSTRAINT_END |
393 | }; | 393 | }; |
394 | 394 | ||
395 | static struct event_constraint intel_westmere_pebs_event_constraints[] = { | 395 | static struct event_constraint intel_westmere_pebs_event_constraints[] = { |
396 | INTEL_EVENT_CONSTRAINT(0x0b, 0xf), /* MEM_INST_RETIRED.* */ | 396 | INTEL_EVENT_CONSTRAINT(0x0b, 0xf), /* MEM_INST_RETIRED.* */ |
397 | INTEL_EVENT_CONSTRAINT(0x0f, 0xf), /* MEM_UNCORE_RETIRED.* */ | 397 | INTEL_EVENT_CONSTRAINT(0x0f, 0xf), /* MEM_UNCORE_RETIRED.* */ |
398 | PEBS_EVENT_CONSTRAINT(0x010c, 0xf), /* MEM_STORE_RETIRED.DTLB_MISS */ | 398 | INTEL_UEVENT_CONSTRAINT(0x010c, 0xf), /* MEM_STORE_RETIRED.DTLB_MISS */ |
399 | INTEL_EVENT_CONSTRAINT(0xc0, 0xf), /* INSTR_RETIRED.* */ | 399 | INTEL_EVENT_CONSTRAINT(0xc0, 0xf), /* INSTR_RETIRED.* */ |
400 | INTEL_EVENT_CONSTRAINT(0xc2, 0xf), /* UOPS_RETIRED.* */ | 400 | INTEL_EVENT_CONSTRAINT(0xc2, 0xf), /* UOPS_RETIRED.* */ |
401 | 401 | INTEL_EVENT_CONSTRAINT(0xc4, 0xf), /* BR_INST_RETIRED.* */ | |
402 | INTEL_EVENT_CONSTRAINT(0xc4, 0xf), /* BR_INST_RETIRED.* */ | 402 | INTEL_EVENT_CONSTRAINT(0xc5, 0xf), /* BR_MISP_RETIRED.* */ |
403 | INTEL_EVENT_CONSTRAINT(0xc5, 0xf), /* BR_MISP_RETIRED.* */ | 403 | INTEL_EVENT_CONSTRAINT(0xc7, 0xf), /* SSEX_UOPS_RETIRED.* */ |
404 | INTEL_EVENT_CONSTRAINT(0xc7, 0xf), /* SSEX_UOPS_RETIRED.* */ | 404 | INTEL_UEVENT_CONSTRAINT(0x20c8, 0xf), /* ITLB_MISS_RETIRED */ |
405 | PEBS_EVENT_CONSTRAINT(0x20c8, 0xf), /* ITLB_MISS_RETIRED */ | 405 | INTEL_EVENT_CONSTRAINT(0xcb, 0xf), /* MEM_LOAD_RETIRED.* */ |
406 | INTEL_EVENT_CONSTRAINT(0xcb, 0xf), /* MEM_LOAD_RETIRED.* */ | 406 | INTEL_EVENT_CONSTRAINT(0xf7, 0xf), /* FP_ASSIST.* */ |
407 | INTEL_EVENT_CONSTRAINT(0xf7, 0xf), /* FP_ASSIST.* */ | ||
408 | EVENT_CONSTRAINT_END | 407 | EVENT_CONSTRAINT_END |
409 | }; | 408 | }; |
410 | 409 | ||
411 | static struct event_constraint intel_snb_pebs_events[] = { | 410 | static struct event_constraint intel_snb_pebs_events[] = { |
412 | PEBS_EVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PRECDIST */ | 411 | INTEL_UEVENT_CONSTRAINT(0x01c0, 0x2), /* INST_RETIRED.PRECDIST */ |
413 | PEBS_EVENT_CONSTRAINT(0x01c2, 0xf), /* UOPS_RETIRED.ALL */ | 412 | INTEL_UEVENT_CONSTRAINT(0x01c2, 0xf), /* UOPS_RETIRED.ALL */ |
414 | PEBS_EVENT_CONSTRAINT(0x02c2, 0xf), /* UOPS_RETIRED.RETIRE_SLOTS */ | 413 | INTEL_UEVENT_CONSTRAINT(0x02c2, 0xf), /* UOPS_RETIRED.RETIRE_SLOTS */ |
415 | PEBS_EVENT_CONSTRAINT(0x01c4, 0xf), /* BR_INST_RETIRED.CONDITIONAL */ | 414 | INTEL_EVENT_CONSTRAINT(0xc4, 0xf), /* BR_INST_RETIRED.* */ |
416 | PEBS_EVENT_CONSTRAINT(0x02c4, 0xf), /* BR_INST_RETIRED.NEAR_CALL */ | 415 | INTEL_EVENT_CONSTRAINT(0xc5, 0xf), /* BR_MISP_RETIRED.* */ |
417 | PEBS_EVENT_CONSTRAINT(0x04c4, 0xf), /* BR_INST_RETIRED.ALL_BRANCHES */ | 416 | INTEL_EVENT_CONSTRAINT(0xcd, 0x8), /* MEM_TRANS_RETIRED.* */ |
418 | PEBS_EVENT_CONSTRAINT(0x08c4, 0xf), /* BR_INST_RETIRED.NEAR_RETURN */ | 417 | INTEL_UEVENT_CONSTRAINT(0x11d0, 0xf), /* MEM_UOP_RETIRED.STLB_MISS_LOADS */ |
419 | PEBS_EVENT_CONSTRAINT(0x10c4, 0xf), /* BR_INST_RETIRED.NOT_TAKEN */ | 418 | INTEL_UEVENT_CONSTRAINT(0x12d0, 0xf), /* MEM_UOP_RETIRED.STLB_MISS_STORES */ |
420 | PEBS_EVENT_CONSTRAINT(0x20c4, 0xf), /* BR_INST_RETIRED.NEAR_TAKEN */ | 419 | INTEL_UEVENT_CONSTRAINT(0x21d0, 0xf), /* MEM_UOP_RETIRED.LOCK_LOADS */ |
421 | PEBS_EVENT_CONSTRAINT(0x40c4, 0xf), /* BR_INST_RETIRED.FAR_BRANCH */ | 420 | INTEL_UEVENT_CONSTRAINT(0x22d0, 0xf), /* MEM_UOP_RETIRED.LOCK_STORES */ |
422 | PEBS_EVENT_CONSTRAINT(0x01c5, 0xf), /* BR_MISP_RETIRED.CONDITIONAL */ | 421 | INTEL_UEVENT_CONSTRAINT(0x41d0, 0xf), /* MEM_UOP_RETIRED.SPLIT_LOADS */ |
423 | PEBS_EVENT_CONSTRAINT(0x02c5, 0xf), /* BR_MISP_RETIRED.NEAR_CALL */ | 422 | INTEL_UEVENT_CONSTRAINT(0x42d0, 0xf), /* MEM_UOP_RETIRED.SPLIT_STORES */ |
424 | PEBS_EVENT_CONSTRAINT(0x04c5, 0xf), /* BR_MISP_RETIRED.ALL_BRANCHES */ | 423 | INTEL_UEVENT_CONSTRAINT(0x81d0, 0xf), /* MEM_UOP_RETIRED.ANY_LOADS */ |
425 | PEBS_EVENT_CONSTRAINT(0x10c5, 0xf), /* BR_MISP_RETIRED.NOT_TAKEN */ | 424 | INTEL_UEVENT_CONSTRAINT(0x82d0, 0xf), /* MEM_UOP_RETIRED.ANY_STORES */ |
426 | PEBS_EVENT_CONSTRAINT(0x20c5, 0xf), /* BR_MISP_RETIRED.TAKEN */ | 425 | INTEL_EVENT_CONSTRAINT(0xd1, 0xf), /* MEM_LOAD_UOPS_RETIRED.* */ |
427 | PEBS_EVENT_CONSTRAINT(0x01cd, 0x8), /* MEM_TRANS_RETIRED.LOAD_LATENCY */ | 426 | INTEL_EVENT_CONSTRAINT(0xd2, 0xf), /* MEM_LOAD_UOPS_LLC_HIT_RETIRED.* */ |
428 | PEBS_EVENT_CONSTRAINT(0x02cd, 0x8), /* MEM_TRANS_RETIRED.PRECISE_STORE */ | 427 | INTEL_UEVENT_CONSTRAINT(0x02d4, 0xf), /* MEM_LOAD_UOPS_MISC_RETIRED.LLC_MISS */ |
429 | PEBS_EVENT_CONSTRAINT(0x11d0, 0xf), /* MEM_UOP_RETIRED.STLB_MISS_LOADS */ | ||
430 | PEBS_EVENT_CONSTRAINT(0x12d0, 0xf), /* MEM_UOP_RETIRED.STLB_MISS_STORES */ | ||
431 | PEBS_EVENT_CONSTRAINT(0x21d0, 0xf), /* MEM_UOP_RETIRED.LOCK_LOADS */ | ||
432 | PEBS_EVENT_CONSTRAINT(0x22d0, 0xf), /* MEM_UOP_RETIRED.LOCK_STORES */ | ||
433 | PEBS_EVENT_CONSTRAINT(0x41d0, 0xf), /* MEM_UOP_RETIRED.SPLIT_LOADS */ | ||
434 | PEBS_EVENT_CONSTRAINT(0x42d0, 0xf), /* MEM_UOP_RETIRED.SPLIT_STORES */ | ||
435 | PEBS_EVENT_CONSTRAINT(0x81d0, 0xf), /* MEM_UOP_RETIRED.ANY_LOADS */ | ||
436 | PEBS_EVENT_CONSTRAINT(0x82d0, 0xf), /* MEM_UOP_RETIRED.ANY_STORES */ | ||
437 | PEBS_EVENT_CONSTRAINT(0x01d1, 0xf), /* MEM_LOAD_UOPS_RETIRED.L1_HIT */ | ||
438 | PEBS_EVENT_CONSTRAINT(0x02d1, 0xf), /* MEM_LOAD_UOPS_RETIRED.L2_HIT */ | ||
439 | PEBS_EVENT_CONSTRAINT(0x04d1, 0xf), /* MEM_LOAD_UOPS_RETIRED.LLC_HIT */ | ||
440 | PEBS_EVENT_CONSTRAINT(0x40d1, 0xf), /* MEM_LOAD_UOPS_RETIRED.HIT_LFB */ | ||
441 | PEBS_EVENT_CONSTRAINT(0x01d2, 0xf), /* MEM_LOAD_UOPS_LLC_HIT_RETIRED.XSNP_MISS */ | ||
442 | PEBS_EVENT_CONSTRAINT(0x02d2, 0xf), /* MEM_LOAD_UOPS_LLC_HIT_RETIRED.XSNP_HIT */ | ||
443 | PEBS_EVENT_CONSTRAINT(0x04d2, 0xf), /* MEM_LOAD_UOPS_LLC_HIT_RETIRED.XSNP_HITM */ | ||
444 | PEBS_EVENT_CONSTRAINT(0x08d2, 0xf), /* MEM_LOAD_UOPS_LLC_HIT_RETIRED.XSNP_NONE */ | ||
445 | PEBS_EVENT_CONSTRAINT(0x02d4, 0xf), /* MEM_LOAD_UOPS_MISC_RETIRED.LLC_MISS */ | ||
446 | EVENT_CONSTRAINT_END | 428 | EVENT_CONSTRAINT_END |
447 | }; | 429 | }; |
448 | 430 | ||