diff options
author | Ingo Molnar <mingo@elte.hu> | 2009-09-21 06:20:38 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-09-21 08:34:11 -0400 |
commit | 57c0c15b5244320065374ad2c54f4fbec77a6428 (patch) | |
tree | 35369d817f5925aca09b083bba47c437b91386d9 | |
parent | cdd6c482c9ff9c55475ee7392ec8f672eddb7be6 (diff) |
perf: Tidy up after the big rename
- provide compatibility Kconfig entry for existing PERF_COUNTERS .config's
- provide courtesy copy of old perf_counter.h, for user-space projects
- small indentation fixups
- fix up MAINTAINERS
- fix small x86 printout fallout
- fix up small PowerPC comment fallout (use 'counter' as in register)
Reviewed-by: Arjan van de Ven <arjan@linux.intel.com>
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
LKML-Reference: <new-submission>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r-- | MAINTAINERS | 2 | ||||
-rw-r--r-- | arch/powerpc/include/asm/paca.h | 2 | ||||
-rw-r--r-- | arch/powerpc/kernel/perf_event.c | 12 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/perf_event.c | 14 | ||||
-rw-r--r-- | include/linux/perf_counter.h | 441 | ||||
-rw-r--r-- | include/linux/perf_event.h | 98 | ||||
-rw-r--r-- | init/Kconfig | 37 | ||||
-rw-r--r-- | kernel/perf_event.c | 4 |
8 files changed, 534 insertions, 76 deletions
diff --git a/MAINTAINERS b/MAINTAINERS index 43761a00e3f..751a307dc44 100644 --- a/MAINTAINERS +++ b/MAINTAINERS | |||
@@ -4000,7 +4000,7 @@ S: Maintained | |||
4000 | F: include/linux/delayacct.h | 4000 | F: include/linux/delayacct.h |
4001 | F: kernel/delayacct.c | 4001 | F: kernel/delayacct.c |
4002 | 4002 | ||
4003 | PERFORMANCE COUNTER SUBSYSTEM | 4003 | PERFORMANCE EVENTS SUBSYSTEM |
4004 | M: Peter Zijlstra <a.p.zijlstra@chello.nl> | 4004 | M: Peter Zijlstra <a.p.zijlstra@chello.nl> |
4005 | M: Paul Mackerras <paulus@samba.org> | 4005 | M: Paul Mackerras <paulus@samba.org> |
4006 | M: Ingo Molnar <mingo@elte.hu> | 4006 | M: Ingo Molnar <mingo@elte.hu> |
diff --git a/arch/powerpc/include/asm/paca.h b/arch/powerpc/include/asm/paca.h index 154f405b642..7d8514cecea 100644 --- a/arch/powerpc/include/asm/paca.h +++ b/arch/powerpc/include/asm/paca.h | |||
@@ -122,7 +122,7 @@ struct paca_struct { | |||
122 | u8 soft_enabled; /* irq soft-enable flag */ | 122 | u8 soft_enabled; /* irq soft-enable flag */ |
123 | u8 hard_enabled; /* set if irqs are enabled in MSR */ | 123 | u8 hard_enabled; /* set if irqs are enabled in MSR */ |
124 | u8 io_sync; /* writel() needs spin_unlock sync */ | 124 | u8 io_sync; /* writel() needs spin_unlock sync */ |
125 | u8 perf_event_pending; /* PM interrupt while soft-disabled */ | 125 | u8 perf_event_pending; /* PM interrupt while soft-disabled */ |
126 | 126 | ||
127 | /* Stuff for accurate time accounting */ | 127 | /* Stuff for accurate time accounting */ |
128 | u64 user_time; /* accumulated usermode TB ticks */ | 128 | u64 user_time; /* accumulated usermode TB ticks */ |
diff --git a/arch/powerpc/kernel/perf_event.c b/arch/powerpc/kernel/perf_event.c index c98321fcb45..197b7d95879 100644 --- a/arch/powerpc/kernel/perf_event.c +++ b/arch/powerpc/kernel/perf_event.c | |||
@@ -41,7 +41,7 @@ DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); | |||
41 | struct power_pmu *ppmu; | 41 | struct power_pmu *ppmu; |
42 | 42 | ||
43 | /* | 43 | /* |
44 | * Normally, to ignore kernel events we set the FCS (freeze events | 44 | * Normally, to ignore kernel events we set the FCS (freeze counters |
45 | * in supervisor mode) bit in MMCR0, but if the kernel runs with the | 45 | * in supervisor mode) bit in MMCR0, but if the kernel runs with the |
46 | * hypervisor bit set in the MSR, or if we are running on a processor | 46 | * hypervisor bit set in the MSR, or if we are running on a processor |
47 | * where the hypervisor bit is forced to 1 (as on Apple G5 processors), | 47 | * where the hypervisor bit is forced to 1 (as on Apple G5 processors), |
@@ -159,7 +159,7 @@ void perf_event_print_debug(void) | |||
159 | } | 159 | } |
160 | 160 | ||
161 | /* | 161 | /* |
162 | * Read one performance monitor event (PMC). | 162 | * Read one performance monitor counter (PMC). |
163 | */ | 163 | */ |
164 | static unsigned long read_pmc(int idx) | 164 | static unsigned long read_pmc(int idx) |
165 | { | 165 | { |
@@ -409,7 +409,7 @@ static void power_pmu_read(struct perf_event *event) | |||
409 | val = read_pmc(event->hw.idx); | 409 | val = read_pmc(event->hw.idx); |
410 | } while (atomic64_cmpxchg(&event->hw.prev_count, prev, val) != prev); | 410 | } while (atomic64_cmpxchg(&event->hw.prev_count, prev, val) != prev); |
411 | 411 | ||
412 | /* The events are only 32 bits wide */ | 412 | /* The counters are only 32 bits wide */ |
413 | delta = (val - prev) & 0xfffffffful; | 413 | delta = (val - prev) & 0xfffffffful; |
414 | atomic64_add(delta, &event->count); | 414 | atomic64_add(delta, &event->count); |
415 | atomic64_sub(delta, &event->hw.period_left); | 415 | atomic64_sub(delta, &event->hw.period_left); |
@@ -543,7 +543,7 @@ void hw_perf_disable(void) | |||
543 | } | 543 | } |
544 | 544 | ||
545 | /* | 545 | /* |
546 | * Set the 'freeze events' bit. | 546 | * Set the 'freeze counters' bit. |
547 | * The barrier is to make sure the mtspr has been | 547 | * The barrier is to make sure the mtspr has been |
548 | * executed and the PMU has frozen the events | 548 | * executed and the PMU has frozen the events |
549 | * before we return. | 549 | * before we return. |
@@ -1124,7 +1124,7 @@ const struct pmu *hw_perf_event_init(struct perf_event *event) | |||
1124 | } | 1124 | } |
1125 | 1125 | ||
1126 | /* | 1126 | /* |
1127 | * A event has overflowed; update its count and record | 1127 | * A counter has overflowed; update its count and record |
1128 | * things if requested. Note that interrupts are hard-disabled | 1128 | * things if requested. Note that interrupts are hard-disabled |
1129 | * here so there is no possibility of being interrupted. | 1129 | * here so there is no possibility of being interrupted. |
1130 | */ | 1130 | */ |
@@ -1271,7 +1271,7 @@ static void perf_event_interrupt(struct pt_regs *regs) | |||
1271 | 1271 | ||
1272 | /* | 1272 | /* |
1273 | * Reset MMCR0 to its normal value. This will set PMXE and | 1273 | * Reset MMCR0 to its normal value. This will set PMXE and |
1274 | * clear FC (freeze events) and PMAO (perf mon alert occurred) | 1274 | * clear FC (freeze counters) and PMAO (perf mon alert occurred) |
1275 | * and thus allow interrupts to occur again. | 1275 | * and thus allow interrupts to occur again. |
1276 | * XXX might want to use MSR.PM to keep the events frozen until | 1276 | * XXX might want to use MSR.PM to keep the events frozen until |
1277 | * we get back out of this interrupt. | 1277 | * we get back out of this interrupt. |
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c index 0d03629fb1a..a3c7adb06b7 100644 --- a/arch/x86/kernel/cpu/perf_event.c +++ b/arch/x86/kernel/cpu/perf_event.c | |||
@@ -2081,13 +2081,13 @@ void __init init_hw_perf_events(void) | |||
2081 | perf_events_lapic_init(); | 2081 | perf_events_lapic_init(); |
2082 | register_die_notifier(&perf_event_nmi_notifier); | 2082 | register_die_notifier(&perf_event_nmi_notifier); |
2083 | 2083 | ||
2084 | pr_info("... version: %d\n", x86_pmu.version); | 2084 | pr_info("... version: %d\n", x86_pmu.version); |
2085 | pr_info("... bit width: %d\n", x86_pmu.event_bits); | 2085 | pr_info("... bit width: %d\n", x86_pmu.event_bits); |
2086 | pr_info("... generic events: %d\n", x86_pmu.num_events); | 2086 | pr_info("... generic registers: %d\n", x86_pmu.num_events); |
2087 | pr_info("... value mask: %016Lx\n", x86_pmu.event_mask); | 2087 | pr_info("... value mask: %016Lx\n", x86_pmu.event_mask); |
2088 | pr_info("... max period: %016Lx\n", x86_pmu.max_period); | 2088 | pr_info("... max period: %016Lx\n", x86_pmu.max_period); |
2089 | pr_info("... fixed-purpose events: %d\n", x86_pmu.num_events_fixed); | 2089 | pr_info("... fixed-purpose events: %d\n", x86_pmu.num_events_fixed); |
2090 | pr_info("... event mask: %016Lx\n", perf_event_mask); | 2090 | pr_info("... event mask: %016Lx\n", perf_event_mask); |
2091 | } | 2091 | } |
2092 | 2092 | ||
2093 | static inline void x86_pmu_read(struct perf_event *event) | 2093 | static inline void x86_pmu_read(struct perf_event *event) |
diff --git a/include/linux/perf_counter.h b/include/linux/perf_counter.h new file mode 100644 index 00000000000..368bd70f1d2 --- /dev/null +++ b/include/linux/perf_counter.h | |||
@@ -0,0 +1,441 @@ | |||
1 | /* | ||
2 | * NOTE: this file will be removed in a future kernel release, it is | ||
3 | * provided as a courtesy copy of user-space code that relies on the | ||
4 | * old (pre-rename) symbols and constants. | ||
5 | * | ||
6 | * Performance events: | ||
7 | * | ||
8 | * Copyright (C) 2008-2009, Thomas Gleixner <tglx@linutronix.de> | ||
9 | * Copyright (C) 2008-2009, Red Hat, Inc., Ingo Molnar | ||
10 | * Copyright (C) 2008-2009, Red Hat, Inc., Peter Zijlstra | ||
11 | * | ||
12 | * Data type definitions, declarations, prototypes. | ||
13 | * | ||
14 | * Started by: Thomas Gleixner and Ingo Molnar | ||
15 | * | ||
16 | * For licencing details see kernel-base/COPYING | ||
17 | */ | ||
18 | #ifndef _LINUX_PERF_COUNTER_H | ||
19 | #define _LINUX_PERF_COUNTER_H | ||
20 | |||
21 | #include <linux/types.h> | ||
22 | #include <linux/ioctl.h> | ||
23 | #include <asm/byteorder.h> | ||
24 | |||
25 | /* | ||
26 | * User-space ABI bits: | ||
27 | */ | ||
28 | |||
29 | /* | ||
30 | * attr.type | ||
31 | */ | ||
32 | enum perf_type_id { | ||
33 | PERF_TYPE_HARDWARE = 0, | ||
34 | PERF_TYPE_SOFTWARE = 1, | ||
35 | PERF_TYPE_TRACEPOINT = 2, | ||
36 | PERF_TYPE_HW_CACHE = 3, | ||
37 | PERF_TYPE_RAW = 4, | ||
38 | |||
39 | PERF_TYPE_MAX, /* non-ABI */ | ||
40 | }; | ||
41 | |||
42 | /* | ||
43 | * Generalized performance counter event types, used by the | ||
44 | * attr.event_id parameter of the sys_perf_counter_open() | ||
45 | * syscall: | ||
46 | */ | ||
47 | enum perf_hw_id { | ||
48 | /* | ||
49 | * Common hardware events, generalized by the kernel: | ||
50 | */ | ||
51 | PERF_COUNT_HW_CPU_CYCLES = 0, | ||
52 | PERF_COUNT_HW_INSTRUCTIONS = 1, | ||
53 | PERF_COUNT_HW_CACHE_REFERENCES = 2, | ||
54 | PERF_COUNT_HW_CACHE_MISSES = 3, | ||
55 | PERF_COUNT_HW_BRANCH_INSTRUCTIONS = 4, | ||
56 | PERF_COUNT_HW_BRANCH_MISSES = 5, | ||
57 | PERF_COUNT_HW_BUS_CYCLES = 6, | ||
58 | |||
59 | PERF_COUNT_HW_MAX, /* non-ABI */ | ||
60 | }; | ||
61 | |||
62 | /* | ||
63 | * Generalized hardware cache counters: | ||
64 | * | ||
65 | * { L1-D, L1-I, LLC, ITLB, DTLB, BPU } x | ||
66 | * { read, write, prefetch } x | ||
67 | * { accesses, misses } | ||
68 | */ | ||
69 | enum perf_hw_cache_id { | ||
70 | PERF_COUNT_HW_CACHE_L1D = 0, | ||
71 | PERF_COUNT_HW_CACHE_L1I = 1, | ||
72 | PERF_COUNT_HW_CACHE_LL = 2, | ||
73 | PERF_COUNT_HW_CACHE_DTLB = 3, | ||
74 | PERF_COUNT_HW_CACHE_ITLB = 4, | ||
75 | PERF_COUNT_HW_CACHE_BPU = 5, | ||
76 | |||
77 | PERF_COUNT_HW_CACHE_MAX, /* non-ABI */ | ||
78 | }; | ||
79 | |||
80 | enum perf_hw_cache_op_id { | ||
81 | PERF_COUNT_HW_CACHE_OP_READ = 0, | ||
82 | PERF_COUNT_HW_CACHE_OP_WRITE = 1, | ||
83 | PERF_COUNT_HW_CACHE_OP_PREFETCH = 2, | ||
84 | |||
85 | PERF_COUNT_HW_CACHE_OP_MAX, /* non-ABI */ | ||
86 | }; | ||
87 | |||
88 | enum perf_hw_cache_op_result_id { | ||
89 | PERF_COUNT_HW_CACHE_RESULT_ACCESS = 0, | ||
90 | PERF_COUNT_HW_CACHE_RESULT_MISS = 1, | ||
91 | |||
92 | PERF_COUNT_HW_CACHE_RESULT_MAX, /* non-ABI */ | ||
93 | }; | ||
94 | |||
95 | /* | ||
96 | * Special "software" counters provided by the kernel, even if the hardware | ||
97 | * does not support performance counters. These counters measure various | ||
98 | * physical and sw events of the kernel (and allow the profiling of them as | ||
99 | * well): | ||
100 | */ | ||
101 | enum perf_sw_ids { | ||
102 | PERF_COUNT_SW_CPU_CLOCK = 0, | ||
103 | PERF_COUNT_SW_TASK_CLOCK = 1, | ||
104 | PERF_COUNT_SW_PAGE_FAULTS = 2, | ||
105 | PERF_COUNT_SW_CONTEXT_SWITCHES = 3, | ||
106 | PERF_COUNT_SW_CPU_MIGRATIONS = 4, | ||
107 | PERF_COUNT_SW_PAGE_FAULTS_MIN = 5, | ||
108 | PERF_COUNT_SW_PAGE_FAULTS_MAJ = 6, | ||
109 | |||
110 | PERF_COUNT_SW_MAX, /* non-ABI */ | ||
111 | }; | ||
112 | |||
113 | /* | ||
114 | * Bits that can be set in attr.sample_type to request information | ||
115 | * in the overflow packets. | ||
116 | */ | ||
117 | enum perf_counter_sample_format { | ||
118 | PERF_SAMPLE_IP = 1U << 0, | ||
119 | PERF_SAMPLE_TID = 1U << 1, | ||
120 | PERF_SAMPLE_TIME = 1U << 2, | ||
121 | PERF_SAMPLE_ADDR = 1U << 3, | ||
122 | PERF_SAMPLE_READ = 1U << 4, | ||
123 | PERF_SAMPLE_CALLCHAIN = 1U << 5, | ||
124 | PERF_SAMPLE_ID = 1U << 6, | ||
125 | PERF_SAMPLE_CPU = 1U << 7, | ||
126 | PERF_SAMPLE_PERIOD = 1U << 8, | ||
127 | PERF_SAMPLE_STREAM_ID = 1U << 9, | ||
128 | PERF_SAMPLE_RAW = 1U << 10, | ||
129 | |||
130 | PERF_SAMPLE_MAX = 1U << 11, /* non-ABI */ | ||
131 | }; | ||
132 | |||
133 | /* | ||
134 | * The format of the data returned by read() on a perf counter fd, | ||
135 | * as specified by attr.read_format: | ||
136 | * | ||
137 | * struct read_format { | ||
138 | * { u64 value; | ||
139 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED | ||
140 | * { u64 time_running; } && PERF_FORMAT_RUNNING | ||
141 | * { u64 id; } && PERF_FORMAT_ID | ||
142 | * } && !PERF_FORMAT_GROUP | ||
143 | * | ||
144 | * { u64 nr; | ||
145 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED | ||
146 | * { u64 time_running; } && PERF_FORMAT_RUNNING | ||
147 | * { u64 value; | ||
148 | * { u64 id; } && PERF_FORMAT_ID | ||
149 | * } cntr[nr]; | ||
150 | * } && PERF_FORMAT_GROUP | ||
151 | * }; | ||
152 | */ | ||
153 | enum perf_counter_read_format { | ||
154 | PERF_FORMAT_TOTAL_TIME_ENABLED = 1U << 0, | ||
155 | PERF_FORMAT_TOTAL_TIME_RUNNING = 1U << 1, | ||
156 | PERF_FORMAT_ID = 1U << 2, | ||
157 | PERF_FORMAT_GROUP = 1U << 3, | ||
158 | |||
159 | PERF_FORMAT_MAX = 1U << 4, /* non-ABI */ | ||
160 | }; | ||
161 | |||
162 | #define PERF_ATTR_SIZE_VER0 64 /* sizeof first published struct */ | ||
163 | |||
164 | /* | ||
165 | * Hardware event to monitor via a performance monitoring counter: | ||
166 | */ | ||
167 | struct perf_counter_attr { | ||
168 | |||
169 | /* | ||
170 | * Major type: hardware/software/tracepoint/etc. | ||
171 | */ | ||
172 | __u32 type; | ||
173 | |||
174 | /* | ||
175 | * Size of the attr structure, for fwd/bwd compat. | ||
176 | */ | ||
177 | __u32 size; | ||
178 | |||
179 | /* | ||
180 | * Type specific configuration information. | ||
181 | */ | ||
182 | __u64 config; | ||
183 | |||
184 | union { | ||
185 | __u64 sample_period; | ||
186 | __u64 sample_freq; | ||
187 | }; | ||
188 | |||
189 | __u64 sample_type; | ||
190 | __u64 read_format; | ||
191 | |||
192 | __u64 disabled : 1, /* off by default */ | ||
193 | inherit : 1, /* children inherit it */ | ||
194 | pinned : 1, /* must always be on PMU */ | ||
195 | exclusive : 1, /* only group on PMU */ | ||
196 | exclude_user : 1, /* don't count user */ | ||
197 | exclude_kernel : 1, /* ditto kernel */ | ||
198 | exclude_hv : 1, /* ditto hypervisor */ | ||
199 | exclude_idle : 1, /* don't count when idle */ | ||
200 | mmap : 1, /* include mmap data */ | ||
201 | comm : 1, /* include comm data */ | ||
202 | freq : 1, /* use freq, not period */ | ||
203 | inherit_stat : 1, /* per task counts */ | ||
204 | enable_on_exec : 1, /* next exec enables */ | ||
205 | task : 1, /* trace fork/exit */ | ||
206 | watermark : 1, /* wakeup_watermark */ | ||
207 | |||
208 | __reserved_1 : 49; | ||
209 | |||
210 | union { | ||
211 | __u32 wakeup_events; /* wakeup every n events */ | ||
212 | __u32 wakeup_watermark; /* bytes before wakeup */ | ||
213 | }; | ||
214 | __u32 __reserved_2; | ||
215 | |||
216 | __u64 __reserved_3; | ||
217 | }; | ||
218 | |||
219 | /* | ||
220 | * Ioctls that can be done on a perf counter fd: | ||
221 | */ | ||
222 | #define PERF_COUNTER_IOC_ENABLE _IO ('$', 0) | ||
223 | #define PERF_COUNTER_IOC_DISABLE _IO ('$', 1) | ||
224 | #define PERF_COUNTER_IOC_REFRESH _IO ('$', 2) | ||
225 | #define PERF_COUNTER_IOC_RESET _IO ('$', 3) | ||
226 | #define PERF_COUNTER_IOC_PERIOD _IOW('$', 4, u64) | ||
227 | #define PERF_COUNTER_IOC_SET_OUTPUT _IO ('$', 5) | ||
228 | |||
229 | enum perf_counter_ioc_flags { | ||
230 | PERF_IOC_FLAG_GROUP = 1U << 0, | ||
231 | }; | ||
232 | |||
233 | /* | ||
234 | * Structure of the page that can be mapped via mmap | ||
235 | */ | ||
236 | struct perf_counter_mmap_page { | ||
237 | __u32 version; /* version number of this structure */ | ||
238 | __u32 compat_version; /* lowest version this is compat with */ | ||
239 | |||
240 | /* | ||
241 | * Bits needed to read the hw counters in user-space. | ||
242 | * | ||
243 | * u32 seq; | ||
244 | * s64 count; | ||
245 | * | ||
246 | * do { | ||
247 | * seq = pc->lock; | ||
248 | * | ||
249 | * barrier() | ||
250 | * if (pc->index) { | ||
251 | * count = pmc_read(pc->index - 1); | ||
252 | * count += pc->offset; | ||
253 | * } else | ||
254 | * goto regular_read; | ||
255 | * | ||
256 | * barrier(); | ||
257 | * } while (pc->lock != seq); | ||
258 | * | ||
259 | * NOTE: for obvious reason this only works on self-monitoring | ||
260 | * processes. | ||
261 | */ | ||
262 | __u32 lock; /* seqlock for synchronization */ | ||
263 | __u32 index; /* hardware counter identifier */ | ||
264 | __s64 offset; /* add to hardware counter value */ | ||
265 | __u64 time_enabled; /* time counter active */ | ||
266 | __u64 time_running; /* time counter on cpu */ | ||
267 | |||
268 | /* | ||
269 | * Hole for extension of the self monitor capabilities | ||
270 | */ | ||
271 | |||
272 | __u64 __reserved[123]; /* align to 1k */ | ||
273 | |||
274 | /* | ||
275 | * Control data for the mmap() data buffer. | ||
276 | * | ||
277 | * User-space reading the @data_head value should issue an rmb(), on | ||
278 | * SMP capable platforms, after reading this value -- see | ||
279 | * perf_counter_wakeup(). | ||
280 | * | ||
281 | * When the mapping is PROT_WRITE the @data_tail value should be | ||
282 | * written by userspace to reflect the last read data. In this case | ||
283 | * the kernel will not over-write unread data. | ||
284 | */ | ||
285 | __u64 data_head; /* head in the data section */ | ||
286 | __u64 data_tail; /* user-space written tail */ | ||
287 | }; | ||
288 | |||
289 | #define PERF_EVENT_MISC_CPUMODE_MASK (3 << 0) | ||
290 | #define PERF_EVENT_MISC_CPUMODE_UNKNOWN (0 << 0) | ||
291 | #define PERF_EVENT_MISC_KERNEL (1 << 0) | ||
292 | #define PERF_EVENT_MISC_USER (2 << 0) | ||
293 | #define PERF_EVENT_MISC_HYPERVISOR (3 << 0) | ||
294 | |||
295 | struct perf_event_header { | ||
296 | __u32 type; | ||
297 | __u16 misc; | ||
298 | __u16 size; | ||
299 | }; | ||
300 | |||
301 | enum perf_event_type { | ||
302 | |||
303 | /* | ||
304 | * The MMAP events record the PROT_EXEC mappings so that we can | ||
305 | * correlate userspace IPs to code. They have the following structure: | ||
306 | * | ||
307 | * struct { | ||
308 | * struct perf_event_header header; | ||
309 | * | ||
310 | * u32 pid, tid; | ||
311 | * u64 addr; | ||
312 | * u64 len; | ||
313 | * u64 pgoff; | ||
314 | * char filename[]; | ||
315 | * }; | ||
316 | */ | ||
317 | PERF_EVENT_MMAP = 1, | ||
318 | |||
319 | /* | ||
320 | * struct { | ||
321 | * struct perf_event_header header; | ||
322 | * u64 id; | ||
323 | * u64 lost; | ||
324 | * }; | ||
325 | */ | ||
326 | PERF_EVENT_LOST = 2, | ||
327 | |||
328 | /* | ||
329 | * struct { | ||
330 | * struct perf_event_header header; | ||
331 | * | ||
332 | * u32 pid, tid; | ||
333 | * char comm[]; | ||
334 | * }; | ||
335 | */ | ||
336 | PERF_EVENT_COMM = 3, | ||
337 | |||
338 | /* | ||
339 | * struct { | ||
340 | * struct perf_event_header header; | ||
341 | * u32 pid, ppid; | ||
342 | * u32 tid, ptid; | ||
343 | * u64 time; | ||
344 | * }; | ||
345 | */ | ||
346 | PERF_EVENT_EXIT = 4, | ||
347 | |||
348 | /* | ||
349 | * struct { | ||
350 | * struct perf_event_header header; | ||
351 | * u64 time; | ||
352 | * u64 id; | ||
353 | * u64 stream_id; | ||
354 | * }; | ||
355 | */ | ||
356 | PERF_EVENT_THROTTLE = 5, | ||
357 | PERF_EVENT_UNTHROTTLE = 6, | ||
358 | |||
359 | /* | ||
360 | * struct { | ||
361 | * struct perf_event_header header; | ||
362 | * u32 pid, ppid; | ||
363 | * u32 tid, ptid; | ||
364 | * { u64 time; } && PERF_SAMPLE_TIME | ||
365 | * }; | ||
366 | */ | ||
367 | PERF_EVENT_FORK = 7, | ||
368 | |||
369 | /* | ||
370 | * struct { | ||
371 | * struct perf_event_header header; | ||
372 | * u32 pid, tid; | ||
373 | * | ||
374 | * struct read_format values; | ||
375 | * }; | ||
376 | */ | ||
377 | PERF_EVENT_READ = 8, | ||
378 | |||
379 | /* | ||
380 | * struct { | ||
381 | * struct perf_event_header header; | ||
382 | * | ||
383 | * { u64 ip; } && PERF_SAMPLE_IP | ||
384 | * { u32 pid, tid; } && PERF_SAMPLE_TID | ||
385 | * { u64 time; } && PERF_SAMPLE_TIME | ||
386 | * { u64 addr; } && PERF_SAMPLE_ADDR | ||
387 | * { u64 id; } && PERF_SAMPLE_ID | ||
388 | * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID | ||
389 | * { u32 cpu, res; } && PERF_SAMPLE_CPU | ||
390 | * { u64 period; } && PERF_SAMPLE_PERIOD | ||
391 | * | ||
392 | * { struct read_format values; } && PERF_SAMPLE_READ | ||
393 | * | ||
394 | * { u64 nr, | ||
395 | * u64 ips[nr]; } && PERF_SAMPLE_CALLCHAIN | ||
396 | * | ||
397 | * # | ||
398 | * # The RAW record below is opaque data wrt the ABI | ||
399 | * # | ||
400 | * # That is, the ABI doesn't make any promises wrt to | ||
401 | * # the stability of its content, it may vary depending | ||
402 | * # on event, hardware, kernel version and phase of | ||
403 | * # the moon. | ||
404 | * # | ||
405 | * # In other words, PERF_SAMPLE_RAW contents are not an ABI. | ||
406 | * # | ||
407 | * | ||
408 | * { u32 size; | ||
409 | * char data[size];}&& PERF_SAMPLE_RAW | ||
410 | * }; | ||
411 | */ | ||
412 | PERF_EVENT_SAMPLE = 9, | ||
413 | |||
414 | PERF_EVENT_MAX, /* non-ABI */ | ||
415 | }; | ||
416 | |||
417 | enum perf_callchain_context { | ||
418 | PERF_CONTEXT_HV = (__u64)-32, | ||
419 | PERF_CONTEXT_KERNEL = (__u64)-128, | ||
420 | PERF_CONTEXT_USER = (__u64)-512, | ||
421 | |||
422 | PERF_CONTEXT_GUEST = (__u64)-2048, | ||
423 | PERF_CONTEXT_GUEST_KERNEL = (__u64)-2176, | ||
424 | PERF_CONTEXT_GUEST_USER = (__u64)-2560, | ||
425 | |||
426 | PERF_CONTEXT_MAX = (__u64)-4095, | ||
427 | }; | ||
428 | |||
429 | #define PERF_FLAG_FD_NO_GROUP (1U << 0) | ||
430 | #define PERF_FLAG_FD_OUTPUT (1U << 1) | ||
431 | |||
432 | /* | ||
433 | * In case some app still references the old symbols: | ||
434 | */ | ||
435 | |||
436 | #define __NR_perf_counter_open __NR_perf_event_open | ||
437 | |||
438 | #define PR_TASK_PERF_COUNTERS_DISABLE PR_TASK_PERF_EVENTS_DISABLE | ||
439 | #define PR_TASK_PERF_COUNTERS_ENABLE PR_TASK_PERF_EVENTS_ENABLE | ||
440 | |||
441 | #endif /* _LINUX_PERF_COUNTER_H */ | ||
diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h index ae9d9ed6df2..acefaf71e6d 100644 --- a/include/linux/perf_event.h +++ b/include/linux/perf_event.h | |||
@@ -1,15 +1,15 @@ | |||
1 | /* | 1 | /* |
2 | * Performance events: | 2 | * Performance events: |
3 | * | 3 | * |
4 | * Copyright (C) 2008-2009, Thomas Gleixner <tglx@linutronix.de> | 4 | * Copyright (C) 2008-2009, Thomas Gleixner <tglx@linutronix.de> |
5 | * Copyright (C) 2008-2009, Red Hat, Inc., Ingo Molnar | 5 | * Copyright (C) 2008-2009, Red Hat, Inc., Ingo Molnar |
6 | * Copyright (C) 2008-2009, Red Hat, Inc., Peter Zijlstra | 6 | * Copyright (C) 2008-2009, Red Hat, Inc., Peter Zijlstra |
7 | * | 7 | * |
8 | * Data type definitions, declarations, prototypes. | 8 | * Data type definitions, declarations, prototypes. |
9 | * | 9 | * |
10 | * Started by: Thomas Gleixner and Ingo Molnar | 10 | * Started by: Thomas Gleixner and Ingo Molnar |
11 | * | 11 | * |
12 | * For licencing details see kernel-base/COPYING | 12 | * For licencing details see kernel-base/COPYING |
13 | */ | 13 | */ |
14 | #ifndef _LINUX_PERF_EVENT_H | 14 | #ifndef _LINUX_PERF_EVENT_H |
15 | #define _LINUX_PERF_EVENT_H | 15 | #define _LINUX_PERF_EVENT_H |
@@ -131,19 +131,19 @@ enum perf_event_sample_format { | |||
131 | * as specified by attr.read_format: | 131 | * as specified by attr.read_format: |
132 | * | 132 | * |
133 | * struct read_format { | 133 | * struct read_format { |
134 | * { u64 value; | 134 | * { u64 value; |
135 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED | 135 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED |
136 | * { u64 time_running; } && PERF_FORMAT_RUNNING | 136 | * { u64 time_running; } && PERF_FORMAT_RUNNING |
137 | * { u64 id; } && PERF_FORMAT_ID | 137 | * { u64 id; } && PERF_FORMAT_ID |
138 | * } && !PERF_FORMAT_GROUP | 138 | * } && !PERF_FORMAT_GROUP |
139 | * | 139 | * |
140 | * { u64 nr; | 140 | * { u64 nr; |
141 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED | 141 | * { u64 time_enabled; } && PERF_FORMAT_ENABLED |
142 | * { u64 time_running; } && PERF_FORMAT_RUNNING | 142 | * { u64 time_running; } && PERF_FORMAT_RUNNING |
143 | * { u64 value; | 143 | * { u64 value; |
144 | * { u64 id; } && PERF_FORMAT_ID | 144 | * { u64 id; } && PERF_FORMAT_ID |
145 | * } cntr[nr]; | 145 | * } cntr[nr]; |
146 | * } && PERF_FORMAT_GROUP | 146 | * } && PERF_FORMAT_GROUP |
147 | * }; | 147 | * }; |
148 | */ | 148 | */ |
149 | enum perf_event_read_format { | 149 | enum perf_event_read_format { |
@@ -152,7 +152,7 @@ enum perf_event_read_format { | |||
152 | PERF_FORMAT_ID = 1U << 2, | 152 | PERF_FORMAT_ID = 1U << 2, |
153 | PERF_FORMAT_GROUP = 1U << 3, | 153 | PERF_FORMAT_GROUP = 1U << 3, |
154 | 154 | ||
155 | PERF_FORMAT_MAX = 1U << 4, /* non-ABI */ | 155 | PERF_FORMAT_MAX = 1U << 4, /* non-ABI */ |
156 | }; | 156 | }; |
157 | 157 | ||
158 | #define PERF_ATTR_SIZE_VER0 64 /* sizeof first published struct */ | 158 | #define PERF_ATTR_SIZE_VER0 64 /* sizeof first published struct */ |
@@ -216,8 +216,8 @@ struct perf_event_attr { | |||
216 | * Ioctls that can be done on a perf event fd: | 216 | * Ioctls that can be done on a perf event fd: |
217 | */ | 217 | */ |
218 | #define PERF_EVENT_IOC_ENABLE _IO ('$', 0) | 218 | #define PERF_EVENT_IOC_ENABLE _IO ('$', 0) |
219 | #define PERF_EVENT_IOC_DISABLE _IO ('$', 1) | 219 | #define PERF_EVENT_IOC_DISABLE _IO ('$', 1) |
220 | #define PERF_EVENT_IOC_REFRESH _IO ('$', 2) | 220 | #define PERF_EVENT_IOC_REFRESH _IO ('$', 2) |
221 | #define PERF_EVENT_IOC_RESET _IO ('$', 3) | 221 | #define PERF_EVENT_IOC_RESET _IO ('$', 3) |
222 | #define PERF_EVENT_IOC_PERIOD _IOW('$', 4, u64) | 222 | #define PERF_EVENT_IOC_PERIOD _IOW('$', 4, u64) |
223 | #define PERF_EVENT_IOC_SET_OUTPUT _IO ('$', 5) | 223 | #define PERF_EVENT_IOC_SET_OUTPUT _IO ('$', 5) |
@@ -314,9 +314,9 @@ enum perf_event_type { | |||
314 | 314 | ||
315 | /* | 315 | /* |
316 | * struct { | 316 | * struct { |
317 | * struct perf_event_header header; | 317 | * struct perf_event_header header; |
318 | * u64 id; | 318 | * u64 id; |
319 | * u64 lost; | 319 | * u64 lost; |
320 | * }; | 320 | * }; |
321 | */ | 321 | */ |
322 | PERF_RECORD_LOST = 2, | 322 | PERF_RECORD_LOST = 2, |
@@ -383,23 +383,23 @@ enum perf_event_type { | |||
383 | * { u64 id; } && PERF_SAMPLE_ID | 383 | * { u64 id; } && PERF_SAMPLE_ID |
384 | * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID | 384 | * { u64 stream_id;} && PERF_SAMPLE_STREAM_ID |
385 | * { u32 cpu, res; } && PERF_SAMPLE_CPU | 385 | * { u32 cpu, res; } && PERF_SAMPLE_CPU |
386 | * { u64 period; } && PERF_SAMPLE_PERIOD | 386 | * { u64 period; } && PERF_SAMPLE_PERIOD |
387 | * | 387 | * |
388 | * { struct read_format values; } && PERF_SAMPLE_READ | 388 | * { struct read_format values; } && PERF_SAMPLE_READ |
389 | * | 389 | * |
390 | * { u64 nr, | 390 | * { u64 nr, |
391 | * u64 ips[nr]; } && PERF_SAMPLE_CALLCHAIN | 391 | * u64 ips[nr]; } && PERF_SAMPLE_CALLCHAIN |
392 | * | 392 | * |
393 | * # | 393 | * # |
394 | * # The RAW record below is opaque data wrt the ABI | 394 | * # The RAW record below is opaque data wrt the ABI |
395 | * # | 395 | * # |
396 | * # That is, the ABI doesn't make any promises wrt to | 396 | * # That is, the ABI doesn't make any promises wrt to |
397 | * # the stability of its content, it may vary depending | 397 | * # the stability of its content, it may vary depending |
398 | * # on event_id, hardware, kernel version and phase of | 398 | * # on event, hardware, kernel version and phase of |
399 | * # the moon. | 399 | * # the moon. |
400 | * # | 400 | * # |
401 | * # In other words, PERF_SAMPLE_RAW contents are not an ABI. | 401 | * # In other words, PERF_SAMPLE_RAW contents are not an ABI. |
402 | * # | 402 | * # |
403 | * | 403 | * |
404 | * { u32 size; | 404 | * { u32 size; |
405 | * char data[size];}&& PERF_SAMPLE_RAW | 405 | * char data[size];}&& PERF_SAMPLE_RAW |
@@ -503,10 +503,10 @@ struct pmu { | |||
503 | * enum perf_event_active_state - the states of a event | 503 | * enum perf_event_active_state - the states of a event |
504 | */ | 504 | */ |
505 | enum perf_event_active_state { | 505 | enum perf_event_active_state { |
506 | PERF_EVENT_STATE_ERROR = -2, | 506 | PERF_EVENT_STATE_ERROR = -2, |
507 | PERF_EVENT_STATE_OFF = -1, | 507 | PERF_EVENT_STATE_OFF = -1, |
508 | PERF_EVENT_STATE_INACTIVE = 0, | 508 | PERF_EVENT_STATE_INACTIVE = 0, |
509 | PERF_EVENT_STATE_ACTIVE = 1, | 509 | PERF_EVENT_STATE_ACTIVE = 1, |
510 | }; | 510 | }; |
511 | 511 | ||
512 | struct file; | 512 | struct file; |
@@ -529,7 +529,7 @@ struct perf_mmap_data { | |||
529 | 529 | ||
530 | long watermark; /* wakeup watermark */ | 530 | long watermark; /* wakeup watermark */ |
531 | 531 | ||
532 | struct perf_event_mmap_page *user_page; | 532 | struct perf_event_mmap_page *user_page; |
533 | void *data_pages[0]; | 533 | void *data_pages[0]; |
534 | }; | 534 | }; |
535 | 535 | ||
@@ -694,14 +694,14 @@ struct perf_cpu_context { | |||
694 | }; | 694 | }; |
695 | 695 | ||
696 | struct perf_output_handle { | 696 | struct perf_output_handle { |
697 | struct perf_event *event; | 697 | struct perf_event *event; |
698 | struct perf_mmap_data *data; | 698 | struct perf_mmap_data *data; |
699 | unsigned long head; | 699 | unsigned long head; |
700 | unsigned long offset; | 700 | unsigned long offset; |
701 | int nmi; | 701 | int nmi; |
702 | int sample; | 702 | int sample; |
703 | int locked; | 703 | int locked; |
704 | unsigned long flags; | 704 | unsigned long flags; |
705 | }; | 705 | }; |
706 | 706 | ||
707 | #ifdef CONFIG_PERF_EVENTS | 707 | #ifdef CONFIG_PERF_EVENTS |
@@ -829,22 +829,22 @@ static inline void | |||
829 | perf_event_task_sched_out(struct task_struct *task, | 829 | perf_event_task_sched_out(struct task_struct *task, |
830 | struct task_struct *next, int cpu) { } | 830 | struct task_struct *next, int cpu) { } |
831 | static inline void | 831 | static inline void |
832 | perf_event_task_tick(struct task_struct *task, int cpu) { } | 832 | perf_event_task_tick(struct task_struct *task, int cpu) { } |
833 | static inline int perf_event_init_task(struct task_struct *child) { return 0; } | 833 | static inline int perf_event_init_task(struct task_struct *child) { return 0; } |
834 | static inline void perf_event_exit_task(struct task_struct *child) { } | 834 | static inline void perf_event_exit_task(struct task_struct *child) { } |
835 | static inline void perf_event_free_task(struct task_struct *task) { } | 835 | static inline void perf_event_free_task(struct task_struct *task) { } |
836 | static inline void perf_event_do_pending(void) { } | 836 | static inline void perf_event_do_pending(void) { } |
837 | static inline void perf_event_print_debug(void) { } | 837 | static inline void perf_event_print_debug(void) { } |
838 | static inline void perf_disable(void) { } | 838 | static inline void perf_disable(void) { } |
839 | static inline void perf_enable(void) { } | 839 | static inline void perf_enable(void) { } |
840 | static inline int perf_event_task_disable(void) { return -EINVAL; } | 840 | static inline int perf_event_task_disable(void) { return -EINVAL; } |
841 | static inline int perf_event_task_enable(void) { return -EINVAL; } | 841 | static inline int perf_event_task_enable(void) { return -EINVAL; } |
842 | 842 | ||
843 | static inline void | 843 | static inline void |
844 | perf_sw_event(u32 event_id, u64 nr, int nmi, | 844 | perf_sw_event(u32 event_id, u64 nr, int nmi, |
845 | struct pt_regs *regs, u64 addr) { } | 845 | struct pt_regs *regs, u64 addr) { } |
846 | 846 | ||
847 | static inline void perf_event_mmap(struct vm_area_struct *vma) { } | 847 | static inline void perf_event_mmap(struct vm_area_struct *vma) { } |
848 | static inline void perf_event_comm(struct task_struct *tsk) { } | 848 | static inline void perf_event_comm(struct task_struct *tsk) { } |
849 | static inline void perf_event_fork(struct task_struct *tsk) { } | 849 | static inline void perf_event_fork(struct task_struct *tsk) { } |
850 | static inline void perf_event_init(void) { } | 850 | static inline void perf_event_init(void) { } |
diff --git a/init/Kconfig b/init/Kconfig index cfdf5c32280..706728be312 100644 --- a/init/Kconfig +++ b/init/Kconfig | |||
@@ -920,26 +920,31 @@ config HAVE_PERF_EVENTS | |||
920 | help | 920 | help |
921 | See tools/perf/design.txt for details. | 921 | See tools/perf/design.txt for details. |
922 | 922 | ||
923 | menu "Performance Counters" | 923 | menu "Kernel Performance Events And Counters" |
924 | 924 | ||
925 | config PERF_EVENTS | 925 | config PERF_EVENTS |
926 | bool "Kernel Performance Counters" | 926 | bool "Kernel performance events and counters" |
927 | default y if PROFILING | 927 | default y if (PROFILING || PERF_COUNTERS) |
928 | depends on HAVE_PERF_EVENTS | 928 | depends on HAVE_PERF_EVENTS |
929 | select ANON_INODES | 929 | select ANON_INODES |
930 | help | 930 | help |
931 | Enable kernel support for performance counter hardware. | 931 | Enable kernel support for various performance events provided |
932 | by software and hardware. | ||
932 | 933 | ||
933 | Performance counters are special hardware registers available | 934 | Software events are supported either build-in or via the |
934 | on most modern CPUs. These registers count the number of certain | 935 | use of generic tracepoints. |
936 | |||
937 | Most modern CPUs support performance events via performance | ||
938 | counter registers. These registers count the number of certain | ||
935 | types of hw events: such as instructions executed, cachemisses | 939 | types of hw events: such as instructions executed, cachemisses |
936 | suffered, or branches mis-predicted - without slowing down the | 940 | suffered, or branches mis-predicted - without slowing down the |
937 | kernel or applications. These registers can also trigger interrupts | 941 | kernel or applications. These registers can also trigger interrupts |
938 | when a threshold number of events have passed - and can thus be | 942 | when a threshold number of events have passed - and can thus be |
939 | used to profile the code that runs on that CPU. | 943 | used to profile the code that runs on that CPU. |
940 | 944 | ||
941 | The Linux Performance Counter subsystem provides an abstraction of | 945 | The Linux Performance Event subsystem provides an abstraction of |
942 | these hardware capabilities, available via a system call. It | 946 | these software and hardware cevent apabilities, available via a |
947 | system call and used by the "perf" utility in tools/perf/. It | ||
943 | provides per task and per CPU counters, and it provides event | 948 | provides per task and per CPU counters, and it provides event |
944 | capabilities on top of those. | 949 | capabilities on top of those. |
945 | 950 | ||
@@ -950,14 +955,26 @@ config EVENT_PROFILE | |||
950 | depends on PERF_EVENTS && EVENT_TRACING | 955 | depends on PERF_EVENTS && EVENT_TRACING |
951 | default y | 956 | default y |
952 | help | 957 | help |
953 | Allow the use of tracepoints as software performance counters. | 958 | Allow the use of tracepoints as software performance events. |
954 | 959 | ||
955 | When this is enabled, you can create perf counters based on | 960 | When this is enabled, you can create perf events based on |
956 | tracepoints using PERF_TYPE_TRACEPOINT and the tracepoint ID | 961 | tracepoints using PERF_TYPE_TRACEPOINT and the tracepoint ID |
957 | found in debugfs://tracing/events/*/*/id. (The -e/--events | 962 | found in debugfs://tracing/events/*/*/id. (The -e/--events |
958 | option to the perf tool can parse and interpret symbolic | 963 | option to the perf tool can parse and interpret symbolic |
959 | tracepoints, in the subsystem:tracepoint_name format.) | 964 | tracepoints, in the subsystem:tracepoint_name format.) |
960 | 965 | ||
966 | config PERF_COUNTERS | ||
967 | bool "Kernel performance counters (old config option)" | ||
968 | depends on HAVE_PERF_EVENTS | ||
969 | help | ||
970 | This config has been obsoleted by the PERF_EVENTS | ||
971 | config option - please see that one for details. | ||
972 | |||
973 | It has no effect on the kernel whether you enable | ||
974 | it or not, it is a compatibility placeholder. | ||
975 | |||
976 | Say N if unsure. | ||
977 | |||
961 | endmenu | 978 | endmenu |
962 | 979 | ||
963 | config VM_EVENT_COUNTERS | 980 | config VM_EVENT_COUNTERS |
diff --git a/kernel/perf_event.c b/kernel/perf_event.c index 6e8b99a04e1..76ac4db405e 100644 --- a/kernel/perf_event.c +++ b/kernel/perf_event.c | |||
@@ -1,12 +1,12 @@ | |||
1 | /* | 1 | /* |
2 | * Performance event core code | 2 | * Performance events core code: |
3 | * | 3 | * |
4 | * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de> | 4 | * Copyright (C) 2008 Thomas Gleixner <tglx@linutronix.de> |
5 | * Copyright (C) 2008-2009 Red Hat, Inc., Ingo Molnar | 5 | * Copyright (C) 2008-2009 Red Hat, Inc., Ingo Molnar |
6 | * Copyright (C) 2008-2009 Red Hat, Inc., Peter Zijlstra <pzijlstr@redhat.com> | 6 | * Copyright (C) 2008-2009 Red Hat, Inc., Peter Zijlstra <pzijlstr@redhat.com> |
7 | * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> | 7 | * Copyright © 2009 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com> |
8 | * | 8 | * |
9 | * For licensing details see kernel-base/COPYING | 9 | * For licensing details see kernel-base/COPYING |
10 | */ | 10 | */ |
11 | 11 | ||
12 | #include <linux/fs.h> | 12 | #include <linux/fs.h> |