aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/perf_counter.c
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2009-06-19 12:27:26 -0400
committerIngo Molnar <mingo@elte.hu>2009-06-23 05:42:44 -0400
commitf29ac756a40d0f1bb07d682ea521e7b666ff06d5 (patch)
tree6fabe2f38ae10f824e57fdd5b035ba81b75a71a8 /kernel/perf_counter.c
parent3d906ef10a539ff336010afab8f6f9c4fe379695 (diff)
perf_counter: Optimize perf_swcounter_event()
Similar to tracepoints, use an enable variable to reduce overhead when unused. Only look for a counter of a particular event type when we know there is at least one in the system. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> LKML-Reference: <new-submission> Cc: Mike Galbraith <efault@gmx.de> Cc: Paul Mackerras <paulus@samba.org> Cc: Arnaldo Carvalho de Melo <acme@redhat.com> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/perf_counter.c')
-rw-r--r--kernel/perf_counter.c18
1 files changed, 15 insertions, 3 deletions
diff --git a/kernel/perf_counter.c b/kernel/perf_counter.c
index 1a933a221ea4..7515c7695428 100644
--- a/kernel/perf_counter.c
+++ b/kernel/perf_counter.c
@@ -3317,8 +3317,8 @@ out:
3317 put_cpu_var(perf_cpu_context); 3317 put_cpu_var(perf_cpu_context);
3318} 3318}
3319 3319
3320void 3320void __perf_swcounter_event(u32 event, u64 nr, int nmi,
3321perf_swcounter_event(u32 event, u64 nr, int nmi, struct pt_regs *regs, u64 addr) 3321 struct pt_regs *regs, u64 addr)
3322{ 3322{
3323 struct perf_sample_data data = { 3323 struct perf_sample_data data = {
3324 .regs = regs, 3324 .regs = regs,
@@ -3509,9 +3509,19 @@ static const struct pmu *tp_perf_counter_init(struct perf_counter *counter)
3509} 3509}
3510#endif 3510#endif
3511 3511
3512atomic_t perf_swcounter_enabled[PERF_COUNT_SW_MAX];
3513
3514static void sw_perf_counter_destroy(struct perf_counter *counter)
3515{
3516 u64 event = counter->attr.config;
3517
3518 atomic_dec(&perf_swcounter_enabled[event]);
3519}
3520
3512static const struct pmu *sw_perf_counter_init(struct perf_counter *counter) 3521static const struct pmu *sw_perf_counter_init(struct perf_counter *counter)
3513{ 3522{
3514 const struct pmu *pmu = NULL; 3523 const struct pmu *pmu = NULL;
3524 u64 event = counter->attr.config;
3515 3525
3516 /* 3526 /*
3517 * Software counters (currently) can't in general distinguish 3527 * Software counters (currently) can't in general distinguish
@@ -3520,7 +3530,7 @@ static const struct pmu *sw_perf_counter_init(struct perf_counter *counter)
3520 * to be kernel events, and page faults are never hypervisor 3530 * to be kernel events, and page faults are never hypervisor
3521 * events. 3531 * events.
3522 */ 3532 */
3523 switch (counter->attr.config) { 3533 switch (event) {
3524 case PERF_COUNT_SW_CPU_CLOCK: 3534 case PERF_COUNT_SW_CPU_CLOCK:
3525 pmu = &perf_ops_cpu_clock; 3535 pmu = &perf_ops_cpu_clock;
3526 3536
@@ -3541,6 +3551,8 @@ static const struct pmu *sw_perf_counter_init(struct perf_counter *counter)
3541 case PERF_COUNT_SW_PAGE_FAULTS_MAJ: 3551 case PERF_COUNT_SW_PAGE_FAULTS_MAJ:
3542 case PERF_COUNT_SW_CONTEXT_SWITCHES: 3552 case PERF_COUNT_SW_CONTEXT_SWITCHES:
3543 case PERF_COUNT_SW_CPU_MIGRATIONS: 3553 case PERF_COUNT_SW_CPU_MIGRATIONS:
3554 atomic_inc(&perf_swcounter_enabled[event]);
3555 counter->destroy = sw_perf_counter_destroy;
3544 pmu = &perf_ops_generic; 3556 pmu = &perf_ops_generic;
3545 break; 3557 break;
3546 } 3558 }