diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2009-06-19 12:27:26 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-06-23 05:42:44 -0400 |
commit | f29ac756a40d0f1bb07d682ea521e7b666ff06d5 (patch) | |
tree | 6fabe2f38ae10f824e57fdd5b035ba81b75a71a8 /kernel | |
parent | 3d906ef10a539ff336010afab8f6f9c4fe379695 (diff) |
perf_counter: Optimize perf_swcounter_event()
Similar to tracepoints, use an enable variable to reduce
overhead when unused.
Only look for a counter of a particular event type when we know
there is at least one in the system.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
LKML-Reference: <new-submission>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/perf_counter.c | 18 |
1 files changed, 15 insertions, 3 deletions
diff --git a/kernel/perf_counter.c b/kernel/perf_counter.c index 1a933a221ea4..7515c7695428 100644 --- a/kernel/perf_counter.c +++ b/kernel/perf_counter.c | |||
@@ -3317,8 +3317,8 @@ out: | |||
3317 | put_cpu_var(perf_cpu_context); | 3317 | put_cpu_var(perf_cpu_context); |
3318 | } | 3318 | } |
3319 | 3319 | ||
3320 | void | 3320 | void __perf_swcounter_event(u32 event, u64 nr, int nmi, |
3321 | perf_swcounter_event(u32 event, u64 nr, int nmi, struct pt_regs *regs, u64 addr) | 3321 | struct pt_regs *regs, u64 addr) |
3322 | { | 3322 | { |
3323 | struct perf_sample_data data = { | 3323 | struct perf_sample_data data = { |
3324 | .regs = regs, | 3324 | .regs = regs, |
@@ -3509,9 +3509,19 @@ static const struct pmu *tp_perf_counter_init(struct perf_counter *counter) | |||
3509 | } | 3509 | } |
3510 | #endif | 3510 | #endif |
3511 | 3511 | ||
3512 | atomic_t perf_swcounter_enabled[PERF_COUNT_SW_MAX]; | ||
3513 | |||
3514 | static void sw_perf_counter_destroy(struct perf_counter *counter) | ||
3515 | { | ||
3516 | u64 event = counter->attr.config; | ||
3517 | |||
3518 | atomic_dec(&perf_swcounter_enabled[event]); | ||
3519 | } | ||
3520 | |||
3512 | static const struct pmu *sw_perf_counter_init(struct perf_counter *counter) | 3521 | static const struct pmu *sw_perf_counter_init(struct perf_counter *counter) |
3513 | { | 3522 | { |
3514 | const struct pmu *pmu = NULL; | 3523 | const struct pmu *pmu = NULL; |
3524 | u64 event = counter->attr.config; | ||
3515 | 3525 | ||
3516 | /* | 3526 | /* |
3517 | * Software counters (currently) can't in general distinguish | 3527 | * Software counters (currently) can't in general distinguish |
@@ -3520,7 +3530,7 @@ static const struct pmu *sw_perf_counter_init(struct perf_counter *counter) | |||
3520 | * to be kernel events, and page faults are never hypervisor | 3530 | * to be kernel events, and page faults are never hypervisor |
3521 | * events. | 3531 | * events. |
3522 | */ | 3532 | */ |
3523 | switch (counter->attr.config) { | 3533 | switch (event) { |
3524 | case PERF_COUNT_SW_CPU_CLOCK: | 3534 | case PERF_COUNT_SW_CPU_CLOCK: |
3525 | pmu = &perf_ops_cpu_clock; | 3535 | pmu = &perf_ops_cpu_clock; |
3526 | 3536 | ||
@@ -3541,6 +3551,8 @@ static const struct pmu *sw_perf_counter_init(struct perf_counter *counter) | |||
3541 | case PERF_COUNT_SW_PAGE_FAULTS_MAJ: | 3551 | case PERF_COUNT_SW_PAGE_FAULTS_MAJ: |
3542 | case PERF_COUNT_SW_CONTEXT_SWITCHES: | 3552 | case PERF_COUNT_SW_CONTEXT_SWITCHES: |
3543 | case PERF_COUNT_SW_CPU_MIGRATIONS: | 3553 | case PERF_COUNT_SW_CPU_MIGRATIONS: |
3554 | atomic_inc(&perf_swcounter_enabled[event]); | ||
3555 | counter->destroy = sw_perf_counter_destroy; | ||
3544 | pmu = &perf_ops_generic; | 3556 | pmu = &perf_ops_generic; |
3545 | break; | 3557 | break; |
3546 | } | 3558 | } |