diff options
author | Stephane Eranian <eranian@google.com> | 2010-02-01 07:50:01 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2010-02-04 03:59:50 -0500 |
commit | 447a194b393f32699607fd99617a40abd6a95114 (patch) | |
tree | 7d202a6ad8f80c913a4e3d439eedf5ba0abbbf39 /arch | |
parent | fce877e3a429940a986e085a41e8b57f2d922e36 (diff) |
perf_events, x86: Fix bug in hw_perf_enable()
We cannot assume that because hwc->idx == assign[i], we can avoid
reprogramming the counter in hw_perf_enable().
The event may have been scheduled out and another event may have been
programmed into this counter. Thus, we need a more robust way of
verifying if the counter still contains config/data related to an event.
This patch adds a generation number to each counter on each cpu. Using
this mechanism we can verify reliabilty whether the content of a counter
corresponds to an event.
Signed-off-by: Stephane Eranian <eranian@google.com>
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Frederic Weisbecker <fweisbec@gmail.com>
LKML-Reference: <4b66dc67.0b38560a.1635.ffffae18@mx.google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r-- | arch/x86/kernel/cpu/perf_event.c | 34 |
1 files changed, 28 insertions, 6 deletions
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c index 96cfc1a4fe9f..a920f173a220 100644 --- a/arch/x86/kernel/cpu/perf_event.c +++ b/arch/x86/kernel/cpu/perf_event.c | |||
@@ -90,6 +90,7 @@ struct cpu_hw_events { | |||
90 | int n_events; | 90 | int n_events; |
91 | int n_added; | 91 | int n_added; |
92 | int assign[X86_PMC_IDX_MAX]; /* event to counter assignment */ | 92 | int assign[X86_PMC_IDX_MAX]; /* event to counter assignment */ |
93 | u64 tags[X86_PMC_IDX_MAX]; | ||
93 | struct perf_event *event_list[X86_PMC_IDX_MAX]; /* in enabled order */ | 94 | struct perf_event *event_list[X86_PMC_IDX_MAX]; /* in enabled order */ |
94 | }; | 95 | }; |
95 | 96 | ||
@@ -1142,6 +1143,8 @@ static int __hw_perf_event_init(struct perf_event *event) | |||
1142 | hwc->config = ARCH_PERFMON_EVENTSEL_INT; | 1143 | hwc->config = ARCH_PERFMON_EVENTSEL_INT; |
1143 | 1144 | ||
1144 | hwc->idx = -1; | 1145 | hwc->idx = -1; |
1146 | hwc->last_cpu = -1; | ||
1147 | hwc->last_tag = ~0ULL; | ||
1145 | 1148 | ||
1146 | /* | 1149 | /* |
1147 | * Count user and OS events unless requested not to. | 1150 | * Count user and OS events unless requested not to. |
@@ -1457,11 +1460,14 @@ static int collect_events(struct cpu_hw_events *cpuc, struct perf_event *leader, | |||
1457 | return n; | 1460 | return n; |
1458 | } | 1461 | } |
1459 | 1462 | ||
1460 | |||
1461 | static inline void x86_assign_hw_event(struct perf_event *event, | 1463 | static inline void x86_assign_hw_event(struct perf_event *event, |
1462 | struct hw_perf_event *hwc, int idx) | 1464 | struct cpu_hw_events *cpuc, int i) |
1463 | { | 1465 | { |
1464 | hwc->idx = idx; | 1466 | struct hw_perf_event *hwc = &event->hw; |
1467 | |||
1468 | hwc->idx = cpuc->assign[i]; | ||
1469 | hwc->last_cpu = smp_processor_id(); | ||
1470 | hwc->last_tag = ++cpuc->tags[i]; | ||
1465 | 1471 | ||
1466 | if (hwc->idx == X86_PMC_IDX_FIXED_BTS) { | 1472 | if (hwc->idx == X86_PMC_IDX_FIXED_BTS) { |
1467 | hwc->config_base = 0; | 1473 | hwc->config_base = 0; |
@@ -1480,6 +1486,15 @@ static inline void x86_assign_hw_event(struct perf_event *event, | |||
1480 | } | 1486 | } |
1481 | } | 1487 | } |
1482 | 1488 | ||
1489 | static inline int match_prev_assignment(struct hw_perf_event *hwc, | ||
1490 | struct cpu_hw_events *cpuc, | ||
1491 | int i) | ||
1492 | { | ||
1493 | return hwc->idx == cpuc->assign[i] && | ||
1494 | hwc->last_cpu == smp_processor_id() && | ||
1495 | hwc->last_tag == cpuc->tags[i]; | ||
1496 | } | ||
1497 | |||
1483 | static void __x86_pmu_disable(struct perf_event *event, struct cpu_hw_events *cpuc); | 1498 | static void __x86_pmu_disable(struct perf_event *event, struct cpu_hw_events *cpuc); |
1484 | 1499 | ||
1485 | void hw_perf_enable(void) | 1500 | void hw_perf_enable(void) |
@@ -1508,7 +1523,14 @@ void hw_perf_enable(void) | |||
1508 | event = cpuc->event_list[i]; | 1523 | event = cpuc->event_list[i]; |
1509 | hwc = &event->hw; | 1524 | hwc = &event->hw; |
1510 | 1525 | ||
1511 | if (hwc->idx == -1 || hwc->idx == cpuc->assign[i]) | 1526 | /* |
1527 | * we can avoid reprogramming counter if: | ||
1528 | * - assigned same counter as last time | ||
1529 | * - running on same CPU as last time | ||
1530 | * - no other event has used the counter since | ||
1531 | */ | ||
1532 | if (hwc->idx == -1 || | ||
1533 | match_prev_assignment(hwc, cpuc, i)) | ||
1512 | continue; | 1534 | continue; |
1513 | 1535 | ||
1514 | __x86_pmu_disable(event, cpuc); | 1536 | __x86_pmu_disable(event, cpuc); |
@@ -1522,12 +1544,12 @@ void hw_perf_enable(void) | |||
1522 | hwc = &event->hw; | 1544 | hwc = &event->hw; |
1523 | 1545 | ||
1524 | if (hwc->idx == -1) { | 1546 | if (hwc->idx == -1) { |
1525 | x86_assign_hw_event(event, hwc, cpuc->assign[i]); | 1547 | x86_assign_hw_event(event, cpuc, i); |
1526 | x86_perf_event_set_period(event, hwc, hwc->idx); | 1548 | x86_perf_event_set_period(event, hwc, hwc->idx); |
1527 | } | 1549 | } |
1528 | /* | 1550 | /* |
1529 | * need to mark as active because x86_pmu_disable() | 1551 | * need to mark as active because x86_pmu_disable() |
1530 | * clear active_mask and eventsp[] yet it preserves | 1552 | * clear active_mask and events[] yet it preserves |
1531 | * idx | 1553 | * idx |
1532 | */ | 1554 | */ |
1533 | set_bit(hwc->idx, cpuc->active_mask); | 1555 | set_bit(hwc->idx, cpuc->active_mask); |