aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/x86/include/asm/perf_event_p4.h33
-rw-r--r--arch/x86/kernel/cpu/perf_event.c7
-rw-r--r--arch/x86/kernel/cpu/perf_event_p4.c135
-rw-r--r--kernel/watchdog.c2
4 files changed, 139 insertions, 38 deletions
diff --git a/arch/x86/include/asm/perf_event_p4.h b/arch/x86/include/asm/perf_event_p4.h
index 56fd9e3abbda..4d86c86178f2 100644
--- a/arch/x86/include/asm/perf_event_p4.h
+++ b/arch/x86/include/asm/perf_event_p4.h
@@ -102,6 +102,14 @@
102#define P4_CONFIG_HT (1ULL << P4_CONFIG_HT_SHIFT) 102#define P4_CONFIG_HT (1ULL << P4_CONFIG_HT_SHIFT)
103 103
104/* 104/*
105 * If an event has alias it should be marked
106 * with a special bit. (Don't forget to check
107 * P4_PEBS_CONFIG_MASK and related bits on
108 * modification.)
109 */
110#define P4_CONFIG_ALIASABLE (1 << 9)
111
112/*
105 * The bits we allow to pass for RAW events 113 * The bits we allow to pass for RAW events
106 */ 114 */
107#define P4_CONFIG_MASK_ESCR \ 115#define P4_CONFIG_MASK_ESCR \
@@ -123,6 +131,31 @@
123 (p4_config_pack_escr(P4_CONFIG_MASK_ESCR)) | \ 131 (p4_config_pack_escr(P4_CONFIG_MASK_ESCR)) | \
124 (p4_config_pack_cccr(P4_CONFIG_MASK_CCCR)) 132 (p4_config_pack_cccr(P4_CONFIG_MASK_CCCR))
125 133
134/*
135 * In case of event aliasing we need to preserve some
136 * caller bits otherwise the mapping won't be complete.
137 */
138#define P4_CONFIG_EVENT_ALIAS_MASK \
139 (p4_config_pack_escr(P4_CONFIG_MASK_ESCR) | \
140 p4_config_pack_cccr(P4_CCCR_EDGE | \
141 P4_CCCR_THRESHOLD_MASK | \
142 P4_CCCR_COMPLEMENT | \
143 P4_CCCR_COMPARE))
144
145#define P4_CONFIG_EVENT_ALIAS_IMMUTABLE_BITS \
146 ((P4_CONFIG_HT) | \
147 p4_config_pack_escr(P4_ESCR_T0_OS | \
148 P4_ESCR_T0_USR | \
149 P4_ESCR_T1_OS | \
150 P4_ESCR_T1_USR) | \
151 p4_config_pack_cccr(P4_CCCR_OVF | \
152 P4_CCCR_CASCADE | \
153 P4_CCCR_FORCE_OVF | \
154 P4_CCCR_THREAD_ANY | \
155 P4_CCCR_OVF_PMI_T0 | \
156 P4_CCCR_OVF_PMI_T1 | \
157 P4_CONFIG_ALIASABLE))
158
126static inline bool p4_is_event_cascaded(u64 config) 159static inline bool p4_is_event_cascaded(u64 config)
127{ 160{
128 u32 cccr = p4_config_unpack_cccr(config); 161 u32 cccr = p4_config_unpack_cccr(config);
diff --git a/arch/x86/kernel/cpu/perf_event.c b/arch/x86/kernel/cpu/perf_event.c
index c53d433c3dde..b7a010fce8c3 100644
--- a/arch/x86/kernel/cpu/perf_event.c
+++ b/arch/x86/kernel/cpu/perf_event.c
@@ -274,7 +274,6 @@ struct x86_pmu {
274 void (*enable_all)(int added); 274 void (*enable_all)(int added);
275 void (*enable)(struct perf_event *); 275 void (*enable)(struct perf_event *);
276 void (*disable)(struct perf_event *); 276 void (*disable)(struct perf_event *);
277 void (*hw_watchdog_set_attr)(struct perf_event_attr *attr);
278 int (*hw_config)(struct perf_event *event); 277 int (*hw_config)(struct perf_event *event);
279 int (*schedule_events)(struct cpu_hw_events *cpuc, int n, int *assign); 278 int (*schedule_events)(struct cpu_hw_events *cpuc, int n, int *assign);
280 unsigned eventsel; 279 unsigned eventsel;
@@ -360,12 +359,6 @@ static u64 __read_mostly hw_cache_extra_regs
360 [PERF_COUNT_HW_CACHE_OP_MAX] 359 [PERF_COUNT_HW_CACHE_OP_MAX]
361 [PERF_COUNT_HW_CACHE_RESULT_MAX]; 360 [PERF_COUNT_HW_CACHE_RESULT_MAX];
362 361
363void hw_nmi_watchdog_set_attr(struct perf_event_attr *wd_attr)
364{
365 if (x86_pmu.hw_watchdog_set_attr)
366 x86_pmu.hw_watchdog_set_attr(wd_attr);
367}
368
369/* 362/*
370 * Propagate event elapsed time into the generic event. 363 * Propagate event elapsed time into the generic event.
371 * Can only be executed on the CPU where the event is active. 364 * Can only be executed on the CPU where the event is active.
diff --git a/arch/x86/kernel/cpu/perf_event_p4.c b/arch/x86/kernel/cpu/perf_event_p4.c
index fb901c5080f7..8b7a0c306784 100644
--- a/arch/x86/kernel/cpu/perf_event_p4.c
+++ b/arch/x86/kernel/cpu/perf_event_p4.c
@@ -570,11 +570,92 @@ static __initconst const u64 p4_hw_cache_event_ids
570 }, 570 },
571}; 571};
572 572
573/*
574 * Because of Netburst being quite restricted in now
575 * many same events can run simultaneously, we use
576 * event aliases, ie different events which have the
577 * same functionallity but use non-intersected resources
578 * (ESCR/CCCR/couter registers). This allow us to run
579 * two or more semi-same events together. It is done
580 * transparently to a user space.
581 *
582 * Never set any cusom internal bits such as P4_CONFIG_HT,
583 * P4_CONFIG_ALIASABLE or bits for P4_PEBS_METRIC, they are
584 * either up-to-dated automatically either not appliable
585 * at all.
586 *
587 * And be really carefull choosing aliases!
588 */
589struct p4_event_alias {
590 u64 orig;
591 u64 alter;
592} p4_event_aliases[] = {
593 {
594 /*
595 * Non-halted cycles can be substituted with
596 * non-sleeping cycles (see Intel SDM Vol3b for
597 * details).
598 */
599 .orig =
600 p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) |
601 P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)),
602 .alter =
603 p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_EXECUTION_EVENT) |
604 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS0)|
605 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS1)|
606 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS2)|
607 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS3)|
608 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS0) |
609 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS1) |
610 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS2) |
611 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS3))|
612 p4_config_pack_cccr(P4_CCCR_THRESHOLD(15) | P4_CCCR_COMPLEMENT |
613 P4_CCCR_COMPARE),
614 },
615};
616
617static u64 p4_get_alias_event(u64 config)
618{
619 u64 config_match;
620 int i;
621
622 /*
623 * Probably we're lucky and don't have to do
624 * matching over all config bits.
625 */
626 if (!(config & P4_CONFIG_ALIASABLE))
627 return 0;
628
629 config_match = config & P4_CONFIG_EVENT_ALIAS_MASK;
630
631 /*
632 * If an event was previously swapped to the alter config
633 * we should swap it back otherwise contnention on registers
634 * will return back.
635 */
636 for (i = 0; i < ARRAY_SIZE(p4_event_aliases); i++) {
637 if (config_match == p4_event_aliases[i].orig) {
638 config_match = p4_event_aliases[i].alter;
639 break;
640 } else if (config_match == p4_event_aliases[i].alter) {
641 config_match = p4_event_aliases[i].orig;
642 break;
643 }
644 }
645
646 if (i >= ARRAY_SIZE(p4_event_aliases))
647 return 0;
648
649 return config_match |
650 (config & P4_CONFIG_EVENT_ALIAS_IMMUTABLE_BITS);
651}
652
573static u64 p4_general_events[PERF_COUNT_HW_MAX] = { 653static u64 p4_general_events[PERF_COUNT_HW_MAX] = {
574 /* non-halted CPU clocks */ 654 /* non-halted CPU clocks */
575 [PERF_COUNT_HW_CPU_CYCLES] = 655 [PERF_COUNT_HW_CPU_CYCLES] =
576 p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) | 656 p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_GLOBAL_POWER_EVENTS) |
577 P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)), 657 P4_ESCR_EMASK_BIT(P4_EVENT_GLOBAL_POWER_EVENTS, RUNNING)) |
658 P4_CONFIG_ALIASABLE,
578 659
579 /* 660 /*
580 * retired instructions 661 * retired instructions
@@ -719,31 +800,6 @@ static int p4_validate_raw_event(struct perf_event *event)
719 return 0; 800 return 0;
720} 801}
721 802
722static void p4_hw_watchdog_set_attr(struct perf_event_attr *wd_attr)
723{
724 /*
725 * Watchdog ticks are special on Netburst, we use
726 * that named "non-sleeping" ticks as recommended
727 * by Intel SDM Vol3b.
728 */
729 WARN_ON_ONCE(wd_attr->type != PERF_TYPE_HARDWARE ||
730 wd_attr->config != PERF_COUNT_HW_CPU_CYCLES);
731
732 wd_attr->type = PERF_TYPE_RAW;
733 wd_attr->config =
734 p4_config_pack_escr(P4_ESCR_EVENT(P4_EVENT_EXECUTION_EVENT) |
735 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS0) |
736 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS1) |
737 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS2) |
738 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, NBOGUS3) |
739 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS0) |
740 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS1) |
741 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS2) |
742 P4_ESCR_EMASK_BIT(P4_EVENT_EXECUTION_EVENT, BOGUS3)) |
743 p4_config_pack_cccr(P4_CCCR_THRESHOLD(15) | P4_CCCR_COMPLEMENT |
744 P4_CCCR_COMPARE);
745}
746
747static int p4_hw_config(struct perf_event *event) 803static int p4_hw_config(struct perf_event *event)
748{ 804{
749 int cpu = get_cpu(); 805 int cpu = get_cpu();
@@ -1159,6 +1215,8 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign
1159 struct p4_event_bind *bind; 1215 struct p4_event_bind *bind;
1160 unsigned int i, thread, num; 1216 unsigned int i, thread, num;
1161 int cntr_idx, escr_idx; 1217 int cntr_idx, escr_idx;
1218 u64 config_alias;
1219 int pass;
1162 1220
1163 bitmap_zero(used_mask, X86_PMC_IDX_MAX); 1221 bitmap_zero(used_mask, X86_PMC_IDX_MAX);
1164 bitmap_zero(escr_mask, P4_ESCR_MSR_TABLE_SIZE); 1222 bitmap_zero(escr_mask, P4_ESCR_MSR_TABLE_SIZE);
@@ -1167,6 +1225,17 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign
1167 1225
1168 hwc = &cpuc->event_list[i]->hw; 1226 hwc = &cpuc->event_list[i]->hw;
1169 thread = p4_ht_thread(cpu); 1227 thread = p4_ht_thread(cpu);
1228 pass = 0;
1229
1230again:
1231 /*
1232 * Aliases are swappable so we may hit circular
1233 * lock if both original config and alias need
1234 * resources (MSR registers) which already busy.
1235 */
1236 if (pass > 2)
1237 goto done;
1238
1170 bind = p4_config_get_bind(hwc->config); 1239 bind = p4_config_get_bind(hwc->config);
1171 escr_idx = p4_get_escr_idx(bind->escr_msr[thread]); 1240 escr_idx = p4_get_escr_idx(bind->escr_msr[thread]);
1172 if (unlikely(escr_idx == -1)) 1241 if (unlikely(escr_idx == -1))
@@ -1180,8 +1249,17 @@ static int p4_pmu_schedule_events(struct cpu_hw_events *cpuc, int n, int *assign
1180 } 1249 }
1181 1250
1182 cntr_idx = p4_next_cntr(thread, used_mask, bind); 1251 cntr_idx = p4_next_cntr(thread, used_mask, bind);
1183 if (cntr_idx == -1 || test_bit(escr_idx, escr_mask)) 1252 if (cntr_idx == -1 || test_bit(escr_idx, escr_mask)) {
1184 goto done; 1253 /*
1254 * Probably an event alias is still available.
1255 */
1256 config_alias = p4_get_alias_event(hwc->config);
1257 if (!config_alias)
1258 goto done;
1259 hwc->config = config_alias;
1260 pass++;
1261 goto again;
1262 }
1185 1263
1186 p4_pmu_swap_config_ts(hwc, cpu); 1264 p4_pmu_swap_config_ts(hwc, cpu);
1187 if (assign) 1265 if (assign)
@@ -1218,7 +1296,6 @@ static __initconst const struct x86_pmu p4_pmu = {
1218 .cntval_bits = ARCH_P4_CNTRVAL_BITS, 1296 .cntval_bits = ARCH_P4_CNTRVAL_BITS,
1219 .cntval_mask = ARCH_P4_CNTRVAL_MASK, 1297 .cntval_mask = ARCH_P4_CNTRVAL_MASK,
1220 .max_period = (1ULL << (ARCH_P4_CNTRVAL_BITS - 1)) - 1, 1298 .max_period = (1ULL << (ARCH_P4_CNTRVAL_BITS - 1)) - 1,
1221 .hw_watchdog_set_attr = p4_hw_watchdog_set_attr,
1222 .hw_config = p4_hw_config, 1299 .hw_config = p4_hw_config,
1223 .schedule_events = p4_pmu_schedule_events, 1300 .schedule_events = p4_pmu_schedule_events,
1224 /* 1301 /*
diff --git a/kernel/watchdog.c b/kernel/watchdog.c
index a933e3a0398b..36491cd5b7d4 100644
--- a/kernel/watchdog.c
+++ b/kernel/watchdog.c
@@ -200,7 +200,6 @@ static int is_softlockup(unsigned long touch_ts)
200} 200}
201 201
202#ifdef CONFIG_HARDLOCKUP_DETECTOR 202#ifdef CONFIG_HARDLOCKUP_DETECTOR
203void __weak hw_nmi_watchdog_set_attr(struct perf_event_attr *wd_attr) { }
204 203
205static struct perf_event_attr wd_hw_attr = { 204static struct perf_event_attr wd_hw_attr = {
206 .type = PERF_TYPE_HARDWARE, 205 .type = PERF_TYPE_HARDWARE,
@@ -372,7 +371,6 @@ static int watchdog_nmi_enable(int cpu)
372 371
373 wd_attr = &wd_hw_attr; 372 wd_attr = &wd_hw_attr;
374 wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); 373 wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh);
375 hw_nmi_watchdog_set_attr(wd_attr);
376 374
377 /* Try to register using hardware perf events */ 375 /* Try to register using hardware perf events */
378 event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback, NULL); 376 event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback, NULL);