diff options
-rw-r--r-- | Documentation/powerpc/00-INDEX | 2 | ||||
-rw-r--r-- | Documentation/powerpc/pmu-ebb.txt | 137 | ||||
-rw-r--r-- | arch/powerpc/include/asm/perf_event_server.h | 6 | ||||
-rw-r--r-- | arch/powerpc/include/asm/processor.h | 3 | ||||
-rw-r--r-- | arch/powerpc/include/asm/reg.h | 8 | ||||
-rw-r--r-- | arch/powerpc/include/asm/switch_to.h | 14 | ||||
-rw-r--r-- | arch/powerpc/kernel/process.c | 4 | ||||
-rw-r--r-- | arch/powerpc/perf/core-book3s.c | 161 |
8 files changed, 321 insertions, 14 deletions
diff --git a/Documentation/powerpc/00-INDEX b/Documentation/powerpc/00-INDEX index dd9e92802ec0..05026ce1875e 100644 --- a/Documentation/powerpc/00-INDEX +++ b/Documentation/powerpc/00-INDEX | |||
@@ -14,6 +14,8 @@ hvcs.txt | |||
14 | - IBM "Hypervisor Virtual Console Server" Installation Guide | 14 | - IBM "Hypervisor Virtual Console Server" Installation Guide |
15 | mpc52xx.txt | 15 | mpc52xx.txt |
16 | - Linux 2.6.x on MPC52xx family | 16 | - Linux 2.6.x on MPC52xx family |
17 | pmu-ebb.txt | ||
18 | - Description of the API for using the PMU with Event Based Branches. | ||
17 | qe_firmware.txt | 19 | qe_firmware.txt |
18 | - describes the layout of firmware binaries for the Freescale QUICC | 20 | - describes the layout of firmware binaries for the Freescale QUICC |
19 | Engine and the code that parses and uploads the microcode therein. | 21 | Engine and the code that parses and uploads the microcode therein. |
diff --git a/Documentation/powerpc/pmu-ebb.txt b/Documentation/powerpc/pmu-ebb.txt new file mode 100644 index 000000000000..73cd163dbfb8 --- /dev/null +++ b/Documentation/powerpc/pmu-ebb.txt | |||
@@ -0,0 +1,137 @@ | |||
1 | PMU Event Based Branches | ||
2 | ======================== | ||
3 | |||
4 | Event Based Branches (EBBs) are a feature which allows the hardware to | ||
5 | branch directly to a specified user space address when certain events occur. | ||
6 | |||
7 | The full specification is available in Power ISA v2.07: | ||
8 | |||
9 | https://www.power.org/documentation/power-isa-version-2-07/ | ||
10 | |||
11 | One type of event for which EBBs can be configured is PMU exceptions. This | ||
12 | document describes the API for configuring the Power PMU to generate EBBs, | ||
13 | using the Linux perf_events API. | ||
14 | |||
15 | |||
16 | Terminology | ||
17 | ----------- | ||
18 | |||
19 | Throughout this document we will refer to an "EBB event" or "EBB events". This | ||
20 | just refers to a struct perf_event which has set the "EBB" flag in its | ||
21 | attr.config. All events which can be configured on the hardware PMU are | ||
22 | possible "EBB events". | ||
23 | |||
24 | |||
25 | Background | ||
26 | ---------- | ||
27 | |||
28 | When a PMU EBB occurs it is delivered to the currently running process. As such | ||
29 | EBBs can only sensibly be used by programs for self-monitoring. | ||
30 | |||
31 | It is a feature of the perf_events API that events can be created on other | ||
32 | processes, subject to standard permission checks. This is also true of EBB | ||
33 | events, however unless the target process enables EBBs (via mtspr(BESCR)) no | ||
34 | EBBs will ever be delivered. | ||
35 | |||
36 | This makes it possible for a process to enable EBBs for itself, but not | ||
37 | actually configure any events. At a later time another process can come along | ||
38 | and attach an EBB event to the process, which will then cause EBBs to be | ||
39 | delivered to the first process. It's not clear if this is actually useful. | ||
40 | |||
41 | |||
42 | When the PMU is configured for EBBs, all PMU interrupts are delivered to the | ||
43 | user process. This means once an EBB event is scheduled on the PMU, no non-EBB | ||
44 | events can be configured. This means that EBB events can not be run | ||
45 | concurrently with regular 'perf' commands, or any other perf events. | ||
46 | |||
47 | It is however safe to run 'perf' commands on a process which is using EBBs. The | ||
48 | kernel will in general schedule the EBB event, and perf will be notified that | ||
49 | its events could not run. | ||
50 | |||
51 | The exclusion between EBB events and regular events is implemented using the | ||
52 | existing "pinned" and "exclusive" attributes of perf_events. This means EBB | ||
53 | events will be given priority over other events, unless they are also pinned. | ||
54 | If an EBB event and a regular event are both pinned, then whichever is enabled | ||
55 | first will be scheduled and the other will be put in error state. See the | ||
56 | section below titled "Enabling an EBB event" for more information. | ||
57 | |||
58 | |||
59 | Creating an EBB event | ||
60 | --------------------- | ||
61 | |||
62 | To request that an event is counted using EBB, the event code should have bit | ||
63 | 63 set. | ||
64 | |||
65 | EBB events must be created with a particular, and restrictive, set of | ||
66 | attributes - this is so that they interoperate correctly with the rest of the | ||
67 | perf_events subsystem. | ||
68 | |||
69 | An EBB event must be created with the "pinned" and "exclusive" attributes set. | ||
70 | Note that if you are creating a group of EBB events, only the leader can have | ||
71 | these attributes set. | ||
72 | |||
73 | An EBB event must NOT set any of the "inherit", "sample_period", "freq" or | ||
74 | "enable_on_exec" attributes. | ||
75 | |||
76 | An EBB event must be attached to a task. This is specified to perf_event_open() | ||
77 | by passing a pid value, typically 0 indicating the current task. | ||
78 | |||
79 | All events in a group must agree on whether they want EBB. That is all events | ||
80 | must request EBB, or none may request EBB. | ||
81 | |||
82 | EBB events must specify the PMC they are to be counted on. This ensures | ||
83 | userspace is able to reliably determine which PMC the event is scheduled on. | ||
84 | |||
85 | |||
86 | Enabling an EBB event | ||
87 | --------------------- | ||
88 | |||
89 | Once an EBB event has been successfully opened, it must be enabled with the | ||
90 | perf_events API. This can be achieved either via the ioctl() interface, or the | ||
91 | prctl() interface. | ||
92 | |||
93 | However, due to the design of the perf_events API, enabling an event does not | ||
94 | guarantee that it has been scheduled on the PMU. To ensure that the EBB event | ||
95 | has been scheduled on the PMU, you must perform a read() on the event. If the | ||
96 | read() returns EOF, then the event has not been scheduled and EBBs are not | ||
97 | enabled. | ||
98 | |||
99 | This behaviour occurs because the EBB event is pinned and exclusive. When the | ||
100 | EBB event is enabled it will force all other non-pinned events off the PMU. In | ||
101 | this case the enable will be successful. However if there is already an event | ||
102 | pinned on the PMU then the enable will not be successful. | ||
103 | |||
104 | |||
105 | Reading an EBB event | ||
106 | -------------------- | ||
107 | |||
108 | It is possible to read() from an EBB event. However the results are | ||
109 | meaningless. Because interrupts are being delivered to the user process the | ||
110 | kernel is not able to count the event, and so will return a junk value. | ||
111 | |||
112 | |||
113 | Closing an EBB event | ||
114 | -------------------- | ||
115 | |||
116 | When an EBB event is finished with, you can close it using close() as for any | ||
117 | regular event. If this is the last EBB event the PMU will be deconfigured and | ||
118 | no further PMU EBBs will be delivered. | ||
119 | |||
120 | |||
121 | EBB Handler | ||
122 | ----------- | ||
123 | |||
124 | The EBB handler is just regular userspace code, however it must be written in | ||
125 | the style of an interrupt handler. When the handler is entered all registers | ||
126 | are live (possibly) and so must be saved somehow before the handler can invoke | ||
127 | other code. | ||
128 | |||
129 | It's up to the program how to handle this. For C programs a relatively simple | ||
130 | option is to create an interrupt frame on the stack and save registers there. | ||
131 | |||
132 | Fork | ||
133 | ---- | ||
134 | |||
135 | EBB events are not inherited across fork. If the child process wishes to use | ||
136 | EBBs it should open a new event for itself. Similarly the EBB state in | ||
137 | BESCR/EBBHR/EBBRR is cleared across fork(). | ||
diff --git a/arch/powerpc/include/asm/perf_event_server.h b/arch/powerpc/include/asm/perf_event_server.h index f265049dd7d6..2dd7bfc459be 100644 --- a/arch/powerpc/include/asm/perf_event_server.h +++ b/arch/powerpc/include/asm/perf_event_server.h | |||
@@ -60,6 +60,7 @@ struct power_pmu { | |||
60 | #define PPMU_HAS_SSLOT 0x00000020 /* Has sampled slot in MMCRA */ | 60 | #define PPMU_HAS_SSLOT 0x00000020 /* Has sampled slot in MMCRA */ |
61 | #define PPMU_HAS_SIER 0x00000040 /* Has SIER */ | 61 | #define PPMU_HAS_SIER 0x00000040 /* Has SIER */ |
62 | #define PPMU_BHRB 0x00000080 /* has BHRB feature enabled */ | 62 | #define PPMU_BHRB 0x00000080 /* has BHRB feature enabled */ |
63 | #define PPMU_EBB 0x00000100 /* supports event based branch */ | ||
63 | 64 | ||
64 | /* | 65 | /* |
65 | * Values for flags to get_alternatives() | 66 | * Values for flags to get_alternatives() |
@@ -68,6 +69,11 @@ struct power_pmu { | |||
68 | #define PPMU_LIMITED_PMC_REQD 2 /* have to put this on a limited PMC */ | 69 | #define PPMU_LIMITED_PMC_REQD 2 /* have to put this on a limited PMC */ |
69 | #define PPMU_ONLY_COUNT_RUN 4 /* only counting in run state */ | 70 | #define PPMU_ONLY_COUNT_RUN 4 /* only counting in run state */ |
70 | 71 | ||
72 | /* | ||
73 | * We use the event config bit 63 as a flag to request EBB. | ||
74 | */ | ||
75 | #define EVENT_CONFIG_EBB_SHIFT 63 | ||
76 | |||
71 | extern int register_power_pmu(struct power_pmu *); | 77 | extern int register_power_pmu(struct power_pmu *); |
72 | 78 | ||
73 | struct pt_regs; | 79 | struct pt_regs; |
diff --git a/arch/powerpc/include/asm/processor.h b/arch/powerpc/include/asm/processor.h index 3f19df3cc7a3..47a35b08b963 100644 --- a/arch/powerpc/include/asm/processor.h +++ b/arch/powerpc/include/asm/processor.h | |||
@@ -287,8 +287,9 @@ struct thread_struct { | |||
287 | unsigned long siar; | 287 | unsigned long siar; |
288 | unsigned long sdar; | 288 | unsigned long sdar; |
289 | unsigned long sier; | 289 | unsigned long sier; |
290 | unsigned long mmcr0; | ||
291 | unsigned long mmcr2; | 290 | unsigned long mmcr2; |
291 | unsigned mmcr0; | ||
292 | unsigned used_ebb; | ||
292 | #endif | 293 | #endif |
293 | }; | 294 | }; |
294 | 295 | ||
diff --git a/arch/powerpc/include/asm/reg.h b/arch/powerpc/include/asm/reg.h index 362142b69d5b..5d7d9c2a5473 100644 --- a/arch/powerpc/include/asm/reg.h +++ b/arch/powerpc/include/asm/reg.h | |||
@@ -621,6 +621,9 @@ | |||
621 | #define MMCR0_PMXE 0x04000000UL /* performance monitor exception enable */ | 621 | #define MMCR0_PMXE 0x04000000UL /* performance monitor exception enable */ |
622 | #define MMCR0_FCECE 0x02000000UL /* freeze ctrs on enabled cond or event */ | 622 | #define MMCR0_FCECE 0x02000000UL /* freeze ctrs on enabled cond or event */ |
623 | #define MMCR0_TBEE 0x00400000UL /* time base exception enable */ | 623 | #define MMCR0_TBEE 0x00400000UL /* time base exception enable */ |
624 | #define MMCR0_EBE 0x00100000UL /* Event based branch enable */ | ||
625 | #define MMCR0_PMCC 0x000c0000UL /* PMC control */ | ||
626 | #define MMCR0_PMCC_U6 0x00080000UL /* PMC1-6 are R/W by user (PR) */ | ||
624 | #define MMCR0_PMC1CE 0x00008000UL /* PMC1 count enable*/ | 627 | #define MMCR0_PMC1CE 0x00008000UL /* PMC1 count enable*/ |
625 | #define MMCR0_PMCjCE 0x00004000UL /* PMCj count enable*/ | 628 | #define MMCR0_PMCjCE 0x00004000UL /* PMCj count enable*/ |
626 | #define MMCR0_TRIGGER 0x00002000UL /* TRIGGER enable */ | 629 | #define MMCR0_TRIGGER 0x00002000UL /* TRIGGER enable */ |
@@ -674,6 +677,11 @@ | |||
674 | #define SIER_SIAR_VALID 0x0400000 /* SIAR contents valid */ | 677 | #define SIER_SIAR_VALID 0x0400000 /* SIAR contents valid */ |
675 | #define SIER_SDAR_VALID 0x0200000 /* SDAR contents valid */ | 678 | #define SIER_SDAR_VALID 0x0200000 /* SDAR contents valid */ |
676 | 679 | ||
680 | /* When EBB is enabled, some of MMCR0/MMCR2/SIER are user accessible */ | ||
681 | #define MMCR0_USER_MASK (MMCR0_FC | MMCR0_PMXE | MMCR0_PMAO) | ||
682 | #define MMCR2_USER_MASK 0x4020100804020000UL /* (FC1P|FC2P|FC3P|FC4P|FC5P|FC6P) */ | ||
683 | #define SIER_USER_MASK 0x7fffffUL | ||
684 | |||
677 | #define SPRN_PA6T_MMCR0 795 | 685 | #define SPRN_PA6T_MMCR0 795 |
678 | #define PA6T_MMCR0_EN0 0x0000000000000001UL | 686 | #define PA6T_MMCR0_EN0 0x0000000000000001UL |
679 | #define PA6T_MMCR0_EN1 0x0000000000000002UL | 687 | #define PA6T_MMCR0_EN1 0x0000000000000002UL |
diff --git a/arch/powerpc/include/asm/switch_to.h b/arch/powerpc/include/asm/switch_to.h index 200d763a0a67..49a13e0ef234 100644 --- a/arch/powerpc/include/asm/switch_to.h +++ b/arch/powerpc/include/asm/switch_to.h | |||
@@ -67,4 +67,18 @@ static inline void flush_spe_to_thread(struct task_struct *t) | |||
67 | } | 67 | } |
68 | #endif | 68 | #endif |
69 | 69 | ||
70 | static inline void clear_task_ebb(struct task_struct *t) | ||
71 | { | ||
72 | #ifdef CONFIG_PPC_BOOK3S_64 | ||
73 | /* EBB perf events are not inherited, so clear all EBB state. */ | ||
74 | t->thread.bescr = 0; | ||
75 | t->thread.mmcr2 = 0; | ||
76 | t->thread.mmcr0 = 0; | ||
77 | t->thread.siar = 0; | ||
78 | t->thread.sdar = 0; | ||
79 | t->thread.sier = 0; | ||
80 | t->thread.used_ebb = 0; | ||
81 | #endif | ||
82 | } | ||
83 | |||
70 | #endif /* _ASM_POWERPC_SWITCH_TO_H */ | 84 | #endif /* _ASM_POWERPC_SWITCH_TO_H */ |
diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c index b0f3e3f77e72..f8a76e6207bd 100644 --- a/arch/powerpc/kernel/process.c +++ b/arch/powerpc/kernel/process.c | |||
@@ -916,7 +916,11 @@ int arch_dup_task_struct(struct task_struct *dst, struct task_struct *src) | |||
916 | flush_altivec_to_thread(src); | 916 | flush_altivec_to_thread(src); |
917 | flush_vsx_to_thread(src); | 917 | flush_vsx_to_thread(src); |
918 | flush_spe_to_thread(src); | 918 | flush_spe_to_thread(src); |
919 | |||
919 | *dst = *src; | 920 | *dst = *src; |
921 | |||
922 | clear_task_ebb(dst); | ||
923 | |||
920 | return 0; | 924 | return 0; |
921 | } | 925 | } |
922 | 926 | ||
diff --git a/arch/powerpc/perf/core-book3s.c b/arch/powerpc/perf/core-book3s.c index c91dc43e04de..a3985aee77fe 100644 --- a/arch/powerpc/perf/core-book3s.c +++ b/arch/powerpc/perf/core-book3s.c | |||
@@ -77,6 +77,9 @@ static unsigned int freeze_events_kernel = MMCR0_FCS; | |||
77 | #define MMCR0_PMCjCE MMCR0_PMCnCE | 77 | #define MMCR0_PMCjCE MMCR0_PMCnCE |
78 | #define MMCR0_FC56 0 | 78 | #define MMCR0_FC56 0 |
79 | #define MMCR0_PMAO 0 | 79 | #define MMCR0_PMAO 0 |
80 | #define MMCR0_EBE 0 | ||
81 | #define MMCR0_PMCC 0 | ||
82 | #define MMCR0_PMCC_U6 0 | ||
80 | 83 | ||
81 | #define SPRN_MMCRA SPRN_MMCR2 | 84 | #define SPRN_MMCRA SPRN_MMCR2 |
82 | #define MMCRA_SAMPLE_ENABLE 0 | 85 | #define MMCRA_SAMPLE_ENABLE 0 |
@@ -104,6 +107,15 @@ static inline int siar_valid(struct pt_regs *regs) | |||
104 | return 1; | 107 | return 1; |
105 | } | 108 | } |
106 | 109 | ||
110 | static bool is_ebb_event(struct perf_event *event) { return false; } | ||
111 | static int ebb_event_check(struct perf_event *event) { return 0; } | ||
112 | static void ebb_event_add(struct perf_event *event) { } | ||
113 | static void ebb_switch_out(unsigned long mmcr0) { } | ||
114 | static unsigned long ebb_switch_in(bool ebb, unsigned long mmcr0) | ||
115 | { | ||
116 | return mmcr0; | ||
117 | } | ||
118 | |||
107 | static inline void power_pmu_bhrb_enable(struct perf_event *event) {} | 119 | static inline void power_pmu_bhrb_enable(struct perf_event *event) {} |
108 | static inline void power_pmu_bhrb_disable(struct perf_event *event) {} | 120 | static inline void power_pmu_bhrb_disable(struct perf_event *event) {} |
109 | void power_pmu_flush_branch_stack(void) {} | 121 | void power_pmu_flush_branch_stack(void) {} |
@@ -464,6 +476,89 @@ void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) | |||
464 | return; | 476 | return; |
465 | } | 477 | } |
466 | 478 | ||
479 | static bool is_ebb_event(struct perf_event *event) | ||
480 | { | ||
481 | /* | ||
482 | * This could be a per-PMU callback, but we'd rather avoid the cost. We | ||
483 | * check that the PMU supports EBB, meaning those that don't can still | ||
484 | * use bit 63 of the event code for something else if they wish. | ||
485 | */ | ||
486 | return (ppmu->flags & PPMU_EBB) && | ||
487 | ((event->attr.config >> EVENT_CONFIG_EBB_SHIFT) & 1); | ||
488 | } | ||
489 | |||
490 | static int ebb_event_check(struct perf_event *event) | ||
491 | { | ||
492 | struct perf_event *leader = event->group_leader; | ||
493 | |||
494 | /* Event and group leader must agree on EBB */ | ||
495 | if (is_ebb_event(leader) != is_ebb_event(event)) | ||
496 | return -EINVAL; | ||
497 | |||
498 | if (is_ebb_event(event)) { | ||
499 | if (!(event->attach_state & PERF_ATTACH_TASK)) | ||
500 | return -EINVAL; | ||
501 | |||
502 | if (!leader->attr.pinned || !leader->attr.exclusive) | ||
503 | return -EINVAL; | ||
504 | |||
505 | if (event->attr.inherit || event->attr.sample_period || | ||
506 | event->attr.enable_on_exec || event->attr.freq) | ||
507 | return -EINVAL; | ||
508 | } | ||
509 | |||
510 | return 0; | ||
511 | } | ||
512 | |||
513 | static void ebb_event_add(struct perf_event *event) | ||
514 | { | ||
515 | if (!is_ebb_event(event) || current->thread.used_ebb) | ||
516 | return; | ||
517 | |||
518 | /* | ||
519 | * IFF this is the first time we've added an EBB event, set | ||
520 | * PMXE in the user MMCR0 so we can detect when it's cleared by | ||
521 | * userspace. We need this so that we can context switch while | ||
522 | * userspace is in the EBB handler (where PMXE is 0). | ||
523 | */ | ||
524 | current->thread.used_ebb = 1; | ||
525 | current->thread.mmcr0 |= MMCR0_PMXE; | ||
526 | } | ||
527 | |||
528 | static void ebb_switch_out(unsigned long mmcr0) | ||
529 | { | ||
530 | if (!(mmcr0 & MMCR0_EBE)) | ||
531 | return; | ||
532 | |||
533 | current->thread.siar = mfspr(SPRN_SIAR); | ||
534 | current->thread.sier = mfspr(SPRN_SIER); | ||
535 | current->thread.sdar = mfspr(SPRN_SDAR); | ||
536 | current->thread.mmcr0 = mmcr0 & MMCR0_USER_MASK; | ||
537 | current->thread.mmcr2 = mfspr(SPRN_MMCR2) & MMCR2_USER_MASK; | ||
538 | } | ||
539 | |||
540 | static unsigned long ebb_switch_in(bool ebb, unsigned long mmcr0) | ||
541 | { | ||
542 | if (!ebb) | ||
543 | goto out; | ||
544 | |||
545 | /* Enable EBB and read/write to all 6 PMCs for userspace */ | ||
546 | mmcr0 |= MMCR0_EBE | MMCR0_PMCC_U6; | ||
547 | |||
548 | /* Add any bits from the user reg, FC or PMAO */ | ||
549 | mmcr0 |= current->thread.mmcr0; | ||
550 | |||
551 | /* Be careful not to set PMXE if userspace had it cleared */ | ||
552 | if (!(current->thread.mmcr0 & MMCR0_PMXE)) | ||
553 | mmcr0 &= ~MMCR0_PMXE; | ||
554 | |||
555 | mtspr(SPRN_SIAR, current->thread.siar); | ||
556 | mtspr(SPRN_SIER, current->thread.sier); | ||
557 | mtspr(SPRN_SDAR, current->thread.sdar); | ||
558 | mtspr(SPRN_MMCR2, current->thread.mmcr2); | ||
559 | out: | ||
560 | return mmcr0; | ||
561 | } | ||
467 | #endif /* CONFIG_PPC64 */ | 562 | #endif /* CONFIG_PPC64 */ |
468 | 563 | ||
469 | static void perf_event_interrupt(struct pt_regs *regs); | 564 | static void perf_event_interrupt(struct pt_regs *regs); |
@@ -734,6 +829,13 @@ static void power_pmu_read(struct perf_event *event) | |||
734 | 829 | ||
735 | if (!event->hw.idx) | 830 | if (!event->hw.idx) |
736 | return; | 831 | return; |
832 | |||
833 | if (is_ebb_event(event)) { | ||
834 | val = read_pmc(event->hw.idx); | ||
835 | local64_set(&event->hw.prev_count, val); | ||
836 | return; | ||
837 | } | ||
838 | |||
737 | /* | 839 | /* |
738 | * Performance monitor interrupts come even when interrupts | 840 | * Performance monitor interrupts come even when interrupts |
739 | * are soft-disabled, as long as interrupts are hard-enabled. | 841 | * are soft-disabled, as long as interrupts are hard-enabled. |
@@ -854,7 +956,7 @@ static void write_mmcr0(struct cpu_hw_events *cpuhw, unsigned long mmcr0) | |||
854 | static void power_pmu_disable(struct pmu *pmu) | 956 | static void power_pmu_disable(struct pmu *pmu) |
855 | { | 957 | { |
856 | struct cpu_hw_events *cpuhw; | 958 | struct cpu_hw_events *cpuhw; |
857 | unsigned long flags, val; | 959 | unsigned long flags, mmcr0, val; |
858 | 960 | ||
859 | if (!ppmu) | 961 | if (!ppmu) |
860 | return; | 962 | return; |
@@ -871,11 +973,11 @@ static void power_pmu_disable(struct pmu *pmu) | |||
871 | } | 973 | } |
872 | 974 | ||
873 | /* | 975 | /* |
874 | * Set the 'freeze counters' bit, clear PMAO/FC56. | 976 | * Set the 'freeze counters' bit, clear EBE/PMCC/PMAO/FC56. |
875 | */ | 977 | */ |
876 | val = mfspr(SPRN_MMCR0); | 978 | val = mmcr0 = mfspr(SPRN_MMCR0); |
877 | val |= MMCR0_FC; | 979 | val |= MMCR0_FC; |
878 | val &= ~(MMCR0_PMAO | MMCR0_FC56); | 980 | val &= ~(MMCR0_EBE | MMCR0_PMCC | MMCR0_PMAO | MMCR0_FC56); |
879 | 981 | ||
880 | /* | 982 | /* |
881 | * The barrier is to make sure the mtspr has been | 983 | * The barrier is to make sure the mtspr has been |
@@ -896,7 +998,10 @@ static void power_pmu_disable(struct pmu *pmu) | |||
896 | 998 | ||
897 | cpuhw->disabled = 1; | 999 | cpuhw->disabled = 1; |
898 | cpuhw->n_added = 0; | 1000 | cpuhw->n_added = 0; |
1001 | |||
1002 | ebb_switch_out(mmcr0); | ||
899 | } | 1003 | } |
1004 | |||
900 | local_irq_restore(flags); | 1005 | local_irq_restore(flags); |
901 | } | 1006 | } |
902 | 1007 | ||
@@ -911,15 +1016,15 @@ static void power_pmu_enable(struct pmu *pmu) | |||
911 | struct cpu_hw_events *cpuhw; | 1016 | struct cpu_hw_events *cpuhw; |
912 | unsigned long flags; | 1017 | unsigned long flags; |
913 | long i; | 1018 | long i; |
914 | unsigned long val; | 1019 | unsigned long val, mmcr0; |
915 | s64 left; | 1020 | s64 left; |
916 | unsigned int hwc_index[MAX_HWEVENTS]; | 1021 | unsigned int hwc_index[MAX_HWEVENTS]; |
917 | int n_lim; | 1022 | int n_lim; |
918 | int idx; | 1023 | int idx; |
1024 | bool ebb; | ||
919 | 1025 | ||
920 | if (!ppmu) | 1026 | if (!ppmu) |
921 | return; | 1027 | return; |
922 | |||
923 | local_irq_save(flags); | 1028 | local_irq_save(flags); |
924 | 1029 | ||
925 | cpuhw = &__get_cpu_var(cpu_hw_events); | 1030 | cpuhw = &__get_cpu_var(cpu_hw_events); |
@@ -934,6 +1039,13 @@ static void power_pmu_enable(struct pmu *pmu) | |||
934 | cpuhw->disabled = 0; | 1039 | cpuhw->disabled = 0; |
935 | 1040 | ||
936 | /* | 1041 | /* |
1042 | * EBB requires an exclusive group and all events must have the EBB | ||
1043 | * flag set, or not set, so we can just check a single event. Also we | ||
1044 | * know we have at least one event. | ||
1045 | */ | ||
1046 | ebb = is_ebb_event(cpuhw->event[0]); | ||
1047 | |||
1048 | /* | ||
937 | * If we didn't change anything, or only removed events, | 1049 | * If we didn't change anything, or only removed events, |
938 | * no need to recalculate MMCR* settings and reset the PMCs. | 1050 | * no need to recalculate MMCR* settings and reset the PMCs. |
939 | * Just reenable the PMU with the current MMCR* settings | 1051 | * Just reenable the PMU with the current MMCR* settings |
@@ -1008,25 +1120,34 @@ static void power_pmu_enable(struct pmu *pmu) | |||
1008 | ++n_lim; | 1120 | ++n_lim; |
1009 | continue; | 1121 | continue; |
1010 | } | 1122 | } |
1011 | val = 0; | 1123 | |
1012 | if (event->hw.sample_period) { | 1124 | if (ebb) |
1013 | left = local64_read(&event->hw.period_left); | 1125 | val = local64_read(&event->hw.prev_count); |
1014 | if (left < 0x80000000L) | 1126 | else { |
1015 | val = 0x80000000L - left; | 1127 | val = 0; |
1128 | if (event->hw.sample_period) { | ||
1129 | left = local64_read(&event->hw.period_left); | ||
1130 | if (left < 0x80000000L) | ||
1131 | val = 0x80000000L - left; | ||
1132 | } | ||
1133 | local64_set(&event->hw.prev_count, val); | ||
1016 | } | 1134 | } |
1017 | local64_set(&event->hw.prev_count, val); | 1135 | |
1018 | event->hw.idx = idx; | 1136 | event->hw.idx = idx; |
1019 | if (event->hw.state & PERF_HES_STOPPED) | 1137 | if (event->hw.state & PERF_HES_STOPPED) |
1020 | val = 0; | 1138 | val = 0; |
1021 | write_pmc(idx, val); | 1139 | write_pmc(idx, val); |
1140 | |||
1022 | perf_event_update_userpage(event); | 1141 | perf_event_update_userpage(event); |
1023 | } | 1142 | } |
1024 | cpuhw->n_limited = n_lim; | 1143 | cpuhw->n_limited = n_lim; |
1025 | cpuhw->mmcr[0] |= MMCR0_PMXE | MMCR0_FCECE; | 1144 | cpuhw->mmcr[0] |= MMCR0_PMXE | MMCR0_FCECE; |
1026 | 1145 | ||
1027 | out_enable: | 1146 | out_enable: |
1147 | mmcr0 = ebb_switch_in(ebb, cpuhw->mmcr[0]); | ||
1148 | |||
1028 | mb(); | 1149 | mb(); |
1029 | write_mmcr0(cpuhw, cpuhw->mmcr[0]); | 1150 | write_mmcr0(cpuhw, mmcr0); |
1030 | 1151 | ||
1031 | /* | 1152 | /* |
1032 | * Enable instruction sampling if necessary | 1153 | * Enable instruction sampling if necessary |
@@ -1124,6 +1245,8 @@ static int power_pmu_add(struct perf_event *event, int ef_flags) | |||
1124 | event->hw.config = cpuhw->events[n0]; | 1245 | event->hw.config = cpuhw->events[n0]; |
1125 | 1246 | ||
1126 | nocheck: | 1247 | nocheck: |
1248 | ebb_event_add(event); | ||
1249 | |||
1127 | ++cpuhw->n_events; | 1250 | ++cpuhw->n_events; |
1128 | ++cpuhw->n_added; | 1251 | ++cpuhw->n_added; |
1129 | 1252 | ||
@@ -1484,6 +1607,11 @@ static int power_pmu_event_init(struct perf_event *event) | |||
1484 | } | 1607 | } |
1485 | } | 1608 | } |
1486 | 1609 | ||
1610 | /* Extra checks for EBB */ | ||
1611 | err = ebb_event_check(event); | ||
1612 | if (err) | ||
1613 | return err; | ||
1614 | |||
1487 | /* | 1615 | /* |
1488 | * If this is in a group, check if it can go on with all the | 1616 | * If this is in a group, check if it can go on with all the |
1489 | * other hardware events in the group. We assume the event | 1617 | * other hardware events in the group. We assume the event |
@@ -1523,6 +1651,13 @@ static int power_pmu_event_init(struct perf_event *event) | |||
1523 | local64_set(&event->hw.period_left, event->hw.last_period); | 1651 | local64_set(&event->hw.period_left, event->hw.last_period); |
1524 | 1652 | ||
1525 | /* | 1653 | /* |
1654 | * For EBB events we just context switch the PMC value, we don't do any | ||
1655 | * of the sample_period logic. We use hw.prev_count for this. | ||
1656 | */ | ||
1657 | if (is_ebb_event(event)) | ||
1658 | local64_set(&event->hw.prev_count, 0); | ||
1659 | |||
1660 | /* | ||
1526 | * See if we need to reserve the PMU. | 1661 | * See if we need to reserve the PMU. |
1527 | * If no events are currently in use, then we have to take a | 1662 | * If no events are currently in use, then we have to take a |
1528 | * mutex to ensure that we don't race with another task doing | 1663 | * mutex to ensure that we don't race with another task doing |