diff options
Diffstat (limited to 'arch/powerpc/perf/core-book3s.c')
-rw-r--r-- | arch/powerpc/perf/core-book3s.c | 18 |
1 files changed, 9 insertions, 9 deletions
diff --git a/arch/powerpc/perf/core-book3s.c b/arch/powerpc/perf/core-book3s.c index b7cd00b0171e..a6995d4e93d4 100644 --- a/arch/powerpc/perf/core-book3s.c +++ b/arch/powerpc/perf/core-book3s.c | |||
@@ -59,9 +59,9 @@ struct cpu_hw_events { | |||
59 | struct perf_branch_entry bhrb_entries[BHRB_MAX_ENTRIES]; | 59 | struct perf_branch_entry bhrb_entries[BHRB_MAX_ENTRIES]; |
60 | }; | 60 | }; |
61 | 61 | ||
62 | DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); | 62 | static DEFINE_PER_CPU(struct cpu_hw_events, cpu_hw_events); |
63 | 63 | ||
64 | struct power_pmu *ppmu; | 64 | static struct power_pmu *ppmu; |
65 | 65 | ||
66 | /* | 66 | /* |
67 | * Normally, to ignore kernel events we set the FCS (freeze counters | 67 | * Normally, to ignore kernel events we set the FCS (freeze counters |
@@ -124,7 +124,7 @@ static unsigned long ebb_switch_in(bool ebb, struct cpu_hw_events *cpuhw) | |||
124 | 124 | ||
125 | static inline void power_pmu_bhrb_enable(struct perf_event *event) {} | 125 | static inline void power_pmu_bhrb_enable(struct perf_event *event) {} |
126 | static inline void power_pmu_bhrb_disable(struct perf_event *event) {} | 126 | static inline void power_pmu_bhrb_disable(struct perf_event *event) {} |
127 | void power_pmu_flush_branch_stack(void) {} | 127 | static void power_pmu_flush_branch_stack(void) {} |
128 | static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} | 128 | static inline void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) {} |
129 | static void pmao_restore_workaround(bool ebb) { } | 129 | static void pmao_restore_workaround(bool ebb) { } |
130 | #endif /* CONFIG_PPC32 */ | 130 | #endif /* CONFIG_PPC32 */ |
@@ -375,7 +375,7 @@ static void power_pmu_bhrb_disable(struct perf_event *event) | |||
375 | /* Called from ctxsw to prevent one process's branch entries to | 375 | /* Called from ctxsw to prevent one process's branch entries to |
376 | * mingle with the other process's entries during context switch. | 376 | * mingle with the other process's entries during context switch. |
377 | */ | 377 | */ |
378 | void power_pmu_flush_branch_stack(void) | 378 | static void power_pmu_flush_branch_stack(void) |
379 | { | 379 | { |
380 | if (ppmu->bhrb_nr) | 380 | if (ppmu->bhrb_nr) |
381 | power_pmu_bhrb_reset(); | 381 | power_pmu_bhrb_reset(); |
@@ -408,7 +408,7 @@ static __u64 power_pmu_bhrb_to(u64 addr) | |||
408 | } | 408 | } |
409 | 409 | ||
410 | /* Processing BHRB entries */ | 410 | /* Processing BHRB entries */ |
411 | void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) | 411 | static void power_pmu_bhrb_read(struct cpu_hw_events *cpuhw) |
412 | { | 412 | { |
413 | u64 val; | 413 | u64 val; |
414 | u64 addr; | 414 | u64 addr; |
@@ -1573,7 +1573,7 @@ static void power_pmu_stop(struct perf_event *event, int ef_flags) | |||
1573 | * Set the flag to make pmu::enable() not perform the | 1573 | * Set the flag to make pmu::enable() not perform the |
1574 | * schedulability test, it will be performed at commit time | 1574 | * schedulability test, it will be performed at commit time |
1575 | */ | 1575 | */ |
1576 | void power_pmu_start_txn(struct pmu *pmu) | 1576 | static void power_pmu_start_txn(struct pmu *pmu) |
1577 | { | 1577 | { |
1578 | struct cpu_hw_events *cpuhw = &__get_cpu_var(cpu_hw_events); | 1578 | struct cpu_hw_events *cpuhw = &__get_cpu_var(cpu_hw_events); |
1579 | 1579 | ||
@@ -1587,7 +1587,7 @@ void power_pmu_start_txn(struct pmu *pmu) | |||
1587 | * Clear the flag and pmu::enable() will perform the | 1587 | * Clear the flag and pmu::enable() will perform the |
1588 | * schedulability test. | 1588 | * schedulability test. |
1589 | */ | 1589 | */ |
1590 | void power_pmu_cancel_txn(struct pmu *pmu) | 1590 | static void power_pmu_cancel_txn(struct pmu *pmu) |
1591 | { | 1591 | { |
1592 | struct cpu_hw_events *cpuhw = &__get_cpu_var(cpu_hw_events); | 1592 | struct cpu_hw_events *cpuhw = &__get_cpu_var(cpu_hw_events); |
1593 | 1593 | ||
@@ -1600,7 +1600,7 @@ void power_pmu_cancel_txn(struct pmu *pmu) | |||
1600 | * Perform the group schedulability test as a whole | 1600 | * Perform the group schedulability test as a whole |
1601 | * Return 0 if success | 1601 | * Return 0 if success |
1602 | */ | 1602 | */ |
1603 | int power_pmu_commit_txn(struct pmu *pmu) | 1603 | static int power_pmu_commit_txn(struct pmu *pmu) |
1604 | { | 1604 | { |
1605 | struct cpu_hw_events *cpuhw; | 1605 | struct cpu_hw_events *cpuhw; |
1606 | long i, n; | 1606 | long i, n; |
@@ -1888,7 +1888,7 @@ ssize_t power_events_sysfs_show(struct device *dev, | |||
1888 | return sprintf(page, "event=0x%02llx\n", pmu_attr->id); | 1888 | return sprintf(page, "event=0x%02llx\n", pmu_attr->id); |
1889 | } | 1889 | } |
1890 | 1890 | ||
1891 | struct pmu power_pmu = { | 1891 | static struct pmu power_pmu = { |
1892 | .pmu_enable = power_pmu_enable, | 1892 | .pmu_enable = power_pmu_enable, |
1893 | .pmu_disable = power_pmu_disable, | 1893 | .pmu_disable = power_pmu_disable, |
1894 | .event_init = power_pmu_event_init, | 1894 | .event_init = power_pmu_event_init, |