aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/trace/trace_event_perf.c
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/trace/trace_event_perf.c')
-rw-r--r--kernel/trace/trace_event_perf.c19
1 files changed, 15 insertions, 4 deletions
diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c
index cb6f365016e4..8a2b73f7c068 100644
--- a/kernel/trace/trace_event_perf.c
+++ b/kernel/trace/trace_event_perf.c
@@ -96,7 +96,9 @@ int perf_trace_init(struct perf_event *p_event)
96 mutex_lock(&event_mutex); 96 mutex_lock(&event_mutex);
97 list_for_each_entry(tp_event, &ftrace_events, list) { 97 list_for_each_entry(tp_event, &ftrace_events, list) {
98 if (tp_event->event.type == event_id && 98 if (tp_event->event.type == event_id &&
99 tp_event->class && tp_event->class->perf_probe && 99 tp_event->class &&
100 (tp_event->class->perf_probe ||
101 tp_event->class->reg) &&
100 try_module_get(tp_event->mod)) { 102 try_module_get(tp_event->mod)) {
101 ret = perf_trace_event_init(tp_event, p_event); 103 ret = perf_trace_event_init(tp_event, p_event);
102 break; 104 break;
@@ -116,7 +118,7 @@ int perf_trace_enable(struct perf_event *p_event)
116 if (WARN_ON_ONCE(!list)) 118 if (WARN_ON_ONCE(!list))
117 return -EINVAL; 119 return -EINVAL;
118 120
119 list = per_cpu_ptr(list, smp_processor_id()); 121 list = this_cpu_ptr(list);
120 hlist_add_head_rcu(&p_event->hlist_entry, list); 122 hlist_add_head_rcu(&p_event->hlist_entry, list);
121 123
122 return 0; 124 return 0;
@@ -132,8 +134,9 @@ void perf_trace_destroy(struct perf_event *p_event)
132 struct ftrace_event_call *tp_event = p_event->tp_event; 134 struct ftrace_event_call *tp_event = p_event->tp_event;
133 int i; 135 int i;
134 136
137 mutex_lock(&event_mutex);
135 if (--tp_event->perf_refcount > 0) 138 if (--tp_event->perf_refcount > 0)
136 return; 139 goto out;
137 140
138 if (tp_event->class->reg) 141 if (tp_event->class->reg)
139 tp_event->class->reg(tp_event, TRACE_REG_PERF_UNREGISTER); 142 tp_event->class->reg(tp_event, TRACE_REG_PERF_UNREGISTER);
@@ -142,6 +145,12 @@ void perf_trace_destroy(struct perf_event *p_event)
142 tp_event->class->perf_probe, 145 tp_event->class->perf_probe,
143 tp_event); 146 tp_event);
144 147
148 /*
149 * Ensure our callback won't be called anymore. See
150 * tracepoint_probe_unregister() and __DO_TRACE().
151 */
152 synchronize_sched();
153
145 free_percpu(tp_event->perf_events); 154 free_percpu(tp_event->perf_events);
146 tp_event->perf_events = NULL; 155 tp_event->perf_events = NULL;
147 156
@@ -151,6 +160,8 @@ void perf_trace_destroy(struct perf_event *p_event)
151 perf_trace_buf[i] = NULL; 160 perf_trace_buf[i] = NULL;
152 } 161 }
153 } 162 }
163out:
164 mutex_unlock(&event_mutex);
154} 165}
155 166
156__kprobes void *perf_trace_buf_prepare(int size, unsigned short type, 167__kprobes void *perf_trace_buf_prepare(int size, unsigned short type,
@@ -169,7 +180,7 @@ __kprobes void *perf_trace_buf_prepare(int size, unsigned short type,
169 if (*rctxp < 0) 180 if (*rctxp < 0)
170 return NULL; 181 return NULL;
171 182
172 raw_data = per_cpu_ptr(perf_trace_buf[*rctxp], smp_processor_id()); 183 raw_data = this_cpu_ptr(perf_trace_buf[*rctxp]);
173 184
174 /* zero the dead bytes from align to not leak stack to user */ 185 /* zero the dead bytes from align to not leak stack to user */
175 memset(&raw_data[size - sizeof(u64)], 0, sizeof(u64)); 186 memset(&raw_data[size - sizeof(u64)], 0, sizeof(u64));