aboutsummaryrefslogtreecommitdiffstats
path: root/include/linux/perf_event.h
diff options
context:
space:
mode:
authorGleb Natapov <gleb@redhat.com>2011-11-27 10:59:09 -0500
committerIngo Molnar <mingo@elte.hu>2011-12-06 02:34:02 -0500
commitb202952075f62603bea9bfb6ebc6b0420db11949 (patch)
tree9c8e0538b455e68b5c371caba5b1585ed0ef9d8a /include/linux/perf_event.h
parentb79387ef185af2323594920923cecba5753c3817 (diff)
perf, core: Rate limit perf_sched_events jump_label patching
jump_lable patching is very expensive operation that involves pausing all cpus. The patching of perf_sched_events jump_label is easily controllable from userspace by unprivileged user. When te user runs a loop like this: "while true; do perf stat -e cycles true; done" ... the performance of my test application that just increments a counter for one second drops by 4%. This is on a 16 cpu box with my test application using only one of them. An impact on a real server doing real work will be worse. Performance of KVM PMU drops nearly 50% due to jump_lable for "perf record" since KVM PMU implementation creates and destroys perf event frequently. This patch introduces a way to rate limit jump_label patching and uses it to fix the above problem. I believe that as jump_label use will spread the problem will become more common and thus solving it in a generic code is appropriate. Also fixing it in the perf code would result in moving jump_label accounting logic to perf code with all the ifdefs in case of JUMP_LABEL=n kernel. With this patch all details are nicely hidden inside jump_label code. Signed-off-by: Gleb Natapov <gleb@redhat.com> Acked-by: Jason Baron <jbaron@redhat.com> Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Link: http://lkml.kernel.org/r/20111127155909.GO2557@redhat.com Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'include/linux/perf_event.h')
-rw-r--r--include/linux/perf_event.h6
1 files changed, 3 insertions, 3 deletions
diff --git a/include/linux/perf_event.h b/include/linux/perf_event.h
index cb44c9e75660..564769cdb473 100644
--- a/include/linux/perf_event.h
+++ b/include/linux/perf_event.h
@@ -1064,12 +1064,12 @@ perf_sw_event(u32 event_id, u64 nr, struct pt_regs *regs, u64 addr)
1064 } 1064 }
1065} 1065}
1066 1066
1067extern struct jump_label_key perf_sched_events; 1067extern struct jump_label_key_deferred perf_sched_events;
1068 1068
1069static inline void perf_event_task_sched_in(struct task_struct *prev, 1069static inline void perf_event_task_sched_in(struct task_struct *prev,
1070 struct task_struct *task) 1070 struct task_struct *task)
1071{ 1071{
1072 if (static_branch(&perf_sched_events)) 1072 if (static_branch(&perf_sched_events.key))
1073 __perf_event_task_sched_in(prev, task); 1073 __perf_event_task_sched_in(prev, task);
1074} 1074}
1075 1075
@@ -1078,7 +1078,7 @@ static inline void perf_event_task_sched_out(struct task_struct *prev,
1078{ 1078{
1079 perf_sw_event(PERF_COUNT_SW_CONTEXT_SWITCHES, 1, NULL, 0); 1079 perf_sw_event(PERF_COUNT_SW_CONTEXT_SWITCHES, 1, NULL, 0);
1080 1080
1081 if (static_branch(&perf_sched_events)) 1081 if (static_branch(&perf_sched_events.key))
1082 __perf_event_task_sched_out(prev, next); 1082 __perf_event_task_sched_out(prev, next);
1083} 1083}
1084 1084