diff options
author | Steven Rostedt (Red Hat) <rostedt@goodmis.org> | 2014-04-29 17:54:37 -0400 |
---|---|---|
committer | Steven Rostedt <rostedt@goodmis.org> | 2014-04-30 08:40:05 -0400 |
commit | b1169cc69ba96b124df820904a6d3eb775491d7f (patch) | |
tree | 95927d3a0c262c94def2d76b7f591e9d8ad46385 /kernel/trace/trace_functions_graph.c | |
parent | f4874261049e3abdd481359d82cafa5068369ebd (diff) |
tracing: Remove mock up poll wait function
Now that the ring buffer has a built in way to wake up readers
when there's data, using irq_work such that it is safe to do it
in any context. But it was still using the old "poor man's"
wait polling that checks every 1/10 of a second to see if it
should wake up a waiter. This makes the latency for a wake up
excruciatingly long. No need to do that anymore.
Completely remove the different wait_poll types from the tracers
and have them all use the default one now.
Reported-by: Johannes Berg <johannes@sipsolutions.net>
Signed-off-by: Steven Rostedt <rostedt@goodmis.org>
Diffstat (limited to 'kernel/trace/trace_functions_graph.c')
-rw-r--r-- | kernel/trace/trace_functions_graph.c | 1 |
1 files changed, 0 insertions, 1 deletions
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index deff11200261..b86dd4d8c6a6 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
@@ -1505,7 +1505,6 @@ static struct tracer graph_trace __tracer_data = { | |||
1505 | .pipe_open = graph_trace_open, | 1505 | .pipe_open = graph_trace_open, |
1506 | .close = graph_trace_close, | 1506 | .close = graph_trace_close, |
1507 | .pipe_close = graph_trace_close, | 1507 | .pipe_close = graph_trace_close, |
1508 | .wait_pipe = poll_wait_pipe, | ||
1509 | .init = graph_trace_init, | 1508 | .init = graph_trace_init, |
1510 | .reset = graph_trace_reset, | 1509 | .reset = graph_trace_reset, |
1511 | .print_line = print_graph_function, | 1510 | .print_line = print_graph_function, |