aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/trace/trace_sched_switch.c
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/trace/trace_sched_switch.c')
-rw-r--r--kernel/trace/trace_sched_switch.c56
1 files changed, 0 insertions, 56 deletions
diff --git a/kernel/trace/trace_sched_switch.c b/kernel/trace/trace_sched_switch.c
index f7c7f4f1544c..2e293beb186e 100644
--- a/kernel/trace/trace_sched_switch.c
+++ b/kernel/trace/trace_sched_switch.c
@@ -17,34 +17,6 @@
17static int sched_ref; 17static int sched_ref;
18static DEFINE_MUTEX(sched_register_mutex); 18static DEFINE_MUTEX(sched_register_mutex);
19 19
20void
21tracing_sched_switch_trace(struct trace_array *tr,
22 struct task_struct *prev,
23 struct task_struct *next,
24 unsigned long flags, int pc)
25{
26 struct ftrace_event_call *call = &event_context_switch;
27 struct ring_buffer *buffer = tr->trace_buffer.buffer;
28 struct ring_buffer_event *event;
29 struct ctx_switch_entry *entry;
30
31 event = trace_buffer_lock_reserve(buffer, TRACE_CTX,
32 sizeof(*entry), flags, pc);
33 if (!event)
34 return;
35 entry = ring_buffer_event_data(event);
36 entry->prev_pid = prev->pid;
37 entry->prev_prio = prev->prio;
38 entry->prev_state = prev->state;
39 entry->next_pid = next->pid;
40 entry->next_prio = next->prio;
41 entry->next_state = next->state;
42 entry->next_cpu = task_cpu(next);
43
44 if (!call_filter_check_discard(call, entry, buffer, event))
45 trace_buffer_unlock_commit(buffer, event, flags, pc);
46}
47
48static void 20static void
49probe_sched_switch(void *ignore, struct task_struct *prev, struct task_struct *next) 21probe_sched_switch(void *ignore, struct task_struct *prev, struct task_struct *next)
50{ 22{
@@ -55,34 +27,6 @@ probe_sched_switch(void *ignore, struct task_struct *prev, struct task_struct *n
55 tracing_record_cmdline(next); 27 tracing_record_cmdline(next);
56} 28}
57 29
58void
59tracing_sched_wakeup_trace(struct trace_array *tr,
60 struct task_struct *wakee,
61 struct task_struct *curr,
62 unsigned long flags, int pc)
63{
64 struct ftrace_event_call *call = &event_wakeup;
65 struct ring_buffer_event *event;
66 struct ctx_switch_entry *entry;
67 struct ring_buffer *buffer = tr->trace_buffer.buffer;
68
69 event = trace_buffer_lock_reserve(buffer, TRACE_WAKE,
70 sizeof(*entry), flags, pc);
71 if (!event)
72 return;
73 entry = ring_buffer_event_data(event);
74 entry->prev_pid = curr->pid;
75 entry->prev_prio = curr->prio;
76 entry->prev_state = curr->state;
77 entry->next_pid = wakee->pid;
78 entry->next_prio = wakee->prio;
79 entry->next_state = wakee->state;
80 entry->next_cpu = task_cpu(wakee);
81
82 if (!call_filter_check_discard(call, entry, buffer, event))
83 trace_buffer_unlock_commit(buffer, event, flags, pc);
84}
85
86static void 30static void
87probe_sched_wakeup(void *ignore, struct task_struct *wakee, int success) 31probe_sched_wakeup(void *ignore, struct task_struct *wakee, int success)
88{ 32{