diff options
Diffstat (limited to 'kernel/trace/trace.h')
-rw-r--r-- | kernel/trace/trace.h | 107 |
1 files changed, 76 insertions, 31 deletions
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 405cb850b75d..4df6a77eb196 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
@@ -11,6 +11,7 @@ | |||
11 | #include <linux/ftrace.h> | 11 | #include <linux/ftrace.h> |
12 | #include <trace/boot.h> | 12 | #include <trace/boot.h> |
13 | #include <linux/kmemtrace.h> | 13 | #include <linux/kmemtrace.h> |
14 | #include <linux/hw_breakpoint.h> | ||
14 | 15 | ||
15 | #include <linux/trace_seq.h> | 16 | #include <linux/trace_seq.h> |
16 | #include <linux/ftrace_event.h> | 17 | #include <linux/ftrace_event.h> |
@@ -37,6 +38,7 @@ enum trace_type { | |||
37 | TRACE_KMEM_ALLOC, | 38 | TRACE_KMEM_ALLOC, |
38 | TRACE_KMEM_FREE, | 39 | TRACE_KMEM_FREE, |
39 | TRACE_BLK, | 40 | TRACE_BLK, |
41 | TRACE_KSYM, | ||
40 | 42 | ||
41 | __TRACE_LAST_TYPE, | 43 | __TRACE_LAST_TYPE, |
42 | }; | 44 | }; |
@@ -98,9 +100,32 @@ struct syscall_trace_enter { | |||
98 | struct syscall_trace_exit { | 100 | struct syscall_trace_exit { |
99 | struct trace_entry ent; | 101 | struct trace_entry ent; |
100 | int nr; | 102 | int nr; |
101 | unsigned long ret; | 103 | long ret; |
102 | }; | 104 | }; |
103 | 105 | ||
106 | struct kprobe_trace_entry { | ||
107 | struct trace_entry ent; | ||
108 | unsigned long ip; | ||
109 | int nargs; | ||
110 | unsigned long args[]; | ||
111 | }; | ||
112 | |||
113 | #define SIZEOF_KPROBE_TRACE_ENTRY(n) \ | ||
114 | (offsetof(struct kprobe_trace_entry, args) + \ | ||
115 | (sizeof(unsigned long) * (n))) | ||
116 | |||
117 | struct kretprobe_trace_entry { | ||
118 | struct trace_entry ent; | ||
119 | unsigned long func; | ||
120 | unsigned long ret_ip; | ||
121 | int nargs; | ||
122 | unsigned long args[]; | ||
123 | }; | ||
124 | |||
125 | #define SIZEOF_KRETPROBE_TRACE_ENTRY(n) \ | ||
126 | (offsetof(struct kretprobe_trace_entry, args) + \ | ||
127 | (sizeof(unsigned long) * (n))) | ||
128 | |||
104 | /* | 129 | /* |
105 | * trace_flag_type is an enumeration that holds different | 130 | * trace_flag_type is an enumeration that holds different |
106 | * states when a trace occurs. These are: | 131 | * states when a trace occurs. These are: |
@@ -209,6 +234,7 @@ extern void __ftrace_bad_type(void); | |||
209 | TRACE_KMEM_ALLOC); \ | 234 | TRACE_KMEM_ALLOC); \ |
210 | IF_ASSIGN(var, ent, struct kmemtrace_free_entry, \ | 235 | IF_ASSIGN(var, ent, struct kmemtrace_free_entry, \ |
211 | TRACE_KMEM_FREE); \ | 236 | TRACE_KMEM_FREE); \ |
237 | IF_ASSIGN(var, ent, struct ksym_trace_entry, TRACE_KSYM);\ | ||
212 | __ftrace_bad_type(); \ | 238 | __ftrace_bad_type(); \ |
213 | } while (0) | 239 | } while (0) |
214 | 240 | ||
@@ -246,6 +272,7 @@ struct tracer_flags { | |||
246 | * @pipe_open: called when the trace_pipe file is opened | 272 | * @pipe_open: called when the trace_pipe file is opened |
247 | * @wait_pipe: override how the user waits for traces on trace_pipe | 273 | * @wait_pipe: override how the user waits for traces on trace_pipe |
248 | * @close: called when the trace file is released | 274 | * @close: called when the trace file is released |
275 | * @pipe_close: called when the trace_pipe file is released | ||
249 | * @read: override the default read callback on trace_pipe | 276 | * @read: override the default read callback on trace_pipe |
250 | * @splice_read: override the default splice_read callback on trace_pipe | 277 | * @splice_read: override the default splice_read callback on trace_pipe |
251 | * @selftest: selftest to run on boot (see trace_selftest.c) | 278 | * @selftest: selftest to run on boot (see trace_selftest.c) |
@@ -264,6 +291,7 @@ struct tracer { | |||
264 | void (*pipe_open)(struct trace_iterator *iter); | 291 | void (*pipe_open)(struct trace_iterator *iter); |
265 | void (*wait_pipe)(struct trace_iterator *iter); | 292 | void (*wait_pipe)(struct trace_iterator *iter); |
266 | void (*close)(struct trace_iterator *iter); | 293 | void (*close)(struct trace_iterator *iter); |
294 | void (*pipe_close)(struct trace_iterator *iter); | ||
267 | ssize_t (*read)(struct trace_iterator *iter, | 295 | ssize_t (*read)(struct trace_iterator *iter, |
268 | struct file *filp, char __user *ubuf, | 296 | struct file *filp, char __user *ubuf, |
269 | size_t cnt, loff_t *ppos); | 297 | size_t cnt, loff_t *ppos); |
@@ -364,6 +392,8 @@ int register_tracer(struct tracer *type); | |||
364 | void unregister_tracer(struct tracer *type); | 392 | void unregister_tracer(struct tracer *type); |
365 | int is_tracing_stopped(void); | 393 | int is_tracing_stopped(void); |
366 | 394 | ||
395 | extern int process_new_ksym_entry(char *ksymname, int op, unsigned long addr); | ||
396 | |||
367 | extern unsigned long nsecs_to_usecs(unsigned long nsecs); | 397 | extern unsigned long nsecs_to_usecs(unsigned long nsecs); |
368 | 398 | ||
369 | #ifdef CONFIG_TRACER_MAX_TRACE | 399 | #ifdef CONFIG_TRACER_MAX_TRACE |
@@ -413,7 +443,7 @@ extern int DYN_FTRACE_TEST_NAME(void); | |||
413 | 443 | ||
414 | extern int ring_buffer_expanded; | 444 | extern int ring_buffer_expanded; |
415 | extern bool tracing_selftest_disabled; | 445 | extern bool tracing_selftest_disabled; |
416 | DECLARE_PER_CPU(local_t, ftrace_cpu_disabled); | 446 | DECLARE_PER_CPU(int, ftrace_cpu_disabled); |
417 | 447 | ||
418 | #ifdef CONFIG_FTRACE_STARTUP_TEST | 448 | #ifdef CONFIG_FTRACE_STARTUP_TEST |
419 | extern int trace_selftest_startup_function(struct tracer *trace, | 449 | extern int trace_selftest_startup_function(struct tracer *trace, |
@@ -438,6 +468,8 @@ extern int trace_selftest_startup_branch(struct tracer *trace, | |||
438 | struct trace_array *tr); | 468 | struct trace_array *tr); |
439 | extern int trace_selftest_startup_hw_branches(struct tracer *trace, | 469 | extern int trace_selftest_startup_hw_branches(struct tracer *trace, |
440 | struct trace_array *tr); | 470 | struct trace_array *tr); |
471 | extern int trace_selftest_startup_ksym(struct tracer *trace, | ||
472 | struct trace_array *tr); | ||
441 | #endif /* CONFIG_FTRACE_STARTUP_TEST */ | 473 | #endif /* CONFIG_FTRACE_STARTUP_TEST */ |
442 | 474 | ||
443 | extern void *head_page(struct trace_array_cpu *data); | 475 | extern void *head_page(struct trace_array_cpu *data); |
@@ -483,10 +515,6 @@ static inline int ftrace_graph_addr(unsigned long addr) | |||
483 | return 0; | 515 | return 0; |
484 | } | 516 | } |
485 | #else | 517 | #else |
486 | static inline int ftrace_trace_addr(unsigned long addr) | ||
487 | { | ||
488 | return 1; | ||
489 | } | ||
490 | static inline int ftrace_graph_addr(unsigned long addr) | 518 | static inline int ftrace_graph_addr(unsigned long addr) |
491 | { | 519 | { |
492 | return 1; | 520 | return 1; |
@@ -500,12 +528,12 @@ print_graph_function(struct trace_iterator *iter) | |||
500 | } | 528 | } |
501 | #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ | 529 | #endif /* CONFIG_FUNCTION_GRAPH_TRACER */ |
502 | 530 | ||
503 | extern struct pid *ftrace_pid_trace; | 531 | extern struct list_head ftrace_pids; |
504 | 532 | ||
505 | #ifdef CONFIG_FUNCTION_TRACER | 533 | #ifdef CONFIG_FUNCTION_TRACER |
506 | static inline int ftrace_trace_task(struct task_struct *task) | 534 | static inline int ftrace_trace_task(struct task_struct *task) |
507 | { | 535 | { |
508 | if (!ftrace_pid_trace) | 536 | if (list_empty(&ftrace_pids)) |
509 | return 1; | 537 | return 1; |
510 | 538 | ||
511 | return test_tsk_trace_trace(task); | 539 | return test_tsk_trace_trace(task); |
@@ -569,18 +597,17 @@ enum trace_iterator_flags { | |||
569 | TRACE_ITER_BIN = 0x40, | 597 | TRACE_ITER_BIN = 0x40, |
570 | TRACE_ITER_BLOCK = 0x80, | 598 | TRACE_ITER_BLOCK = 0x80, |
571 | TRACE_ITER_STACKTRACE = 0x100, | 599 | TRACE_ITER_STACKTRACE = 0x100, |
572 | TRACE_ITER_SCHED_TREE = 0x200, | 600 | TRACE_ITER_PRINTK = 0x200, |
573 | TRACE_ITER_PRINTK = 0x400, | 601 | TRACE_ITER_PREEMPTONLY = 0x400, |
574 | TRACE_ITER_PREEMPTONLY = 0x800, | 602 | TRACE_ITER_BRANCH = 0x800, |
575 | TRACE_ITER_BRANCH = 0x1000, | 603 | TRACE_ITER_ANNOTATE = 0x1000, |
576 | TRACE_ITER_ANNOTATE = 0x2000, | 604 | TRACE_ITER_USERSTACKTRACE = 0x2000, |
577 | TRACE_ITER_USERSTACKTRACE = 0x4000, | 605 | TRACE_ITER_SYM_USEROBJ = 0x4000, |
578 | TRACE_ITER_SYM_USEROBJ = 0x8000, | 606 | TRACE_ITER_PRINTK_MSGONLY = 0x8000, |
579 | TRACE_ITER_PRINTK_MSGONLY = 0x10000, | 607 | TRACE_ITER_CONTEXT_INFO = 0x10000, /* Print pid/cpu/time */ |
580 | TRACE_ITER_CONTEXT_INFO = 0x20000, /* Print pid/cpu/time */ | 608 | TRACE_ITER_LATENCY_FMT = 0x20000, |
581 | TRACE_ITER_LATENCY_FMT = 0x40000, | 609 | TRACE_ITER_SLEEP_TIME = 0x40000, |
582 | TRACE_ITER_SLEEP_TIME = 0x80000, | 610 | TRACE_ITER_GRAPH_TIME = 0x80000, |
583 | TRACE_ITER_GRAPH_TIME = 0x100000, | ||
584 | }; | 611 | }; |
585 | 612 | ||
586 | /* | 613 | /* |
@@ -687,7 +714,6 @@ struct event_filter { | |||
687 | int n_preds; | 714 | int n_preds; |
688 | struct filter_pred **preds; | 715 | struct filter_pred **preds; |
689 | char *filter_string; | 716 | char *filter_string; |
690 | bool no_reset; | ||
691 | }; | 717 | }; |
692 | 718 | ||
693 | struct event_subsystem { | 719 | struct event_subsystem { |
@@ -699,22 +725,40 @@ struct event_subsystem { | |||
699 | }; | 725 | }; |
700 | 726 | ||
701 | struct filter_pred; | 727 | struct filter_pred; |
728 | struct regex; | ||
702 | 729 | ||
703 | typedef int (*filter_pred_fn_t) (struct filter_pred *pred, void *event, | 730 | typedef int (*filter_pred_fn_t) (struct filter_pred *pred, void *event, |
704 | int val1, int val2); | 731 | int val1, int val2); |
705 | 732 | ||
733 | typedef int (*regex_match_func)(char *str, struct regex *r, int len); | ||
734 | |||
735 | enum regex_type { | ||
736 | MATCH_FULL = 0, | ||
737 | MATCH_FRONT_ONLY, | ||
738 | MATCH_MIDDLE_ONLY, | ||
739 | MATCH_END_ONLY, | ||
740 | }; | ||
741 | |||
742 | struct regex { | ||
743 | char pattern[MAX_FILTER_STR_VAL]; | ||
744 | int len; | ||
745 | int field_len; | ||
746 | regex_match_func match; | ||
747 | }; | ||
748 | |||
706 | struct filter_pred { | 749 | struct filter_pred { |
707 | filter_pred_fn_t fn; | 750 | filter_pred_fn_t fn; |
708 | u64 val; | 751 | u64 val; |
709 | char str_val[MAX_FILTER_STR_VAL]; | 752 | struct regex regex; |
710 | int str_len; | 753 | char *field_name; |
711 | char *field_name; | 754 | int offset; |
712 | int offset; | 755 | int not; |
713 | int not; | 756 | int op; |
714 | int op; | 757 | int pop_n; |
715 | int pop_n; | ||
716 | }; | 758 | }; |
717 | 759 | ||
760 | extern enum regex_type | ||
761 | filter_parse_regex(char *buff, int len, char **search, int *not); | ||
718 | extern void print_event_filter(struct ftrace_event_call *call, | 762 | extern void print_event_filter(struct ftrace_event_call *call, |
719 | struct trace_seq *s); | 763 | struct trace_seq *s); |
720 | extern int apply_event_filter(struct ftrace_event_call *call, | 764 | extern int apply_event_filter(struct ftrace_event_call *call, |
@@ -730,7 +774,8 @@ filter_check_discard(struct ftrace_event_call *call, void *rec, | |||
730 | struct ring_buffer *buffer, | 774 | struct ring_buffer *buffer, |
731 | struct ring_buffer_event *event) | 775 | struct ring_buffer_event *event) |
732 | { | 776 | { |
733 | if (unlikely(call->filter_active) && !filter_match_preds(call, rec)) { | 777 | if (unlikely(call->filter_active) && |
778 | !filter_match_preds(call->filter, rec)) { | ||
734 | ring_buffer_discard_commit(buffer, event); | 779 | ring_buffer_discard_commit(buffer, event); |
735 | return 1; | 780 | return 1; |
736 | } | 781 | } |