diff options
author | Ingo Molnar <mingo@elte.hu> | 2008-05-12 15:20:51 -0400 |
---|---|---|
committer | Thomas Gleixner <tglx@linutronix.de> | 2008-05-23 14:58:28 -0400 |
commit | e309b41dd65aa953f86765eeeecc941d8e1e8b8f (patch) | |
tree | 295d4ed6e2a766607f889a04b977ca27cc24929e /kernel/trace | |
parent | b53dde9d34f2df396540988ebc65c33400f57b04 (diff) |
ftrace: remove notrace
now that we have a kbuild method for notrace, no need to pollute the
C code with the annotations.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Diffstat (limited to 'kernel/trace')
-rw-r--r-- | kernel/trace/ftrace.c | 66 | ||||
-rw-r--r-- | kernel/trace/trace.c | 94 | ||||
-rw-r--r-- | kernel/trace/trace.h | 6 | ||||
-rw-r--r-- | kernel/trace/trace_functions.c | 12 | ||||
-rw-r--r-- | kernel/trace/trace_irqsoff.c | 40 | ||||
-rw-r--r-- | kernel/trace/trace_sched_switch.c | 12 | ||||
-rw-r--r-- | kernel/trace/trace_sched_wakeup.c | 28 | ||||
-rw-r--r-- | kernel/trace/trace_selftest.c | 2 |
8 files changed, 130 insertions, 130 deletions
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 57350cbd1f61..281d97a3208c 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
@@ -53,7 +53,7 @@ ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; | |||
53 | /* mcount is defined per arch in assembly */ | 53 | /* mcount is defined per arch in assembly */ |
54 | EXPORT_SYMBOL(mcount); | 54 | EXPORT_SYMBOL(mcount); |
55 | 55 | ||
56 | notrace void ftrace_list_func(unsigned long ip, unsigned long parent_ip) | 56 | void ftrace_list_func(unsigned long ip, unsigned long parent_ip) |
57 | { | 57 | { |
58 | struct ftrace_ops *op = ftrace_list; | 58 | struct ftrace_ops *op = ftrace_list; |
59 | 59 | ||
@@ -79,7 +79,7 @@ void clear_ftrace_function(void) | |||
79 | ftrace_trace_function = ftrace_stub; | 79 | ftrace_trace_function = ftrace_stub; |
80 | } | 80 | } |
81 | 81 | ||
82 | static int notrace __register_ftrace_function(struct ftrace_ops *ops) | 82 | static int __register_ftrace_function(struct ftrace_ops *ops) |
83 | { | 83 | { |
84 | /* Should never be called by interrupts */ | 84 | /* Should never be called by interrupts */ |
85 | spin_lock(&ftrace_lock); | 85 | spin_lock(&ftrace_lock); |
@@ -110,7 +110,7 @@ static int notrace __register_ftrace_function(struct ftrace_ops *ops) | |||
110 | return 0; | 110 | return 0; |
111 | } | 111 | } |
112 | 112 | ||
113 | static int notrace __unregister_ftrace_function(struct ftrace_ops *ops) | 113 | static int __unregister_ftrace_function(struct ftrace_ops *ops) |
114 | { | 114 | { |
115 | struct ftrace_ops **p; | 115 | struct ftrace_ops **p; |
116 | int ret = 0; | 116 | int ret = 0; |
@@ -197,7 +197,7 @@ static int ftrace_record_suspend; | |||
197 | 197 | ||
198 | static struct dyn_ftrace *ftrace_free_records; | 198 | static struct dyn_ftrace *ftrace_free_records; |
199 | 199 | ||
200 | static inline int notrace | 200 | static inline int |
201 | ftrace_ip_in_hash(unsigned long ip, unsigned long key) | 201 | ftrace_ip_in_hash(unsigned long ip, unsigned long key) |
202 | { | 202 | { |
203 | struct dyn_ftrace *p; | 203 | struct dyn_ftrace *p; |
@@ -214,13 +214,13 @@ ftrace_ip_in_hash(unsigned long ip, unsigned long key) | |||
214 | return found; | 214 | return found; |
215 | } | 215 | } |
216 | 216 | ||
217 | static inline void notrace | 217 | static inline void |
218 | ftrace_add_hash(struct dyn_ftrace *node, unsigned long key) | 218 | ftrace_add_hash(struct dyn_ftrace *node, unsigned long key) |
219 | { | 219 | { |
220 | hlist_add_head(&node->node, &ftrace_hash[key]); | 220 | hlist_add_head(&node->node, &ftrace_hash[key]); |
221 | } | 221 | } |
222 | 222 | ||
223 | static notrace void ftrace_free_rec(struct dyn_ftrace *rec) | 223 | static void ftrace_free_rec(struct dyn_ftrace *rec) |
224 | { | 224 | { |
225 | /* no locking, only called from kstop_machine */ | 225 | /* no locking, only called from kstop_machine */ |
226 | 226 | ||
@@ -229,7 +229,7 @@ static notrace void ftrace_free_rec(struct dyn_ftrace *rec) | |||
229 | rec->flags |= FTRACE_FL_FREE; | 229 | rec->flags |= FTRACE_FL_FREE; |
230 | } | 230 | } |
231 | 231 | ||
232 | static notrace struct dyn_ftrace *ftrace_alloc_dyn_node(unsigned long ip) | 232 | static struct dyn_ftrace *ftrace_alloc_dyn_node(unsigned long ip) |
233 | { | 233 | { |
234 | struct dyn_ftrace *rec; | 234 | struct dyn_ftrace *rec; |
235 | 235 | ||
@@ -259,7 +259,7 @@ static notrace struct dyn_ftrace *ftrace_alloc_dyn_node(unsigned long ip) | |||
259 | return &ftrace_pages->records[ftrace_pages->index++]; | 259 | return &ftrace_pages->records[ftrace_pages->index++]; |
260 | } | 260 | } |
261 | 261 | ||
262 | static void notrace | 262 | static void |
263 | ftrace_record_ip(unsigned long ip) | 263 | ftrace_record_ip(unsigned long ip) |
264 | { | 264 | { |
265 | struct dyn_ftrace *node; | 265 | struct dyn_ftrace *node; |
@@ -329,7 +329,7 @@ ftrace_record_ip(unsigned long ip) | |||
329 | #define FTRACE_ADDR ((long)(ftrace_caller)) | 329 | #define FTRACE_ADDR ((long)(ftrace_caller)) |
330 | #define MCOUNT_ADDR ((long)(mcount)) | 330 | #define MCOUNT_ADDR ((long)(mcount)) |
331 | 331 | ||
332 | static void notrace | 332 | static void |
333 | __ftrace_replace_code(struct dyn_ftrace *rec, | 333 | __ftrace_replace_code(struct dyn_ftrace *rec, |
334 | unsigned char *old, unsigned char *new, int enable) | 334 | unsigned char *old, unsigned char *new, int enable) |
335 | { | 335 | { |
@@ -405,7 +405,7 @@ __ftrace_replace_code(struct dyn_ftrace *rec, | |||
405 | } | 405 | } |
406 | } | 406 | } |
407 | 407 | ||
408 | static void notrace ftrace_replace_code(int enable) | 408 | static void ftrace_replace_code(int enable) |
409 | { | 409 | { |
410 | unsigned char *new = NULL, *old = NULL; | 410 | unsigned char *new = NULL, *old = NULL; |
411 | struct dyn_ftrace *rec; | 411 | struct dyn_ftrace *rec; |
@@ -430,7 +430,7 @@ static void notrace ftrace_replace_code(int enable) | |||
430 | } | 430 | } |
431 | } | 431 | } |
432 | 432 | ||
433 | static notrace void ftrace_shutdown_replenish(void) | 433 | static void ftrace_shutdown_replenish(void) |
434 | { | 434 | { |
435 | if (ftrace_pages->next) | 435 | if (ftrace_pages->next) |
436 | return; | 436 | return; |
@@ -439,7 +439,7 @@ static notrace void ftrace_shutdown_replenish(void) | |||
439 | ftrace_pages->next = (void *)get_zeroed_page(GFP_KERNEL); | 439 | ftrace_pages->next = (void *)get_zeroed_page(GFP_KERNEL); |
440 | } | 440 | } |
441 | 441 | ||
442 | static notrace void | 442 | static void |
443 | ftrace_code_disable(struct dyn_ftrace *rec) | 443 | ftrace_code_disable(struct dyn_ftrace *rec) |
444 | { | 444 | { |
445 | unsigned long ip; | 445 | unsigned long ip; |
@@ -458,7 +458,7 @@ ftrace_code_disable(struct dyn_ftrace *rec) | |||
458 | } | 458 | } |
459 | } | 459 | } |
460 | 460 | ||
461 | static int notrace __ftrace_modify_code(void *data) | 461 | static int __ftrace_modify_code(void *data) |
462 | { | 462 | { |
463 | unsigned long addr; | 463 | unsigned long addr; |
464 | int *command = data; | 464 | int *command = data; |
@@ -482,14 +482,14 @@ static int notrace __ftrace_modify_code(void *data) | |||
482 | return 0; | 482 | return 0; |
483 | } | 483 | } |
484 | 484 | ||
485 | static void notrace ftrace_run_update_code(int command) | 485 | static void ftrace_run_update_code(int command) |
486 | { | 486 | { |
487 | stop_machine_run(__ftrace_modify_code, &command, NR_CPUS); | 487 | stop_machine_run(__ftrace_modify_code, &command, NR_CPUS); |
488 | } | 488 | } |
489 | 489 | ||
490 | static ftrace_func_t saved_ftrace_func; | 490 | static ftrace_func_t saved_ftrace_func; |
491 | 491 | ||
492 | static void notrace ftrace_startup(void) | 492 | static void ftrace_startup(void) |
493 | { | 493 | { |
494 | int command = 0; | 494 | int command = 0; |
495 | 495 | ||
@@ -514,7 +514,7 @@ static void notrace ftrace_startup(void) | |||
514 | mutex_unlock(&ftraced_lock); | 514 | mutex_unlock(&ftraced_lock); |
515 | } | 515 | } |
516 | 516 | ||
517 | static void notrace ftrace_shutdown(void) | 517 | static void ftrace_shutdown(void) |
518 | { | 518 | { |
519 | int command = 0; | 519 | int command = 0; |
520 | 520 | ||
@@ -539,7 +539,7 @@ static void notrace ftrace_shutdown(void) | |||
539 | mutex_unlock(&ftraced_lock); | 539 | mutex_unlock(&ftraced_lock); |
540 | } | 540 | } |
541 | 541 | ||
542 | static void notrace ftrace_startup_sysctl(void) | 542 | static void ftrace_startup_sysctl(void) |
543 | { | 543 | { |
544 | int command = FTRACE_ENABLE_MCOUNT; | 544 | int command = FTRACE_ENABLE_MCOUNT; |
545 | 545 | ||
@@ -557,7 +557,7 @@ static void notrace ftrace_startup_sysctl(void) | |||
557 | mutex_unlock(&ftraced_lock); | 557 | mutex_unlock(&ftraced_lock); |
558 | } | 558 | } |
559 | 559 | ||
560 | static void notrace ftrace_shutdown_sysctl(void) | 560 | static void ftrace_shutdown_sysctl(void) |
561 | { | 561 | { |
562 | int command = FTRACE_DISABLE_MCOUNT; | 562 | int command = FTRACE_DISABLE_MCOUNT; |
563 | 563 | ||
@@ -577,7 +577,7 @@ static cycle_t ftrace_update_time; | |||
577 | static unsigned long ftrace_update_cnt; | 577 | static unsigned long ftrace_update_cnt; |
578 | unsigned long ftrace_update_tot_cnt; | 578 | unsigned long ftrace_update_tot_cnt; |
579 | 579 | ||
580 | static int notrace __ftrace_update_code(void *ignore) | 580 | static int __ftrace_update_code(void *ignore) |
581 | { | 581 | { |
582 | struct dyn_ftrace *p; | 582 | struct dyn_ftrace *p; |
583 | struct hlist_head head; | 583 | struct hlist_head head; |
@@ -618,7 +618,7 @@ static int notrace __ftrace_update_code(void *ignore) | |||
618 | return 0; | 618 | return 0; |
619 | } | 619 | } |
620 | 620 | ||
621 | static void notrace ftrace_update_code(void) | 621 | static void ftrace_update_code(void) |
622 | { | 622 | { |
623 | if (unlikely(ftrace_disabled)) | 623 | if (unlikely(ftrace_disabled)) |
624 | return; | 624 | return; |
@@ -626,7 +626,7 @@ static void notrace ftrace_update_code(void) | |||
626 | stop_machine_run(__ftrace_update_code, NULL, NR_CPUS); | 626 | stop_machine_run(__ftrace_update_code, NULL, NR_CPUS); |
627 | } | 627 | } |
628 | 628 | ||
629 | static int notrace ftraced(void *ignore) | 629 | static int ftraced(void *ignore) |
630 | { | 630 | { |
631 | unsigned long usecs; | 631 | unsigned long usecs; |
632 | 632 | ||
@@ -733,7 +733,7 @@ struct ftrace_iterator { | |||
733 | unsigned filtered; | 733 | unsigned filtered; |
734 | }; | 734 | }; |
735 | 735 | ||
736 | static void notrace * | 736 | static void * |
737 | t_next(struct seq_file *m, void *v, loff_t *pos) | 737 | t_next(struct seq_file *m, void *v, loff_t *pos) |
738 | { | 738 | { |
739 | struct ftrace_iterator *iter = m->private; | 739 | struct ftrace_iterator *iter = m->private; |
@@ -806,7 +806,7 @@ static struct seq_operations show_ftrace_seq_ops = { | |||
806 | .show = t_show, | 806 | .show = t_show, |
807 | }; | 807 | }; |
808 | 808 | ||
809 | static int notrace | 809 | static int |
810 | ftrace_avail_open(struct inode *inode, struct file *file) | 810 | ftrace_avail_open(struct inode *inode, struct file *file) |
811 | { | 811 | { |
812 | struct ftrace_iterator *iter; | 812 | struct ftrace_iterator *iter; |
@@ -845,7 +845,7 @@ int ftrace_avail_release(struct inode *inode, struct file *file) | |||
845 | return 0; | 845 | return 0; |
846 | } | 846 | } |
847 | 847 | ||
848 | static void notrace ftrace_filter_reset(void) | 848 | static void ftrace_filter_reset(void) |
849 | { | 849 | { |
850 | struct ftrace_page *pg; | 850 | struct ftrace_page *pg; |
851 | struct dyn_ftrace *rec; | 851 | struct dyn_ftrace *rec; |
@@ -867,7 +867,7 @@ static void notrace ftrace_filter_reset(void) | |||
867 | preempt_enable(); | 867 | preempt_enable(); |
868 | } | 868 | } |
869 | 869 | ||
870 | static int notrace | 870 | static int |
871 | ftrace_filter_open(struct inode *inode, struct file *file) | 871 | ftrace_filter_open(struct inode *inode, struct file *file) |
872 | { | 872 | { |
873 | struct ftrace_iterator *iter; | 873 | struct ftrace_iterator *iter; |
@@ -903,7 +903,7 @@ ftrace_filter_open(struct inode *inode, struct file *file) | |||
903 | return ret; | 903 | return ret; |
904 | } | 904 | } |
905 | 905 | ||
906 | static ssize_t notrace | 906 | static ssize_t |
907 | ftrace_filter_read(struct file *file, char __user *ubuf, | 907 | ftrace_filter_read(struct file *file, char __user *ubuf, |
908 | size_t cnt, loff_t *ppos) | 908 | size_t cnt, loff_t *ppos) |
909 | { | 909 | { |
@@ -913,7 +913,7 @@ ftrace_filter_read(struct file *file, char __user *ubuf, | |||
913 | return -EPERM; | 913 | return -EPERM; |
914 | } | 914 | } |
915 | 915 | ||
916 | static loff_t notrace | 916 | static loff_t |
917 | ftrace_filter_lseek(struct file *file, loff_t offset, int origin) | 917 | ftrace_filter_lseek(struct file *file, loff_t offset, int origin) |
918 | { | 918 | { |
919 | loff_t ret; | 919 | loff_t ret; |
@@ -933,7 +933,7 @@ enum { | |||
933 | MATCH_END_ONLY, | 933 | MATCH_END_ONLY, |
934 | }; | 934 | }; |
935 | 935 | ||
936 | static void notrace | 936 | static void |
937 | ftrace_match(unsigned char *buff, int len) | 937 | ftrace_match(unsigned char *buff, int len) |
938 | { | 938 | { |
939 | char str[KSYM_SYMBOL_LEN]; | 939 | char str[KSYM_SYMBOL_LEN]; |
@@ -1002,7 +1002,7 @@ ftrace_match(unsigned char *buff, int len) | |||
1002 | preempt_enable(); | 1002 | preempt_enable(); |
1003 | } | 1003 | } |
1004 | 1004 | ||
1005 | static ssize_t notrace | 1005 | static ssize_t |
1006 | ftrace_filter_write(struct file *file, const char __user *ubuf, | 1006 | ftrace_filter_write(struct file *file, const char __user *ubuf, |
1007 | size_t cnt, loff_t *ppos) | 1007 | size_t cnt, loff_t *ppos) |
1008 | { | 1008 | { |
@@ -1094,7 +1094,7 @@ ftrace_filter_write(struct file *file, const char __user *ubuf, | |||
1094 | * Filters denote which functions should be enabled when tracing is enabled. | 1094 | * Filters denote which functions should be enabled when tracing is enabled. |
1095 | * If @buf is NULL and reset is set, all functions will be enabled for tracing. | 1095 | * If @buf is NULL and reset is set, all functions will be enabled for tracing. |
1096 | */ | 1096 | */ |
1097 | notrace void ftrace_set_filter(unsigned char *buf, int len, int reset) | 1097 | void ftrace_set_filter(unsigned char *buf, int len, int reset) |
1098 | { | 1098 | { |
1099 | if (unlikely(ftrace_disabled)) | 1099 | if (unlikely(ftrace_disabled)) |
1100 | return; | 1100 | return; |
@@ -1107,7 +1107,7 @@ notrace void ftrace_set_filter(unsigned char *buf, int len, int reset) | |||
1107 | mutex_unlock(&ftrace_filter_lock); | 1107 | mutex_unlock(&ftrace_filter_lock); |
1108 | } | 1108 | } |
1109 | 1109 | ||
1110 | static int notrace | 1110 | static int |
1111 | ftrace_filter_release(struct inode *inode, struct file *file) | 1111 | ftrace_filter_release(struct inode *inode, struct file *file) |
1112 | { | 1112 | { |
1113 | struct seq_file *m = (struct seq_file *)file->private_data; | 1113 | struct seq_file *m = (struct seq_file *)file->private_data; |
@@ -1242,7 +1242,7 @@ static __init int ftrace_init_debugfs(void) | |||
1242 | 1242 | ||
1243 | fs_initcall(ftrace_init_debugfs); | 1243 | fs_initcall(ftrace_init_debugfs); |
1244 | 1244 | ||
1245 | static int __init notrace ftrace_dynamic_init(void) | 1245 | static int __init ftrace_dynamic_init(void) |
1246 | { | 1246 | { |
1247 | struct task_struct *p; | 1247 | struct task_struct *p; |
1248 | unsigned long addr; | 1248 | unsigned long addr; |
@@ -1352,7 +1352,7 @@ int unregister_ftrace_function(struct ftrace_ops *ops) | |||
1352 | return ret; | 1352 | return ret; |
1353 | } | 1353 | } |
1354 | 1354 | ||
1355 | notrace int | 1355 | int |
1356 | ftrace_enable_sysctl(struct ctl_table *table, int write, | 1356 | ftrace_enable_sysctl(struct ctl_table *table, int write, |
1357 | struct file *file, void __user *buffer, size_t *lenp, | 1357 | struct file *file, void __user *buffer, size_t *lenp, |
1358 | loff_t *ppos) | 1358 | loff_t *ppos) |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 9022c357032a..f5898051fdd9 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
@@ -35,7 +35,7 @@ unsigned long __read_mostly tracing_thresh; | |||
35 | 35 | ||
36 | static int tracing_disabled = 1; | 36 | static int tracing_disabled = 1; |
37 | 37 | ||
38 | static long notrace | 38 | static long |
39 | ns2usecs(cycle_t nsec) | 39 | ns2usecs(cycle_t nsec) |
40 | { | 40 | { |
41 | nsec += 500; | 41 | nsec += 500; |
@@ -43,7 +43,7 @@ ns2usecs(cycle_t nsec) | |||
43 | return nsec; | 43 | return nsec; |
44 | } | 44 | } |
45 | 45 | ||
46 | notrace cycle_t ftrace_now(int cpu) | 46 | cycle_t ftrace_now(int cpu) |
47 | { | 47 | { |
48 | return cpu_clock(cpu); | 48 | return cpu_clock(cpu); |
49 | } | 49 | } |
@@ -135,7 +135,7 @@ static DEFINE_SPINLOCK(ftrace_max_lock); | |||
135 | * structure. (this way the maximum trace is permanently saved, | 135 | * structure. (this way the maximum trace is permanently saved, |
136 | * for later retrieval via /debugfs/tracing/latency_trace) | 136 | * for later retrieval via /debugfs/tracing/latency_trace) |
137 | */ | 137 | */ |
138 | static notrace void | 138 | static void |
139 | __update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) | 139 | __update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) |
140 | { | 140 | { |
141 | struct trace_array_cpu *data = tr->data[cpu]; | 141 | struct trace_array_cpu *data = tr->data[cpu]; |
@@ -184,7 +184,7 @@ void *head_page(struct trace_array_cpu *data) | |||
184 | return page_address(page); | 184 | return page_address(page); |
185 | } | 185 | } |
186 | 186 | ||
187 | static notrace int | 187 | static int |
188 | trace_seq_printf(struct trace_seq *s, const char *fmt, ...) | 188 | trace_seq_printf(struct trace_seq *s, const char *fmt, ...) |
189 | { | 189 | { |
190 | int len = (PAGE_SIZE - 1) - s->len; | 190 | int len = (PAGE_SIZE - 1) - s->len; |
@@ -207,7 +207,7 @@ trace_seq_printf(struct trace_seq *s, const char *fmt, ...) | |||
207 | return len; | 207 | return len; |
208 | } | 208 | } |
209 | 209 | ||
210 | static notrace int | 210 | static int |
211 | trace_seq_puts(struct trace_seq *s, const char *str) | 211 | trace_seq_puts(struct trace_seq *s, const char *str) |
212 | { | 212 | { |
213 | int len = strlen(str); | 213 | int len = strlen(str); |
@@ -221,7 +221,7 @@ trace_seq_puts(struct trace_seq *s, const char *str) | |||
221 | return len; | 221 | return len; |
222 | } | 222 | } |
223 | 223 | ||
224 | static notrace int | 224 | static int |
225 | trace_seq_putc(struct trace_seq *s, unsigned char c) | 225 | trace_seq_putc(struct trace_seq *s, unsigned char c) |
226 | { | 226 | { |
227 | if (s->len >= (PAGE_SIZE - 1)) | 227 | if (s->len >= (PAGE_SIZE - 1)) |
@@ -232,7 +232,7 @@ trace_seq_putc(struct trace_seq *s, unsigned char c) | |||
232 | return 1; | 232 | return 1; |
233 | } | 233 | } |
234 | 234 | ||
235 | static notrace int | 235 | static int |
236 | trace_seq_putmem(struct trace_seq *s, void *mem, size_t len) | 236 | trace_seq_putmem(struct trace_seq *s, void *mem, size_t len) |
237 | { | 237 | { |
238 | if (len > ((PAGE_SIZE - 1) - s->len)) | 238 | if (len > ((PAGE_SIZE - 1) - s->len)) |
@@ -246,7 +246,7 @@ trace_seq_putmem(struct trace_seq *s, void *mem, size_t len) | |||
246 | 246 | ||
247 | #define HEX_CHARS 17 | 247 | #define HEX_CHARS 17 |
248 | 248 | ||
249 | static notrace int | 249 | static int |
250 | trace_seq_putmem_hex(struct trace_seq *s, void *mem, size_t len) | 250 | trace_seq_putmem_hex(struct trace_seq *s, void *mem, size_t len) |
251 | { | 251 | { |
252 | unsigned char hex[HEX_CHARS]; | 252 | unsigned char hex[HEX_CHARS]; |
@@ -285,13 +285,13 @@ trace_seq_putmem_hex(struct trace_seq *s, void *mem, size_t len) | |||
285 | return trace_seq_putmem(s, hex, j); | 285 | return trace_seq_putmem(s, hex, j); |
286 | } | 286 | } |
287 | 287 | ||
288 | static notrace void | 288 | static void |
289 | trace_seq_reset(struct trace_seq *s) | 289 | trace_seq_reset(struct trace_seq *s) |
290 | { | 290 | { |
291 | s->len = 0; | 291 | s->len = 0; |
292 | } | 292 | } |
293 | 293 | ||
294 | static notrace void | 294 | static void |
295 | trace_print_seq(struct seq_file *m, struct trace_seq *s) | 295 | trace_print_seq(struct seq_file *m, struct trace_seq *s) |
296 | { | 296 | { |
297 | int len = s->len >= PAGE_SIZE ? PAGE_SIZE - 1 : s->len; | 297 | int len = s->len >= PAGE_SIZE ? PAGE_SIZE - 1 : s->len; |
@@ -302,7 +302,7 @@ trace_print_seq(struct seq_file *m, struct trace_seq *s) | |||
302 | trace_seq_reset(s); | 302 | trace_seq_reset(s); |
303 | } | 303 | } |
304 | 304 | ||
305 | notrace static void | 305 | static void |
306 | flip_trace(struct trace_array_cpu *tr1, struct trace_array_cpu *tr2) | 306 | flip_trace(struct trace_array_cpu *tr1, struct trace_array_cpu *tr2) |
307 | { | 307 | { |
308 | struct list_head flip_pages; | 308 | struct list_head flip_pages; |
@@ -323,7 +323,7 @@ flip_trace(struct trace_array_cpu *tr1, struct trace_array_cpu *tr2) | |||
323 | check_pages(tr2); | 323 | check_pages(tr2); |
324 | } | 324 | } |
325 | 325 | ||
326 | notrace void | 326 | void |
327 | update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) | 327 | update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) |
328 | { | 328 | { |
329 | struct trace_array_cpu *data; | 329 | struct trace_array_cpu *data; |
@@ -348,7 +348,7 @@ update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu) | |||
348 | * @tsk - task with the latency | 348 | * @tsk - task with the latency |
349 | * @cpu - the cpu of the buffer to copy. | 349 | * @cpu - the cpu of the buffer to copy. |
350 | */ | 350 | */ |
351 | notrace void | 351 | void |
352 | update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu) | 352 | update_max_tr_single(struct trace_array *tr, struct task_struct *tsk, int cpu) |
353 | { | 353 | { |
354 | struct trace_array_cpu *data = tr->data[cpu]; | 354 | struct trace_array_cpu *data = tr->data[cpu]; |
@@ -471,7 +471,7 @@ void unregister_tracer(struct tracer *type) | |||
471 | mutex_unlock(&trace_types_lock); | 471 | mutex_unlock(&trace_types_lock); |
472 | } | 472 | } |
473 | 473 | ||
474 | notrace void tracing_reset(struct trace_array_cpu *data) | 474 | void tracing_reset(struct trace_array_cpu *data) |
475 | { | 475 | { |
476 | data->trace_idx = 0; | 476 | data->trace_idx = 0; |
477 | data->trace_head = data->trace_tail = head_page(data); | 477 | data->trace_head = data->trace_tail = head_page(data); |
@@ -494,9 +494,9 @@ static void trace_init_cmdlines(void) | |||
494 | cmdline_idx = 0; | 494 | cmdline_idx = 0; |
495 | } | 495 | } |
496 | 496 | ||
497 | notrace void trace_stop_cmdline_recording(void); | 497 | void trace_stop_cmdline_recording(void); |
498 | 498 | ||
499 | static notrace void trace_save_cmdline(struct task_struct *tsk) | 499 | static void trace_save_cmdline(struct task_struct *tsk) |
500 | { | 500 | { |
501 | unsigned map; | 501 | unsigned map; |
502 | unsigned idx; | 502 | unsigned idx; |
@@ -531,7 +531,7 @@ static notrace void trace_save_cmdline(struct task_struct *tsk) | |||
531 | spin_unlock(&trace_cmdline_lock); | 531 | spin_unlock(&trace_cmdline_lock); |
532 | } | 532 | } |
533 | 533 | ||
534 | static notrace char *trace_find_cmdline(int pid) | 534 | static char *trace_find_cmdline(int pid) |
535 | { | 535 | { |
536 | char *cmdline = "<...>"; | 536 | char *cmdline = "<...>"; |
537 | unsigned map; | 537 | unsigned map; |
@@ -552,7 +552,7 @@ static notrace char *trace_find_cmdline(int pid) | |||
552 | return cmdline; | 552 | return cmdline; |
553 | } | 553 | } |
554 | 554 | ||
555 | notrace void tracing_record_cmdline(struct task_struct *tsk) | 555 | void tracing_record_cmdline(struct task_struct *tsk) |
556 | { | 556 | { |
557 | if (atomic_read(&trace_record_cmdline_disabled)) | 557 | if (atomic_read(&trace_record_cmdline_disabled)) |
558 | return; | 558 | return; |
@@ -560,7 +560,7 @@ notrace void tracing_record_cmdline(struct task_struct *tsk) | |||
560 | trace_save_cmdline(tsk); | 560 | trace_save_cmdline(tsk); |
561 | } | 561 | } |
562 | 562 | ||
563 | static inline notrace struct list_head * | 563 | static inline struct list_head * |
564 | trace_next_list(struct trace_array_cpu *data, struct list_head *next) | 564 | trace_next_list(struct trace_array_cpu *data, struct list_head *next) |
565 | { | 565 | { |
566 | /* | 566 | /* |
@@ -574,7 +574,7 @@ trace_next_list(struct trace_array_cpu *data, struct list_head *next) | |||
574 | return next; | 574 | return next; |
575 | } | 575 | } |
576 | 576 | ||
577 | static inline notrace void * | 577 | static inline void * |
578 | trace_next_page(struct trace_array_cpu *data, void *addr) | 578 | trace_next_page(struct trace_array_cpu *data, void *addr) |
579 | { | 579 | { |
580 | struct list_head *next; | 580 | struct list_head *next; |
@@ -588,7 +588,7 @@ trace_next_page(struct trace_array_cpu *data, void *addr) | |||
588 | return page_address(page); | 588 | return page_address(page); |
589 | } | 589 | } |
590 | 590 | ||
591 | static inline notrace struct trace_entry * | 591 | static inline struct trace_entry * |
592 | tracing_get_trace_entry(struct trace_array *tr, struct trace_array_cpu *data) | 592 | tracing_get_trace_entry(struct trace_array *tr, struct trace_array_cpu *data) |
593 | { | 593 | { |
594 | unsigned long idx, idx_next; | 594 | unsigned long idx, idx_next; |
@@ -623,7 +623,7 @@ tracing_get_trace_entry(struct trace_array *tr, struct trace_array_cpu *data) | |||
623 | return entry; | 623 | return entry; |
624 | } | 624 | } |
625 | 625 | ||
626 | static inline notrace void | 626 | static inline void |
627 | tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags) | 627 | tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags) |
628 | { | 628 | { |
629 | struct task_struct *tsk = current; | 629 | struct task_struct *tsk = current; |
@@ -640,7 +640,7 @@ tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags) | |||
640 | (need_resched() ? TRACE_FLAG_NEED_RESCHED : 0); | 640 | (need_resched() ? TRACE_FLAG_NEED_RESCHED : 0); |
641 | } | 641 | } |
642 | 642 | ||
643 | notrace void | 643 | void |
644 | trace_function(struct trace_array *tr, struct trace_array_cpu *data, | 644 | trace_function(struct trace_array *tr, struct trace_array_cpu *data, |
645 | unsigned long ip, unsigned long parent_ip, unsigned long flags) | 645 | unsigned long ip, unsigned long parent_ip, unsigned long flags) |
646 | { | 646 | { |
@@ -659,7 +659,7 @@ trace_function(struct trace_array *tr, struct trace_array_cpu *data, | |||
659 | wake_up (&trace_wait); | 659 | wake_up (&trace_wait); |
660 | } | 660 | } |
661 | 661 | ||
662 | notrace void | 662 | void |
663 | ftrace(struct trace_array *tr, struct trace_array_cpu *data, | 663 | ftrace(struct trace_array *tr, struct trace_array_cpu *data, |
664 | unsigned long ip, unsigned long parent_ip, unsigned long flags) | 664 | unsigned long ip, unsigned long parent_ip, unsigned long flags) |
665 | { | 665 | { |
@@ -667,7 +667,7 @@ ftrace(struct trace_array *tr, struct trace_array_cpu *data, | |||
667 | trace_function(tr, data, ip, parent_ip, flags); | 667 | trace_function(tr, data, ip, parent_ip, flags); |
668 | } | 668 | } |
669 | 669 | ||
670 | notrace void | 670 | void |
671 | trace_special(struct trace_array *tr, struct trace_array_cpu *data, | 671 | trace_special(struct trace_array *tr, struct trace_array_cpu *data, |
672 | unsigned long arg1, unsigned long arg2, unsigned long arg3) | 672 | unsigned long arg1, unsigned long arg2, unsigned long arg3) |
673 | { | 673 | { |
@@ -687,7 +687,7 @@ trace_special(struct trace_array *tr, struct trace_array_cpu *data, | |||
687 | wake_up (&trace_wait); | 687 | wake_up (&trace_wait); |
688 | } | 688 | } |
689 | 689 | ||
690 | notrace void | 690 | void |
691 | tracing_sched_switch_trace(struct trace_array *tr, | 691 | tracing_sched_switch_trace(struct trace_array *tr, |
692 | struct trace_array_cpu *data, | 692 | struct trace_array_cpu *data, |
693 | struct task_struct *prev, struct task_struct *next, | 693 | struct task_struct *prev, struct task_struct *next, |
@@ -712,7 +712,7 @@ tracing_sched_switch_trace(struct trace_array *tr, | |||
712 | } | 712 | } |
713 | 713 | ||
714 | #ifdef CONFIG_FTRACE | 714 | #ifdef CONFIG_FTRACE |
715 | static notrace void | 715 | static void |
716 | function_trace_call(unsigned long ip, unsigned long parent_ip) | 716 | function_trace_call(unsigned long ip, unsigned long parent_ip) |
717 | { | 717 | { |
718 | struct trace_array *tr = &global_trace; | 718 | struct trace_array *tr = &global_trace; |
@@ -741,12 +741,12 @@ static struct ftrace_ops trace_ops __read_mostly = | |||
741 | .func = function_trace_call, | 741 | .func = function_trace_call, |
742 | }; | 742 | }; |
743 | 743 | ||
744 | notrace void tracing_start_function_trace(void) | 744 | void tracing_start_function_trace(void) |
745 | { | 745 | { |
746 | register_ftrace_function(&trace_ops); | 746 | register_ftrace_function(&trace_ops); |
747 | } | 747 | } |
748 | 748 | ||
749 | notrace void tracing_stop_function_trace(void) | 749 | void tracing_stop_function_trace(void) |
750 | { | 750 | { |
751 | unregister_ftrace_function(&trace_ops); | 751 | unregister_ftrace_function(&trace_ops); |
752 | } | 752 | } |
@@ -786,7 +786,7 @@ trace_entry_idx(struct trace_array *tr, struct trace_array_cpu *data, | |||
786 | return &array[iter->next_page_idx[cpu]]; | 786 | return &array[iter->next_page_idx[cpu]]; |
787 | } | 787 | } |
788 | 788 | ||
789 | static struct trace_entry * notrace | 789 | static struct trace_entry * |
790 | find_next_entry(struct trace_iterator *iter, int *ent_cpu) | 790 | find_next_entry(struct trace_iterator *iter, int *ent_cpu) |
791 | { | 791 | { |
792 | struct trace_array *tr = iter->tr; | 792 | struct trace_array *tr = iter->tr; |
@@ -813,7 +813,7 @@ find_next_entry(struct trace_iterator *iter, int *ent_cpu) | |||
813 | return next; | 813 | return next; |
814 | } | 814 | } |
815 | 815 | ||
816 | static notrace void trace_iterator_increment(struct trace_iterator *iter) | 816 | static void trace_iterator_increment(struct trace_iterator *iter) |
817 | { | 817 | { |
818 | iter->idx++; | 818 | iter->idx++; |
819 | iter->next_idx[iter->cpu]++; | 819 | iter->next_idx[iter->cpu]++; |
@@ -828,7 +828,7 @@ static notrace void trace_iterator_increment(struct trace_iterator *iter) | |||
828 | } | 828 | } |
829 | } | 829 | } |
830 | 830 | ||
831 | static notrace void trace_consume(struct trace_iterator *iter) | 831 | static void trace_consume(struct trace_iterator *iter) |
832 | { | 832 | { |
833 | struct trace_array_cpu *data = iter->tr->data[iter->cpu]; | 833 | struct trace_array_cpu *data = iter->tr->data[iter->cpu]; |
834 | 834 | ||
@@ -844,7 +844,7 @@ static notrace void trace_consume(struct trace_iterator *iter) | |||
844 | data->trace_idx = 0; | 844 | data->trace_idx = 0; |
845 | } | 845 | } |
846 | 846 | ||
847 | static notrace void *find_next_entry_inc(struct trace_iterator *iter) | 847 | static void *find_next_entry_inc(struct trace_iterator *iter) |
848 | { | 848 | { |
849 | struct trace_entry *next; | 849 | struct trace_entry *next; |
850 | int next_cpu = -1; | 850 | int next_cpu = -1; |
@@ -863,7 +863,7 @@ static notrace void *find_next_entry_inc(struct trace_iterator *iter) | |||
863 | return next ? iter : NULL; | 863 | return next ? iter : NULL; |
864 | } | 864 | } |
865 | 865 | ||
866 | static notrace void *s_next(struct seq_file *m, void *v, loff_t *pos) | 866 | static void *s_next(struct seq_file *m, void *v, loff_t *pos) |
867 | { | 867 | { |
868 | struct trace_iterator *iter = m->private; | 868 | struct trace_iterator *iter = m->private; |
869 | void *last_ent = iter->ent; | 869 | void *last_ent = iter->ent; |
@@ -978,7 +978,7 @@ seq_print_sym_offset(struct trace_seq *s, const char *fmt, | |||
978 | # define IP_FMT "%016lx" | 978 | # define IP_FMT "%016lx" |
979 | #endif | 979 | #endif |
980 | 980 | ||
981 | static notrace int | 981 | static int |
982 | seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) | 982 | seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) |
983 | { | 983 | { |
984 | int ret; | 984 | int ret; |
@@ -999,7 +999,7 @@ seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) | |||
999 | return ret; | 999 | return ret; |
1000 | } | 1000 | } |
1001 | 1001 | ||
1002 | static notrace void print_lat_help_header(struct seq_file *m) | 1002 | static void print_lat_help_header(struct seq_file *m) |
1003 | { | 1003 | { |
1004 | seq_puts(m, "# _------=> CPU# \n"); | 1004 | seq_puts(m, "# _------=> CPU# \n"); |
1005 | seq_puts(m, "# / _-----=> irqs-off \n"); | 1005 | seq_puts(m, "# / _-----=> irqs-off \n"); |
@@ -1012,14 +1012,14 @@ static notrace void print_lat_help_header(struct seq_file *m) | |||
1012 | seq_puts(m, "# \\ / ||||| \\ | / \n"); | 1012 | seq_puts(m, "# \\ / ||||| \\ | / \n"); |
1013 | } | 1013 | } |
1014 | 1014 | ||
1015 | static notrace void print_func_help_header(struct seq_file *m) | 1015 | static void print_func_help_header(struct seq_file *m) |
1016 | { | 1016 | { |
1017 | seq_puts(m, "# TASK-PID CPU# TIMESTAMP FUNCTION\n"); | 1017 | seq_puts(m, "# TASK-PID CPU# TIMESTAMP FUNCTION\n"); |
1018 | seq_puts(m, "# | | | | |\n"); | 1018 | seq_puts(m, "# | | | | |\n"); |
1019 | } | 1019 | } |
1020 | 1020 | ||
1021 | 1021 | ||
1022 | static notrace void | 1022 | static void |
1023 | print_trace_header(struct seq_file *m, struct trace_iterator *iter) | 1023 | print_trace_header(struct seq_file *m, struct trace_iterator *iter) |
1024 | { | 1024 | { |
1025 | unsigned long sym_flags = (trace_flags & TRACE_ITER_SYM_MASK); | 1025 | unsigned long sym_flags = (trace_flags & TRACE_ITER_SYM_MASK); |
@@ -1090,7 +1090,7 @@ print_trace_header(struct seq_file *m, struct trace_iterator *iter) | |||
1090 | seq_puts(m, "\n"); | 1090 | seq_puts(m, "\n"); |
1091 | } | 1091 | } |
1092 | 1092 | ||
1093 | static notrace void | 1093 | static void |
1094 | lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) | 1094 | lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) |
1095 | { | 1095 | { |
1096 | int hardirq, softirq; | 1096 | int hardirq, softirq; |
@@ -1127,7 +1127,7 @@ lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) | |||
1127 | 1127 | ||
1128 | unsigned long preempt_mark_thresh = 100; | 1128 | unsigned long preempt_mark_thresh = 100; |
1129 | 1129 | ||
1130 | static notrace void | 1130 | static void |
1131 | lat_print_timestamp(struct trace_seq *s, unsigned long long abs_usecs, | 1131 | lat_print_timestamp(struct trace_seq *s, unsigned long long abs_usecs, |
1132 | unsigned long rel_usecs) | 1132 | unsigned long rel_usecs) |
1133 | { | 1133 | { |
@@ -1142,7 +1142,7 @@ lat_print_timestamp(struct trace_seq *s, unsigned long long abs_usecs, | |||
1142 | 1142 | ||
1143 | static const char state_to_char[] = TASK_STATE_TO_CHAR_STR; | 1143 | static const char state_to_char[] = TASK_STATE_TO_CHAR_STR; |
1144 | 1144 | ||
1145 | static notrace int | 1145 | static int |
1146 | print_lat_fmt(struct trace_iterator *iter, unsigned int trace_idx, int cpu) | 1146 | print_lat_fmt(struct trace_iterator *iter, unsigned int trace_idx, int cpu) |
1147 | { | 1147 | { |
1148 | struct trace_seq *s = &iter->seq; | 1148 | struct trace_seq *s = &iter->seq; |
@@ -1206,7 +1206,7 @@ print_lat_fmt(struct trace_iterator *iter, unsigned int trace_idx, int cpu) | |||
1206 | return 1; | 1206 | return 1; |
1207 | } | 1207 | } |
1208 | 1208 | ||
1209 | static notrace int print_trace_fmt(struct trace_iterator *iter) | 1209 | static int print_trace_fmt(struct trace_iterator *iter) |
1210 | { | 1210 | { |
1211 | struct trace_seq *s = &iter->seq; | 1211 | struct trace_seq *s = &iter->seq; |
1212 | unsigned long sym_flags = (trace_flags & TRACE_ITER_SYM_MASK); | 1212 | unsigned long sym_flags = (trace_flags & TRACE_ITER_SYM_MASK); |
@@ -1279,7 +1279,7 @@ static notrace int print_trace_fmt(struct trace_iterator *iter) | |||
1279 | return 1; | 1279 | return 1; |
1280 | } | 1280 | } |
1281 | 1281 | ||
1282 | static notrace int print_raw_fmt(struct trace_iterator *iter) | 1282 | static int print_raw_fmt(struct trace_iterator *iter) |
1283 | { | 1283 | { |
1284 | struct trace_seq *s = &iter->seq; | 1284 | struct trace_seq *s = &iter->seq; |
1285 | struct trace_entry *entry; | 1285 | struct trace_entry *entry; |
@@ -1336,7 +1336,7 @@ do { \ | |||
1336 | return 0; \ | 1336 | return 0; \ |
1337 | } while (0) | 1337 | } while (0) |
1338 | 1338 | ||
1339 | static notrace int print_hex_fmt(struct trace_iterator *iter) | 1339 | static int print_hex_fmt(struct trace_iterator *iter) |
1340 | { | 1340 | { |
1341 | struct trace_seq *s = &iter->seq; | 1341 | struct trace_seq *s = &iter->seq; |
1342 | unsigned char newline = '\n'; | 1342 | unsigned char newline = '\n'; |
@@ -1375,7 +1375,7 @@ static notrace int print_hex_fmt(struct trace_iterator *iter) | |||
1375 | return 1; | 1375 | return 1; |
1376 | } | 1376 | } |
1377 | 1377 | ||
1378 | static notrace int print_bin_fmt(struct trace_iterator *iter) | 1378 | static int print_bin_fmt(struct trace_iterator *iter) |
1379 | { | 1379 | { |
1380 | struct trace_seq *s = &iter->seq; | 1380 | struct trace_seq *s = &iter->seq; |
1381 | struct trace_entry *entry; | 1381 | struct trace_entry *entry; |
@@ -1475,7 +1475,7 @@ static struct seq_operations tracer_seq_ops = { | |||
1475 | .show = s_show, | 1475 | .show = s_show, |
1476 | }; | 1476 | }; |
1477 | 1477 | ||
1478 | static struct trace_iterator notrace * | 1478 | static struct trace_iterator * |
1479 | __tracing_open(struct inode *inode, struct file *file, int *ret) | 1479 | __tracing_open(struct inode *inode, struct file *file, int *ret) |
1480 | { | 1480 | { |
1481 | struct trace_iterator *iter; | 1481 | struct trace_iterator *iter; |
@@ -1572,7 +1572,7 @@ static int tracing_lt_open(struct inode *inode, struct file *file) | |||
1572 | } | 1572 | } |
1573 | 1573 | ||
1574 | 1574 | ||
1575 | static notrace void * | 1575 | static void * |
1576 | t_next(struct seq_file *m, void *v, loff_t *pos) | 1576 | t_next(struct seq_file *m, void *v, loff_t *pos) |
1577 | { | 1577 | { |
1578 | struct tracer *t = m->private; | 1578 | struct tracer *t = m->private; |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index faf9f67246ac..2b7352bf1ce6 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
@@ -150,7 +150,7 @@ struct trace_iterator { | |||
150 | long idx; | 150 | long idx; |
151 | }; | 151 | }; |
152 | 152 | ||
153 | void notrace tracing_reset(struct trace_array_cpu *data); | 153 | void tracing_reset(struct trace_array_cpu *data); |
154 | int tracing_open_generic(struct inode *inode, struct file *filp); | 154 | int tracing_open_generic(struct inode *inode, struct file *filp); |
155 | struct dentry *tracing_init_dentry(void); | 155 | struct dentry *tracing_init_dentry(void); |
156 | void ftrace(struct trace_array *tr, | 156 | void ftrace(struct trace_array *tr, |
@@ -189,10 +189,10 @@ void update_max_tr(struct trace_array *tr, struct task_struct *tsk, int cpu); | |||
189 | void update_max_tr_single(struct trace_array *tr, | 189 | void update_max_tr_single(struct trace_array *tr, |
190 | struct task_struct *tsk, int cpu); | 190 | struct task_struct *tsk, int cpu); |
191 | 191 | ||
192 | extern notrace cycle_t ftrace_now(int cpu); | 192 | extern cycle_t ftrace_now(int cpu); |
193 | 193 | ||
194 | #ifdef CONFIG_SCHED_TRACER | 194 | #ifdef CONFIG_SCHED_TRACER |
195 | extern void notrace | 195 | extern void |
196 | wakeup_sched_switch(struct task_struct *prev, struct task_struct *next); | 196 | wakeup_sched_switch(struct task_struct *prev, struct task_struct *next); |
197 | #else | 197 | #else |
198 | static inline void | 198 | static inline void |
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c index 69a0eb00a0a5..4165d34bd28a 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c | |||
@@ -16,7 +16,7 @@ | |||
16 | 16 | ||
17 | #include "trace.h" | 17 | #include "trace.h" |
18 | 18 | ||
19 | static notrace void function_reset(struct trace_array *tr) | 19 | static void function_reset(struct trace_array *tr) |
20 | { | 20 | { |
21 | int cpu; | 21 | int cpu; |
22 | 22 | ||
@@ -26,30 +26,30 @@ static notrace void function_reset(struct trace_array *tr) | |||
26 | tracing_reset(tr->data[cpu]); | 26 | tracing_reset(tr->data[cpu]); |
27 | } | 27 | } |
28 | 28 | ||
29 | static notrace void start_function_trace(struct trace_array *tr) | 29 | static void start_function_trace(struct trace_array *tr) |
30 | { | 30 | { |
31 | function_reset(tr); | 31 | function_reset(tr); |
32 | tracing_start_function_trace(); | 32 | tracing_start_function_trace(); |
33 | } | 33 | } |
34 | 34 | ||
35 | static notrace void stop_function_trace(struct trace_array *tr) | 35 | static void stop_function_trace(struct trace_array *tr) |
36 | { | 36 | { |
37 | tracing_stop_function_trace(); | 37 | tracing_stop_function_trace(); |
38 | } | 38 | } |
39 | 39 | ||
40 | static notrace void function_trace_init(struct trace_array *tr) | 40 | static void function_trace_init(struct trace_array *tr) |
41 | { | 41 | { |
42 | if (tr->ctrl) | 42 | if (tr->ctrl) |
43 | start_function_trace(tr); | 43 | start_function_trace(tr); |
44 | } | 44 | } |
45 | 45 | ||
46 | static notrace void function_trace_reset(struct trace_array *tr) | 46 | static void function_trace_reset(struct trace_array *tr) |
47 | { | 47 | { |
48 | if (tr->ctrl) | 48 | if (tr->ctrl) |
49 | stop_function_trace(tr); | 49 | stop_function_trace(tr); |
50 | } | 50 | } |
51 | 51 | ||
52 | static notrace void function_trace_ctrl_update(struct trace_array *tr) | 52 | static void function_trace_ctrl_update(struct trace_array *tr) |
53 | { | 53 | { |
54 | if (tr->ctrl) | 54 | if (tr->ctrl) |
55 | start_function_trace(tr); | 55 | start_function_trace(tr); |
diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c index 2ac0d09db6fb..7a4dc014b8ab 100644 --- a/kernel/trace/trace_irqsoff.c +++ b/kernel/trace/trace_irqsoff.c | |||
@@ -33,7 +33,7 @@ enum { | |||
33 | static int trace_type __read_mostly; | 33 | static int trace_type __read_mostly; |
34 | 34 | ||
35 | #ifdef CONFIG_PREEMPT_TRACER | 35 | #ifdef CONFIG_PREEMPT_TRACER |
36 | static inline int notrace | 36 | static inline int |
37 | preempt_trace(void) | 37 | preempt_trace(void) |
38 | { | 38 | { |
39 | return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); | 39 | return ((trace_type & TRACER_PREEMPT_OFF) && preempt_count()); |
@@ -43,7 +43,7 @@ preempt_trace(void) | |||
43 | #endif | 43 | #endif |
44 | 44 | ||
45 | #ifdef CONFIG_IRQSOFF_TRACER | 45 | #ifdef CONFIG_IRQSOFF_TRACER |
46 | static inline int notrace | 46 | static inline int |
47 | irq_trace(void) | 47 | irq_trace(void) |
48 | { | 48 | { |
49 | return ((trace_type & TRACER_IRQS_OFF) && | 49 | return ((trace_type & TRACER_IRQS_OFF) && |
@@ -67,7 +67,7 @@ static __cacheline_aligned_in_smp unsigned long max_sequence; | |||
67 | /* | 67 | /* |
68 | * irqsoff uses its own tracer function to keep the overhead down: | 68 | * irqsoff uses its own tracer function to keep the overhead down: |
69 | */ | 69 | */ |
70 | static void notrace | 70 | static void |
71 | irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip) | 71 | irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip) |
72 | { | 72 | { |
73 | struct trace_array *tr = irqsoff_trace; | 73 | struct trace_array *tr = irqsoff_trace; |
@@ -109,7 +109,7 @@ static struct ftrace_ops trace_ops __read_mostly = | |||
109 | /* | 109 | /* |
110 | * Should this new latency be reported/recorded? | 110 | * Should this new latency be reported/recorded? |
111 | */ | 111 | */ |
112 | static int notrace report_latency(cycle_t delta) | 112 | static int report_latency(cycle_t delta) |
113 | { | 113 | { |
114 | if (tracing_thresh) { | 114 | if (tracing_thresh) { |
115 | if (delta < tracing_thresh) | 115 | if (delta < tracing_thresh) |
@@ -121,7 +121,7 @@ static int notrace report_latency(cycle_t delta) | |||
121 | return 1; | 121 | return 1; |
122 | } | 122 | } |
123 | 123 | ||
124 | static void notrace | 124 | static void |
125 | check_critical_timing(struct trace_array *tr, | 125 | check_critical_timing(struct trace_array *tr, |
126 | struct trace_array_cpu *data, | 126 | struct trace_array_cpu *data, |
127 | unsigned long parent_ip, | 127 | unsigned long parent_ip, |
@@ -191,7 +191,7 @@ out: | |||
191 | trace_function(tr, data, CALLER_ADDR0, parent_ip, flags); | 191 | trace_function(tr, data, CALLER_ADDR0, parent_ip, flags); |
192 | } | 192 | } |
193 | 193 | ||
194 | static inline void notrace | 194 | static inline void |
195 | start_critical_timing(unsigned long ip, unsigned long parent_ip) | 195 | start_critical_timing(unsigned long ip, unsigned long parent_ip) |
196 | { | 196 | { |
197 | int cpu; | 197 | int cpu; |
@@ -228,7 +228,7 @@ start_critical_timing(unsigned long ip, unsigned long parent_ip) | |||
228 | atomic_dec(&data->disabled); | 228 | atomic_dec(&data->disabled); |
229 | } | 229 | } |
230 | 230 | ||
231 | static inline void notrace | 231 | static inline void |
232 | stop_critical_timing(unsigned long ip, unsigned long parent_ip) | 232 | stop_critical_timing(unsigned long ip, unsigned long parent_ip) |
233 | { | 233 | { |
234 | int cpu; | 234 | int cpu; |
@@ -261,13 +261,13 @@ stop_critical_timing(unsigned long ip, unsigned long parent_ip) | |||
261 | } | 261 | } |
262 | 262 | ||
263 | /* start and stop critical timings used to for stoppage (in idle) */ | 263 | /* start and stop critical timings used to for stoppage (in idle) */ |
264 | void notrace start_critical_timings(void) | 264 | void start_critical_timings(void) |
265 | { | 265 | { |
266 | if (preempt_trace() || irq_trace()) | 266 | if (preempt_trace() || irq_trace()) |
267 | start_critical_timing(CALLER_ADDR0, CALLER_ADDR1); | 267 | start_critical_timing(CALLER_ADDR0, CALLER_ADDR1); |
268 | } | 268 | } |
269 | 269 | ||
270 | void notrace stop_critical_timings(void) | 270 | void stop_critical_timings(void) |
271 | { | 271 | { |
272 | if (preempt_trace() || irq_trace()) | 272 | if (preempt_trace() || irq_trace()) |
273 | stop_critical_timing(CALLER_ADDR0, CALLER_ADDR1); | 273 | stop_critical_timing(CALLER_ADDR0, CALLER_ADDR1); |
@@ -275,13 +275,13 @@ void notrace stop_critical_timings(void) | |||
275 | 275 | ||
276 | #ifdef CONFIG_IRQSOFF_TRACER | 276 | #ifdef CONFIG_IRQSOFF_TRACER |
277 | #ifdef CONFIG_PROVE_LOCKING | 277 | #ifdef CONFIG_PROVE_LOCKING |
278 | void notrace time_hardirqs_on(unsigned long a0, unsigned long a1) | 278 | void time_hardirqs_on(unsigned long a0, unsigned long a1) |
279 | { | 279 | { |
280 | if (!preempt_trace() && irq_trace()) | 280 | if (!preempt_trace() && irq_trace()) |
281 | stop_critical_timing(a0, a1); | 281 | stop_critical_timing(a0, a1); |
282 | } | 282 | } |
283 | 283 | ||
284 | void notrace time_hardirqs_off(unsigned long a0, unsigned long a1) | 284 | void time_hardirqs_off(unsigned long a0, unsigned long a1) |
285 | { | 285 | { |
286 | if (!preempt_trace() && irq_trace()) | 286 | if (!preempt_trace() && irq_trace()) |
287 | start_critical_timing(a0, a1); | 287 | start_critical_timing(a0, a1); |
@@ -309,35 +309,35 @@ void trace_softirqs_off(unsigned long ip) | |||
309 | { | 309 | { |
310 | } | 310 | } |
311 | 311 | ||
312 | inline notrace void print_irqtrace_events(struct task_struct *curr) | 312 | inline void print_irqtrace_events(struct task_struct *curr) |
313 | { | 313 | { |
314 | } | 314 | } |
315 | 315 | ||
316 | /* | 316 | /* |
317 | * We are only interested in hardirq on/off events: | 317 | * We are only interested in hardirq on/off events: |
318 | */ | 318 | */ |
319 | void notrace trace_hardirqs_on(void) | 319 | void trace_hardirqs_on(void) |
320 | { | 320 | { |
321 | if (!preempt_trace() && irq_trace()) | 321 | if (!preempt_trace() && irq_trace()) |
322 | stop_critical_timing(CALLER_ADDR0, CALLER_ADDR1); | 322 | stop_critical_timing(CALLER_ADDR0, CALLER_ADDR1); |
323 | } | 323 | } |
324 | EXPORT_SYMBOL(trace_hardirqs_on); | 324 | EXPORT_SYMBOL(trace_hardirqs_on); |
325 | 325 | ||
326 | void notrace trace_hardirqs_off(void) | 326 | void trace_hardirqs_off(void) |
327 | { | 327 | { |
328 | if (!preempt_trace() && irq_trace()) | 328 | if (!preempt_trace() && irq_trace()) |
329 | start_critical_timing(CALLER_ADDR0, CALLER_ADDR1); | 329 | start_critical_timing(CALLER_ADDR0, CALLER_ADDR1); |
330 | } | 330 | } |
331 | EXPORT_SYMBOL(trace_hardirqs_off); | 331 | EXPORT_SYMBOL(trace_hardirqs_off); |
332 | 332 | ||
333 | void notrace trace_hardirqs_on_caller(unsigned long caller_addr) | 333 | void trace_hardirqs_on_caller(unsigned long caller_addr) |
334 | { | 334 | { |
335 | if (!preempt_trace() && irq_trace()) | 335 | if (!preempt_trace() && irq_trace()) |
336 | stop_critical_timing(CALLER_ADDR0, caller_addr); | 336 | stop_critical_timing(CALLER_ADDR0, caller_addr); |
337 | } | 337 | } |
338 | EXPORT_SYMBOL(trace_hardirqs_on_caller); | 338 | EXPORT_SYMBOL(trace_hardirqs_on_caller); |
339 | 339 | ||
340 | void notrace trace_hardirqs_off_caller(unsigned long caller_addr) | 340 | void trace_hardirqs_off_caller(unsigned long caller_addr) |
341 | { | 341 | { |
342 | if (!preempt_trace() && irq_trace()) | 342 | if (!preempt_trace() && irq_trace()) |
343 | start_critical_timing(CALLER_ADDR0, caller_addr); | 343 | start_critical_timing(CALLER_ADDR0, caller_addr); |
@@ -348,12 +348,12 @@ EXPORT_SYMBOL(trace_hardirqs_off_caller); | |||
348 | #endif /* CONFIG_IRQSOFF_TRACER */ | 348 | #endif /* CONFIG_IRQSOFF_TRACER */ |
349 | 349 | ||
350 | #ifdef CONFIG_PREEMPT_TRACER | 350 | #ifdef CONFIG_PREEMPT_TRACER |
351 | void notrace trace_preempt_on(unsigned long a0, unsigned long a1) | 351 | void trace_preempt_on(unsigned long a0, unsigned long a1) |
352 | { | 352 | { |
353 | stop_critical_timing(a0, a1); | 353 | stop_critical_timing(a0, a1); |
354 | } | 354 | } |
355 | 355 | ||
356 | void notrace trace_preempt_off(unsigned long a0, unsigned long a1) | 356 | void trace_preempt_off(unsigned long a0, unsigned long a1) |
357 | { | 357 | { |
358 | start_critical_timing(a0, a1); | 358 | start_critical_timing(a0, a1); |
359 | } | 359 | } |
@@ -395,14 +395,14 @@ static void irqsoff_tracer_ctrl_update(struct trace_array *tr) | |||
395 | stop_irqsoff_tracer(tr); | 395 | stop_irqsoff_tracer(tr); |
396 | } | 396 | } |
397 | 397 | ||
398 | static void notrace irqsoff_tracer_open(struct trace_iterator *iter) | 398 | static void irqsoff_tracer_open(struct trace_iterator *iter) |
399 | { | 399 | { |
400 | /* stop the trace while dumping */ | 400 | /* stop the trace while dumping */ |
401 | if (iter->tr->ctrl) | 401 | if (iter->tr->ctrl) |
402 | stop_irqsoff_tracer(iter->tr); | 402 | stop_irqsoff_tracer(iter->tr); |
403 | } | 403 | } |
404 | 404 | ||
405 | static void notrace irqsoff_tracer_close(struct trace_iterator *iter) | 405 | static void irqsoff_tracer_close(struct trace_iterator *iter) |
406 | { | 406 | { |
407 | if (iter->tr->ctrl) | 407 | if (iter->tr->ctrl) |
408 | start_irqsoff_tracer(iter->tr); | 408 | start_irqsoff_tracer(iter->tr); |
diff --git a/kernel/trace/trace_sched_switch.c b/kernel/trace/trace_sched_switch.c index 8d656672da93..b738eaca1dbe 100644 --- a/kernel/trace/trace_sched_switch.c +++ b/kernel/trace/trace_sched_switch.c | |||
@@ -17,7 +17,7 @@ | |||
17 | static struct trace_array *ctx_trace; | 17 | static struct trace_array *ctx_trace; |
18 | static int __read_mostly tracer_enabled; | 18 | static int __read_mostly tracer_enabled; |
19 | 19 | ||
20 | static void notrace | 20 | static void |
21 | ctx_switch_func(struct task_struct *prev, struct task_struct *next) | 21 | ctx_switch_func(struct task_struct *prev, struct task_struct *next) |
22 | { | 22 | { |
23 | struct trace_array *tr = ctx_trace; | 23 | struct trace_array *tr = ctx_trace; |
@@ -57,7 +57,7 @@ void ftrace_ctx_switch(struct task_struct *prev, struct task_struct *next) | |||
57 | wakeup_sched_switch(prev, next); | 57 | wakeup_sched_switch(prev, next); |
58 | } | 58 | } |
59 | 59 | ||
60 | static notrace void sched_switch_reset(struct trace_array *tr) | 60 | static void sched_switch_reset(struct trace_array *tr) |
61 | { | 61 | { |
62 | int cpu; | 62 | int cpu; |
63 | 63 | ||
@@ -67,18 +67,18 @@ static notrace void sched_switch_reset(struct trace_array *tr) | |||
67 | tracing_reset(tr->data[cpu]); | 67 | tracing_reset(tr->data[cpu]); |
68 | } | 68 | } |
69 | 69 | ||
70 | static notrace void start_sched_trace(struct trace_array *tr) | 70 | static void start_sched_trace(struct trace_array *tr) |
71 | { | 71 | { |
72 | sched_switch_reset(tr); | 72 | sched_switch_reset(tr); |
73 | tracer_enabled = 1; | 73 | tracer_enabled = 1; |
74 | } | 74 | } |
75 | 75 | ||
76 | static notrace void stop_sched_trace(struct trace_array *tr) | 76 | static void stop_sched_trace(struct trace_array *tr) |
77 | { | 77 | { |
78 | tracer_enabled = 0; | 78 | tracer_enabled = 0; |
79 | } | 79 | } |
80 | 80 | ||
81 | static notrace void sched_switch_trace_init(struct trace_array *tr) | 81 | static void sched_switch_trace_init(struct trace_array *tr) |
82 | { | 82 | { |
83 | ctx_trace = tr; | 83 | ctx_trace = tr; |
84 | 84 | ||
@@ -86,7 +86,7 @@ static notrace void sched_switch_trace_init(struct trace_array *tr) | |||
86 | start_sched_trace(tr); | 86 | start_sched_trace(tr); |
87 | } | 87 | } |
88 | 88 | ||
89 | static notrace void sched_switch_trace_reset(struct trace_array *tr) | 89 | static void sched_switch_trace_reset(struct trace_array *tr) |
90 | { | 90 | { |
91 | if (tr->ctrl) | 91 | if (tr->ctrl) |
92 | stop_sched_trace(tr); | 92 | stop_sched_trace(tr); |
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index 3549e4154f1f..662679c78b66 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c | |||
@@ -27,12 +27,12 @@ static unsigned wakeup_prio = -1; | |||
27 | 27 | ||
28 | static DEFINE_SPINLOCK(wakeup_lock); | 28 | static DEFINE_SPINLOCK(wakeup_lock); |
29 | 29 | ||
30 | static void notrace __wakeup_reset(struct trace_array *tr); | 30 | static void __wakeup_reset(struct trace_array *tr); |
31 | 31 | ||
32 | /* | 32 | /* |
33 | * Should this new latency be reported/recorded? | 33 | * Should this new latency be reported/recorded? |
34 | */ | 34 | */ |
35 | static int notrace report_latency(cycle_t delta) | 35 | static int report_latency(cycle_t delta) |
36 | { | 36 | { |
37 | if (tracing_thresh) { | 37 | if (tracing_thresh) { |
38 | if (delta < tracing_thresh) | 38 | if (delta < tracing_thresh) |
@@ -44,7 +44,7 @@ static int notrace report_latency(cycle_t delta) | |||
44 | return 1; | 44 | return 1; |
45 | } | 45 | } |
46 | 46 | ||
47 | void notrace | 47 | void |
48 | wakeup_sched_switch(struct task_struct *prev, struct task_struct *next) | 48 | wakeup_sched_switch(struct task_struct *prev, struct task_struct *next) |
49 | { | 49 | { |
50 | unsigned long latency = 0, t0 = 0, t1 = 0; | 50 | unsigned long latency = 0, t0 = 0, t1 = 0; |
@@ -126,7 +126,7 @@ out: | |||
126 | atomic_dec(&tr->data[cpu]->disabled); | 126 | atomic_dec(&tr->data[cpu]->disabled); |
127 | } | 127 | } |
128 | 128 | ||
129 | static void notrace __wakeup_reset(struct trace_array *tr) | 129 | static void __wakeup_reset(struct trace_array *tr) |
130 | { | 130 | { |
131 | struct trace_array_cpu *data; | 131 | struct trace_array_cpu *data; |
132 | int cpu; | 132 | int cpu; |
@@ -147,7 +147,7 @@ static void notrace __wakeup_reset(struct trace_array *tr) | |||
147 | wakeup_task = NULL; | 147 | wakeup_task = NULL; |
148 | } | 148 | } |
149 | 149 | ||
150 | static void notrace wakeup_reset(struct trace_array *tr) | 150 | static void wakeup_reset(struct trace_array *tr) |
151 | { | 151 | { |
152 | unsigned long flags; | 152 | unsigned long flags; |
153 | 153 | ||
@@ -156,7 +156,7 @@ static void notrace wakeup_reset(struct trace_array *tr) | |||
156 | spin_unlock_irqrestore(&wakeup_lock, flags); | 156 | spin_unlock_irqrestore(&wakeup_lock, flags); |
157 | } | 157 | } |
158 | 158 | ||
159 | static notrace void | 159 | static void |
160 | wakeup_check_start(struct trace_array *tr, struct task_struct *p, | 160 | wakeup_check_start(struct trace_array *tr, struct task_struct *p, |
161 | struct task_struct *curr) | 161 | struct task_struct *curr) |
162 | { | 162 | { |
@@ -201,7 +201,7 @@ out: | |||
201 | atomic_dec(&tr->data[cpu]->disabled); | 201 | atomic_dec(&tr->data[cpu]->disabled); |
202 | } | 202 | } |
203 | 203 | ||
204 | notrace void | 204 | void |
205 | ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr) | 205 | ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr) |
206 | { | 206 | { |
207 | if (likely(!tracer_enabled)) | 207 | if (likely(!tracer_enabled)) |
@@ -210,7 +210,7 @@ ftrace_wake_up_task(struct task_struct *wakee, struct task_struct *curr) | |||
210 | wakeup_check_start(wakeup_trace, wakee, curr); | 210 | wakeup_check_start(wakeup_trace, wakee, curr); |
211 | } | 211 | } |
212 | 212 | ||
213 | notrace void | 213 | void |
214 | ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) | 214 | ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) |
215 | { | 215 | { |
216 | if (likely(!tracer_enabled)) | 216 | if (likely(!tracer_enabled)) |
@@ -219,7 +219,7 @@ ftrace_wake_up_new_task(struct task_struct *wakee, struct task_struct *curr) | |||
219 | wakeup_check_start(wakeup_trace, wakee, curr); | 219 | wakeup_check_start(wakeup_trace, wakee, curr); |
220 | } | 220 | } |
221 | 221 | ||
222 | static notrace void start_wakeup_tracer(struct trace_array *tr) | 222 | static void start_wakeup_tracer(struct trace_array *tr) |
223 | { | 223 | { |
224 | wakeup_reset(tr); | 224 | wakeup_reset(tr); |
225 | 225 | ||
@@ -237,12 +237,12 @@ static notrace void start_wakeup_tracer(struct trace_array *tr) | |||
237 | return; | 237 | return; |
238 | } | 238 | } |
239 | 239 | ||
240 | static notrace void stop_wakeup_tracer(struct trace_array *tr) | 240 | static void stop_wakeup_tracer(struct trace_array *tr) |
241 | { | 241 | { |
242 | tracer_enabled = 0; | 242 | tracer_enabled = 0; |
243 | } | 243 | } |
244 | 244 | ||
245 | static notrace void wakeup_tracer_init(struct trace_array *tr) | 245 | static void wakeup_tracer_init(struct trace_array *tr) |
246 | { | 246 | { |
247 | wakeup_trace = tr; | 247 | wakeup_trace = tr; |
248 | 248 | ||
@@ -250,7 +250,7 @@ static notrace void wakeup_tracer_init(struct trace_array *tr) | |||
250 | start_wakeup_tracer(tr); | 250 | start_wakeup_tracer(tr); |
251 | } | 251 | } |
252 | 252 | ||
253 | static notrace void wakeup_tracer_reset(struct trace_array *tr) | 253 | static void wakeup_tracer_reset(struct trace_array *tr) |
254 | { | 254 | { |
255 | if (tr->ctrl) { | 255 | if (tr->ctrl) { |
256 | stop_wakeup_tracer(tr); | 256 | stop_wakeup_tracer(tr); |
@@ -267,14 +267,14 @@ static void wakeup_tracer_ctrl_update(struct trace_array *tr) | |||
267 | stop_wakeup_tracer(tr); | 267 | stop_wakeup_tracer(tr); |
268 | } | 268 | } |
269 | 269 | ||
270 | static void notrace wakeup_tracer_open(struct trace_iterator *iter) | 270 | static void wakeup_tracer_open(struct trace_iterator *iter) |
271 | { | 271 | { |
272 | /* stop the trace while dumping */ | 272 | /* stop the trace while dumping */ |
273 | if (iter->tr->ctrl) | 273 | if (iter->tr->ctrl) |
274 | stop_wakeup_tracer(iter->tr); | 274 | stop_wakeup_tracer(iter->tr); |
275 | } | 275 | } |
276 | 276 | ||
277 | static void notrace wakeup_tracer_close(struct trace_iterator *iter) | 277 | static void wakeup_tracer_close(struct trace_iterator *iter) |
278 | { | 278 | { |
279 | /* forget about any processes we were recording */ | 279 | /* forget about any processes we were recording */ |
280 | if (iter->tr->ctrl) | 280 | if (iter->tr->ctrl) |
diff --git a/kernel/trace/trace_selftest.c b/kernel/trace/trace_selftest.c index 546307de6e3d..85715b86a342 100644 --- a/kernel/trace/trace_selftest.c +++ b/kernel/trace/trace_selftest.c | |||
@@ -3,7 +3,7 @@ | |||
3 | #include <linux/kthread.h> | 3 | #include <linux/kthread.h> |
4 | #include <linux/delay.h> | 4 | #include <linux/delay.h> |
5 | 5 | ||
6 | static notrace inline int trace_valid_entry(struct trace_entry *entry) | 6 | static inline int trace_valid_entry(struct trace_entry *entry) |
7 | { | 7 | { |
8 | switch (entry->type) { | 8 | switch (entry->type) { |
9 | case TRACE_FN: | 9 | case TRACE_FN: |