diff options
-rw-r--r-- | include/linux/tracepoint.h | 16 | ||||
-rw-r--r-- | kernel/trace/blktrace.c | 10 | ||||
-rw-r--r-- | kernel/trace/trace.c | 2 | ||||
-rw-r--r-- | kernel/trace/trace_branch.c | 2 | ||||
-rw-r--r-- | kernel/trace/trace_events_stage_3.h | 98 | ||||
-rw-r--r-- | kernel/trace/trace_export.c | 2 | ||||
-rw-r--r-- | kernel/trace/trace_functions_graph.c | 6 | ||||
-rw-r--r-- | kernel/trace/trace_output.c | 14 | ||||
-rw-r--r-- | kernel/trace/trace_workqueue.c | 6 |
9 files changed, 78 insertions, 78 deletions
diff --git a/include/linux/tracepoint.h b/include/linux/tracepoint.h index 119ece224c21..d35a7ee7611f 100644 --- a/include/linux/tracepoint.h +++ b/include/linux/tracepoint.h | |||
@@ -178,8 +178,8 @@ static inline void tracepoint_synchronize_unregister(void) | |||
178 | * * prototype, declare it via TP_PROTO(): | 178 | * * prototype, declare it via TP_PROTO(): |
179 | * * | 179 | * * |
180 | * | 180 | * |
181 | * TP_PROTO(struct rq *rq, struct task_struct *prev, | 181 | * TP_PROTO(struct rq *rq, struct task_struct *prev, |
182 | * struct task_struct *next), | 182 | * struct task_struct *next), |
183 | * | 183 | * |
184 | * * | 184 | * * |
185 | * * Define the call signature of the 'function'. | 185 | * * Define the call signature of the 'function'. |
@@ -187,7 +187,7 @@ static inline void tracepoint_synchronize_unregister(void) | |||
187 | * * TP_PROTO1/TP_PROTO2/TP_PROTO3 ugliness.) | 187 | * * TP_PROTO1/TP_PROTO2/TP_PROTO3 ugliness.) |
188 | * * | 188 | * * |
189 | * | 189 | * |
190 | * TP_ARGS(rq, prev, next), | 190 | * TP_ARGS(rq, prev, next), |
191 | * | 191 | * |
192 | * * | 192 | * * |
193 | * * Fast binary tracing: define the trace record via | 193 | * * Fast binary tracing: define the trace record via |
@@ -229,13 +229,13 @@ static inline void tracepoint_synchronize_unregister(void) | |||
229 | * * happens, on an active tracepoint. | 229 | * * happens, on an active tracepoint. |
230 | * * | 230 | * * |
231 | * | 231 | * |
232 | * TP_fast_assign( | 232 | * TP_fast_assign( |
233 | * memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN); | 233 | * memcpy(__entry->next_comm, next->comm, TASK_COMM_LEN); |
234 | * __entry->prev_pid = prev->pid; | 234 | * __entry->prev_pid = prev->pid; |
235 | * __entry->prev_prio = prev->prio; | 235 | * __entry->prev_prio = prev->prio; |
236 | * memcpy(__entry->prev_comm, prev->comm, TASK_COMM_LEN); | 236 | * memcpy(__entry->prev_comm, prev->comm, TASK_COMM_LEN); |
237 | * __entry->next_pid = next->pid; | 237 | * __entry->next_pid = next->pid; |
238 | * __entry->next_prio = next->prio; | 238 | * __entry->next_prio = next->prio; |
239 | * ) | 239 | * ) |
240 | * | 240 | * |
241 | * * | 241 | * * |
diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c index e39679a72a3b..bec69d3678c1 100644 --- a/kernel/trace/blktrace.c +++ b/kernel/trace/blktrace.c | |||
@@ -33,7 +33,7 @@ static struct trace_array *blk_tr; | |||
33 | static int __read_mostly blk_tracer_enabled; | 33 | static int __read_mostly blk_tracer_enabled; |
34 | 34 | ||
35 | /* Select an alternative, minimalistic output than the original one */ | 35 | /* Select an alternative, minimalistic output than the original one */ |
36 | #define TRACE_BLK_OPT_CLASSIC 0x1 | 36 | #define TRACE_BLK_OPT_CLASSIC 0x1 |
37 | 37 | ||
38 | static struct tracer_opt blk_tracer_opts[] = { | 38 | static struct tracer_opt blk_tracer_opts[] = { |
39 | /* Default disable the minimalistic output */ | 39 | /* Default disable the minimalistic output */ |
@@ -564,7 +564,7 @@ EXPORT_SYMBOL_GPL(blk_trace_startstop); | |||
564 | /** | 564 | /** |
565 | * blk_trace_ioctl: - handle the ioctls associated with tracing | 565 | * blk_trace_ioctl: - handle the ioctls associated with tracing |
566 | * @bdev: the block device | 566 | * @bdev: the block device |
567 | * @cmd: the ioctl cmd | 567 | * @cmd: the ioctl cmd |
568 | * @arg: the argument data, if any | 568 | * @arg: the argument data, if any |
569 | * | 569 | * |
570 | **/ | 570 | **/ |
@@ -1128,9 +1128,9 @@ static void blk_tracer_reset(struct trace_array *tr) | |||
1128 | 1128 | ||
1129 | static struct { | 1129 | static struct { |
1130 | const char *act[2]; | 1130 | const char *act[2]; |
1131 | int (*print)(struct trace_seq *s, const struct trace_entry *ent); | 1131 | int (*print)(struct trace_seq *s, const struct trace_entry *ent); |
1132 | } what2act[] __read_mostly = { | 1132 | } what2act[] __read_mostly = { |
1133 | [__BLK_TA_QUEUE] = {{ "Q", "queue" }, blk_log_generic }, | 1133 | [__BLK_TA_QUEUE] = {{ "Q", "queue" }, blk_log_generic }, |
1134 | [__BLK_TA_BACKMERGE] = {{ "M", "backmerge" }, blk_log_generic }, | 1134 | [__BLK_TA_BACKMERGE] = {{ "M", "backmerge" }, blk_log_generic }, |
1135 | [__BLK_TA_FRONTMERGE] = {{ "F", "frontmerge" }, blk_log_generic }, | 1135 | [__BLK_TA_FRONTMERGE] = {{ "F", "frontmerge" }, blk_log_generic }, |
1136 | [__BLK_TA_GETRQ] = {{ "G", "getrq" }, blk_log_generic }, | 1136 | [__BLK_TA_GETRQ] = {{ "G", "getrq" }, blk_log_generic }, |
@@ -1229,7 +1229,7 @@ static struct tracer blk_tracer __read_mostly = { | |||
1229 | }; | 1229 | }; |
1230 | 1230 | ||
1231 | static struct trace_event trace_blk_event = { | 1231 | static struct trace_event trace_blk_event = { |
1232 | .type = TRACE_BLK, | 1232 | .type = TRACE_BLK, |
1233 | .trace = blk_trace_event_print, | 1233 | .trace = blk_trace_event_print, |
1234 | .binary = blk_trace_event_print_binary, | 1234 | .binary = blk_trace_event_print_binary, |
1235 | }; | 1235 | }; |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index cc94f8642485..8c6a902db40a 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
@@ -799,7 +799,7 @@ tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags, | |||
799 | 799 | ||
800 | entry->preempt_count = pc & 0xff; | 800 | entry->preempt_count = pc & 0xff; |
801 | entry->pid = (tsk) ? tsk->pid : 0; | 801 | entry->pid = (tsk) ? tsk->pid : 0; |
802 | entry->tgid = (tsk) ? tsk->tgid : 0; | 802 | entry->tgid = (tsk) ? tsk->tgid : 0; |
803 | entry->flags = | 803 | entry->flags = |
804 | #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT | 804 | #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT |
805 | (irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) | | 805 | (irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) | |
diff --git a/kernel/trace/trace_branch.c b/kernel/trace/trace_branch.c index aaa0755268b9..ad8c22efff41 100644 --- a/kernel/trace/trace_branch.c +++ b/kernel/trace/trace_branch.c | |||
@@ -157,7 +157,7 @@ static enum print_line_t trace_branch_print(struct trace_iterator *iter, | |||
157 | 157 | ||
158 | 158 | ||
159 | static struct trace_event trace_branch_event = { | 159 | static struct trace_event trace_branch_event = { |
160 | .type = TRACE_BRANCH, | 160 | .type = TRACE_BRANCH, |
161 | .trace = trace_branch_print, | 161 | .trace = trace_branch_print, |
162 | }; | 162 | }; |
163 | 163 | ||
diff --git a/kernel/trace/trace_events_stage_3.h b/kernel/trace/trace_events_stage_3.h index 6ee1de59f19d..ae2e323df0c7 100644 --- a/kernel/trace/trace_events_stage_3.h +++ b/kernel/trace/trace_events_stage_3.h | |||
@@ -5,23 +5,23 @@ | |||
5 | * | 5 | * |
6 | * static void ftrace_event_<call>(proto) | 6 | * static void ftrace_event_<call>(proto) |
7 | * { | 7 | * { |
8 | * event_trace_printk(_RET_IP_, "<call>: " <fmt>); | 8 | * event_trace_printk(_RET_IP_, "<call>: " <fmt>); |
9 | * } | 9 | * } |
10 | * | 10 | * |
11 | * static int ftrace_reg_event_<call>(void) | 11 | * static int ftrace_reg_event_<call>(void) |
12 | * { | 12 | * { |
13 | * int ret; | 13 | * int ret; |
14 | * | 14 | * |
15 | * ret = register_trace_<call>(ftrace_event_<call>); | 15 | * ret = register_trace_<call>(ftrace_event_<call>); |
16 | * if (!ret) | 16 | * if (!ret) |
17 | * pr_info("event trace: Could not activate trace point " | 17 | * pr_info("event trace: Could not activate trace point " |
18 | * "probe to <call>"); | 18 | * "probe to <call>"); |
19 | * return ret; | 19 | * return ret; |
20 | * } | 20 | * } |
21 | * | 21 | * |
22 | * static void ftrace_unreg_event_<call>(void) | 22 | * static void ftrace_unreg_event_<call>(void) |
23 | * { | 23 | * { |
24 | * unregister_trace_<call>(ftrace_event_<call>); | 24 | * unregister_trace_<call>(ftrace_event_<call>); |
25 | * } | 25 | * } |
26 | * | 26 | * |
27 | * For those macros defined with TRACE_FORMAT: | 27 | * For those macros defined with TRACE_FORMAT: |
@@ -29,9 +29,9 @@ | |||
29 | * static struct ftrace_event_call __used | 29 | * static struct ftrace_event_call __used |
30 | * __attribute__((__aligned__(4))) | 30 | * __attribute__((__aligned__(4))) |
31 | * __attribute__((section("_ftrace_events"))) event_<call> = { | 31 | * __attribute__((section("_ftrace_events"))) event_<call> = { |
32 | * .name = "<call>", | 32 | * .name = "<call>", |
33 | * .regfunc = ftrace_reg_event_<call>, | 33 | * .regfunc = ftrace_reg_event_<call>, |
34 | * .unregfunc = ftrace_unreg_event_<call>, | 34 | * .unregfunc = ftrace_unreg_event_<call>, |
35 | * } | 35 | * } |
36 | * | 36 | * |
37 | * | 37 | * |
@@ -41,66 +41,66 @@ | |||
41 | * | 41 | * |
42 | * static void ftrace_raw_event_<call>(proto) | 42 | * static void ftrace_raw_event_<call>(proto) |
43 | * { | 43 | * { |
44 | * struct ring_buffer_event *event; | 44 | * struct ring_buffer_event *event; |
45 | * struct ftrace_raw_<call> *entry; <-- defined in stage 1 | 45 | * struct ftrace_raw_<call> *entry; <-- defined in stage 1 |
46 | * unsigned long irq_flags; | 46 | * unsigned long irq_flags; |
47 | * int pc; | 47 | * int pc; |
48 | * | 48 | * |
49 | * local_save_flags(irq_flags); | 49 | * local_save_flags(irq_flags); |
50 | * pc = preempt_count(); | 50 | * pc = preempt_count(); |
51 | * | 51 | * |
52 | * event = trace_current_buffer_lock_reserve(event_<call>.id, | 52 | * event = trace_current_buffer_lock_reserve(event_<call>.id, |
53 | * sizeof(struct ftrace_raw_<call>), | 53 | * sizeof(struct ftrace_raw_<call>), |
54 | * irq_flags, pc); | 54 | * irq_flags, pc); |
55 | * if (!event) | 55 | * if (!event) |
56 | * return; | 56 | * return; |
57 | * entry = ring_buffer_event_data(event); | 57 | * entry = ring_buffer_event_data(event); |
58 | * | 58 | * |
59 | * <assign>; <-- Here we assign the entries by the __field and | 59 | * <assign>; <-- Here we assign the entries by the __field and |
60 | * __array macros. | 60 | * __array macros. |
61 | * | 61 | * |
62 | * trace_current_buffer_unlock_commit(event, irq_flags, pc); | 62 | * trace_current_buffer_unlock_commit(event, irq_flags, pc); |
63 | * } | 63 | * } |
64 | * | 64 | * |
65 | * static int ftrace_raw_reg_event_<call>(void) | 65 | * static int ftrace_raw_reg_event_<call>(void) |
66 | * { | 66 | * { |
67 | * int ret; | 67 | * int ret; |
68 | * | 68 | * |
69 | * ret = register_trace_<call>(ftrace_raw_event_<call>); | 69 | * ret = register_trace_<call>(ftrace_raw_event_<call>); |
70 | * if (!ret) | 70 | * if (!ret) |
71 | * pr_info("event trace: Could not activate trace point " | 71 | * pr_info("event trace: Could not activate trace point " |
72 | * "probe to <call>"); | 72 | * "probe to <call>"); |
73 | * return ret; | 73 | * return ret; |
74 | * } | 74 | * } |
75 | * | 75 | * |
76 | * static void ftrace_unreg_event_<call>(void) | 76 | * static void ftrace_unreg_event_<call>(void) |
77 | * { | 77 | * { |
78 | * unregister_trace_<call>(ftrace_raw_event_<call>); | 78 | * unregister_trace_<call>(ftrace_raw_event_<call>); |
79 | * } | 79 | * } |
80 | * | 80 | * |
81 | * static struct trace_event ftrace_event_type_<call> = { | 81 | * static struct trace_event ftrace_event_type_<call> = { |
82 | * .trace = ftrace_raw_output_<call>, <-- stage 2 | 82 | * .trace = ftrace_raw_output_<call>, <-- stage 2 |
83 | * }; | 83 | * }; |
84 | * | 84 | * |
85 | * static int ftrace_raw_init_event_<call>(void) | 85 | * static int ftrace_raw_init_event_<call>(void) |
86 | * { | 86 | * { |
87 | * int id; | 87 | * int id; |
88 | * | 88 | * |
89 | * id = register_ftrace_event(&ftrace_event_type_<call>); | 89 | * id = register_ftrace_event(&ftrace_event_type_<call>); |
90 | * if (!id) | 90 | * if (!id) |
91 | * return -ENODEV; | 91 | * return -ENODEV; |
92 | * event_<call>.id = id; | 92 | * event_<call>.id = id; |
93 | * return 0; | 93 | * return 0; |
94 | * } | 94 | * } |
95 | * | 95 | * |
96 | * static struct ftrace_event_call __used | 96 | * static struct ftrace_event_call __used |
97 | * __attribute__((__aligned__(4))) | 97 | * __attribute__((__aligned__(4))) |
98 | * __attribute__((section("_ftrace_events"))) event_<call> = { | 98 | * __attribute__((section("_ftrace_events"))) event_<call> = { |
99 | * .name = "<call>", | 99 | * .name = "<call>", |
100 | * .system = "<system>", | 100 | * .system = "<system>", |
101 | * .raw_init = ftrace_raw_init_event_<call>, | 101 | * .raw_init = ftrace_raw_init_event_<call>, |
102 | * .regfunc = ftrace_reg_event_<call>, | 102 | * .regfunc = ftrace_reg_event_<call>, |
103 | * .unregfunc = ftrace_unreg_event_<call>, | 103 | * .unregfunc = ftrace_unreg_event_<call>, |
104 | * .show_format = ftrace_format_<call>, | 104 | * .show_format = ftrace_format_<call>, |
105 | * } | 105 | * } |
106 | * | 106 | * |
@@ -138,7 +138,7 @@ _TRACE_FORMAT(call, PARAMS(proto), PARAMS(args), PARAMS(fmt)) \ | |||
138 | static struct ftrace_event_call __used \ | 138 | static struct ftrace_event_call __used \ |
139 | __attribute__((__aligned__(4))) \ | 139 | __attribute__((__aligned__(4))) \ |
140 | __attribute__((section("_ftrace_events"))) event_##call = { \ | 140 | __attribute__((section("_ftrace_events"))) event_##call = { \ |
141 | .name = #call, \ | 141 | .name = #call, \ |
142 | .system = __stringify(TRACE_SYSTEM), \ | 142 | .system = __stringify(TRACE_SYSTEM), \ |
143 | .regfunc = ftrace_reg_event_##call, \ | 143 | .regfunc = ftrace_reg_event_##call, \ |
144 | .unregfunc = ftrace_unreg_event_##call, \ | 144 | .unregfunc = ftrace_unreg_event_##call, \ |
@@ -163,7 +163,7 @@ static void ftrace_raw_event_##call(proto) \ | |||
163 | pc = preempt_count(); \ | 163 | pc = preempt_count(); \ |
164 | \ | 164 | \ |
165 | event = trace_current_buffer_lock_reserve(event_##call.id, \ | 165 | event = trace_current_buffer_lock_reserve(event_##call.id, \ |
166 | sizeof(struct ftrace_raw_##call), \ | 166 | sizeof(struct ftrace_raw_##call), \ |
167 | irq_flags, pc); \ | 167 | irq_flags, pc); \ |
168 | if (!event) \ | 168 | if (!event) \ |
169 | return; \ | 169 | return; \ |
@@ -208,7 +208,7 @@ static int ftrace_raw_init_event_##call(void) \ | |||
208 | static struct ftrace_event_call __used \ | 208 | static struct ftrace_event_call __used \ |
209 | __attribute__((__aligned__(4))) \ | 209 | __attribute__((__aligned__(4))) \ |
210 | __attribute__((section("_ftrace_events"))) event_##call = { \ | 210 | __attribute__((section("_ftrace_events"))) event_##call = { \ |
211 | .name = #call, \ | 211 | .name = #call, \ |
212 | .system = __stringify(TRACE_SYSTEM), \ | 212 | .system = __stringify(TRACE_SYSTEM), \ |
213 | .raw_init = ftrace_raw_init_event_##call, \ | 213 | .raw_init = ftrace_raw_init_event_##call, \ |
214 | .regfunc = ftrace_raw_reg_event_##call, \ | 214 | .regfunc = ftrace_raw_reg_event_##call, \ |
diff --git a/kernel/trace/trace_export.c b/kernel/trace/trace_export.c index 23ae78430d58..4d9952d3df50 100644 --- a/kernel/trace/trace_export.c +++ b/kernel/trace/trace_export.c | |||
@@ -94,7 +94,7 @@ ftrace_format_##call(struct trace_seq *s) \ | |||
94 | static struct ftrace_event_call __used \ | 94 | static struct ftrace_event_call __used \ |
95 | __attribute__((__aligned__(4))) \ | 95 | __attribute__((__aligned__(4))) \ |
96 | __attribute__((section("_ftrace_events"))) event_##call = { \ | 96 | __attribute__((section("_ftrace_events"))) event_##call = { \ |
97 | .name = #call, \ | 97 | .name = #call, \ |
98 | .id = proto, \ | 98 | .id = proto, \ |
99 | .system = __stringify(TRACE_SYSTEM), \ | 99 | .system = __stringify(TRACE_SYSTEM), \ |
100 | .show_format = ftrace_format_##call, \ | 100 | .show_format = ftrace_format_##call, \ |
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index 453ebd3b636e..d1493b853e41 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
@@ -841,12 +841,12 @@ static void graph_trace_close(struct trace_iterator *iter) | |||
841 | } | 841 | } |
842 | 842 | ||
843 | static struct tracer graph_trace __read_mostly = { | 843 | static struct tracer graph_trace __read_mostly = { |
844 | .name = "function_graph", | 844 | .name = "function_graph", |
845 | .open = graph_trace_open, | 845 | .open = graph_trace_open, |
846 | .close = graph_trace_close, | 846 | .close = graph_trace_close, |
847 | .wait_pipe = poll_wait_pipe, | 847 | .wait_pipe = poll_wait_pipe, |
848 | .init = graph_trace_init, | 848 | .init = graph_trace_init, |
849 | .reset = graph_trace_reset, | 849 | .reset = graph_trace_reset, |
850 | .print_line = print_graph_function, | 850 | .print_line = print_graph_function, |
851 | .print_header = print_graph_headers, | 851 | .print_header = print_graph_headers, |
852 | .flags = &tracer_flags, | 852 | .flags = &tracer_flags, |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index ef8fd661b217..491832af9ba1 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
@@ -565,7 +565,7 @@ static enum print_line_t trace_fn_bin(struct trace_iterator *iter, int flags) | |||
565 | } | 565 | } |
566 | 566 | ||
567 | static struct trace_event trace_fn_event = { | 567 | static struct trace_event trace_fn_event = { |
568 | .type = TRACE_FN, | 568 | .type = TRACE_FN, |
569 | .trace = trace_fn_trace, | 569 | .trace = trace_fn_trace, |
570 | .raw = trace_fn_raw, | 570 | .raw = trace_fn_raw, |
571 | .hex = trace_fn_hex, | 571 | .hex = trace_fn_hex, |
@@ -696,7 +696,7 @@ static enum print_line_t trace_ctxwake_bin(struct trace_iterator *iter, | |||
696 | } | 696 | } |
697 | 697 | ||
698 | static struct trace_event trace_ctx_event = { | 698 | static struct trace_event trace_ctx_event = { |
699 | .type = TRACE_CTX, | 699 | .type = TRACE_CTX, |
700 | .trace = trace_ctx_print, | 700 | .trace = trace_ctx_print, |
701 | .raw = trace_ctx_raw, | 701 | .raw = trace_ctx_raw, |
702 | .hex = trace_ctx_hex, | 702 | .hex = trace_ctx_hex, |
@@ -704,7 +704,7 @@ static struct trace_event trace_ctx_event = { | |||
704 | }; | 704 | }; |
705 | 705 | ||
706 | static struct trace_event trace_wake_event = { | 706 | static struct trace_event trace_wake_event = { |
707 | .type = TRACE_WAKE, | 707 | .type = TRACE_WAKE, |
708 | .trace = trace_wake_print, | 708 | .trace = trace_wake_print, |
709 | .raw = trace_wake_raw, | 709 | .raw = trace_wake_raw, |
710 | .hex = trace_wake_hex, | 710 | .hex = trace_wake_hex, |
@@ -759,7 +759,7 @@ static enum print_line_t trace_special_bin(struct trace_iterator *iter, | |||
759 | } | 759 | } |
760 | 760 | ||
761 | static struct trace_event trace_special_event = { | 761 | static struct trace_event trace_special_event = { |
762 | .type = TRACE_SPECIAL, | 762 | .type = TRACE_SPECIAL, |
763 | .trace = trace_special_print, | 763 | .trace = trace_special_print, |
764 | .raw = trace_special_print, | 764 | .raw = trace_special_print, |
765 | .hex = trace_special_hex, | 765 | .hex = trace_special_hex, |
@@ -796,7 +796,7 @@ static enum print_line_t trace_stack_print(struct trace_iterator *iter, | |||
796 | } | 796 | } |
797 | 797 | ||
798 | static struct trace_event trace_stack_event = { | 798 | static struct trace_event trace_stack_event = { |
799 | .type = TRACE_STACK, | 799 | .type = TRACE_STACK, |
800 | .trace = trace_stack_print, | 800 | .trace = trace_stack_print, |
801 | .raw = trace_special_print, | 801 | .raw = trace_special_print, |
802 | .hex = trace_special_hex, | 802 | .hex = trace_special_hex, |
@@ -825,7 +825,7 @@ static enum print_line_t trace_user_stack_print(struct trace_iterator *iter, | |||
825 | } | 825 | } |
826 | 826 | ||
827 | static struct trace_event trace_user_stack_event = { | 827 | static struct trace_event trace_user_stack_event = { |
828 | .type = TRACE_USER_STACK, | 828 | .type = TRACE_USER_STACK, |
829 | .trace = trace_user_stack_print, | 829 | .trace = trace_user_stack_print, |
830 | .raw = trace_special_print, | 830 | .raw = trace_special_print, |
831 | .hex = trace_special_hex, | 831 | .hex = trace_special_hex, |
@@ -879,7 +879,7 @@ static enum print_line_t trace_print_raw(struct trace_iterator *iter, int flags) | |||
879 | 879 | ||
880 | 880 | ||
881 | static struct trace_event trace_print_event = { | 881 | static struct trace_event trace_print_event = { |
882 | .type = TRACE_PRINT, | 882 | .type = TRACE_PRINT, |
883 | .trace = trace_print_print, | 883 | .trace = trace_print_print, |
884 | .raw = trace_print_raw, | 884 | .raw = trace_print_raw, |
885 | }; | 885 | }; |
diff --git a/kernel/trace/trace_workqueue.c b/kernel/trace/trace_workqueue.c index 4664990fe9c5..e542483df623 100644 --- a/kernel/trace/trace_workqueue.c +++ b/kernel/trace/trace_workqueue.c | |||
@@ -19,14 +19,14 @@ struct cpu_workqueue_stats { | |||
19 | /* Useful to know if we print the cpu headers */ | 19 | /* Useful to know if we print the cpu headers */ |
20 | bool first_entry; | 20 | bool first_entry; |
21 | int cpu; | 21 | int cpu; |
22 | pid_t pid; | 22 | pid_t pid; |
23 | /* Can be inserted from interrupt or user context, need to be atomic */ | 23 | /* Can be inserted from interrupt or user context, need to be atomic */ |
24 | atomic_t inserted; | 24 | atomic_t inserted; |
25 | /* | 25 | /* |
26 | * Don't need to be atomic, works are serialized in a single workqueue thread | 26 | * Don't need to be atomic, works are serialized in a single workqueue thread |
27 | * on a single CPU. | 27 | * on a single CPU. |
28 | */ | 28 | */ |
29 | unsigned int executed; | 29 | unsigned int executed; |
30 | }; | 30 | }; |
31 | 31 | ||
32 | /* List of workqueue threads on one cpu */ | 32 | /* List of workqueue threads on one cpu */ |