diff options
Diffstat (limited to 'kernel/trace/trace.c')
| -rw-r--r-- | kernel/trace/trace.c | 117 |
1 files changed, 67 insertions, 50 deletions
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 0418e2650d41..0c61836e30e7 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -169,10 +169,11 @@ static struct trace_array global_trace; | |||
| 169 | 169 | ||
| 170 | static DEFINE_PER_CPU(struct trace_array_cpu, global_trace_cpu); | 170 | static DEFINE_PER_CPU(struct trace_array_cpu, global_trace_cpu); |
| 171 | 171 | ||
| 172 | int filter_current_check_discard(struct ftrace_event_call *call, void *rec, | 172 | int filter_current_check_discard(struct ring_buffer *buffer, |
| 173 | struct ftrace_event_call *call, void *rec, | ||
| 173 | struct ring_buffer_event *event) | 174 | struct ring_buffer_event *event) |
| 174 | { | 175 | { |
| 175 | return filter_check_discard(call, rec, global_trace.buffer, event); | 176 | return filter_check_discard(call, rec, buffer, event); |
| 176 | } | 177 | } |
| 177 | EXPORT_SYMBOL_GPL(filter_current_check_discard); | 178 | EXPORT_SYMBOL_GPL(filter_current_check_discard); |
| 178 | 179 | ||
| @@ -887,14 +888,15 @@ tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags, | |||
| 887 | } | 888 | } |
| 888 | EXPORT_SYMBOL_GPL(tracing_generic_entry_update); | 889 | EXPORT_SYMBOL_GPL(tracing_generic_entry_update); |
| 889 | 890 | ||
| 890 | struct ring_buffer_event *trace_buffer_lock_reserve(struct trace_array *tr, | 891 | struct ring_buffer_event * |
| 891 | int type, | 892 | trace_buffer_lock_reserve(struct ring_buffer *buffer, |
| 892 | unsigned long len, | 893 | int type, |
| 893 | unsigned long flags, int pc) | 894 | unsigned long len, |
| 895 | unsigned long flags, int pc) | ||
| 894 | { | 896 | { |
| 895 | struct ring_buffer_event *event; | 897 | struct ring_buffer_event *event; |
| 896 | 898 | ||
| 897 | event = ring_buffer_lock_reserve(tr->buffer, len); | 899 | event = ring_buffer_lock_reserve(buffer, len); |
| 898 | if (event != NULL) { | 900 | if (event != NULL) { |
| 899 | struct trace_entry *ent = ring_buffer_event_data(event); | 901 | struct trace_entry *ent = ring_buffer_event_data(event); |
| 900 | 902 | ||
| @@ -905,53 +907,59 @@ struct ring_buffer_event *trace_buffer_lock_reserve(struct trace_array *tr, | |||
| 905 | return event; | 907 | return event; |
| 906 | } | 908 | } |
| 907 | 909 | ||
| 908 | static inline void __trace_buffer_unlock_commit(struct trace_array *tr, | 910 | static inline void |
| 909 | struct ring_buffer_event *event, | 911 | __trace_buffer_unlock_commit(struct ring_buffer *buffer, |
| 910 | unsigned long flags, int pc, | 912 | struct ring_buffer_event *event, |
| 911 | int wake) | 913 | unsigned long flags, int pc, |
| 914 | int wake) | ||
| 912 | { | 915 | { |
| 913 | ring_buffer_unlock_commit(tr->buffer, event); | 916 | ring_buffer_unlock_commit(buffer, event); |
| 914 | 917 | ||
| 915 | ftrace_trace_stack(tr, flags, 6, pc); | 918 | ftrace_trace_stack(buffer, flags, 6, pc); |
| 916 | ftrace_trace_userstack(tr, flags, pc); | 919 | ftrace_trace_userstack(buffer, flags, pc); |
| 917 | 920 | ||
| 918 | if (wake) | 921 | if (wake) |
| 919 | trace_wake_up(); | 922 | trace_wake_up(); |
| 920 | } | 923 | } |
| 921 | 924 | ||
| 922 | void trace_buffer_unlock_commit(struct trace_array *tr, | 925 | void trace_buffer_unlock_commit(struct ring_buffer *buffer, |
| 923 | struct ring_buffer_event *event, | 926 | struct ring_buffer_event *event, |
| 924 | unsigned long flags, int pc) | 927 | unsigned long flags, int pc) |
| 925 | { | 928 | { |
| 926 | __trace_buffer_unlock_commit(tr, event, flags, pc, 1); | 929 | __trace_buffer_unlock_commit(buffer, event, flags, pc, 1); |
| 927 | } | 930 | } |
| 928 | 931 | ||
| 929 | struct ring_buffer_event * | 932 | struct ring_buffer_event * |
| 930 | trace_current_buffer_lock_reserve(int type, unsigned long len, | 933 | trace_current_buffer_lock_reserve(struct ring_buffer **current_rb, |
| 934 | int type, unsigned long len, | ||
| 931 | unsigned long flags, int pc) | 935 | unsigned long flags, int pc) |
| 932 | { | 936 | { |
| 933 | return trace_buffer_lock_reserve(&global_trace, | 937 | *current_rb = global_trace.buffer; |
| 938 | return trace_buffer_lock_reserve(*current_rb, | ||
| 934 | type, len, flags, pc); | 939 | type, len, flags, pc); |
| 935 | } | 940 | } |
| 936 | EXPORT_SYMBOL_GPL(trace_current_buffer_lock_reserve); | 941 | EXPORT_SYMBOL_GPL(trace_current_buffer_lock_reserve); |
| 937 | 942 | ||
| 938 | void trace_current_buffer_unlock_commit(struct ring_buffer_event *event, | 943 | void trace_current_buffer_unlock_commit(struct ring_buffer *buffer, |
| 944 | struct ring_buffer_event *event, | ||
| 939 | unsigned long flags, int pc) | 945 | unsigned long flags, int pc) |
| 940 | { | 946 | { |
| 941 | __trace_buffer_unlock_commit(&global_trace, event, flags, pc, 1); | 947 | __trace_buffer_unlock_commit(buffer, event, flags, pc, 1); |
| 942 | } | 948 | } |
| 943 | EXPORT_SYMBOL_GPL(trace_current_buffer_unlock_commit); | 949 | EXPORT_SYMBOL_GPL(trace_current_buffer_unlock_commit); |
| 944 | 950 | ||
| 945 | void trace_nowake_buffer_unlock_commit(struct ring_buffer_event *event, | 951 | void trace_nowake_buffer_unlock_commit(struct ring_buffer *buffer, |
| 946 | unsigned long flags, int pc) | 952 | struct ring_buffer_event *event, |
| 953 | unsigned long flags, int pc) | ||
| 947 | { | 954 | { |
| 948 | __trace_buffer_unlock_commit(&global_trace, event, flags, pc, 0); | 955 | __trace_buffer_unlock_commit(buffer, event, flags, pc, 0); |
| 949 | } | 956 | } |
| 950 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); | 957 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); |
| 951 | 958 | ||
| 952 | void trace_current_buffer_discard_commit(struct ring_buffer_event *event) | 959 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, |
| 960 | struct ring_buffer_event *event) | ||
| 953 | { | 961 | { |
| 954 | ring_buffer_discard_commit(global_trace.buffer, event); | 962 | ring_buffer_discard_commit(buffer, event); |
| 955 | } | 963 | } |
| 956 | EXPORT_SYMBOL_GPL(trace_current_buffer_discard_commit); | 964 | EXPORT_SYMBOL_GPL(trace_current_buffer_discard_commit); |
| 957 | 965 | ||
| @@ -961,6 +969,7 @@ trace_function(struct trace_array *tr, | |||
| 961 | int pc) | 969 | int pc) |
| 962 | { | 970 | { |
| 963 | struct ftrace_event_call *call = &event_function; | 971 | struct ftrace_event_call *call = &event_function; |
| 972 | struct ring_buffer *buffer = tr->buffer; | ||
| 964 | struct ring_buffer_event *event; | 973 | struct ring_buffer_event *event; |
| 965 | struct ftrace_entry *entry; | 974 | struct ftrace_entry *entry; |
| 966 | 975 | ||
| @@ -968,7 +977,7 @@ trace_function(struct trace_array *tr, | |||
| 968 | if (unlikely(local_read(&__get_cpu_var(ftrace_cpu_disabled)))) | 977 | if (unlikely(local_read(&__get_cpu_var(ftrace_cpu_disabled)))) |
| 969 | return; | 978 | return; |
| 970 | 979 | ||
| 971 | event = trace_buffer_lock_reserve(tr, TRACE_FN, sizeof(*entry), | 980 | event = trace_buffer_lock_reserve(buffer, TRACE_FN, sizeof(*entry), |
| 972 | flags, pc); | 981 | flags, pc); |
| 973 | if (!event) | 982 | if (!event) |
| 974 | return; | 983 | return; |
| @@ -976,8 +985,8 @@ trace_function(struct trace_array *tr, | |||
| 976 | entry->ip = ip; | 985 | entry->ip = ip; |
| 977 | entry->parent_ip = parent_ip; | 986 | entry->parent_ip = parent_ip; |
| 978 | 987 | ||
| 979 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 988 | if (!filter_check_discard(call, entry, buffer, event)) |
| 980 | ring_buffer_unlock_commit(tr->buffer, event); | 989 | ring_buffer_unlock_commit(buffer, event); |
| 981 | } | 990 | } |
| 982 | 991 | ||
| 983 | void | 992 | void |
| @@ -990,7 +999,7 @@ ftrace(struct trace_array *tr, struct trace_array_cpu *data, | |||
| 990 | } | 999 | } |
| 991 | 1000 | ||
| 992 | #ifdef CONFIG_STACKTRACE | 1001 | #ifdef CONFIG_STACKTRACE |
| 993 | static void __ftrace_trace_stack(struct trace_array *tr, | 1002 | static void __ftrace_trace_stack(struct ring_buffer *buffer, |
| 994 | unsigned long flags, | 1003 | unsigned long flags, |
| 995 | int skip, int pc) | 1004 | int skip, int pc) |
| 996 | { | 1005 | { |
| @@ -999,7 +1008,7 @@ static void __ftrace_trace_stack(struct trace_array *tr, | |||
| 999 | struct stack_entry *entry; | 1008 | struct stack_entry *entry; |
| 1000 | struct stack_trace trace; | 1009 | struct stack_trace trace; |
| 1001 | 1010 | ||
| 1002 | event = trace_buffer_lock_reserve(tr, TRACE_STACK, | 1011 | event = trace_buffer_lock_reserve(buffer, TRACE_STACK, |
| 1003 | sizeof(*entry), flags, pc); | 1012 | sizeof(*entry), flags, pc); |
| 1004 | if (!event) | 1013 | if (!event) |
| 1005 | return; | 1014 | return; |
| @@ -1012,26 +1021,27 @@ static void __ftrace_trace_stack(struct trace_array *tr, | |||
| 1012 | trace.entries = entry->caller; | 1021 | trace.entries = entry->caller; |
| 1013 | 1022 | ||
| 1014 | save_stack_trace(&trace); | 1023 | save_stack_trace(&trace); |
| 1015 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 1024 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1016 | ring_buffer_unlock_commit(tr->buffer, event); | 1025 | ring_buffer_unlock_commit(buffer, event); |
| 1017 | } | 1026 | } |
| 1018 | 1027 | ||
| 1019 | void ftrace_trace_stack(struct trace_array *tr, unsigned long flags, int skip, | 1028 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, |
| 1020 | int pc) | 1029 | int skip, int pc) |
| 1021 | { | 1030 | { |
| 1022 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) | 1031 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) |
| 1023 | return; | 1032 | return; |
| 1024 | 1033 | ||
| 1025 | __ftrace_trace_stack(tr, flags, skip, pc); | 1034 | __ftrace_trace_stack(buffer, flags, skip, pc); |
| 1026 | } | 1035 | } |
| 1027 | 1036 | ||
| 1028 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, | 1037 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, |
| 1029 | int pc) | 1038 | int pc) |
| 1030 | { | 1039 | { |
| 1031 | __ftrace_trace_stack(tr, flags, skip, pc); | 1040 | __ftrace_trace_stack(tr->buffer, flags, skip, pc); |
| 1032 | } | 1041 | } |
| 1033 | 1042 | ||
| 1034 | void ftrace_trace_userstack(struct trace_array *tr, unsigned long flags, int pc) | 1043 | void |
| 1044 | ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) | ||
| 1035 | { | 1045 | { |
| 1036 | struct ftrace_event_call *call = &event_user_stack; | 1046 | struct ftrace_event_call *call = &event_user_stack; |
| 1037 | struct ring_buffer_event *event; | 1047 | struct ring_buffer_event *event; |
| @@ -1041,7 +1051,7 @@ void ftrace_trace_userstack(struct trace_array *tr, unsigned long flags, int pc) | |||
| 1041 | if (!(trace_flags & TRACE_ITER_USERSTACKTRACE)) | 1051 | if (!(trace_flags & TRACE_ITER_USERSTACKTRACE)) |
| 1042 | return; | 1052 | return; |
| 1043 | 1053 | ||
| 1044 | event = trace_buffer_lock_reserve(tr, TRACE_USER_STACK, | 1054 | event = trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, |
| 1045 | sizeof(*entry), flags, pc); | 1055 | sizeof(*entry), flags, pc); |
| 1046 | if (!event) | 1056 | if (!event) |
| 1047 | return; | 1057 | return; |
| @@ -1055,8 +1065,8 @@ void ftrace_trace_userstack(struct trace_array *tr, unsigned long flags, int pc) | |||
| 1055 | trace.entries = entry->caller; | 1065 | trace.entries = entry->caller; |
| 1056 | 1066 | ||
| 1057 | save_stack_trace_user(&trace); | 1067 | save_stack_trace_user(&trace); |
| 1058 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 1068 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1059 | ring_buffer_unlock_commit(tr->buffer, event); | 1069 | ring_buffer_unlock_commit(buffer, event); |
| 1060 | } | 1070 | } |
| 1061 | 1071 | ||
| 1062 | #ifdef UNUSED | 1072 | #ifdef UNUSED |
| @@ -1075,9 +1085,10 @@ ftrace_trace_special(void *__tr, | |||
| 1075 | { | 1085 | { |
| 1076 | struct ring_buffer_event *event; | 1086 | struct ring_buffer_event *event; |
| 1077 | struct trace_array *tr = __tr; | 1087 | struct trace_array *tr = __tr; |
| 1088 | struct ring_buffer *buffer = tr->buffer; | ||
| 1078 | struct special_entry *entry; | 1089 | struct special_entry *entry; |
| 1079 | 1090 | ||
| 1080 | event = trace_buffer_lock_reserve(tr, TRACE_SPECIAL, | 1091 | event = trace_buffer_lock_reserve(buffer, TRACE_SPECIAL, |
| 1081 | sizeof(*entry), 0, pc); | 1092 | sizeof(*entry), 0, pc); |
| 1082 | if (!event) | 1093 | if (!event) |
| 1083 | return; | 1094 | return; |
| @@ -1085,7 +1096,7 @@ ftrace_trace_special(void *__tr, | |||
| 1085 | entry->arg1 = arg1; | 1096 | entry->arg1 = arg1; |
| 1086 | entry->arg2 = arg2; | 1097 | entry->arg2 = arg2; |
| 1087 | entry->arg3 = arg3; | 1098 | entry->arg3 = arg3; |
| 1088 | trace_buffer_unlock_commit(tr, event, 0, pc); | 1099 | trace_buffer_unlock_commit(buffer, event, 0, pc); |
| 1089 | } | 1100 | } |
| 1090 | 1101 | ||
| 1091 | void | 1102 | void |
| @@ -1131,6 +1142,7 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1131 | 1142 | ||
| 1132 | struct ftrace_event_call *call = &event_bprint; | 1143 | struct ftrace_event_call *call = &event_bprint; |
| 1133 | struct ring_buffer_event *event; | 1144 | struct ring_buffer_event *event; |
| 1145 | struct ring_buffer *buffer; | ||
| 1134 | struct trace_array *tr = &global_trace; | 1146 | struct trace_array *tr = &global_trace; |
| 1135 | struct trace_array_cpu *data; | 1147 | struct trace_array_cpu *data; |
| 1136 | struct bprint_entry *entry; | 1148 | struct bprint_entry *entry; |
| @@ -1163,7 +1175,9 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1163 | goto out_unlock; | 1175 | goto out_unlock; |
| 1164 | 1176 | ||
| 1165 | size = sizeof(*entry) + sizeof(u32) * len; | 1177 | size = sizeof(*entry) + sizeof(u32) * len; |
| 1166 | event = trace_buffer_lock_reserve(tr, TRACE_BPRINT, size, flags, pc); | 1178 | buffer = tr->buffer; |
| 1179 | event = trace_buffer_lock_reserve(buffer, TRACE_BPRINT, size, | ||
| 1180 | flags, pc); | ||
| 1167 | if (!event) | 1181 | if (!event) |
| 1168 | goto out_unlock; | 1182 | goto out_unlock; |
| 1169 | entry = ring_buffer_event_data(event); | 1183 | entry = ring_buffer_event_data(event); |
| @@ -1171,8 +1185,8 @@ int trace_vbprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1171 | entry->fmt = fmt; | 1185 | entry->fmt = fmt; |
| 1172 | 1186 | ||
| 1173 | memcpy(entry->buf, trace_buf, sizeof(u32) * len); | 1187 | memcpy(entry->buf, trace_buf, sizeof(u32) * len); |
| 1174 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 1188 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1175 | ring_buffer_unlock_commit(tr->buffer, event); | 1189 | ring_buffer_unlock_commit(buffer, event); |
| 1176 | 1190 | ||
| 1177 | out_unlock: | 1191 | out_unlock: |
| 1178 | __raw_spin_unlock(&trace_buf_lock); | 1192 | __raw_spin_unlock(&trace_buf_lock); |
| @@ -1194,6 +1208,7 @@ int trace_vprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1194 | 1208 | ||
| 1195 | struct ftrace_event_call *call = &event_print; | 1209 | struct ftrace_event_call *call = &event_print; |
| 1196 | struct ring_buffer_event *event; | 1210 | struct ring_buffer_event *event; |
| 1211 | struct ring_buffer *buffer; | ||
| 1197 | struct trace_array *tr = &global_trace; | 1212 | struct trace_array *tr = &global_trace; |
| 1198 | struct trace_array_cpu *data; | 1213 | struct trace_array_cpu *data; |
| 1199 | int cpu, len = 0, size, pc; | 1214 | int cpu, len = 0, size, pc; |
| @@ -1222,7 +1237,9 @@ int trace_vprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1222 | trace_buf[len] = 0; | 1237 | trace_buf[len] = 0; |
| 1223 | 1238 | ||
| 1224 | size = sizeof(*entry) + len + 1; | 1239 | size = sizeof(*entry) + len + 1; |
| 1225 | event = trace_buffer_lock_reserve(tr, TRACE_PRINT, size, irq_flags, pc); | 1240 | buffer = tr->buffer; |
| 1241 | event = trace_buffer_lock_reserve(buffer, TRACE_PRINT, size, | ||
| 1242 | irq_flags, pc); | ||
| 1226 | if (!event) | 1243 | if (!event) |
| 1227 | goto out_unlock; | 1244 | goto out_unlock; |
| 1228 | entry = ring_buffer_event_data(event); | 1245 | entry = ring_buffer_event_data(event); |
| @@ -1230,8 +1247,8 @@ int trace_vprintk(unsigned long ip, const char *fmt, va_list args) | |||
| 1230 | 1247 | ||
| 1231 | memcpy(&entry->buf, trace_buf, len); | 1248 | memcpy(&entry->buf, trace_buf, len); |
| 1232 | entry->buf[len] = 0; | 1249 | entry->buf[len] = 0; |
| 1233 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 1250 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1234 | ring_buffer_unlock_commit(tr->buffer, event); | 1251 | ring_buffer_unlock_commit(buffer, event); |
| 1235 | 1252 | ||
| 1236 | out_unlock: | 1253 | out_unlock: |
| 1237 | __raw_spin_unlock(&trace_buf_lock); | 1254 | __raw_spin_unlock(&trace_buf_lock); |
