diff options
| author | Frederic Weisbecker <fweisbec@gmail.com> | 2009-09-23 17:08:43 -0400 |
|---|---|---|
| committer | Frederic Weisbecker <fweisbec@gmail.com> | 2009-09-23 17:08:43 -0400 |
| commit | d7a4b414eed51f1653bb05ebe84122bf9a7ae18b (patch) | |
| tree | bd6603a0c27de4c138a1767871897e9cd3e1a1d2 /kernel/trace | |
| parent | 1f0ab40976460bc4673fa204ce917a725185d8f2 (diff) | |
| parent | a724eada8c2a7b62463b73ccf73fd0bb6e928aeb (diff) | |
Merge commit 'linus/master' into tracing/kprobes
Conflicts:
kernel/trace/Makefile
kernel/trace/trace.h
kernel/trace/trace_event_types.h
kernel/trace/trace_export.c
Merge reason:
Sync with latest significant tracing core changes.
Diffstat (limited to 'kernel/trace')
27 files changed, 1256 insertions, 1114 deletions
diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig index e78dcbde1a81..15372a9f2399 100644 --- a/kernel/trace/Kconfig +++ b/kernel/trace/Kconfig | |||
| @@ -11,12 +11,18 @@ config NOP_TRACER | |||
| 11 | 11 | ||
| 12 | config HAVE_FTRACE_NMI_ENTER | 12 | config HAVE_FTRACE_NMI_ENTER |
| 13 | bool | 13 | bool |
| 14 | help | ||
| 15 | See Documentation/trace/ftrace-implementation.txt | ||
| 14 | 16 | ||
| 15 | config HAVE_FUNCTION_TRACER | 17 | config HAVE_FUNCTION_TRACER |
| 16 | bool | 18 | bool |
| 19 | help | ||
| 20 | See Documentation/trace/ftrace-implementation.txt | ||
| 17 | 21 | ||
| 18 | config HAVE_FUNCTION_GRAPH_TRACER | 22 | config HAVE_FUNCTION_GRAPH_TRACER |
| 19 | bool | 23 | bool |
| 24 | help | ||
| 25 | See Documentation/trace/ftrace-implementation.txt | ||
| 20 | 26 | ||
| 21 | config HAVE_FUNCTION_GRAPH_FP_TEST | 27 | config HAVE_FUNCTION_GRAPH_FP_TEST |
| 22 | bool | 28 | bool |
| @@ -28,21 +34,25 @@ config HAVE_FUNCTION_GRAPH_FP_TEST | |||
| 28 | config HAVE_FUNCTION_TRACE_MCOUNT_TEST | 34 | config HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 29 | bool | 35 | bool |
| 30 | help | 36 | help |
| 31 | This gets selected when the arch tests the function_trace_stop | 37 | See Documentation/trace/ftrace-implementation.txt |
| 32 | variable at the mcount call site. Otherwise, this variable | ||
| 33 | is tested by the called function. | ||
| 34 | 38 | ||
| 35 | config HAVE_DYNAMIC_FTRACE | 39 | config HAVE_DYNAMIC_FTRACE |
| 36 | bool | 40 | bool |
| 41 | help | ||
| 42 | See Documentation/trace/ftrace-implementation.txt | ||
| 37 | 43 | ||
| 38 | config HAVE_FTRACE_MCOUNT_RECORD | 44 | config HAVE_FTRACE_MCOUNT_RECORD |
| 39 | bool | 45 | bool |
| 46 | help | ||
| 47 | See Documentation/trace/ftrace-implementation.txt | ||
| 40 | 48 | ||
| 41 | config HAVE_HW_BRANCH_TRACER | 49 | config HAVE_HW_BRANCH_TRACER |
| 42 | bool | 50 | bool |
| 43 | 51 | ||
| 44 | config HAVE_SYSCALL_TRACEPOINTS | 52 | config HAVE_SYSCALL_TRACEPOINTS |
| 45 | bool | 53 | bool |
| 54 | help | ||
| 55 | See Documentation/trace/ftrace-implementation.txt | ||
| 46 | 56 | ||
| 47 | config TRACER_MAX_TRACE | 57 | config TRACER_MAX_TRACE |
| 48 | bool | 58 | bool |
| @@ -73,7 +83,7 @@ config RING_BUFFER_ALLOW_SWAP | |||
| 73 | # This allows those options to appear when no other tracer is selected. But the | 83 | # This allows those options to appear when no other tracer is selected. But the |
| 74 | # options do not appear when something else selects it. We need the two options | 84 | # options do not appear when something else selects it. We need the two options |
| 75 | # GENERIC_TRACER and TRACING to avoid circular dependencies to accomplish the | 85 | # GENERIC_TRACER and TRACING to avoid circular dependencies to accomplish the |
| 76 | # hidding of the automatic options options. | 86 | # hidding of the automatic options. |
| 77 | 87 | ||
| 78 | config TRACING | 88 | config TRACING |
| 79 | bool | 89 | bool |
| @@ -481,6 +491,18 @@ config FTRACE_STARTUP_TEST | |||
| 481 | functioning properly. It will do tests on all the configured | 491 | functioning properly. It will do tests on all the configured |
| 482 | tracers of ftrace. | 492 | tracers of ftrace. |
| 483 | 493 | ||
| 494 | config EVENT_TRACE_TEST_SYSCALLS | ||
| 495 | bool "Run selftest on syscall events" | ||
| 496 | depends on FTRACE_STARTUP_TEST | ||
| 497 | help | ||
| 498 | This option will also enable testing every syscall event. | ||
| 499 | It only enables the event and disables it and runs various loads | ||
| 500 | with the event enabled. This adds a bit more time for kernel boot | ||
| 501 | up since it runs this on every system call defined. | ||
| 502 | |||
| 503 | TBD - enable a way to actually call the syscalls as we test their | ||
| 504 | events | ||
| 505 | |||
| 484 | config MMIOTRACE | 506 | config MMIOTRACE |
| 485 | bool "Memory mapped IO tracing" | 507 | bool "Memory mapped IO tracing" |
| 486 | depends on HAVE_MMIOTRACE_SUPPORT && PCI | 508 | depends on HAVE_MMIOTRACE_SUPPORT && PCI |
diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile index 7c00a1ec1496..c8cb75d7f280 100644 --- a/kernel/trace/Makefile +++ b/kernel/trace/Makefile | |||
| @@ -42,7 +42,6 @@ obj-$(CONFIG_BOOT_TRACER) += trace_boot.o | |||
| 42 | obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o | 42 | obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o |
| 43 | obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o | 43 | obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o |
| 44 | obj-$(CONFIG_HW_BRANCH_TRACER) += trace_hw_branches.o | 44 | obj-$(CONFIG_HW_BRANCH_TRACER) += trace_hw_branches.o |
| 45 | obj-$(CONFIG_POWER_TRACER) += trace_power.o | ||
| 46 | obj-$(CONFIG_KMEMTRACE) += kmemtrace.o | 45 | obj-$(CONFIG_KMEMTRACE) += kmemtrace.o |
| 47 | obj-$(CONFIG_WORKQUEUE_TRACER) += trace_workqueue.o | 46 | obj-$(CONFIG_WORKQUEUE_TRACER) += trace_workqueue.o |
| 48 | obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o | 47 | obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o |
| @@ -55,5 +54,6 @@ obj-$(CONFIG_FTRACE_SYSCALLS) += trace_syscalls.o | |||
| 55 | obj-$(CONFIG_EVENT_PROFILE) += trace_event_profile.o | 54 | obj-$(CONFIG_EVENT_PROFILE) += trace_event_profile.o |
| 56 | obj-$(CONFIG_EVENT_TRACING) += trace_events_filter.o | 55 | obj-$(CONFIG_EVENT_TRACING) += trace_events_filter.o |
| 57 | obj-$(CONFIG_KPROBE_TRACER) += trace_kprobe.o | 56 | obj-$(CONFIG_KPROBE_TRACER) += trace_kprobe.o |
| 57 | obj-$(CONFIG_EVENT_TRACING) += power-traces.o | ||
| 58 | 58 | ||
| 59 | libftrace-y := ftrace.o | 59 | libftrace-y := ftrace.o |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 8c804e24f96f..23df7771c937 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -1323,11 +1323,10 @@ static int __init ftrace_dyn_table_alloc(unsigned long num_to_init) | |||
| 1323 | 1323 | ||
| 1324 | enum { | 1324 | enum { |
| 1325 | FTRACE_ITER_FILTER = (1 << 0), | 1325 | FTRACE_ITER_FILTER = (1 << 0), |
| 1326 | FTRACE_ITER_CONT = (1 << 1), | 1326 | FTRACE_ITER_NOTRACE = (1 << 1), |
| 1327 | FTRACE_ITER_NOTRACE = (1 << 2), | 1327 | FTRACE_ITER_FAILURES = (1 << 2), |
| 1328 | FTRACE_ITER_FAILURES = (1 << 3), | 1328 | FTRACE_ITER_PRINTALL = (1 << 3), |
| 1329 | FTRACE_ITER_PRINTALL = (1 << 4), | 1329 | FTRACE_ITER_HASH = (1 << 4), |
| 1330 | FTRACE_ITER_HASH = (1 << 5), | ||
| 1331 | }; | 1330 | }; |
| 1332 | 1331 | ||
| 1333 | #define FTRACE_BUFF_MAX (KSYM_SYMBOL_LEN+4) /* room for wildcards */ | 1332 | #define FTRACE_BUFF_MAX (KSYM_SYMBOL_LEN+4) /* room for wildcards */ |
| @@ -1337,8 +1336,7 @@ struct ftrace_iterator { | |||
| 1337 | int hidx; | 1336 | int hidx; |
| 1338 | int idx; | 1337 | int idx; |
| 1339 | unsigned flags; | 1338 | unsigned flags; |
| 1340 | unsigned char buffer[FTRACE_BUFF_MAX+1]; | 1339 | struct trace_parser parser; |
| 1341 | unsigned buffer_idx; | ||
| 1342 | }; | 1340 | }; |
| 1343 | 1341 | ||
| 1344 | static void * | 1342 | static void * |
| @@ -1407,7 +1405,7 @@ static int t_hash_show(struct seq_file *m, void *v) | |||
| 1407 | if (rec->ops->print) | 1405 | if (rec->ops->print) |
| 1408 | return rec->ops->print(m, rec->ip, rec->ops, rec->data); | 1406 | return rec->ops->print(m, rec->ip, rec->ops, rec->data); |
| 1409 | 1407 | ||
| 1410 | seq_printf(m, "%pf:%pf", (void *)rec->ip, (void *)rec->ops->func); | 1408 | seq_printf(m, "%ps:%ps", (void *)rec->ip, (void *)rec->ops->func); |
| 1411 | 1409 | ||
| 1412 | if (rec->data) | 1410 | if (rec->data) |
| 1413 | seq_printf(m, ":%p", rec->data); | 1411 | seq_printf(m, ":%p", rec->data); |
| @@ -1517,12 +1515,12 @@ static int t_show(struct seq_file *m, void *v) | |||
| 1517 | if (!rec) | 1515 | if (!rec) |
| 1518 | return 0; | 1516 | return 0; |
| 1519 | 1517 | ||
| 1520 | seq_printf(m, "%pf\n", (void *)rec->ip); | 1518 | seq_printf(m, "%ps\n", (void *)rec->ip); |
| 1521 | 1519 | ||
| 1522 | return 0; | 1520 | return 0; |
| 1523 | } | 1521 | } |
| 1524 | 1522 | ||
| 1525 | static struct seq_operations show_ftrace_seq_ops = { | 1523 | static const struct seq_operations show_ftrace_seq_ops = { |
| 1526 | .start = t_start, | 1524 | .start = t_start, |
| 1527 | .next = t_next, | 1525 | .next = t_next, |
| 1528 | .stop = t_stop, | 1526 | .stop = t_stop, |
| @@ -1604,6 +1602,11 @@ ftrace_regex_open(struct inode *inode, struct file *file, int enable) | |||
| 1604 | if (!iter) | 1602 | if (!iter) |
| 1605 | return -ENOMEM; | 1603 | return -ENOMEM; |
| 1606 | 1604 | ||
| 1605 | if (trace_parser_get_init(&iter->parser, FTRACE_BUFF_MAX)) { | ||
| 1606 | kfree(iter); | ||
| 1607 | return -ENOMEM; | ||
| 1608 | } | ||
| 1609 | |||
| 1607 | mutex_lock(&ftrace_regex_lock); | 1610 | mutex_lock(&ftrace_regex_lock); |
| 1608 | if ((file->f_mode & FMODE_WRITE) && | 1611 | if ((file->f_mode & FMODE_WRITE) && |
| 1609 | (file->f_flags & O_TRUNC)) | 1612 | (file->f_flags & O_TRUNC)) |
| @@ -2059,9 +2062,9 @@ __unregister_ftrace_function_probe(char *glob, struct ftrace_probe_ops *ops, | |||
| 2059 | int i, len = 0; | 2062 | int i, len = 0; |
| 2060 | char *search; | 2063 | char *search; |
| 2061 | 2064 | ||
| 2062 | if (glob && (strcmp(glob, "*") || !strlen(glob))) | 2065 | if (glob && (strcmp(glob, "*") == 0 || !strlen(glob))) |
| 2063 | glob = NULL; | 2066 | glob = NULL; |
| 2064 | else { | 2067 | else if (glob) { |
| 2065 | int not; | 2068 | int not; |
| 2066 | 2069 | ||
| 2067 | type = ftrace_setup_glob(glob, strlen(glob), &search, ¬); | 2070 | type = ftrace_setup_glob(glob, strlen(glob), &search, ¬); |
| @@ -2196,9 +2199,8 @@ ftrace_regex_write(struct file *file, const char __user *ubuf, | |||
| 2196 | size_t cnt, loff_t *ppos, int enable) | 2199 | size_t cnt, loff_t *ppos, int enable) |
| 2197 | { | 2200 | { |
| 2198 | struct ftrace_iterator *iter; | 2201 | struct ftrace_iterator *iter; |
| 2199 | char ch; | 2202 | struct trace_parser *parser; |
| 2200 | size_t read = 0; | 2203 | ssize_t ret, read; |
| 2201 | ssize_t ret; | ||
| 2202 | 2204 | ||
| 2203 | if (!cnt || cnt < 0) | 2205 | if (!cnt || cnt < 0) |
| 2204 | return 0; | 2206 | return 0; |
| @@ -2211,72 +2213,23 @@ ftrace_regex_write(struct file *file, const char __user *ubuf, | |||
| 2211 | } else | 2213 | } else |
| 2212 | iter = file->private_data; | 2214 | iter = file->private_data; |
| 2213 | 2215 | ||
| 2214 | if (!*ppos) { | 2216 | parser = &iter->parser; |
| 2215 | iter->flags &= ~FTRACE_ITER_CONT; | 2217 | read = trace_get_user(parser, ubuf, cnt, ppos); |
| 2216 | iter->buffer_idx = 0; | ||
| 2217 | } | ||
| 2218 | 2218 | ||
| 2219 | ret = get_user(ch, ubuf++); | 2219 | if (trace_parser_loaded(parser) && |
| 2220 | if (ret) | 2220 | !trace_parser_cont(parser)) { |
| 2221 | goto out; | 2221 | ret = ftrace_process_regex(parser->buffer, |
| 2222 | read++; | 2222 | parser->idx, enable); |
| 2223 | cnt--; | ||
| 2224 | |||
| 2225 | /* | ||
| 2226 | * If the parser haven't finished with the last write, | ||
| 2227 | * continue reading the user input without skipping spaces. | ||
| 2228 | */ | ||
| 2229 | if (!(iter->flags & FTRACE_ITER_CONT)) { | ||
| 2230 | /* skip white space */ | ||
| 2231 | while (cnt && isspace(ch)) { | ||
| 2232 | ret = get_user(ch, ubuf++); | ||
| 2233 | if (ret) | ||
| 2234 | goto out; | ||
| 2235 | read++; | ||
| 2236 | cnt--; | ||
| 2237 | } | ||
| 2238 | |||
| 2239 | /* only spaces were written */ | ||
| 2240 | if (isspace(ch)) { | ||
| 2241 | *ppos += read; | ||
| 2242 | ret = read; | ||
| 2243 | goto out; | ||
| 2244 | } | ||
| 2245 | |||
| 2246 | iter->buffer_idx = 0; | ||
| 2247 | } | ||
| 2248 | |||
| 2249 | while (cnt && !isspace(ch)) { | ||
| 2250 | if (iter->buffer_idx < FTRACE_BUFF_MAX) | ||
| 2251 | iter->buffer[iter->buffer_idx++] = ch; | ||
| 2252 | else { | ||
| 2253 | ret = -EINVAL; | ||
| 2254 | goto out; | ||
| 2255 | } | ||
| 2256 | ret = get_user(ch, ubuf++); | ||
| 2257 | if (ret) | 2223 | if (ret) |
| 2258 | goto out; | 2224 | goto out; |
| 2259 | read++; | ||
| 2260 | cnt--; | ||
| 2261 | } | ||
| 2262 | 2225 | ||
| 2263 | if (isspace(ch)) { | 2226 | trace_parser_clear(parser); |
| 2264 | iter->buffer[iter->buffer_idx] = 0; | ||
| 2265 | ret = ftrace_process_regex(iter->buffer, | ||
| 2266 | iter->buffer_idx, enable); | ||
| 2267 | if (ret) | ||
| 2268 | goto out; | ||
| 2269 | iter->buffer_idx = 0; | ||
| 2270 | } else { | ||
| 2271 | iter->flags |= FTRACE_ITER_CONT; | ||
| 2272 | iter->buffer[iter->buffer_idx++] = ch; | ||
| 2273 | } | 2227 | } |
| 2274 | 2228 | ||
| 2275 | *ppos += read; | ||
| 2276 | ret = read; | 2229 | ret = read; |
| 2277 | out: | ||
| 2278 | mutex_unlock(&ftrace_regex_lock); | ||
| 2279 | 2230 | ||
| 2231 | mutex_unlock(&ftrace_regex_lock); | ||
| 2232 | out: | ||
| 2280 | return ret; | 2233 | return ret; |
| 2281 | } | 2234 | } |
| 2282 | 2235 | ||
| @@ -2381,6 +2334,7 @@ ftrace_regex_release(struct inode *inode, struct file *file, int enable) | |||
| 2381 | { | 2334 | { |
| 2382 | struct seq_file *m = (struct seq_file *)file->private_data; | 2335 | struct seq_file *m = (struct seq_file *)file->private_data; |
| 2383 | struct ftrace_iterator *iter; | 2336 | struct ftrace_iterator *iter; |
| 2337 | struct trace_parser *parser; | ||
| 2384 | 2338 | ||
| 2385 | mutex_lock(&ftrace_regex_lock); | 2339 | mutex_lock(&ftrace_regex_lock); |
| 2386 | if (file->f_mode & FMODE_READ) { | 2340 | if (file->f_mode & FMODE_READ) { |
| @@ -2390,9 +2344,10 @@ ftrace_regex_release(struct inode *inode, struct file *file, int enable) | |||
| 2390 | } else | 2344 | } else |
| 2391 | iter = file->private_data; | 2345 | iter = file->private_data; |
| 2392 | 2346 | ||
| 2393 | if (iter->buffer_idx) { | 2347 | parser = &iter->parser; |
| 2394 | iter->buffer[iter->buffer_idx] = 0; | 2348 | if (trace_parser_loaded(parser)) { |
| 2395 | ftrace_match_records(iter->buffer, iter->buffer_idx, enable); | 2349 | parser->buffer[parser->idx] = 0; |
| 2350 | ftrace_match_records(parser->buffer, parser->idx, enable); | ||
| 2396 | } | 2351 | } |
| 2397 | 2352 | ||
| 2398 | mutex_lock(&ftrace_lock); | 2353 | mutex_lock(&ftrace_lock); |
| @@ -2400,7 +2355,9 @@ ftrace_regex_release(struct inode *inode, struct file *file, int enable) | |||
| 2400 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); | 2355 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); |
| 2401 | mutex_unlock(&ftrace_lock); | 2356 | mutex_unlock(&ftrace_lock); |
| 2402 | 2357 | ||
| 2358 | trace_parser_put(parser); | ||
| 2403 | kfree(iter); | 2359 | kfree(iter); |
| 2360 | |||
| 2404 | mutex_unlock(&ftrace_regex_lock); | 2361 | mutex_unlock(&ftrace_regex_lock); |
| 2405 | return 0; | 2362 | return 0; |
| 2406 | } | 2363 | } |
| @@ -2457,11 +2414,9 @@ unsigned long ftrace_graph_funcs[FTRACE_GRAPH_MAX_FUNCS] __read_mostly; | |||
| 2457 | static void * | 2414 | static void * |
| 2458 | __g_next(struct seq_file *m, loff_t *pos) | 2415 | __g_next(struct seq_file *m, loff_t *pos) |
| 2459 | { | 2416 | { |
| 2460 | unsigned long *array = m->private; | ||
| 2461 | |||
| 2462 | if (*pos >= ftrace_graph_count) | 2417 | if (*pos >= ftrace_graph_count) |
| 2463 | return NULL; | 2418 | return NULL; |
| 2464 | return &array[*pos]; | 2419 | return &ftrace_graph_funcs[*pos]; |
| 2465 | } | 2420 | } |
| 2466 | 2421 | ||
| 2467 | static void * | 2422 | static void * |
| @@ -2499,12 +2454,12 @@ static int g_show(struct seq_file *m, void *v) | |||
| 2499 | return 0; | 2454 | return 0; |
| 2500 | } | 2455 | } |
| 2501 | 2456 | ||
| 2502 | seq_printf(m, "%pf\n", v); | 2457 | seq_printf(m, "%ps\n", (void *)*ptr); |
| 2503 | 2458 | ||
| 2504 | return 0; | 2459 | return 0; |
| 2505 | } | 2460 | } |
| 2506 | 2461 | ||
| 2507 | static struct seq_operations ftrace_graph_seq_ops = { | 2462 | static const struct seq_operations ftrace_graph_seq_ops = { |
| 2508 | .start = g_start, | 2463 | .start = g_start, |
| 2509 | .next = g_next, | 2464 | .next = g_next, |
| 2510 | .stop = g_stop, | 2465 | .stop = g_stop, |
| @@ -2525,16 +2480,10 @@ ftrace_graph_open(struct inode *inode, struct file *file) | |||
| 2525 | ftrace_graph_count = 0; | 2480 | ftrace_graph_count = 0; |
| 2526 | memset(ftrace_graph_funcs, 0, sizeof(ftrace_graph_funcs)); | 2481 | memset(ftrace_graph_funcs, 0, sizeof(ftrace_graph_funcs)); |
| 2527 | } | 2482 | } |
| 2483 | mutex_unlock(&graph_lock); | ||
| 2528 | 2484 | ||
| 2529 | if (file->f_mode & FMODE_READ) { | 2485 | if (file->f_mode & FMODE_READ) |
| 2530 | ret = seq_open(file, &ftrace_graph_seq_ops); | 2486 | ret = seq_open(file, &ftrace_graph_seq_ops); |
| 2531 | if (!ret) { | ||
| 2532 | struct seq_file *m = file->private_data; | ||
| 2533 | m->private = ftrace_graph_funcs; | ||
| 2534 | } | ||
| 2535 | } else | ||
| 2536 | file->private_data = ftrace_graph_funcs; | ||
| 2537 | mutex_unlock(&graph_lock); | ||
| 2538 | 2487 | ||
| 2539 | return ret; | 2488 | return ret; |
| 2540 | } | 2489 | } |
| @@ -2602,12 +2551,9 @@ static ssize_t | |||
| 2602 | ftrace_graph_write(struct file *file, const char __user *ubuf, | 2551 | ftrace_graph_write(struct file *file, const char __user *ubuf, |
| 2603 | size_t cnt, loff_t *ppos) | 2552 | size_t cnt, loff_t *ppos) |
| 2604 | { | 2553 | { |
| 2605 | unsigned char buffer[FTRACE_BUFF_MAX+1]; | 2554 | struct trace_parser parser; |
| 2606 | unsigned long *array; | ||
| 2607 | size_t read = 0; | 2555 | size_t read = 0; |
| 2608 | ssize_t ret; | 2556 | ssize_t ret; |
| 2609 | int index = 0; | ||
| 2610 | char ch; | ||
| 2611 | 2557 | ||
| 2612 | if (!cnt || cnt < 0) | 2558 | if (!cnt || cnt < 0) |
| 2613 | return 0; | 2559 | return 0; |
| @@ -2619,57 +2565,26 @@ ftrace_graph_write(struct file *file, const char __user *ubuf, | |||
| 2619 | goto out; | 2565 | goto out; |
| 2620 | } | 2566 | } |
| 2621 | 2567 | ||
| 2622 | if (file->f_mode & FMODE_READ) { | 2568 | if (trace_parser_get_init(&parser, FTRACE_BUFF_MAX)) { |
| 2623 | struct seq_file *m = file->private_data; | 2569 | ret = -ENOMEM; |
| 2624 | array = m->private; | ||
| 2625 | } else | ||
| 2626 | array = file->private_data; | ||
| 2627 | |||
| 2628 | ret = get_user(ch, ubuf++); | ||
| 2629 | if (ret) | ||
| 2630 | goto out; | 2570 | goto out; |
| 2631 | read++; | ||
| 2632 | cnt--; | ||
| 2633 | |||
| 2634 | /* skip white space */ | ||
| 2635 | while (cnt && isspace(ch)) { | ||
| 2636 | ret = get_user(ch, ubuf++); | ||
| 2637 | if (ret) | ||
| 2638 | goto out; | ||
| 2639 | read++; | ||
| 2640 | cnt--; | ||
| 2641 | } | 2571 | } |
| 2642 | 2572 | ||
| 2643 | if (isspace(ch)) { | 2573 | read = trace_get_user(&parser, ubuf, cnt, ppos); |
| 2644 | *ppos += read; | ||
| 2645 | ret = read; | ||
| 2646 | goto out; | ||
| 2647 | } | ||
| 2648 | 2574 | ||
| 2649 | while (cnt && !isspace(ch)) { | 2575 | if (trace_parser_loaded((&parser))) { |
| 2650 | if (index < FTRACE_BUFF_MAX) | 2576 | parser.buffer[parser.idx] = 0; |
| 2651 | buffer[index++] = ch; | 2577 | |
| 2652 | else { | 2578 | /* we allow only one expression at a time */ |
| 2653 | ret = -EINVAL; | 2579 | ret = ftrace_set_func(ftrace_graph_funcs, &ftrace_graph_count, |
| 2654 | goto out; | 2580 | parser.buffer); |
| 2655 | } | ||
| 2656 | ret = get_user(ch, ubuf++); | ||
| 2657 | if (ret) | 2581 | if (ret) |
| 2658 | goto out; | 2582 | goto out; |
| 2659 | read++; | ||
| 2660 | cnt--; | ||
| 2661 | } | 2583 | } |
| 2662 | buffer[index] = 0; | ||
| 2663 | |||
| 2664 | /* we allow only one expression at a time */ | ||
| 2665 | ret = ftrace_set_func(array, &ftrace_graph_count, buffer); | ||
| 2666 | if (ret) | ||
| 2667 | goto out; | ||
| 2668 | |||
| 2669 | file->f_pos += read; | ||
| 2670 | 2584 | ||
| 2671 | ret = read; | 2585 | ret = read; |
| 2672 | out: | 2586 | out: |
| 2587 | trace_parser_put(&parser); | ||
| 2673 | mutex_unlock(&graph_lock); | 2588 | mutex_unlock(&graph_lock); |
| 2674 | 2589 | ||
| 2675 | return ret; | 2590 | return ret; |
diff --git a/kernel/trace/power-traces.c b/kernel/trace/power-traces.c new file mode 100644 index 000000000000..e06c6e3d56a3 --- /dev/null +++ b/kernel/trace/power-traces.c | |||
| @@ -0,0 +1,20 @@ | |||
| 1 | /* | ||
| 2 | * Power trace points | ||
| 3 | * | ||
| 4 | * Copyright (C) 2009 Arjan van de Ven <arjan@linux.intel.com> | ||
| 5 | */ | ||
| 6 | |||
| 7 | #include <linux/string.h> | ||
| 8 | #include <linux/types.h> | ||
| 9 | #include <linux/workqueue.h> | ||
| 10 | #include <linux/sched.h> | ||
| 11 | #include <linux/module.h> | ||
| 12 | #include <linux/slab.h> | ||
| 13 | |||
| 14 | #define CREATE_TRACE_POINTS | ||
| 15 | #include <trace/events/power.h> | ||
| 16 | |||
| 17 | EXPORT_TRACEPOINT_SYMBOL_GPL(power_start); | ||
| 18 | EXPORT_TRACEPOINT_SYMBOL_GPL(power_end); | ||
| 19 | EXPORT_TRACEPOINT_SYMBOL_GPL(power_frequency); | ||
| 20 | |||
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index 454e74e718cf..d4ff01970547 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -201,8 +201,6 @@ int tracing_is_on(void) | |||
| 201 | } | 201 | } |
| 202 | EXPORT_SYMBOL_GPL(tracing_is_on); | 202 | EXPORT_SYMBOL_GPL(tracing_is_on); |
| 203 | 203 | ||
| 204 | #include "trace.h" | ||
| 205 | |||
| 206 | #define RB_EVNT_HDR_SIZE (offsetof(struct ring_buffer_event, array)) | 204 | #define RB_EVNT_HDR_SIZE (offsetof(struct ring_buffer_event, array)) |
| 207 | #define RB_ALIGNMENT 4U | 205 | #define RB_ALIGNMENT 4U |
| 208 | #define RB_MAX_SMALL_DATA (RB_ALIGNMENT * RINGBUF_TYPE_DATA_TYPE_LEN_MAX) | 206 | #define RB_MAX_SMALL_DATA (RB_ALIGNMENT * RINGBUF_TYPE_DATA_TYPE_LEN_MAX) |
| @@ -701,8 +699,8 @@ static int rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 701 | 699 | ||
| 702 | val &= ~RB_FLAG_MASK; | 700 | val &= ~RB_FLAG_MASK; |
| 703 | 701 | ||
| 704 | ret = (unsigned long)cmpxchg(&list->next, | 702 | ret = cmpxchg((unsigned long *)&list->next, |
| 705 | val | old_flag, val | new_flag); | 703 | val | old_flag, val | new_flag); |
| 706 | 704 | ||
| 707 | /* check if the reader took the page */ | 705 | /* check if the reader took the page */ |
| 708 | if ((ret & ~RB_FLAG_MASK) != val) | 706 | if ((ret & ~RB_FLAG_MASK) != val) |
| @@ -794,7 +792,7 @@ static int rb_head_page_replace(struct buffer_page *old, | |||
| 794 | val = *ptr & ~RB_FLAG_MASK; | 792 | val = *ptr & ~RB_FLAG_MASK; |
| 795 | val |= RB_PAGE_HEAD; | 793 | val |= RB_PAGE_HEAD; |
| 796 | 794 | ||
| 797 | ret = cmpxchg(ptr, val, &new->list); | 795 | ret = cmpxchg(ptr, val, (unsigned long)&new->list); |
| 798 | 796 | ||
| 799 | return ret == val; | 797 | return ret == val; |
| 800 | } | 798 | } |
| @@ -2997,15 +2995,12 @@ static void rb_advance_iter(struct ring_buffer_iter *iter) | |||
| 2997 | } | 2995 | } |
| 2998 | 2996 | ||
| 2999 | static struct ring_buffer_event * | 2997 | static struct ring_buffer_event * |
| 3000 | rb_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts) | 2998 | rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts) |
| 3001 | { | 2999 | { |
| 3002 | struct ring_buffer_per_cpu *cpu_buffer; | ||
| 3003 | struct ring_buffer_event *event; | 3000 | struct ring_buffer_event *event; |
| 3004 | struct buffer_page *reader; | 3001 | struct buffer_page *reader; |
| 3005 | int nr_loops = 0; | 3002 | int nr_loops = 0; |
| 3006 | 3003 | ||
| 3007 | cpu_buffer = buffer->buffers[cpu]; | ||
| 3008 | |||
| 3009 | again: | 3004 | again: |
| 3010 | /* | 3005 | /* |
| 3011 | * We repeat when a timestamp is encountered. It is possible | 3006 | * We repeat when a timestamp is encountered. It is possible |
| @@ -3049,7 +3044,7 @@ rb_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts) | |||
| 3049 | case RINGBUF_TYPE_DATA: | 3044 | case RINGBUF_TYPE_DATA: |
| 3050 | if (ts) { | 3045 | if (ts) { |
| 3051 | *ts = cpu_buffer->read_stamp + event->time_delta; | 3046 | *ts = cpu_buffer->read_stamp + event->time_delta; |
| 3052 | ring_buffer_normalize_time_stamp(buffer, | 3047 | ring_buffer_normalize_time_stamp(cpu_buffer->buffer, |
| 3053 | cpu_buffer->cpu, ts); | 3048 | cpu_buffer->cpu, ts); |
| 3054 | } | 3049 | } |
| 3055 | return event; | 3050 | return event; |
| @@ -3168,7 +3163,7 @@ ring_buffer_peek(struct ring_buffer *buffer, int cpu, u64 *ts) | |||
| 3168 | local_irq_save(flags); | 3163 | local_irq_save(flags); |
| 3169 | if (dolock) | 3164 | if (dolock) |
| 3170 | spin_lock(&cpu_buffer->reader_lock); | 3165 | spin_lock(&cpu_buffer->reader_lock); |
| 3171 | event = rb_buffer_peek(buffer, cpu, ts); | 3166 | event = rb_buffer_peek(cpu_buffer, ts); |
| 3172 | if (event && event->type_len == RINGBUF_TYPE_PADDING) | 3167 | if (event && event->type_len == RINGBUF_TYPE_PADDING) |
| 3173 | rb_advance_reader(cpu_buffer); | 3168 | rb_advance_reader(cpu_buffer); |
| 3174 | if (dolock) | 3169 | if (dolock) |
| @@ -3237,7 +3232,7 @@ ring_buffer_consume(struct ring_buffer *buffer, int cpu, u64 *ts) | |||
| 3237 | if (dolock) | 3232 | if (dolock) |
| 3238 | spin_lock(&cpu_buffer->reader_lock); | 3233 | spin_lock(&cpu_buffer->reader_lock); |
| 3239 | 3234 | ||
| 3240 | event = rb_buffer_peek(buffer, cpu, ts); | 3235 | event = rb_buffer_peek(cpu_buffer, ts); |
| 3241 | if (event) | 3236 | if (event) |
| 3242 | rb_advance_reader(cpu_buffer); | 3237 | rb_advance_reader(cpu_buffer); |
| 3243 | 3238 | ||
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 5c75deeefe30..6c0f6a8a22eb 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -125,13 +125,13 @@ int ftrace_dump_on_oops; | |||
| 125 | 125 | ||
| 126 | static int tracing_set_tracer(const char *buf); | 126 | static int tracing_set_tracer(const char *buf); |
| 127 | 127 | ||
| 128 | #define BOOTUP_TRACER_SIZE 100 | 128 | #define MAX_TRACER_SIZE 100 |
| 129 | static char bootup_tracer_buf[BOOTUP_TRACER_SIZE] __initdata; | 129 | static char bootup_tracer_buf[MAX_TRACER_SIZE] __initdata; |
| 130 | static char *default_bootup_tracer; | 130 | static char *default_bootup_tracer; |
| 131 | 131 | ||
| 132 | static int __init set_ftrace(char *str) | 132 | static int __init set_ftrace(char *str) |
| 133 | { | 133 | { |
| 134 | strncpy(bootup_tracer_buf, str, BOOTUP_TRACER_SIZE); | 134 | strncpy(bootup_tracer_buf, str, MAX_TRACER_SIZE); |
| 135 | default_bootup_tracer = bootup_tracer_buf; | 135 | default_bootup_tracer = bootup_tracer_buf; |
| 136 | /* We are using ftrace early, expand it */ | 136 | /* We are using ftrace early, expand it */ |
| 137 | ring_buffer_expanded = 1; | 137 | ring_buffer_expanded = 1; |
| @@ -242,13 +242,6 @@ static struct tracer *trace_types __read_mostly; | |||
| 242 | static struct tracer *current_trace __read_mostly; | 242 | static struct tracer *current_trace __read_mostly; |
| 243 | 243 | ||
| 244 | /* | 244 | /* |
| 245 | * max_tracer_type_len is used to simplify the allocating of | ||
| 246 | * buffers to read userspace tracer names. We keep track of | ||
| 247 | * the longest tracer name registered. | ||
| 248 | */ | ||
| 249 | static int max_tracer_type_len; | ||
| 250 | |||
| 251 | /* | ||
| 252 | * trace_types_lock is used to protect the trace_types list. | 245 | * trace_types_lock is used to protect the trace_types list. |
| 253 | * This lock is also used to keep user access serialized. | 246 | * This lock is also used to keep user access serialized. |
| 254 | * Accesses from userspace will grab this lock while userspace | 247 | * Accesses from userspace will grab this lock while userspace |
| @@ -275,12 +268,18 @@ static DEFINE_SPINLOCK(tracing_start_lock); | |||
| 275 | */ | 268 | */ |
| 276 | void trace_wake_up(void) | 269 | void trace_wake_up(void) |
| 277 | { | 270 | { |
| 271 | int cpu; | ||
| 272 | |||
| 273 | if (trace_flags & TRACE_ITER_BLOCK) | ||
| 274 | return; | ||
| 278 | /* | 275 | /* |
| 279 | * The runqueue_is_locked() can fail, but this is the best we | 276 | * The runqueue_is_locked() can fail, but this is the best we |
| 280 | * have for now: | 277 | * have for now: |
| 281 | */ | 278 | */ |
| 282 | if (!(trace_flags & TRACE_ITER_BLOCK) && !runqueue_is_locked()) | 279 | cpu = get_cpu(); |
| 280 | if (!runqueue_is_locked(cpu)) | ||
| 283 | wake_up(&trace_wait); | 281 | wake_up(&trace_wait); |
| 282 | put_cpu(); | ||
| 284 | } | 283 | } |
| 285 | 284 | ||
| 286 | static int __init set_buf_size(char *str) | 285 | static int __init set_buf_size(char *str) |
| @@ -339,6 +338,112 @@ static struct { | |||
| 339 | 338 | ||
| 340 | int trace_clock_id; | 339 | int trace_clock_id; |
| 341 | 340 | ||
| 341 | /* | ||
| 342 | * trace_parser_get_init - gets the buffer for trace parser | ||
| 343 | */ | ||
| 344 | int trace_parser_get_init(struct trace_parser *parser, int size) | ||
| 345 | { | ||
| 346 | memset(parser, 0, sizeof(*parser)); | ||
| 347 | |||
| 348 | parser->buffer = kmalloc(size, GFP_KERNEL); | ||
| 349 | if (!parser->buffer) | ||
| 350 | return 1; | ||
| 351 | |||
| 352 | parser->size = size; | ||
| 353 | return 0; | ||
| 354 | } | ||
| 355 | |||
| 356 | /* | ||
| 357 | * trace_parser_put - frees the buffer for trace parser | ||
| 358 | */ | ||
| 359 | void trace_parser_put(struct trace_parser *parser) | ||
| 360 | { | ||
| 361 | kfree(parser->buffer); | ||
| 362 | } | ||
| 363 | |||
| 364 | /* | ||
| 365 | * trace_get_user - reads the user input string separated by space | ||
| 366 | * (matched by isspace(ch)) | ||
| 367 | * | ||
| 368 | * For each string found the 'struct trace_parser' is updated, | ||
| 369 | * and the function returns. | ||
| 370 | * | ||
| 371 | * Returns number of bytes read. | ||
| 372 | * | ||
| 373 | * See kernel/trace/trace.h for 'struct trace_parser' details. | ||
| 374 | */ | ||
| 375 | int trace_get_user(struct trace_parser *parser, const char __user *ubuf, | ||
| 376 | size_t cnt, loff_t *ppos) | ||
| 377 | { | ||
| 378 | char ch; | ||
| 379 | size_t read = 0; | ||
| 380 | ssize_t ret; | ||
| 381 | |||
| 382 | if (!*ppos) | ||
| 383 | trace_parser_clear(parser); | ||
| 384 | |||
| 385 | ret = get_user(ch, ubuf++); | ||
| 386 | if (ret) | ||
| 387 | goto out; | ||
| 388 | |||
| 389 | read++; | ||
| 390 | cnt--; | ||
| 391 | |||
| 392 | /* | ||
| 393 | * The parser is not finished with the last write, | ||
| 394 | * continue reading the user input without skipping spaces. | ||
| 395 | */ | ||
| 396 | if (!parser->cont) { | ||
| 397 | /* skip white space */ | ||
| 398 | while (cnt && isspace(ch)) { | ||
| 399 | ret = get_user(ch, ubuf++); | ||
| 400 | if (ret) | ||
| 401 | goto out; | ||
| 402 | read++; | ||
| 403 | cnt--; | ||
| 404 | } | ||
| 405 | |||
| 406 | /* only spaces were written */ | ||
| 407 | if (isspace(ch)) { | ||
| 408 | *ppos += read; | ||
| 409 | ret = read; | ||
| 410 | goto out; | ||
| 411 | } | ||
| 412 | |||
| 413 | parser->idx = 0; | ||
| 414 | } | ||
| 415 | |||
| 416 | /* read the non-space input */ | ||
| 417 | while (cnt && !isspace(ch)) { | ||
| 418 | if (parser->idx < parser->size) | ||
| 419 | parser->buffer[parser->idx++] = ch; | ||
| 420 | else { | ||
| 421 | ret = -EINVAL; | ||
| 422 | goto out; | ||
| 423 | } | ||
| 424 | ret = get_user(ch, ubuf++); | ||
| 425 | if (ret) | ||
| 426 | goto out; | ||
| 427 | read++; | ||
| 428 | cnt--; | ||
| 429 | } | ||
| 430 | |||
| 431 | /* We either got finished input or we have to wait for another call. */ | ||
| 432 | if (isspace(ch)) { | ||
| 433 | parser->buffer[parser->idx] = 0; | ||
| 434 | parser->cont = false; | ||
| 435 | } else { | ||
| 436 | parser->cont = true; | ||
| 437 | parser->buffer[parser->idx++] = ch; | ||
| 438 | } | ||
| 439 | |||
| 440 | *ppos += read; | ||
| 441 | ret = read; | ||
| 442 | |||
| 443 | out: | ||
| 444 | return ret; | ||
| 445 | } | ||
| 446 | |||
| 342 | ssize_t trace_seq_to_user(struct trace_seq *s, char __user *ubuf, size_t cnt) | 447 | ssize_t trace_seq_to_user(struct trace_seq *s, char __user *ubuf, size_t cnt) |
| 343 | { | 448 | { |
| 344 | int len; | 449 | int len; |
| @@ -513,7 +618,6 @@ __releases(kernel_lock) | |||
| 513 | __acquires(kernel_lock) | 618 | __acquires(kernel_lock) |
| 514 | { | 619 | { |
| 515 | struct tracer *t; | 620 | struct tracer *t; |
| 516 | int len; | ||
| 517 | int ret = 0; | 621 | int ret = 0; |
| 518 | 622 | ||
| 519 | if (!type->name) { | 623 | if (!type->name) { |
| @@ -521,6 +625,11 @@ __acquires(kernel_lock) | |||
| 521 | return -1; | 625 | return -1; |
| 522 | } | 626 | } |
| 523 | 627 | ||
| 628 | if (strlen(type->name) > MAX_TRACER_SIZE) { | ||
| 629 | pr_info("Tracer has a name longer than %d\n", MAX_TRACER_SIZE); | ||
| 630 | return -1; | ||
| 631 | } | ||
| 632 | |||
| 524 | /* | 633 | /* |
| 525 | * When this gets called we hold the BKL which means that | 634 | * When this gets called we hold the BKL which means that |
| 526 | * preemption is disabled. Various trace selftests however | 635 | * preemption is disabled. Various trace selftests however |
| @@ -535,7 +644,7 @@ __acquires(kernel_lock) | |||
| 535 | for (t = trace_types; t; t = t->next) { | 644 | for (t = trace_types; t; t = t->next) { |
| 536 | if (strcmp(type->name, t->name) == 0) { | 645 | if (strcmp(type->name, t->name) == 0) { |
| 537 | /* already found */ | 646 | /* already found */ |
| 538 | pr_info("Trace %s already registered\n", | 647 | pr_info("Tracer %s already registered\n", |
| 539 | type->name); | 648 | type->name); |
| 540 | ret = -1; | 649 | ret = -1; |
| 541 | goto out; | 650 | goto out; |
| @@ -586,9 +695,6 @@ __acquires(kernel_lock) | |||
| 586 | 695 | ||
| 587 | type->next = trace_types; | 696 | type->next = trace_types; |
| 588 | trace_types = type; | 697 | trace_types = type; |
| 589 | len = strlen(type->name); | ||
| 590 | if (len > max_tracer_type_len) | ||
| 591 | max_tracer_type_len = len; | ||
| 592 | 698 | ||
| 593 | out: | 699 | out: |
| 594 | tracing_selftest_running = false; | 700 | tracing_selftest_running = false; |
| @@ -597,7 +703,7 @@ __acquires(kernel_lock) | |||
| 597 | if (ret || !default_bootup_tracer) | 703 | if (ret || !default_bootup_tracer) |
| 598 | goto out_unlock; | 704 | goto out_unlock; |
| 599 | 705 | ||
| 600 | if (strncmp(default_bootup_tracer, type->name, BOOTUP_TRACER_SIZE)) | 706 | if (strncmp(default_bootup_tracer, type->name, MAX_TRACER_SIZE)) |
| 601 | goto out_unlock; | 707 | goto out_unlock; |
| 602 | 708 | ||
| 603 | printk(KERN_INFO "Starting tracer '%s'\n", type->name); | 709 | printk(KERN_INFO "Starting tracer '%s'\n", type->name); |
| @@ -619,14 +725,13 @@ __acquires(kernel_lock) | |||
| 619 | void unregister_tracer(struct tracer *type) | 725 | void unregister_tracer(struct tracer *type) |
| 620 | { | 726 | { |
| 621 | struct tracer **t; | 727 | struct tracer **t; |
| 622 | int len; | ||
| 623 | 728 | ||
| 624 | mutex_lock(&trace_types_lock); | 729 | mutex_lock(&trace_types_lock); |
| 625 | for (t = &trace_types; *t; t = &(*t)->next) { | 730 | for (t = &trace_types; *t; t = &(*t)->next) { |
| 626 | if (*t == type) | 731 | if (*t == type) |
| 627 | goto found; | 732 | goto found; |
| 628 | } | 733 | } |
| 629 | pr_info("Trace %s not registered\n", type->name); | 734 | pr_info("Tracer %s not registered\n", type->name); |
| 630 | goto out; | 735 | goto out; |
| 631 | 736 | ||
| 632 | found: | 737 | found: |
| @@ -639,17 +744,7 @@ void unregister_tracer(struct tracer *type) | |||
| 639 | current_trace->stop(&global_trace); | 744 | current_trace->stop(&global_trace); |
| 640 | current_trace = &nop_trace; | 745 | current_trace = &nop_trace; |
| 641 | } | 746 | } |
| 642 | 747 | out: | |
| 643 | if (strlen(type->name) != max_tracer_type_len) | ||
| 644 | goto out; | ||
| 645 | |||
| 646 | max_tracer_type_len = 0; | ||
| 647 | for (t = &trace_types; *t; t = &(*t)->next) { | ||
| 648 | len = strlen((*t)->name); | ||
| 649 | if (len > max_tracer_type_len) | ||
| 650 | max_tracer_type_len = len; | ||
| 651 | } | ||
| 652 | out: | ||
| 653 | mutex_unlock(&trace_types_lock); | 748 | mutex_unlock(&trace_types_lock); |
| 654 | } | 749 | } |
| 655 | 750 | ||
| @@ -719,6 +814,11 @@ static void trace_init_cmdlines(void) | |||
| 719 | cmdline_idx = 0; | 814 | cmdline_idx = 0; |
| 720 | } | 815 | } |
| 721 | 816 | ||
| 817 | int is_tracing_stopped(void) | ||
| 818 | { | ||
| 819 | return trace_stop_count; | ||
| 820 | } | ||
| 821 | |||
| 722 | /** | 822 | /** |
| 723 | * ftrace_off_permanent - disable all ftrace code permanently | 823 | * ftrace_off_permanent - disable all ftrace code permanently |
| 724 | * | 824 | * |
| @@ -886,7 +986,7 @@ tracing_generic_entry_update(struct trace_entry *entry, unsigned long flags, | |||
| 886 | 986 | ||
| 887 | entry->preempt_count = pc & 0xff; | 987 | entry->preempt_count = pc & 0xff; |
| 888 | entry->pid = (tsk) ? tsk->pid : 0; | 988 | entry->pid = (tsk) ? tsk->pid : 0; |
| 889 | entry->tgid = (tsk) ? tsk->tgid : 0; | 989 | entry->lock_depth = (tsk) ? tsk->lock_depth : 0; |
| 890 | entry->flags = | 990 | entry->flags = |
| 891 | #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT | 991 | #ifdef CONFIG_TRACE_IRQFLAGS_SUPPORT |
| 892 | (irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) | | 992 | (irqs_disabled_flags(flags) ? TRACE_FLAG_IRQS_OFF : 0) | |
| @@ -1068,6 +1168,7 @@ ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) | |||
| 1068 | return; | 1168 | return; |
| 1069 | entry = ring_buffer_event_data(event); | 1169 | entry = ring_buffer_event_data(event); |
| 1070 | 1170 | ||
| 1171 | entry->tgid = current->tgid; | ||
| 1071 | memset(&entry->caller, 0, sizeof(entry->caller)); | 1172 | memset(&entry->caller, 0, sizeof(entry->caller)); |
| 1072 | 1173 | ||
| 1073 | trace.nr_entries = 0; | 1174 | trace.nr_entries = 0; |
| @@ -1094,6 +1195,7 @@ ftrace_trace_special(void *__tr, | |||
| 1094 | unsigned long arg1, unsigned long arg2, unsigned long arg3, | 1195 | unsigned long arg1, unsigned long arg2, unsigned long arg3, |
| 1095 | int pc) | 1196 | int pc) |
| 1096 | { | 1197 | { |
| 1198 | struct ftrace_event_call *call = &event_special; | ||
| 1097 | struct ring_buffer_event *event; | 1199 | struct ring_buffer_event *event; |
| 1098 | struct trace_array *tr = __tr; | 1200 | struct trace_array *tr = __tr; |
| 1099 | struct ring_buffer *buffer = tr->buffer; | 1201 | struct ring_buffer *buffer = tr->buffer; |
| @@ -1107,7 +1209,9 @@ ftrace_trace_special(void *__tr, | |||
| 1107 | entry->arg1 = arg1; | 1209 | entry->arg1 = arg1; |
| 1108 | entry->arg2 = arg2; | 1210 | entry->arg2 = arg2; |
| 1109 | entry->arg3 = arg3; | 1211 | entry->arg3 = arg3; |
| 1110 | trace_buffer_unlock_commit(buffer, event, 0, pc); | 1212 | |
| 1213 | if (!filter_check_discard(call, entry, buffer, event)) | ||
| 1214 | trace_buffer_unlock_commit(buffer, event, 0, pc); | ||
| 1111 | } | 1215 | } |
| 1112 | 1216 | ||
| 1113 | void | 1217 | void |
| @@ -1530,10 +1634,10 @@ static void print_lat_help_header(struct seq_file *m) | |||
| 1530 | seq_puts(m, "# | / _----=> need-resched \n"); | 1634 | seq_puts(m, "# | / _----=> need-resched \n"); |
| 1531 | seq_puts(m, "# || / _---=> hardirq/softirq \n"); | 1635 | seq_puts(m, "# || / _---=> hardirq/softirq \n"); |
| 1532 | seq_puts(m, "# ||| / _--=> preempt-depth \n"); | 1636 | seq_puts(m, "# ||| / _--=> preempt-depth \n"); |
| 1533 | seq_puts(m, "# |||| / \n"); | 1637 | seq_puts(m, "# |||| /_--=> lock-depth \n"); |
| 1534 | seq_puts(m, "# ||||| delay \n"); | 1638 | seq_puts(m, "# |||||/ delay \n"); |
| 1535 | seq_puts(m, "# cmd pid ||||| time | caller \n"); | 1639 | seq_puts(m, "# cmd pid |||||| time | caller \n"); |
| 1536 | seq_puts(m, "# \\ / ||||| \\ | / \n"); | 1640 | seq_puts(m, "# \\ / |||||| \\ | / \n"); |
| 1537 | } | 1641 | } |
| 1538 | 1642 | ||
| 1539 | static void print_func_help_header(struct seq_file *m) | 1643 | static void print_func_help_header(struct seq_file *m) |
| @@ -1845,7 +1949,7 @@ static int s_show(struct seq_file *m, void *v) | |||
| 1845 | return 0; | 1949 | return 0; |
| 1846 | } | 1950 | } |
| 1847 | 1951 | ||
| 1848 | static struct seq_operations tracer_seq_ops = { | 1952 | static const struct seq_operations tracer_seq_ops = { |
| 1849 | .start = s_start, | 1953 | .start = s_start, |
| 1850 | .next = s_next, | 1954 | .next = s_next, |
| 1851 | .stop = s_stop, | 1955 | .stop = s_stop, |
| @@ -2059,7 +2163,7 @@ static int t_show(struct seq_file *m, void *v) | |||
| 2059 | return 0; | 2163 | return 0; |
| 2060 | } | 2164 | } |
| 2061 | 2165 | ||
| 2062 | static struct seq_operations show_traces_seq_ops = { | 2166 | static const struct seq_operations show_traces_seq_ops = { |
| 2063 | .start = t_start, | 2167 | .start = t_start, |
| 2064 | .next = t_next, | 2168 | .next = t_next, |
| 2065 | .stop = t_stop, | 2169 | .stop = t_stop, |
| @@ -2489,7 +2593,7 @@ static ssize_t | |||
| 2489 | tracing_set_trace_read(struct file *filp, char __user *ubuf, | 2593 | tracing_set_trace_read(struct file *filp, char __user *ubuf, |
| 2490 | size_t cnt, loff_t *ppos) | 2594 | size_t cnt, loff_t *ppos) |
| 2491 | { | 2595 | { |
| 2492 | char buf[max_tracer_type_len+2]; | 2596 | char buf[MAX_TRACER_SIZE+2]; |
| 2493 | int r; | 2597 | int r; |
| 2494 | 2598 | ||
| 2495 | mutex_lock(&trace_types_lock); | 2599 | mutex_lock(&trace_types_lock); |
| @@ -2639,15 +2743,15 @@ static ssize_t | |||
| 2639 | tracing_set_trace_write(struct file *filp, const char __user *ubuf, | 2743 | tracing_set_trace_write(struct file *filp, const char __user *ubuf, |
| 2640 | size_t cnt, loff_t *ppos) | 2744 | size_t cnt, loff_t *ppos) |
| 2641 | { | 2745 | { |
| 2642 | char buf[max_tracer_type_len+1]; | 2746 | char buf[MAX_TRACER_SIZE+1]; |
| 2643 | int i; | 2747 | int i; |
| 2644 | size_t ret; | 2748 | size_t ret; |
| 2645 | int err; | 2749 | int err; |
| 2646 | 2750 | ||
| 2647 | ret = cnt; | 2751 | ret = cnt; |
| 2648 | 2752 | ||
| 2649 | if (cnt > max_tracer_type_len) | 2753 | if (cnt > MAX_TRACER_SIZE) |
| 2650 | cnt = max_tracer_type_len; | 2754 | cnt = MAX_TRACER_SIZE; |
| 2651 | 2755 | ||
| 2652 | if (copy_from_user(&buf, ubuf, cnt)) | 2756 | if (copy_from_user(&buf, ubuf, cnt)) |
| 2653 | return -EFAULT; | 2757 | return -EFAULT; |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 821064914c80..104c1a72418f 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
| @@ -7,10 +7,10 @@ | |||
| 7 | #include <linux/clocksource.h> | 7 | #include <linux/clocksource.h> |
| 8 | #include <linux/ring_buffer.h> | 8 | #include <linux/ring_buffer.h> |
| 9 | #include <linux/mmiotrace.h> | 9 | #include <linux/mmiotrace.h> |
| 10 | #include <linux/tracepoint.h> | ||
| 10 | #include <linux/ftrace.h> | 11 | #include <linux/ftrace.h> |
| 11 | #include <trace/boot.h> | 12 | #include <trace/boot.h> |
| 12 | #include <linux/kmemtrace.h> | 13 | #include <linux/kmemtrace.h> |
| 13 | #include <trace/power.h> | ||
| 14 | 14 | ||
| 15 | #include <linux/trace_seq.h> | 15 | #include <linux/trace_seq.h> |
| 16 | #include <linux/ftrace_event.h> | 16 | #include <linux/ftrace_event.h> |
| @@ -36,163 +36,59 @@ enum trace_type { | |||
| 36 | TRACE_HW_BRANCHES, | 36 | TRACE_HW_BRANCHES, |
| 37 | TRACE_KMEM_ALLOC, | 37 | TRACE_KMEM_ALLOC, |
| 38 | TRACE_KMEM_FREE, | 38 | TRACE_KMEM_FREE, |
| 39 | TRACE_POWER, | ||
| 40 | TRACE_BLK, | 39 | TRACE_BLK, |
| 41 | 40 | ||
| 42 | __TRACE_LAST_TYPE, | 41 | __TRACE_LAST_TYPE, |
| 43 | }; | 42 | }; |
| 44 | 43 | ||
| 45 | /* | 44 | enum kmemtrace_type_id { |
| 46 | * Function trace entry - function address and parent function addres: | 45 | KMEMTRACE_TYPE_KMALLOC = 0, /* kmalloc() or kfree(). */ |
| 47 | */ | 46 | KMEMTRACE_TYPE_CACHE, /* kmem_cache_*(). */ |
| 48 | struct ftrace_entry { | 47 | KMEMTRACE_TYPE_PAGES, /* __get_free_pages() and friends. */ |
| 49 | struct trace_entry ent; | ||
| 50 | unsigned long ip; | ||
| 51 | unsigned long parent_ip; | ||
| 52 | }; | ||
| 53 | |||
| 54 | /* Function call entry */ | ||
| 55 | struct ftrace_graph_ent_entry { | ||
| 56 | struct trace_entry ent; | ||
| 57 | struct ftrace_graph_ent graph_ent; | ||
| 58 | }; | 48 | }; |
| 59 | 49 | ||
| 60 | /* Function return entry */ | ||
| 61 | struct ftrace_graph_ret_entry { | ||
| 62 | struct trace_entry ent; | ||
| 63 | struct ftrace_graph_ret ret; | ||
| 64 | }; | ||
| 65 | extern struct tracer boot_tracer; | 50 | extern struct tracer boot_tracer; |
| 66 | 51 | ||
| 67 | /* | 52 | #undef __field |
| 68 | * Context switch trace entry - which task (and prio) we switched from/to: | 53 | #define __field(type, item) type item; |
| 69 | */ | ||
| 70 | struct ctx_switch_entry { | ||
| 71 | struct trace_entry ent; | ||
| 72 | unsigned int prev_pid; | ||
| 73 | unsigned char prev_prio; | ||
| 74 | unsigned char prev_state; | ||
| 75 | unsigned int next_pid; | ||
| 76 | unsigned char next_prio; | ||
| 77 | unsigned char next_state; | ||
| 78 | unsigned int next_cpu; | ||
| 79 | }; | ||
| 80 | |||
| 81 | /* | ||
| 82 | * Special (free-form) trace entry: | ||
| 83 | */ | ||
| 84 | struct special_entry { | ||
| 85 | struct trace_entry ent; | ||
| 86 | unsigned long arg1; | ||
| 87 | unsigned long arg2; | ||
| 88 | unsigned long arg3; | ||
| 89 | }; | ||
| 90 | |||
| 91 | /* | ||
| 92 | * Stack-trace entry: | ||
| 93 | */ | ||
| 94 | |||
| 95 | #define FTRACE_STACK_ENTRIES 8 | ||
| 96 | 54 | ||
| 97 | struct stack_entry { | 55 | #undef __field_struct |
| 98 | struct trace_entry ent; | 56 | #define __field_struct(type, item) __field(type, item) |
| 99 | unsigned long caller[FTRACE_STACK_ENTRIES]; | ||
| 100 | }; | ||
| 101 | 57 | ||
| 102 | struct userstack_entry { | 58 | #undef __field_desc |
| 103 | struct trace_entry ent; | 59 | #define __field_desc(type, container, item) |
| 104 | unsigned long caller[FTRACE_STACK_ENTRIES]; | ||
| 105 | }; | ||
| 106 | 60 | ||
| 107 | /* | 61 | #undef __array |
| 108 | * trace_printk entry: | 62 | #define __array(type, item, size) type item[size]; |
| 109 | */ | ||
| 110 | struct bprint_entry { | ||
| 111 | struct trace_entry ent; | ||
| 112 | unsigned long ip; | ||
| 113 | const char *fmt; | ||
| 114 | u32 buf[]; | ||
| 115 | }; | ||
| 116 | 63 | ||
| 117 | struct print_entry { | 64 | #undef __array_desc |
| 118 | struct trace_entry ent; | 65 | #define __array_desc(type, container, item, size) |
| 119 | unsigned long ip; | ||
| 120 | char buf[]; | ||
| 121 | }; | ||
| 122 | |||
| 123 | #define TRACE_OLD_SIZE 88 | ||
| 124 | |||
| 125 | struct trace_field_cont { | ||
| 126 | unsigned char type; | ||
| 127 | /* Temporary till we get rid of this completely */ | ||
| 128 | char buf[TRACE_OLD_SIZE - 1]; | ||
| 129 | }; | ||
| 130 | 66 | ||
| 131 | struct trace_mmiotrace_rw { | 67 | #undef __dynamic_array |
| 132 | struct trace_entry ent; | 68 | #define __dynamic_array(type, item) type item[]; |
| 133 | struct mmiotrace_rw rw; | ||
| 134 | }; | ||
| 135 | 69 | ||
| 136 | struct trace_mmiotrace_map { | 70 | #undef F_STRUCT |
| 137 | struct trace_entry ent; | 71 | #define F_STRUCT(args...) args |
| 138 | struct mmiotrace_map map; | ||
| 139 | }; | ||
| 140 | |||
| 141 | struct trace_boot_call { | ||
| 142 | struct trace_entry ent; | ||
| 143 | struct boot_trace_call boot_call; | ||
| 144 | }; | ||
| 145 | 72 | ||
| 146 | struct trace_boot_ret { | 73 | #undef FTRACE_ENTRY |
| 147 | struct trace_entry ent; | 74 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ |
| 148 | struct boot_trace_ret boot_ret; | 75 | struct struct_name { \ |
| 149 | }; | 76 | struct trace_entry ent; \ |
| 77 | tstruct \ | ||
| 78 | } | ||
| 150 | 79 | ||
| 151 | #define TRACE_FUNC_SIZE 30 | 80 | #undef TP_ARGS |
| 152 | #define TRACE_FILE_SIZE 20 | 81 | #define TP_ARGS(args...) args |
| 153 | struct trace_branch { | ||
| 154 | struct trace_entry ent; | ||
| 155 | unsigned line; | ||
| 156 | char func[TRACE_FUNC_SIZE+1]; | ||
| 157 | char file[TRACE_FILE_SIZE+1]; | ||
| 158 | char correct; | ||
| 159 | }; | ||
| 160 | 82 | ||
| 161 | struct hw_branch_entry { | 83 | #undef FTRACE_ENTRY_DUP |
| 162 | struct trace_entry ent; | 84 | #define FTRACE_ENTRY_DUP(name, name_struct, id, tstruct, printk) |
| 163 | u64 from; | ||
| 164 | u64 to; | ||
| 165 | }; | ||
| 166 | 85 | ||
| 167 | struct trace_power { | 86 | #include "trace_entries.h" |
| 168 | struct trace_entry ent; | ||
| 169 | struct power_trace state_data; | ||
| 170 | }; | ||
| 171 | |||
| 172 | enum kmemtrace_type_id { | ||
| 173 | KMEMTRACE_TYPE_KMALLOC = 0, /* kmalloc() or kfree(). */ | ||
| 174 | KMEMTRACE_TYPE_CACHE, /* kmem_cache_*(). */ | ||
| 175 | KMEMTRACE_TYPE_PAGES, /* __get_free_pages() and friends. */ | ||
| 176 | }; | ||
| 177 | |||
| 178 | struct kmemtrace_alloc_entry { | ||
| 179 | struct trace_entry ent; | ||
| 180 | enum kmemtrace_type_id type_id; | ||
| 181 | unsigned long call_site; | ||
| 182 | const void *ptr; | ||
| 183 | size_t bytes_req; | ||
| 184 | size_t bytes_alloc; | ||
| 185 | gfp_t gfp_flags; | ||
| 186 | int node; | ||
| 187 | }; | ||
| 188 | |||
| 189 | struct kmemtrace_free_entry { | ||
| 190 | struct trace_entry ent; | ||
| 191 | enum kmemtrace_type_id type_id; | ||
| 192 | unsigned long call_site; | ||
| 193 | const void *ptr; | ||
| 194 | }; | ||
| 195 | 87 | ||
| 88 | /* | ||
| 89 | * syscalls are special, and need special handling, this is why | ||
| 90 | * they are not included in trace_entries.h | ||
| 91 | */ | ||
| 196 | struct syscall_trace_enter { | 92 | struct syscall_trace_enter { |
| 197 | struct trace_entry ent; | 93 | struct trace_entry ent; |
| 198 | int nr; | 94 | int nr; |
| @@ -228,14 +124,12 @@ struct kretprobe_trace_entry { | |||
| 228 | (offsetof(struct kretprobe_trace_entry, args) + \ | 124 | (offsetof(struct kretprobe_trace_entry, args) + \ |
| 229 | (sizeof(unsigned long) * (n))) | 125 | (sizeof(unsigned long) * (n))) |
| 230 | 126 | ||
| 231 | |||
| 232 | |||
| 233 | /* | 127 | /* |
| 234 | * trace_flag_type is an enumeration that holds different | 128 | * trace_flag_type is an enumeration that holds different |
| 235 | * states when a trace occurs. These are: | 129 | * states when a trace occurs. These are: |
| 236 | * IRQS_OFF - interrupts were disabled | 130 | * IRQS_OFF - interrupts were disabled |
| 237 | * IRQS_NOSUPPORT - arch does not support irqs_disabled_flags | 131 | * IRQS_NOSUPPORT - arch does not support irqs_disabled_flags |
| 238 | * NEED_RESCED - reschedule is requested | 132 | * NEED_RESCHED - reschedule is requested |
| 239 | * HARDIRQ - inside an interrupt handler | 133 | * HARDIRQ - inside an interrupt handler |
| 240 | * SOFTIRQ - inside a softirq handler | 134 | * SOFTIRQ - inside a softirq handler |
| 241 | */ | 135 | */ |
| @@ -334,7 +228,6 @@ extern void __ftrace_bad_type(void); | |||
| 334 | IF_ASSIGN(var, ent, struct ftrace_graph_ret_entry, \ | 228 | IF_ASSIGN(var, ent, struct ftrace_graph_ret_entry, \ |
| 335 | TRACE_GRAPH_RET); \ | 229 | TRACE_GRAPH_RET); \ |
| 336 | IF_ASSIGN(var, ent, struct hw_branch_entry, TRACE_HW_BRANCHES);\ | 230 | IF_ASSIGN(var, ent, struct hw_branch_entry, TRACE_HW_BRANCHES);\ |
| 337 | IF_ASSIGN(var, ent, struct trace_power, TRACE_POWER); \ | ||
| 338 | IF_ASSIGN(var, ent, struct kmemtrace_alloc_entry, \ | 231 | IF_ASSIGN(var, ent, struct kmemtrace_alloc_entry, \ |
| 339 | TRACE_KMEM_ALLOC); \ | 232 | TRACE_KMEM_ALLOC); \ |
| 340 | IF_ASSIGN(var, ent, struct kmemtrace_free_entry, \ | 233 | IF_ASSIGN(var, ent, struct kmemtrace_free_entry, \ |
| @@ -414,7 +307,6 @@ struct tracer { | |||
| 414 | struct tracer *next; | 307 | struct tracer *next; |
| 415 | int print_max; | 308 | int print_max; |
| 416 | struct tracer_flags *flags; | 309 | struct tracer_flags *flags; |
| 417 | struct tracer_stat *stats; | ||
| 418 | }; | 310 | }; |
| 419 | 311 | ||
| 420 | 312 | ||
| @@ -493,6 +385,7 @@ void tracing_stop_sched_switch_record(void); | |||
| 493 | void tracing_start_sched_switch_record(void); | 385 | void tracing_start_sched_switch_record(void); |
| 494 | int register_tracer(struct tracer *type); | 386 | int register_tracer(struct tracer *type); |
| 495 | void unregister_tracer(struct tracer *type); | 387 | void unregister_tracer(struct tracer *type); |
| 388 | int is_tracing_stopped(void); | ||
| 496 | 389 | ||
| 497 | extern unsigned long nsecs_to_usecs(unsigned long nsecs); | 390 | extern unsigned long nsecs_to_usecs(unsigned long nsecs); |
| 498 | 391 | ||
| @@ -533,20 +426,6 @@ static inline void __trace_stack(struct trace_array *tr, unsigned long flags, | |||
| 533 | 426 | ||
| 534 | extern cycle_t ftrace_now(int cpu); | 427 | extern cycle_t ftrace_now(int cpu); |
| 535 | 428 | ||
| 536 | #ifdef CONFIG_CONTEXT_SWITCH_TRACER | ||
| 537 | typedef void | ||
| 538 | (*tracer_switch_func_t)(void *private, | ||
| 539 | void *__rq, | ||
| 540 | struct task_struct *prev, | ||
| 541 | struct task_struct *next); | ||
| 542 | |||
| 543 | struct tracer_switch_ops { | ||
| 544 | tracer_switch_func_t func; | ||
| 545 | void *private; | ||
| 546 | struct tracer_switch_ops *next; | ||
| 547 | }; | ||
| 548 | #endif /* CONFIG_CONTEXT_SWITCH_TRACER */ | ||
| 549 | |||
| 550 | extern void trace_find_cmdline(int pid, char comm[]); | 429 | extern void trace_find_cmdline(int pid, char comm[]); |
| 551 | 430 | ||
| 552 | #ifdef CONFIG_DYNAMIC_FTRACE | 431 | #ifdef CONFIG_DYNAMIC_FTRACE |
| @@ -662,6 +541,41 @@ static inline int ftrace_trace_task(struct task_struct *task) | |||
| 662 | #endif | 541 | #endif |
| 663 | 542 | ||
| 664 | /* | 543 | /* |
| 544 | * struct trace_parser - servers for reading the user input separated by spaces | ||
| 545 | * @cont: set if the input is not complete - no final space char was found | ||
| 546 | * @buffer: holds the parsed user input | ||
| 547 | * @idx: user input lenght | ||
| 548 | * @size: buffer size | ||
| 549 | */ | ||
| 550 | struct trace_parser { | ||
| 551 | bool cont; | ||
| 552 | char *buffer; | ||
| 553 | unsigned idx; | ||
| 554 | unsigned size; | ||
| 555 | }; | ||
| 556 | |||
| 557 | static inline bool trace_parser_loaded(struct trace_parser *parser) | ||
| 558 | { | ||
| 559 | return (parser->idx != 0); | ||
| 560 | } | ||
| 561 | |||
| 562 | static inline bool trace_parser_cont(struct trace_parser *parser) | ||
| 563 | { | ||
| 564 | return parser->cont; | ||
| 565 | } | ||
| 566 | |||
| 567 | static inline void trace_parser_clear(struct trace_parser *parser) | ||
| 568 | { | ||
| 569 | parser->cont = false; | ||
| 570 | parser->idx = 0; | ||
| 571 | } | ||
| 572 | |||
| 573 | extern int trace_parser_get_init(struct trace_parser *parser, int size); | ||
| 574 | extern void trace_parser_put(struct trace_parser *parser); | ||
| 575 | extern int trace_get_user(struct trace_parser *parser, const char __user *ubuf, | ||
| 576 | size_t cnt, loff_t *ppos); | ||
| 577 | |||
| 578 | /* | ||
| 665 | * trace_iterator_flags is an enumeration that defines bit | 579 | * trace_iterator_flags is an enumeration that defines bit |
| 666 | * positions into trace_flags that controls the output. | 580 | * positions into trace_flags that controls the output. |
| 667 | * | 581 | * |
| @@ -847,58 +761,18 @@ filter_check_discard(struct ftrace_event_call *call, void *rec, | |||
| 847 | return 0; | 761 | return 0; |
| 848 | } | 762 | } |
| 849 | 763 | ||
| 850 | #define DEFINE_COMPARISON_PRED(type) \ | ||
| 851 | static int filter_pred_##type(struct filter_pred *pred, void *event, \ | ||
| 852 | int val1, int val2) \ | ||
| 853 | { \ | ||
| 854 | type *addr = (type *)(event + pred->offset); \ | ||
| 855 | type val = (type)pred->val; \ | ||
| 856 | int match = 0; \ | ||
| 857 | \ | ||
| 858 | switch (pred->op) { \ | ||
| 859 | case OP_LT: \ | ||
| 860 | match = (*addr < val); \ | ||
| 861 | break; \ | ||
| 862 | case OP_LE: \ | ||
| 863 | match = (*addr <= val); \ | ||
| 864 | break; \ | ||
| 865 | case OP_GT: \ | ||
| 866 | match = (*addr > val); \ | ||
| 867 | break; \ | ||
| 868 | case OP_GE: \ | ||
| 869 | match = (*addr >= val); \ | ||
| 870 | break; \ | ||
| 871 | default: \ | ||
| 872 | break; \ | ||
| 873 | } \ | ||
| 874 | \ | ||
| 875 | return match; \ | ||
| 876 | } | ||
| 877 | |||
| 878 | #define DEFINE_EQUALITY_PRED(size) \ | ||
| 879 | static int filter_pred_##size(struct filter_pred *pred, void *event, \ | ||
| 880 | int val1, int val2) \ | ||
| 881 | { \ | ||
| 882 | u##size *addr = (u##size *)(event + pred->offset); \ | ||
| 883 | u##size val = (u##size)pred->val; \ | ||
| 884 | int match; \ | ||
| 885 | \ | ||
| 886 | match = (val == *addr) ^ pred->not; \ | ||
| 887 | \ | ||
| 888 | return match; \ | ||
| 889 | } | ||
| 890 | |||
| 891 | extern struct mutex event_mutex; | 764 | extern struct mutex event_mutex; |
| 892 | extern struct list_head ftrace_events; | 765 | extern struct list_head ftrace_events; |
| 893 | 766 | ||
| 894 | extern const char *__start___trace_bprintk_fmt[]; | 767 | extern const char *__start___trace_bprintk_fmt[]; |
| 895 | extern const char *__stop___trace_bprintk_fmt[]; | 768 | extern const char *__stop___trace_bprintk_fmt[]; |
| 896 | 769 | ||
| 897 | #undef TRACE_EVENT_FORMAT | 770 | #undef FTRACE_ENTRY |
| 898 | #define TRACE_EVENT_FORMAT(call, proto, args, fmt, tstruct, tpfmt) \ | 771 | #define FTRACE_ENTRY(call, struct_name, id, tstruct, print) \ |
| 899 | extern struct ftrace_event_call event_##call; | 772 | extern struct ftrace_event_call event_##call; |
| 900 | #undef TRACE_EVENT_FORMAT_NOFILTER | 773 | #undef FTRACE_ENTRY_DUP |
| 901 | #define TRACE_EVENT_FORMAT_NOFILTER(call, proto, args, fmt, tstruct, tpfmt) | 774 | #define FTRACE_ENTRY_DUP(call, struct_name, id, tstruct, print) \ |
| 902 | #include "trace_event_types.h" | 775 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) |
| 776 | #include "trace_entries.h" | ||
| 903 | 777 | ||
| 904 | #endif /* _LINUX_KERNEL_TRACE_H */ | 778 | #endif /* _LINUX_KERNEL_TRACE_H */ |
diff --git a/kernel/trace/trace_boot.c b/kernel/trace/trace_boot.c index 19bfc75d467e..c21d5f3956ad 100644 --- a/kernel/trace/trace_boot.c +++ b/kernel/trace/trace_boot.c | |||
| @@ -129,6 +129,7 @@ struct tracer boot_tracer __read_mostly = | |||
| 129 | 129 | ||
| 130 | void trace_boot_call(struct boot_trace_call *bt, initcall_t fn) | 130 | void trace_boot_call(struct boot_trace_call *bt, initcall_t fn) |
| 131 | { | 131 | { |
| 132 | struct ftrace_event_call *call = &event_boot_call; | ||
| 132 | struct ring_buffer_event *event; | 133 | struct ring_buffer_event *event; |
| 133 | struct ring_buffer *buffer; | 134 | struct ring_buffer *buffer; |
| 134 | struct trace_boot_call *entry; | 135 | struct trace_boot_call *entry; |
| @@ -150,13 +151,15 @@ void trace_boot_call(struct boot_trace_call *bt, initcall_t fn) | |||
| 150 | goto out; | 151 | goto out; |
| 151 | entry = ring_buffer_event_data(event); | 152 | entry = ring_buffer_event_data(event); |
| 152 | entry->boot_call = *bt; | 153 | entry->boot_call = *bt; |
| 153 | trace_buffer_unlock_commit(buffer, event, 0, 0); | 154 | if (!filter_check_discard(call, entry, buffer, event)) |
| 155 | trace_buffer_unlock_commit(buffer, event, 0, 0); | ||
| 154 | out: | 156 | out: |
| 155 | preempt_enable(); | 157 | preempt_enable(); |
| 156 | } | 158 | } |
| 157 | 159 | ||
| 158 | void trace_boot_ret(struct boot_trace_ret *bt, initcall_t fn) | 160 | void trace_boot_ret(struct boot_trace_ret *bt, initcall_t fn) |
| 159 | { | 161 | { |
| 162 | struct ftrace_event_call *call = &event_boot_ret; | ||
| 160 | struct ring_buffer_event *event; | 163 | struct ring_buffer_event *event; |
| 161 | struct ring_buffer *buffer; | 164 | struct ring_buffer *buffer; |
| 162 | struct trace_boot_ret *entry; | 165 | struct trace_boot_ret *entry; |
| @@ -175,7 +178,8 @@ void trace_boot_ret(struct boot_trace_ret *bt, initcall_t fn) | |||
| 175 | goto out; | 178 | goto out; |
| 176 | entry = ring_buffer_event_data(event); | 179 | entry = ring_buffer_event_data(event); |
| 177 | entry->boot_ret = *bt; | 180 | entry->boot_ret = *bt; |
| 178 | trace_buffer_unlock_commit(buffer, event, 0, 0); | 181 | if (!filter_check_discard(call, entry, buffer, event)) |
| 182 | trace_buffer_unlock_commit(buffer, event, 0, 0); | ||
| 179 | out: | 183 | out: |
| 180 | preempt_enable(); | 184 | preempt_enable(); |
| 181 | } | 185 | } |
diff --git a/kernel/trace/trace_clock.c b/kernel/trace/trace_clock.c index b588fd81f7f9..20c5f92e28a8 100644 --- a/kernel/trace/trace_clock.c +++ b/kernel/trace/trace_clock.c | |||
| @@ -66,10 +66,14 @@ u64 notrace trace_clock(void) | |||
| 66 | * Used by plugins that need globally coherent timestamps. | 66 | * Used by plugins that need globally coherent timestamps. |
| 67 | */ | 67 | */ |
| 68 | 68 | ||
| 69 | static u64 prev_trace_clock_time; | 69 | /* keep prev_time and lock in the same cacheline. */ |
| 70 | 70 | static struct { | |
| 71 | static raw_spinlock_t trace_clock_lock ____cacheline_aligned_in_smp = | 71 | u64 prev_time; |
| 72 | (raw_spinlock_t)__RAW_SPIN_LOCK_UNLOCKED; | 72 | raw_spinlock_t lock; |
| 73 | } trace_clock_struct ____cacheline_aligned_in_smp = | ||
| 74 | { | ||
| 75 | .lock = (raw_spinlock_t)__RAW_SPIN_LOCK_UNLOCKED, | ||
| 76 | }; | ||
| 73 | 77 | ||
| 74 | u64 notrace trace_clock_global(void) | 78 | u64 notrace trace_clock_global(void) |
| 75 | { | 79 | { |
| @@ -88,19 +92,19 @@ u64 notrace trace_clock_global(void) | |||
| 88 | if (unlikely(in_nmi())) | 92 | if (unlikely(in_nmi())) |
| 89 | goto out; | 93 | goto out; |
| 90 | 94 | ||
| 91 | __raw_spin_lock(&trace_clock_lock); | 95 | __raw_spin_lock(&trace_clock_struct.lock); |
| 92 | 96 | ||
| 93 | /* | 97 | /* |
| 94 | * TODO: if this happens often then maybe we should reset | 98 | * TODO: if this happens often then maybe we should reset |
| 95 | * my_scd->clock to prev_trace_clock_time+1, to make sure | 99 | * my_scd->clock to prev_time+1, to make sure |
| 96 | * we start ticking with the local clock from now on? | 100 | * we start ticking with the local clock from now on? |
| 97 | */ | 101 | */ |
| 98 | if ((s64)(now - prev_trace_clock_time) < 0) | 102 | if ((s64)(now - trace_clock_struct.prev_time) < 0) |
| 99 | now = prev_trace_clock_time + 1; | 103 | now = trace_clock_struct.prev_time + 1; |
| 100 | 104 | ||
| 101 | prev_trace_clock_time = now; | 105 | trace_clock_struct.prev_time = now; |
| 102 | 106 | ||
| 103 | __raw_spin_unlock(&trace_clock_lock); | 107 | __raw_spin_unlock(&trace_clock_struct.lock); |
| 104 | 108 | ||
| 105 | out: | 109 | out: |
| 106 | raw_local_irq_restore(flags); | 110 | raw_local_irq_restore(flags); |
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h new file mode 100644 index 000000000000..ead3d724599d --- /dev/null +++ b/kernel/trace/trace_entries.h | |||
| @@ -0,0 +1,366 @@ | |||
| 1 | /* | ||
| 2 | * This file defines the trace event structures that go into the ring | ||
| 3 | * buffer directly. They are created via macros so that changes for them | ||
| 4 | * appear in the format file. Using macros will automate this process. | ||
| 5 | * | ||
| 6 | * The macro used to create a ftrace data structure is: | ||
| 7 | * | ||
| 8 | * FTRACE_ENTRY( name, struct_name, id, structure, print ) | ||
| 9 | * | ||
| 10 | * @name: the name used the event name, as well as the name of | ||
| 11 | * the directory that holds the format file. | ||
| 12 | * | ||
| 13 | * @struct_name: the name of the structure that is created. | ||
| 14 | * | ||
| 15 | * @id: The event identifier that is used to detect what event | ||
| 16 | * this is from the ring buffer. | ||
| 17 | * | ||
| 18 | * @structure: the structure layout | ||
| 19 | * | ||
| 20 | * - __field( type, item ) | ||
| 21 | * This is equivalent to declaring | ||
| 22 | * type item; | ||
| 23 | * in the structure. | ||
| 24 | * - __array( type, item, size ) | ||
| 25 | * This is equivalent to declaring | ||
| 26 | * type item[size]; | ||
| 27 | * in the structure. | ||
| 28 | * | ||
| 29 | * * for structures within structures, the format of the internal | ||
| 30 | * structure is layed out. This allows the internal structure | ||
| 31 | * to be deciphered for the format file. Although these macros | ||
| 32 | * may become out of sync with the internal structure, they | ||
| 33 | * will create a compile error if it happens. Since the | ||
| 34 | * internel structures are just tracing helpers, this is not | ||
| 35 | * an issue. | ||
| 36 | * | ||
| 37 | * When an internal structure is used, it should use: | ||
| 38 | * | ||
| 39 | * __field_struct( type, item ) | ||
| 40 | * | ||
| 41 | * instead of __field. This will prevent it from being shown in | ||
| 42 | * the output file. The fields in the structure should use. | ||
| 43 | * | ||
| 44 | * __field_desc( type, container, item ) | ||
| 45 | * __array_desc( type, container, item, len ) | ||
| 46 | * | ||
| 47 | * type, item and len are the same as __field and __array, but | ||
| 48 | * container is added. This is the name of the item in | ||
| 49 | * __field_struct that this is describing. | ||
| 50 | * | ||
| 51 | * | ||
| 52 | * @print: the print format shown to users in the format file. | ||
| 53 | */ | ||
| 54 | |||
| 55 | /* | ||
| 56 | * Function trace entry - function address and parent function addres: | ||
| 57 | */ | ||
| 58 | FTRACE_ENTRY(function, ftrace_entry, | ||
| 59 | |||
| 60 | TRACE_FN, | ||
| 61 | |||
| 62 | F_STRUCT( | ||
| 63 | __field( unsigned long, ip ) | ||
| 64 | __field( unsigned long, parent_ip ) | ||
| 65 | ), | ||
| 66 | |||
| 67 | F_printk(" %lx <-- %lx", __entry->ip, __entry->parent_ip) | ||
| 68 | ); | ||
| 69 | |||
| 70 | /* Function call entry */ | ||
| 71 | FTRACE_ENTRY(funcgraph_entry, ftrace_graph_ent_entry, | ||
| 72 | |||
| 73 | TRACE_GRAPH_ENT, | ||
| 74 | |||
| 75 | F_STRUCT( | ||
| 76 | __field_struct( struct ftrace_graph_ent, graph_ent ) | ||
| 77 | __field_desc( unsigned long, graph_ent, func ) | ||
| 78 | __field_desc( int, graph_ent, depth ) | ||
| 79 | ), | ||
| 80 | |||
| 81 | F_printk("--> %lx (%d)", __entry->func, __entry->depth) | ||
| 82 | ); | ||
| 83 | |||
| 84 | /* Function return entry */ | ||
| 85 | FTRACE_ENTRY(funcgraph_exit, ftrace_graph_ret_entry, | ||
| 86 | |||
| 87 | TRACE_GRAPH_RET, | ||
| 88 | |||
| 89 | F_STRUCT( | ||
| 90 | __field_struct( struct ftrace_graph_ret, ret ) | ||
| 91 | __field_desc( unsigned long, ret, func ) | ||
| 92 | __field_desc( unsigned long long, ret, calltime) | ||
| 93 | __field_desc( unsigned long long, ret, rettime ) | ||
| 94 | __field_desc( unsigned long, ret, overrun ) | ||
| 95 | __field_desc( int, ret, depth ) | ||
| 96 | ), | ||
| 97 | |||
| 98 | F_printk("<-- %lx (%d) (start: %llx end: %llx) over: %d", | ||
| 99 | __entry->func, __entry->depth, | ||
| 100 | __entry->calltime, __entry->rettime, | ||
| 101 | __entry->depth) | ||
| 102 | ); | ||
| 103 | |||
| 104 | /* | ||
| 105 | * Context switch trace entry - which task (and prio) we switched from/to: | ||
| 106 | * | ||
| 107 | * This is used for both wakeup and context switches. We only want | ||
| 108 | * to create one structure, but we need two outputs for it. | ||
| 109 | */ | ||
| 110 | #define FTRACE_CTX_FIELDS \ | ||
| 111 | __field( unsigned int, prev_pid ) \ | ||
| 112 | __field( unsigned char, prev_prio ) \ | ||
| 113 | __field( unsigned char, prev_state ) \ | ||
| 114 | __field( unsigned int, next_pid ) \ | ||
| 115 | __field( unsigned char, next_prio ) \ | ||
| 116 | __field( unsigned char, next_state ) \ | ||
| 117 | __field( unsigned int, next_cpu ) | ||
| 118 | |||
| 119 | FTRACE_ENTRY(context_switch, ctx_switch_entry, | ||
| 120 | |||
| 121 | TRACE_CTX, | ||
| 122 | |||
| 123 | F_STRUCT( | ||
| 124 | FTRACE_CTX_FIELDS | ||
| 125 | ), | ||
| 126 | |||
| 127 | F_printk("%u:%u:%u ==> %u:%u:%u [%03u]", | ||
| 128 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, | ||
| 129 | __entry->next_pid, __entry->next_prio, __entry->next_state, | ||
| 130 | __entry->next_cpu | ||
| 131 | ) | ||
| 132 | ); | ||
| 133 | |||
| 134 | /* | ||
| 135 | * FTRACE_ENTRY_DUP only creates the format file, it will not | ||
| 136 | * create another structure. | ||
| 137 | */ | ||
| 138 | FTRACE_ENTRY_DUP(wakeup, ctx_switch_entry, | ||
| 139 | |||
| 140 | TRACE_WAKE, | ||
| 141 | |||
| 142 | F_STRUCT( | ||
| 143 | FTRACE_CTX_FIELDS | ||
| 144 | ), | ||
| 145 | |||
| 146 | F_printk("%u:%u:%u ==+ %u:%u:%u [%03u]", | ||
| 147 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, | ||
| 148 | __entry->next_pid, __entry->next_prio, __entry->next_state, | ||
| 149 | __entry->next_cpu | ||
| 150 | ) | ||
| 151 | ); | ||
| 152 | |||
| 153 | /* | ||
| 154 | * Special (free-form) trace entry: | ||
| 155 | */ | ||
| 156 | FTRACE_ENTRY(special, special_entry, | ||
| 157 | |||
| 158 | TRACE_SPECIAL, | ||
| 159 | |||
| 160 | F_STRUCT( | ||
| 161 | __field( unsigned long, arg1 ) | ||
| 162 | __field( unsigned long, arg2 ) | ||
| 163 | __field( unsigned long, arg3 ) | ||
| 164 | ), | ||
| 165 | |||
| 166 | F_printk("(%08lx) (%08lx) (%08lx)", | ||
| 167 | __entry->arg1, __entry->arg2, __entry->arg3) | ||
| 168 | ); | ||
| 169 | |||
| 170 | /* | ||
| 171 | * Stack-trace entry: | ||
| 172 | */ | ||
| 173 | |||
| 174 | #define FTRACE_STACK_ENTRIES 8 | ||
| 175 | |||
| 176 | FTRACE_ENTRY(kernel_stack, stack_entry, | ||
| 177 | |||
| 178 | TRACE_STACK, | ||
| 179 | |||
| 180 | F_STRUCT( | ||
| 181 | __array( unsigned long, caller, FTRACE_STACK_ENTRIES ) | ||
| 182 | ), | ||
| 183 | |||
| 184 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | ||
| 185 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", | ||
| 186 | __entry->caller[0], __entry->caller[1], __entry->caller[2], | ||
| 187 | __entry->caller[3], __entry->caller[4], __entry->caller[5], | ||
| 188 | __entry->caller[6], __entry->caller[7]) | ||
| 189 | ); | ||
| 190 | |||
| 191 | FTRACE_ENTRY(user_stack, userstack_entry, | ||
| 192 | |||
| 193 | TRACE_USER_STACK, | ||
| 194 | |||
| 195 | F_STRUCT( | ||
| 196 | __field( unsigned int, tgid ) | ||
| 197 | __array( unsigned long, caller, FTRACE_STACK_ENTRIES ) | ||
| 198 | ), | ||
| 199 | |||
| 200 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | ||
| 201 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", | ||
| 202 | __entry->caller[0], __entry->caller[1], __entry->caller[2], | ||
| 203 | __entry->caller[3], __entry->caller[4], __entry->caller[5], | ||
| 204 | __entry->caller[6], __entry->caller[7]) | ||
| 205 | ); | ||
| 206 | |||
| 207 | /* | ||
| 208 | * trace_printk entry: | ||
| 209 | */ | ||
| 210 | FTRACE_ENTRY(bprint, bprint_entry, | ||
| 211 | |||
| 212 | TRACE_BPRINT, | ||
| 213 | |||
| 214 | F_STRUCT( | ||
| 215 | __field( unsigned long, ip ) | ||
| 216 | __field( const char *, fmt ) | ||
| 217 | __dynamic_array( u32, buf ) | ||
| 218 | ), | ||
| 219 | |||
| 220 | F_printk("%08lx fmt:%p", | ||
| 221 | __entry->ip, __entry->fmt) | ||
| 222 | ); | ||
| 223 | |||
| 224 | FTRACE_ENTRY(print, print_entry, | ||
| 225 | |||
| 226 | TRACE_PRINT, | ||
| 227 | |||
| 228 | F_STRUCT( | ||
| 229 | __field( unsigned long, ip ) | ||
| 230 | __dynamic_array( char, buf ) | ||
| 231 | ), | ||
| 232 | |||
| 233 | F_printk("%08lx %s", | ||
| 234 | __entry->ip, __entry->buf) | ||
| 235 | ); | ||
| 236 | |||
| 237 | FTRACE_ENTRY(mmiotrace_rw, trace_mmiotrace_rw, | ||
| 238 | |||
| 239 | TRACE_MMIO_RW, | ||
| 240 | |||
| 241 | F_STRUCT( | ||
| 242 | __field_struct( struct mmiotrace_rw, rw ) | ||
| 243 | __field_desc( resource_size_t, rw, phys ) | ||
| 244 | __field_desc( unsigned long, rw, value ) | ||
| 245 | __field_desc( unsigned long, rw, pc ) | ||
| 246 | __field_desc( int, rw, map_id ) | ||
| 247 | __field_desc( unsigned char, rw, opcode ) | ||
| 248 | __field_desc( unsigned char, rw, width ) | ||
| 249 | ), | ||
| 250 | |||
| 251 | F_printk("%lx %lx %lx %d %x %x", | ||
| 252 | (unsigned long)__entry->phys, __entry->value, __entry->pc, | ||
| 253 | __entry->map_id, __entry->opcode, __entry->width) | ||
| 254 | ); | ||
| 255 | |||
| 256 | FTRACE_ENTRY(mmiotrace_map, trace_mmiotrace_map, | ||
| 257 | |||
| 258 | TRACE_MMIO_MAP, | ||
| 259 | |||
| 260 | F_STRUCT( | ||
| 261 | __field_struct( struct mmiotrace_map, map ) | ||
| 262 | __field_desc( resource_size_t, map, phys ) | ||
| 263 | __field_desc( unsigned long, map, virt ) | ||
| 264 | __field_desc( unsigned long, map, len ) | ||
| 265 | __field_desc( int, map, map_id ) | ||
| 266 | __field_desc( unsigned char, map, opcode ) | ||
| 267 | ), | ||
| 268 | |||
| 269 | F_printk("%lx %lx %lx %d %x", | ||
| 270 | (unsigned long)__entry->phys, __entry->virt, __entry->len, | ||
| 271 | __entry->map_id, __entry->opcode) | ||
| 272 | ); | ||
| 273 | |||
| 274 | FTRACE_ENTRY(boot_call, trace_boot_call, | ||
| 275 | |||
| 276 | TRACE_BOOT_CALL, | ||
| 277 | |||
| 278 | F_STRUCT( | ||
| 279 | __field_struct( struct boot_trace_call, boot_call ) | ||
| 280 | __field_desc( pid_t, boot_call, caller ) | ||
| 281 | __array_desc( char, boot_call, func, KSYM_SYMBOL_LEN) | ||
| 282 | ), | ||
| 283 | |||
| 284 | F_printk("%d %s", __entry->caller, __entry->func) | ||
| 285 | ); | ||
| 286 | |||
| 287 | FTRACE_ENTRY(boot_ret, trace_boot_ret, | ||
| 288 | |||
| 289 | TRACE_BOOT_RET, | ||
| 290 | |||
| 291 | F_STRUCT( | ||
| 292 | __field_struct( struct boot_trace_ret, boot_ret ) | ||
| 293 | __array_desc( char, boot_ret, func, KSYM_SYMBOL_LEN) | ||
| 294 | __field_desc( int, boot_ret, result ) | ||
| 295 | __field_desc( unsigned long, boot_ret, duration ) | ||
| 296 | ), | ||
| 297 | |||
| 298 | F_printk("%s %d %lx", | ||
| 299 | __entry->func, __entry->result, __entry->duration) | ||
| 300 | ); | ||
| 301 | |||
| 302 | #define TRACE_FUNC_SIZE 30 | ||
| 303 | #define TRACE_FILE_SIZE 20 | ||
| 304 | |||
| 305 | FTRACE_ENTRY(branch, trace_branch, | ||
| 306 | |||
| 307 | TRACE_BRANCH, | ||
| 308 | |||
| 309 | F_STRUCT( | ||
| 310 | __field( unsigned int, line ) | ||
| 311 | __array( char, func, TRACE_FUNC_SIZE+1 ) | ||
| 312 | __array( char, file, TRACE_FILE_SIZE+1 ) | ||
| 313 | __field( char, correct ) | ||
| 314 | ), | ||
| 315 | |||
| 316 | F_printk("%u:%s:%s (%u)", | ||
| 317 | __entry->line, | ||
| 318 | __entry->func, __entry->file, __entry->correct) | ||
| 319 | ); | ||
| 320 | |||
| 321 | FTRACE_ENTRY(hw_branch, hw_branch_entry, | ||
| 322 | |||
| 323 | TRACE_HW_BRANCHES, | ||
| 324 | |||
| 325 | F_STRUCT( | ||
| 326 | __field( u64, from ) | ||
| 327 | __field( u64, to ) | ||
| 328 | ), | ||
| 329 | |||
| 330 | F_printk("from: %llx to: %llx", __entry->from, __entry->to) | ||
| 331 | ); | ||
| 332 | |||
| 333 | FTRACE_ENTRY(kmem_alloc, kmemtrace_alloc_entry, | ||
| 334 | |||
| 335 | TRACE_KMEM_ALLOC, | ||
| 336 | |||
| 337 | F_STRUCT( | ||
| 338 | __field( enum kmemtrace_type_id, type_id ) | ||
| 339 | __field( unsigned long, call_site ) | ||
| 340 | __field( const void *, ptr ) | ||
| 341 | __field( size_t, bytes_req ) | ||
| 342 | __field( size_t, bytes_alloc ) | ||
| 343 | __field( gfp_t, gfp_flags ) | ||
| 344 | __field( int, node ) | ||
| 345 | ), | ||
| 346 | |||
| 347 | F_printk("type:%u call_site:%lx ptr:%p req:%zi alloc:%zi" | ||
| 348 | " flags:%x node:%d", | ||
| 349 | __entry->type_id, __entry->call_site, __entry->ptr, | ||
| 350 | __entry->bytes_req, __entry->bytes_alloc, | ||
| 351 | __entry->gfp_flags, __entry->node) | ||
| 352 | ); | ||
| 353 | |||
| 354 | FTRACE_ENTRY(kmem_free, kmemtrace_free_entry, | ||
| 355 | |||
| 356 | TRACE_KMEM_FREE, | ||
| 357 | |||
| 358 | F_STRUCT( | ||
| 359 | __field( enum kmemtrace_type_id, type_id ) | ||
| 360 | __field( unsigned long, call_site ) | ||
| 361 | __field( const void *, ptr ) | ||
| 362 | ), | ||
| 363 | |||
| 364 | F_printk("type:%u call_site:%lx ptr:%p", | ||
| 365 | __entry->type_id, __entry->call_site, __entry->ptr) | ||
| 366 | ); | ||
diff --git a/kernel/trace/trace_event_profile.c b/kernel/trace/trace_event_profile.c index 11ba5bb4ed0a..e812f1c1264c 100644 --- a/kernel/trace/trace_event_profile.c +++ b/kernel/trace/trace_event_profile.c | |||
| @@ -5,8 +5,60 @@ | |||
| 5 | * | 5 | * |
| 6 | */ | 6 | */ |
| 7 | 7 | ||
| 8 | #include <linux/module.h> | ||
| 8 | #include "trace.h" | 9 | #include "trace.h" |
| 9 | 10 | ||
| 11 | /* | ||
| 12 | * We can't use a size but a type in alloc_percpu() | ||
| 13 | * So let's create a dummy type that matches the desired size | ||
| 14 | */ | ||
| 15 | typedef struct {char buf[FTRACE_MAX_PROFILE_SIZE];} profile_buf_t; | ||
| 16 | |||
| 17 | char *trace_profile_buf; | ||
| 18 | EXPORT_SYMBOL_GPL(trace_profile_buf); | ||
| 19 | |||
| 20 | char *trace_profile_buf_nmi; | ||
| 21 | EXPORT_SYMBOL_GPL(trace_profile_buf_nmi); | ||
| 22 | |||
| 23 | /* Count the events in use (per event id, not per instance) */ | ||
| 24 | static int total_profile_count; | ||
| 25 | |||
| 26 | static int ftrace_profile_enable_event(struct ftrace_event_call *event) | ||
| 27 | { | ||
| 28 | char *buf; | ||
| 29 | int ret = -ENOMEM; | ||
| 30 | |||
| 31 | if (atomic_inc_return(&event->profile_count)) | ||
| 32 | return 0; | ||
| 33 | |||
| 34 | if (!total_profile_count++) { | ||
| 35 | buf = (char *)alloc_percpu(profile_buf_t); | ||
| 36 | if (!buf) | ||
| 37 | goto fail_buf; | ||
| 38 | |||
| 39 | rcu_assign_pointer(trace_profile_buf, buf); | ||
| 40 | |||
| 41 | buf = (char *)alloc_percpu(profile_buf_t); | ||
| 42 | if (!buf) | ||
| 43 | goto fail_buf_nmi; | ||
| 44 | |||
| 45 | rcu_assign_pointer(trace_profile_buf_nmi, buf); | ||
| 46 | } | ||
| 47 | |||
| 48 | ret = event->profile_enable(event); | ||
| 49 | if (!ret) | ||
| 50 | return 0; | ||
| 51 | |||
| 52 | kfree(trace_profile_buf_nmi); | ||
| 53 | fail_buf_nmi: | ||
| 54 | kfree(trace_profile_buf); | ||
| 55 | fail_buf: | ||
| 56 | total_profile_count--; | ||
| 57 | atomic_dec(&event->profile_count); | ||
| 58 | |||
| 59 | return ret; | ||
| 60 | } | ||
| 61 | |||
| 10 | int ftrace_profile_enable(int event_id) | 62 | int ftrace_profile_enable(int event_id) |
| 11 | { | 63 | { |
| 12 | struct ftrace_event_call *event; | 64 | struct ftrace_event_call *event; |
| @@ -14,8 +66,9 @@ int ftrace_profile_enable(int event_id) | |||
| 14 | 66 | ||
| 15 | mutex_lock(&event_mutex); | 67 | mutex_lock(&event_mutex); |
| 16 | list_for_each_entry(event, &ftrace_events, list) { | 68 | list_for_each_entry(event, &ftrace_events, list) { |
| 17 | if (event->id == event_id && event->profile_enable) { | 69 | if (event->id == event_id && event->profile_enable && |
| 18 | ret = event->profile_enable(event); | 70 | try_module_get(event->mod)) { |
| 71 | ret = ftrace_profile_enable_event(event); | ||
| 19 | break; | 72 | break; |
| 20 | } | 73 | } |
| 21 | } | 74 | } |
| @@ -24,6 +77,33 @@ int ftrace_profile_enable(int event_id) | |||
| 24 | return ret; | 77 | return ret; |
| 25 | } | 78 | } |
| 26 | 79 | ||
| 80 | static void ftrace_profile_disable_event(struct ftrace_event_call *event) | ||
| 81 | { | ||
| 82 | char *buf, *nmi_buf; | ||
| 83 | |||
| 84 | if (!atomic_add_negative(-1, &event->profile_count)) | ||
| 85 | return; | ||
| 86 | |||
| 87 | event->profile_disable(event); | ||
| 88 | |||
| 89 | if (!--total_profile_count) { | ||
| 90 | buf = trace_profile_buf; | ||
| 91 | rcu_assign_pointer(trace_profile_buf, NULL); | ||
| 92 | |||
| 93 | nmi_buf = trace_profile_buf_nmi; | ||
| 94 | rcu_assign_pointer(trace_profile_buf_nmi, NULL); | ||
| 95 | |||
| 96 | /* | ||
| 97 | * Ensure every events in profiling have finished before | ||
| 98 | * releasing the buffers | ||
| 99 | */ | ||
| 100 | synchronize_sched(); | ||
| 101 | |||
| 102 | free_percpu(buf); | ||
| 103 | free_percpu(nmi_buf); | ||
| 104 | } | ||
| 105 | } | ||
| 106 | |||
| 27 | void ftrace_profile_disable(int event_id) | 107 | void ftrace_profile_disable(int event_id) |
| 28 | { | 108 | { |
| 29 | struct ftrace_event_call *event; | 109 | struct ftrace_event_call *event; |
| @@ -31,7 +111,8 @@ void ftrace_profile_disable(int event_id) | |||
| 31 | mutex_lock(&event_mutex); | 111 | mutex_lock(&event_mutex); |
| 32 | list_for_each_entry(event, &ftrace_events, list) { | 112 | list_for_each_entry(event, &ftrace_events, list) { |
| 33 | if (event->id == event_id) { | 113 | if (event->id == event_id) { |
| 34 | event->profile_disable(event); | 114 | ftrace_profile_disable_event(event); |
| 115 | module_put(event->mod); | ||
| 35 | break; | 116 | break; |
| 36 | } | 117 | } |
| 37 | } | 118 | } |
diff --git a/kernel/trace/trace_event_types.h b/kernel/trace/trace_event_types.h deleted file mode 100644 index e74f0906ab1a..000000000000 --- a/kernel/trace/trace_event_types.h +++ /dev/null | |||
| @@ -1,178 +0,0 @@ | |||
| 1 | #undef TRACE_SYSTEM | ||
| 2 | #define TRACE_SYSTEM ftrace | ||
| 3 | |||
| 4 | /* | ||
| 5 | * We cheat and use the proto type field as the ID | ||
| 6 | * and args as the entry type (minus 'struct') | ||
| 7 | */ | ||
| 8 | TRACE_EVENT_FORMAT(function, TRACE_FN, ftrace_entry, ignore, | ||
| 9 | TRACE_STRUCT( | ||
| 10 | TRACE_FIELD(unsigned long, ip, ip) | ||
| 11 | TRACE_FIELD(unsigned long, parent_ip, parent_ip) | ||
| 12 | ), | ||
| 13 | TP_RAW_FMT(" %lx <-- %lx") | ||
| 14 | ); | ||
| 15 | |||
| 16 | TRACE_EVENT_FORMAT(funcgraph_entry, TRACE_GRAPH_ENT, | ||
| 17 | ftrace_graph_ent_entry, ignore, | ||
| 18 | TRACE_STRUCT( | ||
| 19 | TRACE_FIELD(unsigned long, graph_ent.func, func) | ||
| 20 | TRACE_FIELD(int, graph_ent.depth, depth) | ||
| 21 | ), | ||
| 22 | TP_RAW_FMT("--> %lx (%d)") | ||
| 23 | ); | ||
| 24 | |||
| 25 | TRACE_EVENT_FORMAT(funcgraph_exit, TRACE_GRAPH_RET, | ||
| 26 | ftrace_graph_ret_entry, ignore, | ||
| 27 | TRACE_STRUCT( | ||
| 28 | TRACE_FIELD(unsigned long, ret.func, func) | ||
| 29 | TRACE_FIELD(unsigned long long, ret.calltime, calltime) | ||
| 30 | TRACE_FIELD(unsigned long long, ret.rettime, rettime) | ||
| 31 | TRACE_FIELD(unsigned long, ret.overrun, overrun) | ||
| 32 | TRACE_FIELD(int, ret.depth, depth) | ||
| 33 | ), | ||
| 34 | TP_RAW_FMT("<-- %lx (%d)") | ||
| 35 | ); | ||
| 36 | |||
| 37 | TRACE_EVENT_FORMAT(wakeup, TRACE_WAKE, ctx_switch_entry, ignore, | ||
| 38 | TRACE_STRUCT( | ||
| 39 | TRACE_FIELD(unsigned int, prev_pid, prev_pid) | ||
| 40 | TRACE_FIELD(unsigned char, prev_prio, prev_prio) | ||
| 41 | TRACE_FIELD(unsigned char, prev_state, prev_state) | ||
| 42 | TRACE_FIELD(unsigned int, next_pid, next_pid) | ||
| 43 | TRACE_FIELD(unsigned char, next_prio, next_prio) | ||
| 44 | TRACE_FIELD(unsigned char, next_state, next_state) | ||
| 45 | TRACE_FIELD(unsigned int, next_cpu, next_cpu) | ||
| 46 | ), | ||
| 47 | TP_RAW_FMT("%u:%u:%u ==+ %u:%u:%u [%03u]") | ||
| 48 | ); | ||
| 49 | |||
| 50 | TRACE_EVENT_FORMAT(context_switch, TRACE_CTX, ctx_switch_entry, ignore, | ||
| 51 | TRACE_STRUCT( | ||
| 52 | TRACE_FIELD(unsigned int, prev_pid, prev_pid) | ||
| 53 | TRACE_FIELD(unsigned char, prev_prio, prev_prio) | ||
| 54 | TRACE_FIELD(unsigned char, prev_state, prev_state) | ||
| 55 | TRACE_FIELD(unsigned int, next_pid, next_pid) | ||
| 56 | TRACE_FIELD(unsigned char, next_prio, next_prio) | ||
| 57 | TRACE_FIELD(unsigned char, next_state, next_state) | ||
| 58 | TRACE_FIELD(unsigned int, next_cpu, next_cpu) | ||
| 59 | ), | ||
| 60 | TP_RAW_FMT("%u:%u:%u ==+ %u:%u:%u [%03u]") | ||
| 61 | ); | ||
| 62 | |||
| 63 | TRACE_EVENT_FORMAT_NOFILTER(special, TRACE_SPECIAL, special_entry, ignore, | ||
| 64 | TRACE_STRUCT( | ||
| 65 | TRACE_FIELD(unsigned long, arg1, arg1) | ||
| 66 | TRACE_FIELD(unsigned long, arg2, arg2) | ||
| 67 | TRACE_FIELD(unsigned long, arg3, arg3) | ||
| 68 | ), | ||
| 69 | TP_RAW_FMT("(%08lx) (%08lx) (%08lx)") | ||
| 70 | ); | ||
| 71 | |||
| 72 | /* | ||
| 73 | * Stack-trace entry: | ||
| 74 | */ | ||
| 75 | |||
| 76 | /* #define FTRACE_STACK_ENTRIES 8 */ | ||
| 77 | |||
| 78 | TRACE_EVENT_FORMAT(kernel_stack, TRACE_STACK, stack_entry, ignore, | ||
| 79 | TRACE_STRUCT( | ||
| 80 | TRACE_FIELD(unsigned long, caller[0], stack0) | ||
| 81 | TRACE_FIELD(unsigned long, caller[1], stack1) | ||
| 82 | TRACE_FIELD(unsigned long, caller[2], stack2) | ||
| 83 | TRACE_FIELD(unsigned long, caller[3], stack3) | ||
| 84 | TRACE_FIELD(unsigned long, caller[4], stack4) | ||
| 85 | TRACE_FIELD(unsigned long, caller[5], stack5) | ||
| 86 | TRACE_FIELD(unsigned long, caller[6], stack6) | ||
| 87 | TRACE_FIELD(unsigned long, caller[7], stack7) | ||
| 88 | ), | ||
| 89 | TP_RAW_FMT("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | ||
| 90 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n") | ||
| 91 | ); | ||
| 92 | |||
| 93 | TRACE_EVENT_FORMAT(user_stack, TRACE_USER_STACK, userstack_entry, ignore, | ||
| 94 | TRACE_STRUCT( | ||
| 95 | TRACE_FIELD(unsigned long, caller[0], stack0) | ||
| 96 | TRACE_FIELD(unsigned long, caller[1], stack1) | ||
| 97 | TRACE_FIELD(unsigned long, caller[2], stack2) | ||
| 98 | TRACE_FIELD(unsigned long, caller[3], stack3) | ||
| 99 | TRACE_FIELD(unsigned long, caller[4], stack4) | ||
| 100 | TRACE_FIELD(unsigned long, caller[5], stack5) | ||
| 101 | TRACE_FIELD(unsigned long, caller[6], stack6) | ||
| 102 | TRACE_FIELD(unsigned long, caller[7], stack7) | ||
| 103 | ), | ||
| 104 | TP_RAW_FMT("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | ||
| 105 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n") | ||
| 106 | ); | ||
| 107 | |||
| 108 | TRACE_EVENT_FORMAT(bprint, TRACE_BPRINT, bprint_entry, ignore, | ||
| 109 | TRACE_STRUCT( | ||
| 110 | TRACE_FIELD(unsigned long, ip, ip) | ||
| 111 | TRACE_FIELD(char *, fmt, fmt) | ||
| 112 | TRACE_FIELD_ZERO(char, buf) | ||
| 113 | ), | ||
| 114 | TP_RAW_FMT("%08lx (%d) fmt:%p %s") | ||
| 115 | ); | ||
| 116 | |||
| 117 | TRACE_EVENT_FORMAT(print, TRACE_PRINT, print_entry, ignore, | ||
| 118 | TRACE_STRUCT( | ||
| 119 | TRACE_FIELD(unsigned long, ip, ip) | ||
| 120 | TRACE_FIELD_ZERO(char, buf) | ||
| 121 | ), | ||
| 122 | TP_RAW_FMT("%08lx (%d) fmt:%p %s") | ||
| 123 | ); | ||
| 124 | |||
| 125 | TRACE_EVENT_FORMAT(branch, TRACE_BRANCH, trace_branch, ignore, | ||
| 126 | TRACE_STRUCT( | ||
| 127 | TRACE_FIELD(unsigned int, line, line) | ||
| 128 | TRACE_FIELD_SPECIAL(char func[TRACE_FUNC_SIZE+1], func, | ||
| 129 | TRACE_FUNC_SIZE+1, func) | ||
| 130 | TRACE_FIELD_SPECIAL(char file[TRACE_FUNC_SIZE+1], file, | ||
| 131 | TRACE_FUNC_SIZE+1, file) | ||
| 132 | TRACE_FIELD(char, correct, correct) | ||
| 133 | ), | ||
| 134 | TP_RAW_FMT("%u:%s:%s (%u)") | ||
| 135 | ); | ||
| 136 | |||
| 137 | TRACE_EVENT_FORMAT(hw_branch, TRACE_HW_BRANCHES, hw_branch_entry, ignore, | ||
| 138 | TRACE_STRUCT( | ||
| 139 | TRACE_FIELD(u64, from, from) | ||
| 140 | TRACE_FIELD(u64, to, to) | ||
| 141 | ), | ||
| 142 | TP_RAW_FMT("from: %llx to: %llx") | ||
| 143 | ); | ||
| 144 | |||
| 145 | TRACE_EVENT_FORMAT(power, TRACE_POWER, trace_power, ignore, | ||
| 146 | TRACE_STRUCT( | ||
| 147 | TRACE_FIELD_SIGN(ktime_t, state_data.stamp, stamp, 1) | ||
| 148 | TRACE_FIELD_SIGN(ktime_t, state_data.end, end, 1) | ||
| 149 | TRACE_FIELD(int, state_data.type, type) | ||
| 150 | TRACE_FIELD(int, state_data.state, state) | ||
| 151 | ), | ||
| 152 | TP_RAW_FMT("%llx->%llx type:%u state:%u") | ||
| 153 | ); | ||
| 154 | |||
| 155 | TRACE_EVENT_FORMAT(kmem_alloc, TRACE_KMEM_ALLOC, kmemtrace_alloc_entry, ignore, | ||
| 156 | TRACE_STRUCT( | ||
| 157 | TRACE_FIELD(enum kmemtrace_type_id, type_id, type_id) | ||
| 158 | TRACE_FIELD(unsigned long, call_site, call_site) | ||
| 159 | TRACE_FIELD(const void *, ptr, ptr) | ||
| 160 | TRACE_FIELD(size_t, bytes_req, bytes_req) | ||
| 161 | TRACE_FIELD(size_t, bytes_alloc, bytes_alloc) | ||
| 162 | TRACE_FIELD(gfp_t, gfp_flags, gfp_flags) | ||
| 163 | TRACE_FIELD(int, node, node) | ||
| 164 | ), | ||
| 165 | TP_RAW_FMT("type:%u call_site:%lx ptr:%p req:%lu alloc:%lu" | ||
| 166 | " flags:%x node:%d") | ||
| 167 | ); | ||
| 168 | |||
| 169 | TRACE_EVENT_FORMAT(kmem_free, TRACE_KMEM_FREE, kmemtrace_free_entry, ignore, | ||
| 170 | TRACE_STRUCT( | ||
| 171 | TRACE_FIELD(enum kmemtrace_type_id, type_id, type_id) | ||
| 172 | TRACE_FIELD(unsigned long, call_site, call_site) | ||
| 173 | TRACE_FIELD(const void *, ptr, ptr) | ||
| 174 | ), | ||
| 175 | TP_RAW_FMT("type:%u call_site:%lx ptr:%p") | ||
| 176 | ); | ||
| 177 | |||
| 178 | #undef TRACE_SYSTEM | ||
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index f85b0f1cb942..a4b7c9a9130c 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
| @@ -21,6 +21,7 @@ | |||
| 21 | 21 | ||
| 22 | #include "trace_output.h" | 22 | #include "trace_output.h" |
| 23 | 23 | ||
| 24 | #undef TRACE_SYSTEM | ||
| 24 | #define TRACE_SYSTEM "TRACE_SYSTEM" | 25 | #define TRACE_SYSTEM "TRACE_SYSTEM" |
| 25 | 26 | ||
| 26 | DEFINE_MUTEX(event_mutex); | 27 | DEFINE_MUTEX(event_mutex); |
| @@ -86,7 +87,7 @@ int trace_define_common_fields(struct ftrace_event_call *call) | |||
| 86 | __common_field(unsigned char, flags); | 87 | __common_field(unsigned char, flags); |
| 87 | __common_field(unsigned char, preempt_count); | 88 | __common_field(unsigned char, preempt_count); |
| 88 | __common_field(int, pid); | 89 | __common_field(int, pid); |
| 89 | __common_field(int, tgid); | 90 | __common_field(int, lock_depth); |
| 90 | 91 | ||
| 91 | return ret; | 92 | return ret; |
| 92 | } | 93 | } |
| @@ -226,11 +227,9 @@ static ssize_t | |||
| 226 | ftrace_event_write(struct file *file, const char __user *ubuf, | 227 | ftrace_event_write(struct file *file, const char __user *ubuf, |
| 227 | size_t cnt, loff_t *ppos) | 228 | size_t cnt, loff_t *ppos) |
| 228 | { | 229 | { |
| 230 | struct trace_parser parser; | ||
| 229 | size_t read = 0; | 231 | size_t read = 0; |
| 230 | int i, set = 1; | ||
| 231 | ssize_t ret; | 232 | ssize_t ret; |
| 232 | char *buf; | ||
| 233 | char ch; | ||
| 234 | 233 | ||
| 235 | if (!cnt || cnt < 0) | 234 | if (!cnt || cnt < 0) |
| 236 | return 0; | 235 | return 0; |
| @@ -239,60 +238,28 @@ ftrace_event_write(struct file *file, const char __user *ubuf, | |||
| 239 | if (ret < 0) | 238 | if (ret < 0) |
| 240 | return ret; | 239 | return ret; |
| 241 | 240 | ||
| 242 | ret = get_user(ch, ubuf++); | 241 | if (trace_parser_get_init(&parser, EVENT_BUF_SIZE + 1)) |
| 243 | if (ret) | ||
| 244 | return ret; | ||
| 245 | read++; | ||
| 246 | cnt--; | ||
| 247 | |||
| 248 | /* skip white space */ | ||
| 249 | while (cnt && isspace(ch)) { | ||
| 250 | ret = get_user(ch, ubuf++); | ||
| 251 | if (ret) | ||
| 252 | return ret; | ||
| 253 | read++; | ||
| 254 | cnt--; | ||
| 255 | } | ||
| 256 | |||
| 257 | /* Only white space found? */ | ||
| 258 | if (isspace(ch)) { | ||
| 259 | file->f_pos += read; | ||
| 260 | ret = read; | ||
| 261 | return ret; | ||
| 262 | } | ||
| 263 | |||
| 264 | buf = kmalloc(EVENT_BUF_SIZE+1, GFP_KERNEL); | ||
| 265 | if (!buf) | ||
| 266 | return -ENOMEM; | 242 | return -ENOMEM; |
| 267 | 243 | ||
| 268 | if (cnt > EVENT_BUF_SIZE) | 244 | read = trace_get_user(&parser, ubuf, cnt, ppos); |
| 269 | cnt = EVENT_BUF_SIZE; | 245 | |
| 246 | if (trace_parser_loaded((&parser))) { | ||
| 247 | int set = 1; | ||
| 270 | 248 | ||
| 271 | i = 0; | 249 | if (*parser.buffer == '!') |
| 272 | while (cnt && !isspace(ch)) { | ||
| 273 | if (!i && ch == '!') | ||
| 274 | set = 0; | 250 | set = 0; |
| 275 | else | ||
| 276 | buf[i++] = ch; | ||
| 277 | 251 | ||
| 278 | ret = get_user(ch, ubuf++); | 252 | parser.buffer[parser.idx] = 0; |
| 253 | |||
| 254 | ret = ftrace_set_clr_event(parser.buffer + !set, set); | ||
| 279 | if (ret) | 255 | if (ret) |
| 280 | goto out_free; | 256 | goto out_put; |
| 281 | read++; | ||
| 282 | cnt--; | ||
| 283 | } | 257 | } |
| 284 | buf[i] = 0; | ||
| 285 | |||
| 286 | file->f_pos += read; | ||
| 287 | |||
| 288 | ret = ftrace_set_clr_event(buf, set); | ||
| 289 | if (ret) | ||
| 290 | goto out_free; | ||
| 291 | 258 | ||
| 292 | ret = read; | 259 | ret = read; |
| 293 | 260 | ||
| 294 | out_free: | 261 | out_put: |
| 295 | kfree(buf); | 262 | trace_parser_put(&parser); |
| 296 | 263 | ||
| 297 | return ret; | 264 | return ret; |
| 298 | } | 265 | } |
| @@ -300,42 +267,32 @@ ftrace_event_write(struct file *file, const char __user *ubuf, | |||
| 300 | static void * | 267 | static void * |
| 301 | t_next(struct seq_file *m, void *v, loff_t *pos) | 268 | t_next(struct seq_file *m, void *v, loff_t *pos) |
| 302 | { | 269 | { |
| 303 | struct list_head *list = m->private; | 270 | struct ftrace_event_call *call = v; |
| 304 | struct ftrace_event_call *call; | ||
| 305 | 271 | ||
| 306 | (*pos)++; | 272 | (*pos)++; |
| 307 | 273 | ||
| 308 | for (;;) { | 274 | list_for_each_entry_continue(call, &ftrace_events, list) { |
| 309 | if (list == &ftrace_events) | ||
| 310 | return NULL; | ||
| 311 | |||
| 312 | call = list_entry(list, struct ftrace_event_call, list); | ||
| 313 | |||
| 314 | /* | 275 | /* |
| 315 | * The ftrace subsystem is for showing formats only. | 276 | * The ftrace subsystem is for showing formats only. |
| 316 | * They can not be enabled or disabled via the event files. | 277 | * They can not be enabled or disabled via the event files. |
| 317 | */ | 278 | */ |
| 318 | if (call->regfunc) | 279 | if (call->regfunc) |
| 319 | break; | 280 | return call; |
| 320 | |||
| 321 | list = list->next; | ||
| 322 | } | 281 | } |
| 323 | 282 | ||
| 324 | m->private = list->next; | 283 | return NULL; |
| 325 | |||
| 326 | return call; | ||
| 327 | } | 284 | } |
| 328 | 285 | ||
| 329 | static void *t_start(struct seq_file *m, loff_t *pos) | 286 | static void *t_start(struct seq_file *m, loff_t *pos) |
| 330 | { | 287 | { |
| 331 | struct ftrace_event_call *call = NULL; | 288 | struct ftrace_event_call *call; |
| 332 | loff_t l; | 289 | loff_t l; |
| 333 | 290 | ||
| 334 | mutex_lock(&event_mutex); | 291 | mutex_lock(&event_mutex); |
| 335 | 292 | ||
| 336 | m->private = ftrace_events.next; | 293 | call = list_entry(&ftrace_events, struct ftrace_event_call, list); |
| 337 | for (l = 0; l <= *pos; ) { | 294 | for (l = 0; l <= *pos; ) { |
| 338 | call = t_next(m, NULL, &l); | 295 | call = t_next(m, call, &l); |
| 339 | if (!call) | 296 | if (!call) |
| 340 | break; | 297 | break; |
| 341 | } | 298 | } |
| @@ -345,37 +302,28 @@ static void *t_start(struct seq_file *m, loff_t *pos) | |||
| 345 | static void * | 302 | static void * |
| 346 | s_next(struct seq_file *m, void *v, loff_t *pos) | 303 | s_next(struct seq_file *m, void *v, loff_t *pos) |
| 347 | { | 304 | { |
| 348 | struct list_head *list = m->private; | 305 | struct ftrace_event_call *call = v; |
| 349 | struct ftrace_event_call *call; | ||
| 350 | 306 | ||
| 351 | (*pos)++; | 307 | (*pos)++; |
| 352 | 308 | ||
| 353 | retry: | 309 | list_for_each_entry_continue(call, &ftrace_events, list) { |
| 354 | if (list == &ftrace_events) | 310 | if (call->enabled) |
| 355 | return NULL; | 311 | return call; |
| 356 | |||
| 357 | call = list_entry(list, struct ftrace_event_call, list); | ||
| 358 | |||
| 359 | if (!call->enabled) { | ||
| 360 | list = list->next; | ||
| 361 | goto retry; | ||
| 362 | } | 312 | } |
| 363 | 313 | ||
| 364 | m->private = list->next; | 314 | return NULL; |
| 365 | |||
| 366 | return call; | ||
| 367 | } | 315 | } |
| 368 | 316 | ||
| 369 | static void *s_start(struct seq_file *m, loff_t *pos) | 317 | static void *s_start(struct seq_file *m, loff_t *pos) |
| 370 | { | 318 | { |
| 371 | struct ftrace_event_call *call = NULL; | 319 | struct ftrace_event_call *call; |
| 372 | loff_t l; | 320 | loff_t l; |
| 373 | 321 | ||
| 374 | mutex_lock(&event_mutex); | 322 | mutex_lock(&event_mutex); |
| 375 | 323 | ||
| 376 | m->private = ftrace_events.next; | 324 | call = list_entry(&ftrace_events, struct ftrace_event_call, list); |
| 377 | for (l = 0; l <= *pos; ) { | 325 | for (l = 0; l <= *pos; ) { |
| 378 | call = s_next(m, NULL, &l); | 326 | call = s_next(m, call, &l); |
| 379 | if (!call) | 327 | if (!call) |
| 380 | break; | 328 | break; |
| 381 | } | 329 | } |
| @@ -574,7 +522,7 @@ static int trace_write_header(struct trace_seq *s) | |||
| 574 | FIELD(unsigned char, flags), | 522 | FIELD(unsigned char, flags), |
| 575 | FIELD(unsigned char, preempt_count), | 523 | FIELD(unsigned char, preempt_count), |
| 576 | FIELD(int, pid), | 524 | FIELD(int, pid), |
| 577 | FIELD(int, tgid)); | 525 | FIELD(int, lock_depth)); |
| 578 | } | 526 | } |
| 579 | 527 | ||
| 580 | static ssize_t | 528 | static ssize_t |
| @@ -1242,7 +1190,7 @@ static int trace_module_notify(struct notifier_block *self, | |||
| 1242 | } | 1190 | } |
| 1243 | #endif /* CONFIG_MODULES */ | 1191 | #endif /* CONFIG_MODULES */ |
| 1244 | 1192 | ||
| 1245 | struct notifier_block trace_module_nb = { | 1193 | static struct notifier_block trace_module_nb = { |
| 1246 | .notifier_call = trace_module_notify, | 1194 | .notifier_call = trace_module_notify, |
| 1247 | .priority = 0, | 1195 | .priority = 0, |
| 1248 | }; | 1196 | }; |
| @@ -1414,6 +1362,18 @@ static __init void event_trace_self_tests(void) | |||
| 1414 | if (!call->regfunc) | 1362 | if (!call->regfunc) |
| 1415 | continue; | 1363 | continue; |
| 1416 | 1364 | ||
| 1365 | /* | ||
| 1366 | * Testing syscall events here is pretty useless, but | ||
| 1367 | * we still do it if configured. But this is time consuming. | ||
| 1368 | * What we really need is a user thread to perform the | ||
| 1369 | * syscalls as we test. | ||
| 1370 | */ | ||
| 1371 | #ifndef CONFIG_EVENT_TRACE_TEST_SYSCALLS | ||
| 1372 | if (call->system && | ||
| 1373 | strcmp(call->system, "syscalls") == 0) | ||
| 1374 | continue; | ||
| 1375 | #endif | ||
| 1376 | |||
| 1417 | pr_info("Testing event %s: ", call->name); | 1377 | pr_info("Testing event %s: ", call->name); |
| 1418 | 1378 | ||
| 1419 | /* | 1379 | /* |
| @@ -1487,7 +1447,7 @@ static __init void event_trace_self_tests(void) | |||
| 1487 | 1447 | ||
| 1488 | #ifdef CONFIG_FUNCTION_TRACER | 1448 | #ifdef CONFIG_FUNCTION_TRACER |
| 1489 | 1449 | ||
| 1490 | static DEFINE_PER_CPU(atomic_t, test_event_disable); | 1450 | static DEFINE_PER_CPU(atomic_t, ftrace_test_event_disable); |
| 1491 | 1451 | ||
| 1492 | static void | 1452 | static void |
| 1493 | function_test_events_call(unsigned long ip, unsigned long parent_ip) | 1453 | function_test_events_call(unsigned long ip, unsigned long parent_ip) |
| @@ -1504,7 +1464,7 @@ function_test_events_call(unsigned long ip, unsigned long parent_ip) | |||
| 1504 | pc = preempt_count(); | 1464 | pc = preempt_count(); |
| 1505 | resched = ftrace_preempt_disable(); | 1465 | resched = ftrace_preempt_disable(); |
| 1506 | cpu = raw_smp_processor_id(); | 1466 | cpu = raw_smp_processor_id(); |
| 1507 | disabled = atomic_inc_return(&per_cpu(test_event_disable, cpu)); | 1467 | disabled = atomic_inc_return(&per_cpu(ftrace_test_event_disable, cpu)); |
| 1508 | 1468 | ||
| 1509 | if (disabled != 1) | 1469 | if (disabled != 1) |
| 1510 | goto out; | 1470 | goto out; |
| @@ -1523,7 +1483,7 @@ function_test_events_call(unsigned long ip, unsigned long parent_ip) | |||
| 1523 | trace_nowake_buffer_unlock_commit(buffer, event, flags, pc); | 1483 | trace_nowake_buffer_unlock_commit(buffer, event, flags, pc); |
| 1524 | 1484 | ||
| 1525 | out: | 1485 | out: |
| 1526 | atomic_dec(&per_cpu(test_event_disable, cpu)); | 1486 | atomic_dec(&per_cpu(ftrace_test_event_disable, cpu)); |
| 1527 | ftrace_preempt_enable(resched); | 1487 | ftrace_preempt_enable(resched); |
| 1528 | } | 1488 | } |
| 1529 | 1489 | ||
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 93660fbbf629..23245785927f 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
| @@ -121,6 +121,47 @@ struct filter_parse_state { | |||
| 121 | } operand; | 121 | } operand; |
| 122 | }; | 122 | }; |
| 123 | 123 | ||
| 124 | #define DEFINE_COMPARISON_PRED(type) \ | ||
| 125 | static int filter_pred_##type(struct filter_pred *pred, void *event, \ | ||
| 126 | int val1, int val2) \ | ||
| 127 | { \ | ||
| 128 | type *addr = (type *)(event + pred->offset); \ | ||
| 129 | type val = (type)pred->val; \ | ||
| 130 | int match = 0; \ | ||
| 131 | \ | ||
| 132 | switch (pred->op) { \ | ||
| 133 | case OP_LT: \ | ||
| 134 | match = (*addr < val); \ | ||
| 135 | break; \ | ||
| 136 | case OP_LE: \ | ||
| 137 | match = (*addr <= val); \ | ||
| 138 | break; \ | ||
| 139 | case OP_GT: \ | ||
| 140 | match = (*addr > val); \ | ||
| 141 | break; \ | ||
| 142 | case OP_GE: \ | ||
| 143 | match = (*addr >= val); \ | ||
| 144 | break; \ | ||
| 145 | default: \ | ||
| 146 | break; \ | ||
| 147 | } \ | ||
| 148 | \ | ||
| 149 | return match; \ | ||
| 150 | } | ||
| 151 | |||
| 152 | #define DEFINE_EQUALITY_PRED(size) \ | ||
| 153 | static int filter_pred_##size(struct filter_pred *pred, void *event, \ | ||
| 154 | int val1, int val2) \ | ||
| 155 | { \ | ||
| 156 | u##size *addr = (u##size *)(event + pred->offset); \ | ||
| 157 | u##size val = (u##size)pred->val; \ | ||
| 158 | int match; \ | ||
| 159 | \ | ||
| 160 | match = (val == *addr) ^ pred->not; \ | ||
| 161 | \ | ||
| 162 | return match; \ | ||
| 163 | } | ||
| 164 | |||
| 124 | DEFINE_COMPARISON_PRED(s64); | 165 | DEFINE_COMPARISON_PRED(s64); |
| 125 | DEFINE_COMPARISON_PRED(u64); | 166 | DEFINE_COMPARISON_PRED(u64); |
| 126 | DEFINE_COMPARISON_PRED(s32); | 167 | DEFINE_COMPARISON_PRED(s32); |
diff --git a/kernel/trace/trace_export.c b/kernel/trace/trace_export.c index a79ef6f193c0..ed7d48083520 100644 --- a/kernel/trace/trace_export.c +++ b/kernel/trace/trace_export.c | |||
| @@ -15,147 +15,124 @@ | |||
| 15 | 15 | ||
| 16 | #include "trace_output.h" | 16 | #include "trace_output.h" |
| 17 | 17 | ||
| 18 | #undef TRACE_SYSTEM | ||
| 19 | #define TRACE_SYSTEM ftrace | ||
| 18 | 20 | ||
| 19 | #undef TRACE_STRUCT | 21 | /* not needed for this file */ |
| 20 | #define TRACE_STRUCT(args...) args | 22 | #undef __field_struct |
| 23 | #define __field_struct(type, item) | ||
| 21 | 24 | ||
| 22 | extern void __bad_type_size(void); | 25 | #undef __field |
| 26 | #define __field(type, item) type item; | ||
| 23 | 27 | ||
| 24 | #undef TRACE_FIELD | 28 | #undef __field_desc |
| 25 | #define TRACE_FIELD(type, item, assign) \ | 29 | #define __field_desc(type, container, item) type item; |
| 26 | if (sizeof(type) != sizeof(field.item)) \ | 30 | |
| 27 | __bad_type_size(); \ | 31 | #undef __array |
| 32 | #define __array(type, item, size) type item[size]; | ||
| 33 | |||
| 34 | #undef __array_desc | ||
| 35 | #define __array_desc(type, container, item, size) type item[size]; | ||
| 36 | |||
| 37 | #undef __dynamic_array | ||
| 38 | #define __dynamic_array(type, item) type item[]; | ||
| 39 | |||
| 40 | #undef F_STRUCT | ||
| 41 | #define F_STRUCT(args...) args | ||
| 42 | |||
| 43 | #undef F_printk | ||
| 44 | #define F_printk(fmt, args...) fmt, args | ||
| 45 | |||
| 46 | #undef FTRACE_ENTRY | ||
| 47 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ | ||
| 48 | struct ____ftrace_##name { \ | ||
| 49 | tstruct \ | ||
| 50 | }; \ | ||
| 51 | static void __used ____ftrace_check_##name(void) \ | ||
| 52 | { \ | ||
| 53 | struct ____ftrace_##name *__entry = NULL; \ | ||
| 54 | \ | ||
| 55 | /* force cmpile-time check on F_printk() */ \ | ||
| 56 | printk(print); \ | ||
| 57 | } | ||
| 58 | |||
| 59 | #undef FTRACE_ENTRY_DUP | ||
| 60 | #define FTRACE_ENTRY_DUP(name, struct_name, id, tstruct, print) \ | ||
| 61 | FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print)) | ||
| 62 | |||
| 63 | #include "trace_entries.h" | ||
| 64 | |||
| 65 | |||
| 66 | #undef __field | ||
| 67 | #define __field(type, item) \ | ||
| 28 | ret = trace_seq_printf(s, "\tfield:" #type " " #item ";\t" \ | 68 | ret = trace_seq_printf(s, "\tfield:" #type " " #item ";\t" \ |
| 29 | "offset:%u;\tsize:%u;\n", \ | 69 | "offset:%zu;\tsize:%zu;\n", \ |
| 30 | (unsigned int)offsetof(typeof(field), item), \ | 70 | offsetof(typeof(field), item), \ |
| 31 | (unsigned int)sizeof(field.item)); \ | 71 | sizeof(field.item)); \ |
| 32 | if (!ret) \ | 72 | if (!ret) \ |
| 33 | return 0; | 73 | return 0; |
| 34 | 74 | ||
| 75 | #undef __field_desc | ||
| 76 | #define __field_desc(type, container, item) \ | ||
| 77 | ret = trace_seq_printf(s, "\tfield:" #type " " #item ";\t" \ | ||
| 78 | "offset:%zu;\tsize:%zu;\n", \ | ||
| 79 | offsetof(typeof(field), container.item), \ | ||
| 80 | sizeof(field.container.item)); \ | ||
| 81 | if (!ret) \ | ||
| 82 | return 0; | ||
| 35 | 83 | ||
| 36 | #undef TRACE_FIELD_SPECIAL | 84 | #undef __array |
| 37 | #define TRACE_FIELD_SPECIAL(type_item, item, len, cmd) \ | 85 | #define __array(type, item, len) \ |
| 38 | ret = trace_seq_printf(s, "\tfield special:" #type_item ";\t" \ | 86 | ret = trace_seq_printf(s, "\tfield:" #type " " #item "[" #len "];\t" \ |
| 39 | "offset:%u;\tsize:%u;\n", \ | 87 | "offset:%zu;\tsize:%zu;\n", \ |
| 40 | (unsigned int)offsetof(typeof(field), item), \ | 88 | offsetof(typeof(field), item), \ |
| 41 | (unsigned int)sizeof(field.item)); \ | 89 | sizeof(field.item)); \ |
| 42 | if (!ret) \ | 90 | if (!ret) \ |
| 43 | return 0; | 91 | return 0; |
| 44 | 92 | ||
| 45 | #undef TRACE_FIELD_ZERO | 93 | #undef __array_desc |
| 46 | #define TRACE_FIELD_ZERO(type, item) \ | 94 | #define __array_desc(type, container, item, len) \ |
| 47 | ret = trace_seq_printf(s, "\tfield:" #type " " #item ";\t" \ | 95 | ret = trace_seq_printf(s, "\tfield:" #type " " #item "[" #len "];\t" \ |
| 48 | "offset:%u;\tsize:0;\n", \ | 96 | "offset:%zu;\tsize:%zu;\n", \ |
| 49 | (unsigned int)offsetof(typeof(field), item)); \ | 97 | offsetof(typeof(field), container.item), \ |
| 98 | sizeof(field.container.item)); \ | ||
| 50 | if (!ret) \ | 99 | if (!ret) \ |
| 51 | return 0; | 100 | return 0; |
| 52 | 101 | ||
| 53 | #undef TRACE_FIELD_SIGN | 102 | #undef __dynamic_array |
| 54 | #define TRACE_FIELD_SIGN(type, item, assign, is_signed) \ | 103 | #define __dynamic_array(type, item) \ |
| 55 | TRACE_FIELD(type, item, assign) | 104 | ret = trace_seq_printf(s, "\tfield:" #type " " #item ";\t" \ |
| 105 | "offset:%zu;\tsize:0;\n", \ | ||
| 106 | offsetof(typeof(field), item)); \ | ||
| 107 | if (!ret) \ | ||
| 108 | return 0; | ||
| 56 | 109 | ||
| 57 | #undef TP_RAW_FMT | 110 | #undef F_printk |
| 58 | #define TP_RAW_FMT(args...) args | 111 | #define F_printk(fmt, args...) "%s, %s\n", #fmt, __stringify(args) |
| 59 | 112 | ||
| 60 | #undef TRACE_EVENT_FORMAT | 113 | #undef __entry |
| 61 | #define TRACE_EVENT_FORMAT(call, proto, args, fmt, tstruct, tpfmt) \ | 114 | #define __entry REC |
| 62 | static int \ | ||
| 63 | ftrace_format_##call(struct ftrace_event_call *unused, \ | ||
| 64 | struct trace_seq *s) \ | ||
| 65 | { \ | ||
| 66 | struct args field; \ | ||
| 67 | int ret; \ | ||
| 68 | \ | ||
| 69 | tstruct; \ | ||
| 70 | \ | ||
| 71 | trace_seq_printf(s, "\nprint fmt: \"%s\"\n", tpfmt); \ | ||
| 72 | \ | ||
| 73 | return ret; \ | ||
| 74 | } | ||
| 75 | 115 | ||
| 76 | #undef TRACE_EVENT_FORMAT_NOFILTER | 116 | #undef FTRACE_ENTRY |
| 77 | #define TRACE_EVENT_FORMAT_NOFILTER(call, proto, args, fmt, tstruct, \ | 117 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ |
| 78 | tpfmt) \ | ||
| 79 | static int \ | 118 | static int \ |
| 80 | ftrace_format_##call(struct ftrace_event_call *unused, \ | 119 | ftrace_format_##name(struct ftrace_event_call *unused, \ |
| 81 | struct trace_seq *s) \ | 120 | struct trace_seq *s) \ |
| 82 | { \ | 121 | { \ |
| 83 | struct args field; \ | 122 | struct struct_name field __attribute__((unused)); \ |
| 84 | int ret; \ | 123 | int ret = 0; \ |
| 85 | \ | 124 | \ |
| 86 | tstruct; \ | 125 | tstruct; \ |
| 87 | \ | 126 | \ |
| 88 | trace_seq_printf(s, "\nprint fmt: \"%s\"\n", tpfmt); \ | 127 | trace_seq_printf(s, "\nprint fmt: " print); \ |
| 89 | \ | 128 | \ |
| 90 | return ret; \ | 129 | return ret; \ |
| 91 | } | 130 | } |
| 92 | 131 | ||
| 93 | #include "trace_event_types.h" | 132 | #include "trace_entries.h" |
| 94 | |||
| 95 | #undef TRACE_FIELD | ||
| 96 | #define TRACE_FIELD(type, item, assign)\ | ||
| 97 | entry->item = assign; | ||
| 98 | |||
| 99 | #undef TRACE_FIELD | ||
| 100 | #define TRACE_FIELD(type, item, assign)\ | ||
| 101 | entry->item = assign; | ||
| 102 | |||
| 103 | #undef TRACE_FIELD_SIGN | ||
| 104 | #define TRACE_FIELD_SIGN(type, item, assign, is_signed) \ | ||
| 105 | TRACE_FIELD(type, item, assign) | ||
| 106 | |||
| 107 | #undef TRACE_FIELD_ZERO | ||
| 108 | #define TRACE_FIELD_ZERO(type, item) | ||
| 109 | |||
| 110 | #undef TP_CMD | ||
| 111 | #define TP_CMD(cmd...) cmd | ||
| 112 | |||
| 113 | #undef TRACE_ENTRY | ||
| 114 | #define TRACE_ENTRY entry | ||
| 115 | |||
| 116 | #undef TRACE_FIELD_SPECIAL | ||
| 117 | #define TRACE_FIELD_SPECIAL(type_item, item, len, cmd) \ | ||
| 118 | cmd; | ||
| 119 | |||
| 120 | static int ftrace_raw_init_event(struct ftrace_event_call *event_call) | ||
| 121 | { | ||
| 122 | INIT_LIST_HEAD(&event_call->fields); | ||
| 123 | |||
| 124 | return 0; | ||
| 125 | } | ||
| 126 | |||
| 127 | #undef TRACE_EVENT_FORMAT | ||
| 128 | #define TRACE_EVENT_FORMAT(call, proto, args, fmt, tstruct, tpfmt) \ | ||
| 129 | int ftrace_define_fields_##call(struct ftrace_event_call *event_call); \ | ||
| 130 | \ | ||
| 131 | struct ftrace_event_call __used \ | ||
| 132 | __attribute__((__aligned__(4))) \ | ||
| 133 | __attribute__((section("_ftrace_events"))) event_##call = { \ | ||
| 134 | .name = #call, \ | ||
| 135 | .id = proto, \ | ||
| 136 | .system = __stringify(TRACE_SYSTEM), \ | ||
| 137 | .raw_init = ftrace_raw_init_event, \ | ||
| 138 | .show_format = ftrace_format_##call, \ | ||
| 139 | .define_fields = ftrace_define_fields_##call, \ | ||
| 140 | }; \ | ||
| 141 | |||
| 142 | #undef TRACE_EVENT_FORMAT_NOFILTER | ||
| 143 | #define TRACE_EVENT_FORMAT_NOFILTER(call, proto, args, fmt, tstruct, \ | ||
| 144 | tpfmt) \ | ||
| 145 | \ | ||
| 146 | struct ftrace_event_call __used \ | ||
| 147 | __attribute__((__aligned__(4))) \ | ||
| 148 | __attribute__((section("_ftrace_events"))) event_##call = { \ | ||
| 149 | .name = #call, \ | ||
| 150 | .id = proto, \ | ||
| 151 | .system = __stringify(TRACE_SYSTEM), \ | ||
| 152 | .show_format = ftrace_format_##call, \ | ||
| 153 | }; | ||
| 154 | |||
| 155 | #include "trace_event_types.h" | ||
| 156 | 133 | ||
| 157 | #undef TRACE_FIELD | 134 | #undef __field |
| 158 | #define TRACE_FIELD(type, item, assign) \ | 135 | #define __field(type, item) \ |
| 159 | ret = trace_define_field(event_call, #type, #item, \ | 136 | ret = trace_define_field(event_call, #type, #item, \ |
| 160 | offsetof(typeof(field), item), \ | 137 | offsetof(typeof(field), item), \ |
| 161 | sizeof(field.item), \ | 138 | sizeof(field.item), \ |
| @@ -163,32 +140,45 @@ __attribute__((section("_ftrace_events"))) event_##call = { \ | |||
| 163 | if (ret) \ | 140 | if (ret) \ |
| 164 | return ret; | 141 | return ret; |
| 165 | 142 | ||
| 166 | #undef TRACE_FIELD_SPECIAL | 143 | #undef __field_desc |
| 167 | #define TRACE_FIELD_SPECIAL(type, item, len, cmd) \ | 144 | #define __field_desc(type, container, item) \ |
| 145 | ret = trace_define_field(event_call, #type, #item, \ | ||
| 146 | offsetof(typeof(field), \ | ||
| 147 | container.item), \ | ||
| 148 | sizeof(field.container.item), \ | ||
| 149 | is_signed_type(type), FILTER_OTHER); \ | ||
| 150 | if (ret) \ | ||
| 151 | return ret; | ||
| 152 | |||
| 153 | #undef __array | ||
| 154 | #define __array(type, item, len) \ | ||
| 155 | BUILD_BUG_ON(len > MAX_FILTER_STR_VAL); \ | ||
| 168 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ | 156 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ |
| 169 | offsetof(typeof(field), item), \ | 157 | offsetof(typeof(field), item), \ |
| 170 | sizeof(field.item), 0, FILTER_OTHER); \ | 158 | sizeof(field.item), 0, FILTER_OTHER); \ |
| 171 | if (ret) \ | 159 | if (ret) \ |
| 172 | return ret; | 160 | return ret; |
| 173 | 161 | ||
| 174 | #undef TRACE_FIELD_SIGN | 162 | #undef __array_desc |
| 175 | #define TRACE_FIELD_SIGN(type, item, assign, is_signed) \ | 163 | #define __array_desc(type, container, item, len) \ |
| 176 | ret = trace_define_field(event_call, #type, #item, \ | 164 | BUILD_BUG_ON(len > MAX_FILTER_STR_VAL); \ |
| 177 | offsetof(typeof(field), item), \ | 165 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ |
| 178 | sizeof(field.item), is_signed, \ | 166 | offsetof(typeof(field), \ |
| 167 | container.item), \ | ||
| 168 | sizeof(field.container.item), 0, \ | ||
| 179 | FILTER_OTHER); \ | 169 | FILTER_OTHER); \ |
| 180 | if (ret) \ | 170 | if (ret) \ |
| 181 | return ret; | 171 | return ret; |
| 182 | 172 | ||
| 183 | #undef TRACE_FIELD_ZERO | 173 | #undef __dynamic_array |
| 184 | #define TRACE_FIELD_ZERO(type, item) | 174 | #define __dynamic_array(type, item) |
| 185 | 175 | ||
| 186 | #undef TRACE_EVENT_FORMAT | 176 | #undef FTRACE_ENTRY |
| 187 | #define TRACE_EVENT_FORMAT(call, proto, args, fmt, tstruct, tpfmt) \ | 177 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ |
| 188 | int \ | 178 | int \ |
| 189 | ftrace_define_fields_##call(struct ftrace_event_call *event_call) \ | 179 | ftrace_define_fields_##name(struct ftrace_event_call *event_call) \ |
| 190 | { \ | 180 | { \ |
| 191 | struct args field; \ | 181 | struct struct_name field; \ |
| 192 | int ret; \ | 182 | int ret; \ |
| 193 | \ | 183 | \ |
| 194 | ret = trace_define_common_fields(event_call); \ | 184 | ret = trace_define_common_fields(event_call); \ |
| @@ -200,8 +190,41 @@ ftrace_define_fields_##call(struct ftrace_event_call *event_call) \ | |||
| 200 | return ret; \ | 190 | return ret; \ |
| 201 | } | 191 | } |
| 202 | 192 | ||
| 203 | #undef TRACE_EVENT_FORMAT_NOFILTER | 193 | #include "trace_entries.h" |
| 204 | #define TRACE_EVENT_FORMAT_NOFILTER(call, proto, args, fmt, tstruct, \ | 194 | |
| 205 | tpfmt) | 195 | static int ftrace_raw_init_event(struct ftrace_event_call *call) |
| 196 | { | ||
| 197 | INIT_LIST_HEAD(&call->fields); | ||
| 198 | return 0; | ||
| 199 | } | ||
| 200 | |||
| 201 | #undef __field | ||
| 202 | #define __field(type, item) | ||
| 203 | |||
| 204 | #undef __field_desc | ||
| 205 | #define __field_desc(type, container, item) | ||
| 206 | |||
| 207 | #undef __array | ||
| 208 | #define __array(type, item, len) | ||
| 209 | |||
| 210 | #undef __array_desc | ||
| 211 | #define __array_desc(type, container, item, len) | ||
| 212 | |||
| 213 | #undef __dynamic_array | ||
| 214 | #define __dynamic_array(type, item) | ||
| 215 | |||
| 216 | #undef FTRACE_ENTRY | ||
| 217 | #define FTRACE_ENTRY(call, struct_name, type, tstruct, print) \ | ||
| 218 | \ | ||
| 219 | struct ftrace_event_call __used \ | ||
| 220 | __attribute__((__aligned__(4))) \ | ||
| 221 | __attribute__((section("_ftrace_events"))) event_##call = { \ | ||
| 222 | .name = #call, \ | ||
| 223 | .id = type, \ | ||
| 224 | .system = __stringify(TRACE_SYSTEM), \ | ||
| 225 | .raw_init = ftrace_raw_init_event, \ | ||
| 226 | .show_format = ftrace_format_##call, \ | ||
| 227 | .define_fields = ftrace_define_fields_##call, \ | ||
| 228 | }; \ | ||
| 206 | 229 | ||
| 207 | #include "trace_event_types.h" | 230 | #include "trace_entries.h" |
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c index 5b01b94518fc..b3f3776b0cd6 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c | |||
| @@ -290,7 +290,7 @@ ftrace_trace_onoff_print(struct seq_file *m, unsigned long ip, | |||
| 290 | { | 290 | { |
| 291 | long count = (long)data; | 291 | long count = (long)data; |
| 292 | 292 | ||
| 293 | seq_printf(m, "%pf:", (void *)ip); | 293 | seq_printf(m, "%ps:", (void *)ip); |
| 294 | 294 | ||
| 295 | if (ops == &traceon_probe_ops) | 295 | if (ops == &traceon_probe_ops) |
| 296 | seq_printf(m, "traceon"); | 296 | seq_printf(m, "traceon"); |
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index b3749a2c3132..45e6c01b2e4d 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
| @@ -124,7 +124,7 @@ ftrace_pop_return_trace(struct ftrace_graph_ret *trace, unsigned long *ret, | |||
| 124 | if (unlikely(current->ret_stack[index].fp != frame_pointer)) { | 124 | if (unlikely(current->ret_stack[index].fp != frame_pointer)) { |
| 125 | ftrace_graph_stop(); | 125 | ftrace_graph_stop(); |
| 126 | WARN(1, "Bad frame pointer: expected %lx, received %lx\n" | 126 | WARN(1, "Bad frame pointer: expected %lx, received %lx\n" |
| 127 | " from func %pF return to %lx\n", | 127 | " from func %ps return to %lx\n", |
| 128 | current->ret_stack[index].fp, | 128 | current->ret_stack[index].fp, |
| 129 | frame_pointer, | 129 | frame_pointer, |
| 130 | (void *)current->ret_stack[index].func, | 130 | (void *)current->ret_stack[index].func, |
| @@ -364,6 +364,15 @@ print_graph_proc(struct trace_seq *s, pid_t pid) | |||
| 364 | } | 364 | } |
| 365 | 365 | ||
| 366 | 366 | ||
| 367 | static enum print_line_t | ||
| 368 | print_graph_lat_fmt(struct trace_seq *s, struct trace_entry *entry) | ||
| 369 | { | ||
| 370 | if (!trace_seq_putc(s, ' ')) | ||
| 371 | return 0; | ||
| 372 | |||
| 373 | return trace_print_lat_fmt(s, entry); | ||
| 374 | } | ||
| 375 | |||
| 367 | /* If the pid changed since the last trace, output this event */ | 376 | /* If the pid changed since the last trace, output this event */ |
| 368 | static enum print_line_t | 377 | static enum print_line_t |
| 369 | verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) | 378 | verif_pid(struct trace_seq *s, pid_t pid, int cpu, struct fgraph_data *data) |
| @@ -521,6 +530,7 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
| 521 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 530 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 522 | return TRACE_TYPE_PARTIAL_LINE; | 531 | return TRACE_TYPE_PARTIAL_LINE; |
| 523 | } | 532 | } |
| 533 | |||
| 524 | /* Proc */ | 534 | /* Proc */ |
| 525 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) { | 535 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) { |
| 526 | ret = print_graph_proc(s, pid); | 536 | ret = print_graph_proc(s, pid); |
| @@ -659,7 +669,7 @@ print_graph_entry_leaf(struct trace_iterator *iter, | |||
| 659 | return TRACE_TYPE_PARTIAL_LINE; | 669 | return TRACE_TYPE_PARTIAL_LINE; |
| 660 | } | 670 | } |
| 661 | 671 | ||
| 662 | ret = trace_seq_printf(s, "%pf();\n", (void *)call->func); | 672 | ret = trace_seq_printf(s, "%ps();\n", (void *)call->func); |
| 663 | if (!ret) | 673 | if (!ret) |
| 664 | return TRACE_TYPE_PARTIAL_LINE; | 674 | return TRACE_TYPE_PARTIAL_LINE; |
| 665 | 675 | ||
| @@ -702,7 +712,7 @@ print_graph_entry_nested(struct trace_iterator *iter, | |||
| 702 | return TRACE_TYPE_PARTIAL_LINE; | 712 | return TRACE_TYPE_PARTIAL_LINE; |
| 703 | } | 713 | } |
| 704 | 714 | ||
| 705 | ret = trace_seq_printf(s, "%pf() {\n", (void *)call->func); | 715 | ret = trace_seq_printf(s, "%ps() {\n", (void *)call->func); |
| 706 | if (!ret) | 716 | if (!ret) |
| 707 | return TRACE_TYPE_PARTIAL_LINE; | 717 | return TRACE_TYPE_PARTIAL_LINE; |
| 708 | 718 | ||
| @@ -758,6 +768,13 @@ print_graph_prologue(struct trace_iterator *iter, struct trace_seq *s, | |||
| 758 | return TRACE_TYPE_PARTIAL_LINE; | 768 | return TRACE_TYPE_PARTIAL_LINE; |
| 759 | } | 769 | } |
| 760 | 770 | ||
| 771 | /* Latency format */ | ||
| 772 | if (trace_flags & TRACE_ITER_LATENCY_FMT) { | ||
| 773 | ret = print_graph_lat_fmt(s, ent); | ||
| 774 | if (ret == TRACE_TYPE_PARTIAL_LINE) | ||
| 775 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 776 | } | ||
| 777 | |||
| 761 | return 0; | 778 | return 0; |
| 762 | } | 779 | } |
| 763 | 780 | ||
| @@ -952,28 +969,59 @@ print_graph_function(struct trace_iterator *iter) | |||
| 952 | return TRACE_TYPE_HANDLED; | 969 | return TRACE_TYPE_HANDLED; |
| 953 | } | 970 | } |
| 954 | 971 | ||
| 972 | static void print_lat_header(struct seq_file *s) | ||
| 973 | { | ||
| 974 | static const char spaces[] = " " /* 16 spaces */ | ||
| 975 | " " /* 4 spaces */ | ||
| 976 | " "; /* 17 spaces */ | ||
| 977 | int size = 0; | ||
| 978 | |||
| 979 | if (tracer_flags.val & TRACE_GRAPH_PRINT_ABS_TIME) | ||
| 980 | size += 16; | ||
| 981 | if (tracer_flags.val & TRACE_GRAPH_PRINT_CPU) | ||
| 982 | size += 4; | ||
| 983 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) | ||
| 984 | size += 17; | ||
| 985 | |||
| 986 | seq_printf(s, "#%.*s _-----=> irqs-off \n", size, spaces); | ||
| 987 | seq_printf(s, "#%.*s / _----=> need-resched \n", size, spaces); | ||
| 988 | seq_printf(s, "#%.*s| / _---=> hardirq/softirq \n", size, spaces); | ||
| 989 | seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); | ||
| 990 | seq_printf(s, "#%.*s||| / _-=> lock-depth \n", size, spaces); | ||
| 991 | seq_printf(s, "#%.*s|||| / \n", size, spaces); | ||
| 992 | } | ||
| 993 | |||
| 955 | static void print_graph_headers(struct seq_file *s) | 994 | static void print_graph_headers(struct seq_file *s) |
| 956 | { | 995 | { |
| 996 | int lat = trace_flags & TRACE_ITER_LATENCY_FMT; | ||
| 997 | |||
| 998 | if (lat) | ||
| 999 | print_lat_header(s); | ||
| 1000 | |||
| 957 | /* 1st line */ | 1001 | /* 1st line */ |
| 958 | seq_printf(s, "# "); | 1002 | seq_printf(s, "#"); |
| 959 | if (tracer_flags.val & TRACE_GRAPH_PRINT_ABS_TIME) | 1003 | if (tracer_flags.val & TRACE_GRAPH_PRINT_ABS_TIME) |
| 960 | seq_printf(s, " TIME "); | 1004 | seq_printf(s, " TIME "); |
| 961 | if (tracer_flags.val & TRACE_GRAPH_PRINT_CPU) | 1005 | if (tracer_flags.val & TRACE_GRAPH_PRINT_CPU) |
| 962 | seq_printf(s, "CPU"); | 1006 | seq_printf(s, " CPU"); |
| 963 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) | 1007 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) |
| 964 | seq_printf(s, " TASK/PID "); | 1008 | seq_printf(s, " TASK/PID "); |
| 1009 | if (lat) | ||
| 1010 | seq_printf(s, "|||||"); | ||
| 965 | if (tracer_flags.val & TRACE_GRAPH_PRINT_DURATION) | 1011 | if (tracer_flags.val & TRACE_GRAPH_PRINT_DURATION) |
| 966 | seq_printf(s, " DURATION "); | 1012 | seq_printf(s, " DURATION "); |
| 967 | seq_printf(s, " FUNCTION CALLS\n"); | 1013 | seq_printf(s, " FUNCTION CALLS\n"); |
| 968 | 1014 | ||
| 969 | /* 2nd line */ | 1015 | /* 2nd line */ |
| 970 | seq_printf(s, "# "); | 1016 | seq_printf(s, "#"); |
| 971 | if (tracer_flags.val & TRACE_GRAPH_PRINT_ABS_TIME) | 1017 | if (tracer_flags.val & TRACE_GRAPH_PRINT_ABS_TIME) |
| 972 | seq_printf(s, " | "); | 1018 | seq_printf(s, " | "); |
| 973 | if (tracer_flags.val & TRACE_GRAPH_PRINT_CPU) | 1019 | if (tracer_flags.val & TRACE_GRAPH_PRINT_CPU) |
| 974 | seq_printf(s, "| "); | 1020 | seq_printf(s, " | "); |
| 975 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) | 1021 | if (tracer_flags.val & TRACE_GRAPH_PRINT_PROC) |
| 976 | seq_printf(s, " | | "); | 1022 | seq_printf(s, " | | "); |
| 1023 | if (lat) | ||
| 1024 | seq_printf(s, "|||||"); | ||
| 977 | if (tracer_flags.val & TRACE_GRAPH_PRINT_DURATION) | 1025 | if (tracer_flags.val & TRACE_GRAPH_PRINT_DURATION) |
| 978 | seq_printf(s, " | | "); | 1026 | seq_printf(s, " | | "); |
| 979 | seq_printf(s, " | | | |\n"); | 1027 | seq_printf(s, " | | | |\n"); |
diff --git a/kernel/trace/trace_hw_branches.c b/kernel/trace/trace_hw_branches.c index ca7d7c4d0c2a..23b63859130e 100644 --- a/kernel/trace/trace_hw_branches.c +++ b/kernel/trace/trace_hw_branches.c | |||
| @@ -155,7 +155,7 @@ static enum print_line_t bts_trace_print_line(struct trace_iterator *iter) | |||
| 155 | seq_print_ip_sym(seq, it->from, symflags) && | 155 | seq_print_ip_sym(seq, it->from, symflags) && |
| 156 | trace_seq_printf(seq, "\n")) | 156 | trace_seq_printf(seq, "\n")) |
| 157 | return TRACE_TYPE_HANDLED; | 157 | return TRACE_TYPE_HANDLED; |
| 158 | return TRACE_TYPE_PARTIAL_LINE;; | 158 | return TRACE_TYPE_PARTIAL_LINE; |
| 159 | } | 159 | } |
| 160 | return TRACE_TYPE_UNHANDLED; | 160 | return TRACE_TYPE_UNHANDLED; |
| 161 | } | 161 | } |
diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c index 5555b75a0d12..3aa7eaa2114c 100644 --- a/kernel/trace/trace_irqsoff.c +++ b/kernel/trace/trace_irqsoff.c | |||
| @@ -129,15 +129,10 @@ check_critical_timing(struct trace_array *tr, | |||
| 129 | unsigned long parent_ip, | 129 | unsigned long parent_ip, |
| 130 | int cpu) | 130 | int cpu) |
| 131 | { | 131 | { |
| 132 | unsigned long latency, t0, t1; | ||
| 133 | cycle_t T0, T1, delta; | 132 | cycle_t T0, T1, delta; |
| 134 | unsigned long flags; | 133 | unsigned long flags; |
| 135 | int pc; | 134 | int pc; |
| 136 | 135 | ||
| 137 | /* | ||
| 138 | * usecs conversion is slow so we try to delay the conversion | ||
| 139 | * as long as possible: | ||
| 140 | */ | ||
| 141 | T0 = data->preempt_timestamp; | 136 | T0 = data->preempt_timestamp; |
| 142 | T1 = ftrace_now(cpu); | 137 | T1 = ftrace_now(cpu); |
| 143 | delta = T1-T0; | 138 | delta = T1-T0; |
| @@ -157,18 +152,15 @@ check_critical_timing(struct trace_array *tr, | |||
| 157 | 152 | ||
| 158 | trace_function(tr, CALLER_ADDR0, parent_ip, flags, pc); | 153 | trace_function(tr, CALLER_ADDR0, parent_ip, flags, pc); |
| 159 | 154 | ||
| 160 | latency = nsecs_to_usecs(delta); | ||
| 161 | |||
| 162 | if (data->critical_sequence != max_sequence) | 155 | if (data->critical_sequence != max_sequence) |
| 163 | goto out_unlock; | 156 | goto out_unlock; |
| 164 | 157 | ||
| 165 | tracing_max_latency = delta; | ||
| 166 | t0 = nsecs_to_usecs(T0); | ||
| 167 | t1 = nsecs_to_usecs(T1); | ||
| 168 | |||
| 169 | data->critical_end = parent_ip; | 158 | data->critical_end = parent_ip; |
| 170 | 159 | ||
| 171 | update_max_tr_single(tr, current, cpu); | 160 | if (likely(!is_tracing_stopped())) { |
| 161 | tracing_max_latency = delta; | ||
| 162 | update_max_tr_single(tr, current, cpu); | ||
| 163 | } | ||
| 172 | 164 | ||
| 173 | max_sequence++; | 165 | max_sequence++; |
| 174 | 166 | ||
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index f6821f16227e..09cba270392d 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
| @@ -28,7 +28,7 @@ | |||
| 28 | #include <linux/string.h> | 28 | #include <linux/string.h> |
| 29 | #include <linux/ctype.h> | 29 | #include <linux/ctype.h> |
| 30 | #include <linux/ptrace.h> | 30 | #include <linux/ptrace.h> |
| 31 | #include <linux/perf_counter.h> | 31 | #include <linux/perf_event.h> |
| 32 | 32 | ||
| 33 | #include "trace.h" | 33 | #include "trace.h" |
| 34 | #include "trace_output.h" | 34 | #include "trace_output.h" |
| @@ -1176,7 +1176,7 @@ static __kprobes int kprobe_profile_func(struct kprobe *kp, | |||
| 1176 | entry->ip = (unsigned long)kp->addr; | 1176 | entry->ip = (unsigned long)kp->addr; |
| 1177 | for (i = 0; i < tp->nr_args; i++) | 1177 | for (i = 0; i < tp->nr_args; i++) |
| 1178 | entry->args[i] = call_fetch(&tp->args[i].fetch, regs); | 1178 | entry->args[i] = call_fetch(&tp->args[i].fetch, regs); |
| 1179 | perf_tpcounter_event(call->id, entry->ip, 1, entry, size); | 1179 | perf_tp_event(call->id, entry->ip, 1, entry, size); |
| 1180 | } while (0); | 1180 | } while (0); |
| 1181 | return 0; | 1181 | return 0; |
| 1182 | } | 1182 | } |
| @@ -1213,7 +1213,7 @@ static __kprobes int kretprobe_profile_func(struct kretprobe_instance *ri, | |||
| 1213 | entry->ret_ip = (unsigned long)ri->ret_addr; | 1213 | entry->ret_ip = (unsigned long)ri->ret_addr; |
| 1214 | for (i = 0; i < tp->nr_args; i++) | 1214 | for (i = 0; i < tp->nr_args; i++) |
| 1215 | entry->args[i] = call_fetch(&tp->args[i].fetch, regs); | 1215 | entry->args[i] = call_fetch(&tp->args[i].fetch, regs); |
| 1216 | perf_tpcounter_event(call->id, entry->ret_ip, 1, entry, size); | 1216 | perf_tp_event(call->id, entry->ret_ip, 1, entry, size); |
| 1217 | } while (0); | 1217 | } while (0); |
| 1218 | return 0; | 1218 | return 0; |
| 1219 | } | 1219 | } |
| @@ -1222,10 +1222,8 @@ static int probe_profile_enable(struct ftrace_event_call *call) | |||
| 1222 | { | 1222 | { |
| 1223 | struct trace_probe *tp = (struct trace_probe *)call->data; | 1223 | struct trace_probe *tp = (struct trace_probe *)call->data; |
| 1224 | 1224 | ||
| 1225 | if (atomic_inc_return(&call->profile_count)) | ||
| 1226 | return 0; | ||
| 1227 | |||
| 1228 | tp->flags |= TP_FLAG_PROFILE; | 1225 | tp->flags |= TP_FLAG_PROFILE; |
| 1226 | |||
| 1229 | if (probe_is_return(tp)) | 1227 | if (probe_is_return(tp)) |
| 1230 | return enable_kretprobe(&tp->rp); | 1228 | return enable_kretprobe(&tp->rp); |
| 1231 | else | 1229 | else |
| @@ -1236,10 +1234,9 @@ static void probe_profile_disable(struct ftrace_event_call *call) | |||
| 1236 | { | 1234 | { |
| 1237 | struct trace_probe *tp = (struct trace_probe *)call->data; | 1235 | struct trace_probe *tp = (struct trace_probe *)call->data; |
| 1238 | 1236 | ||
| 1239 | if (atomic_add_negative(-1, &call->profile_count)) | 1237 | tp->flags &= ~TP_FLAG_PROFILE; |
| 1240 | tp->flags &= ~TP_FLAG_PROFILE; | ||
| 1241 | 1238 | ||
| 1242 | if (!(tp->flags & (TP_FLAG_TRACE | TP_FLAG_PROFILE))) { | 1239 | if (!(tp->flags & TP_FLAG_TRACE)) { |
| 1243 | if (probe_is_return(tp)) | 1240 | if (probe_is_return(tp)) |
| 1244 | disable_kretprobe(&tp->rp); | 1241 | disable_kretprobe(&tp->rp); |
| 1245 | else | 1242 | else |
diff --git a/kernel/trace/trace_mmiotrace.c b/kernel/trace/trace_mmiotrace.c index c4c9bbda53d3..0acd834659ed 100644 --- a/kernel/trace/trace_mmiotrace.c +++ b/kernel/trace/trace_mmiotrace.c | |||
| @@ -307,6 +307,7 @@ static void __trace_mmiotrace_rw(struct trace_array *tr, | |||
| 307 | struct trace_array_cpu *data, | 307 | struct trace_array_cpu *data, |
| 308 | struct mmiotrace_rw *rw) | 308 | struct mmiotrace_rw *rw) |
| 309 | { | 309 | { |
| 310 | struct ftrace_event_call *call = &event_mmiotrace_rw; | ||
| 310 | struct ring_buffer *buffer = tr->buffer; | 311 | struct ring_buffer *buffer = tr->buffer; |
| 311 | struct ring_buffer_event *event; | 312 | struct ring_buffer_event *event; |
| 312 | struct trace_mmiotrace_rw *entry; | 313 | struct trace_mmiotrace_rw *entry; |
| @@ -320,7 +321,9 @@ static void __trace_mmiotrace_rw(struct trace_array *tr, | |||
| 320 | } | 321 | } |
| 321 | entry = ring_buffer_event_data(event); | 322 | entry = ring_buffer_event_data(event); |
| 322 | entry->rw = *rw; | 323 | entry->rw = *rw; |
| 323 | trace_buffer_unlock_commit(buffer, event, 0, pc); | 324 | |
| 325 | if (!filter_check_discard(call, entry, buffer, event)) | ||
| 326 | trace_buffer_unlock_commit(buffer, event, 0, pc); | ||
| 324 | } | 327 | } |
| 325 | 328 | ||
| 326 | void mmio_trace_rw(struct mmiotrace_rw *rw) | 329 | void mmio_trace_rw(struct mmiotrace_rw *rw) |
| @@ -334,6 +337,7 @@ static void __trace_mmiotrace_map(struct trace_array *tr, | |||
| 334 | struct trace_array_cpu *data, | 337 | struct trace_array_cpu *data, |
| 335 | struct mmiotrace_map *map) | 338 | struct mmiotrace_map *map) |
| 336 | { | 339 | { |
| 340 | struct ftrace_event_call *call = &event_mmiotrace_map; | ||
| 337 | struct ring_buffer *buffer = tr->buffer; | 341 | struct ring_buffer *buffer = tr->buffer; |
| 338 | struct ring_buffer_event *event; | 342 | struct ring_buffer_event *event; |
| 339 | struct trace_mmiotrace_map *entry; | 343 | struct trace_mmiotrace_map *entry; |
| @@ -347,7 +351,9 @@ static void __trace_mmiotrace_map(struct trace_array *tr, | |||
| 347 | } | 351 | } |
| 348 | entry = ring_buffer_event_data(event); | 352 | entry = ring_buffer_event_data(event); |
| 349 | entry->map = *map; | 353 | entry->map = *map; |
| 350 | trace_buffer_unlock_commit(buffer, event, 0, pc); | 354 | |
| 355 | if (!filter_check_discard(call, entry, buffer, event)) | ||
| 356 | trace_buffer_unlock_commit(buffer, event, 0, pc); | ||
| 351 | } | 357 | } |
| 352 | 358 | ||
| 353 | void mmio_trace_mapping(struct mmiotrace_map *map) | 359 | void mmio_trace_mapping(struct mmiotrace_map *map) |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index e0c2545622e8..f572f44c6e1e 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
| @@ -407,7 +407,7 @@ seq_print_userip_objs(const struct userstack_entry *entry, struct trace_seq *s, | |||
| 407 | * since individual threads might have already quit! | 407 | * since individual threads might have already quit! |
| 408 | */ | 408 | */ |
| 409 | rcu_read_lock(); | 409 | rcu_read_lock(); |
| 410 | task = find_task_by_vpid(entry->ent.tgid); | 410 | task = find_task_by_vpid(entry->tgid); |
| 411 | if (task) | 411 | if (task) |
| 412 | mm = get_task_mm(task); | 412 | mm = get_task_mm(task); |
| 413 | rcu_read_unlock(); | 413 | rcu_read_unlock(); |
| @@ -460,18 +460,23 @@ seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) | |||
| 460 | return ret; | 460 | return ret; |
| 461 | } | 461 | } |
| 462 | 462 | ||
| 463 | static int | 463 | /** |
| 464 | lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) | 464 | * trace_print_lat_fmt - print the irq, preempt and lockdep fields |
| 465 | * @s: trace seq struct to write to | ||
| 466 | * @entry: The trace entry field from the ring buffer | ||
| 467 | * | ||
| 468 | * Prints the generic fields of irqs off, in hard or softirq, preempt | ||
| 469 | * count and lock depth. | ||
| 470 | */ | ||
| 471 | int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry) | ||
| 465 | { | 472 | { |
| 466 | int hardirq, softirq; | 473 | int hardirq, softirq; |
| 467 | char comm[TASK_COMM_LEN]; | 474 | int ret; |
| 468 | 475 | ||
| 469 | trace_find_cmdline(entry->pid, comm); | ||
| 470 | hardirq = entry->flags & TRACE_FLAG_HARDIRQ; | 476 | hardirq = entry->flags & TRACE_FLAG_HARDIRQ; |
| 471 | softirq = entry->flags & TRACE_FLAG_SOFTIRQ; | 477 | softirq = entry->flags & TRACE_FLAG_SOFTIRQ; |
| 472 | 478 | ||
| 473 | if (!trace_seq_printf(s, "%8.8s-%-5d %3d%c%c%c", | 479 | if (!trace_seq_printf(s, "%c%c%c", |
| 474 | comm, entry->pid, cpu, | ||
| 475 | (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' : | 480 | (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' : |
| 476 | (entry->flags & TRACE_FLAG_IRQS_NOSUPPORT) ? | 481 | (entry->flags & TRACE_FLAG_IRQS_NOSUPPORT) ? |
| 477 | 'X' : '.', | 482 | 'X' : '.', |
| @@ -481,9 +486,30 @@ lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) | |||
| 481 | hardirq ? 'h' : softirq ? 's' : '.')) | 486 | hardirq ? 'h' : softirq ? 's' : '.')) |
| 482 | return 0; | 487 | return 0; |
| 483 | 488 | ||
| 489 | if (entry->lock_depth < 0) | ||
| 490 | ret = trace_seq_putc(s, '.'); | ||
| 491 | else | ||
| 492 | ret = trace_seq_printf(s, "%d", entry->lock_depth); | ||
| 493 | if (!ret) | ||
| 494 | return 0; | ||
| 495 | |||
| 484 | if (entry->preempt_count) | 496 | if (entry->preempt_count) |
| 485 | return trace_seq_printf(s, "%x", entry->preempt_count); | 497 | return trace_seq_printf(s, "%x", entry->preempt_count); |
| 486 | return trace_seq_puts(s, "."); | 498 | return trace_seq_putc(s, '.'); |
| 499 | } | ||
| 500 | |||
| 501 | static int | ||
| 502 | lat_print_generic(struct trace_seq *s, struct trace_entry *entry, int cpu) | ||
| 503 | { | ||
| 504 | char comm[TASK_COMM_LEN]; | ||
| 505 | |||
| 506 | trace_find_cmdline(entry->pid, comm); | ||
| 507 | |||
| 508 | if (!trace_seq_printf(s, "%8.8s-%-5d %3d", | ||
| 509 | comm, entry->pid, cpu)) | ||
| 510 | return 0; | ||
| 511 | |||
| 512 | return trace_print_lat_fmt(s, entry); | ||
| 487 | } | 513 | } |
| 488 | 514 | ||
| 489 | static unsigned long preempt_mark_thresh = 100; | 515 | static unsigned long preempt_mark_thresh = 100; |
diff --git a/kernel/trace/trace_output.h b/kernel/trace/trace_output.h index d38bec4a9c30..9d91c72ba38b 100644 --- a/kernel/trace/trace_output.h +++ b/kernel/trace/trace_output.h | |||
| @@ -26,6 +26,8 @@ extern struct trace_event *ftrace_find_event(int type); | |||
| 26 | 26 | ||
| 27 | extern enum print_line_t trace_nop_print(struct trace_iterator *iter, | 27 | extern enum print_line_t trace_nop_print(struct trace_iterator *iter, |
| 28 | int flags); | 28 | int flags); |
| 29 | extern int | ||
| 30 | trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry); | ||
| 29 | 31 | ||
| 30 | /* used by module unregistering */ | 32 | /* used by module unregistering */ |
| 31 | extern int __unregister_ftrace_event(struct trace_event *event); | 33 | extern int __unregister_ftrace_event(struct trace_event *event); |
diff --git a/kernel/trace/trace_power.c b/kernel/trace/trace_power.c deleted file mode 100644 index fe1a00f1445a..000000000000 --- a/kernel/trace/trace_power.c +++ /dev/null | |||
| @@ -1,218 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * ring buffer based C-state tracer | ||
| 3 | * | ||
| 4 | * Arjan van de Ven <arjan@linux.intel.com> | ||
| 5 | * Copyright (C) 2008 Intel Corporation | ||
| 6 | * | ||
| 7 | * Much is borrowed from trace_boot.c which is | ||
| 8 | * Copyright (C) 2008 Frederic Weisbecker <fweisbec@gmail.com> | ||
| 9 | * | ||
| 10 | */ | ||
| 11 | |||
| 12 | #include <linux/init.h> | ||
| 13 | #include <linux/debugfs.h> | ||
| 14 | #include <trace/power.h> | ||
| 15 | #include <linux/kallsyms.h> | ||
| 16 | #include <linux/module.h> | ||
| 17 | |||
| 18 | #include "trace.h" | ||
| 19 | #include "trace_output.h" | ||
| 20 | |||
| 21 | static struct trace_array *power_trace; | ||
| 22 | static int __read_mostly trace_power_enabled; | ||
| 23 | |||
| 24 | static void probe_power_start(struct power_trace *it, unsigned int type, | ||
| 25 | unsigned int level) | ||
| 26 | { | ||
| 27 | if (!trace_power_enabled) | ||
| 28 | return; | ||
| 29 | |||
| 30 | memset(it, 0, sizeof(struct power_trace)); | ||
| 31 | it->state = level; | ||
| 32 | it->type = type; | ||
| 33 | it->stamp = ktime_get(); | ||
| 34 | } | ||
| 35 | |||
| 36 | |||
| 37 | static void probe_power_end(struct power_trace *it) | ||
| 38 | { | ||
| 39 | struct ftrace_event_call *call = &event_power; | ||
| 40 | struct ring_buffer_event *event; | ||
| 41 | struct ring_buffer *buffer; | ||
| 42 | struct trace_power *entry; | ||
| 43 | struct trace_array_cpu *data; | ||
| 44 | struct trace_array *tr = power_trace; | ||
| 45 | |||
| 46 | if (!trace_power_enabled) | ||
| 47 | return; | ||
| 48 | |||
| 49 | buffer = tr->buffer; | ||
| 50 | |||
| 51 | preempt_disable(); | ||
| 52 | it->end = ktime_get(); | ||
| 53 | data = tr->data[smp_processor_id()]; | ||
| 54 | |||
| 55 | event = trace_buffer_lock_reserve(buffer, TRACE_POWER, | ||
| 56 | sizeof(*entry), 0, 0); | ||
| 57 | if (!event) | ||
| 58 | goto out; | ||
| 59 | entry = ring_buffer_event_data(event); | ||
| 60 | entry->state_data = *it; | ||
| 61 | if (!filter_check_discard(call, entry, buffer, event)) | ||
| 62 | trace_buffer_unlock_commit(buffer, event, 0, 0); | ||
| 63 | out: | ||
| 64 | preempt_enable(); | ||
| 65 | } | ||
| 66 | |||
| 67 | static void probe_power_mark(struct power_trace *it, unsigned int type, | ||
| 68 | unsigned int level) | ||
| 69 | { | ||
| 70 | struct ftrace_event_call *call = &event_power; | ||
| 71 | struct ring_buffer_event *event; | ||
| 72 | struct ring_buffer *buffer; | ||
| 73 | struct trace_power *entry; | ||
| 74 | struct trace_array_cpu *data; | ||
| 75 | struct trace_array *tr = power_trace; | ||
| 76 | |||
| 77 | if (!trace_power_enabled) | ||
| 78 | return; | ||
| 79 | |||
| 80 | buffer = tr->buffer; | ||
| 81 | |||
| 82 | memset(it, 0, sizeof(struct power_trace)); | ||
| 83 | it->state = level; | ||
| 84 | it->type = type; | ||
| 85 | it->stamp = ktime_get(); | ||
| 86 | preempt_disable(); | ||
| 87 | it->end = it->stamp; | ||
| 88 | data = tr->data[smp_processor_id()]; | ||
| 89 | |||
| 90 | event = trace_buffer_lock_reserve(buffer, TRACE_POWER, | ||
| 91 | sizeof(*entry), 0, 0); | ||
| 92 | if (!event) | ||
| 93 | goto out; | ||
| 94 | entry = ring_buffer_event_data(event); | ||
| 95 | entry->state_data = *it; | ||
| 96 | if (!filter_check_discard(call, entry, buffer, event)) | ||
| 97 | trace_buffer_unlock_commit(buffer, event, 0, 0); | ||
| 98 | out: | ||
| 99 | preempt_enable(); | ||
| 100 | } | ||
| 101 | |||
| 102 | static int tracing_power_register(void) | ||
| 103 | { | ||
| 104 | int ret; | ||
| 105 | |||
| 106 | ret = register_trace_power_start(probe_power_start); | ||
| 107 | if (ret) { | ||
| 108 | pr_info("power trace: Couldn't activate tracepoint" | ||
| 109 | " probe to trace_power_start\n"); | ||
| 110 | return ret; | ||
| 111 | } | ||
| 112 | ret = register_trace_power_end(probe_power_end); | ||
| 113 | if (ret) { | ||
| 114 | pr_info("power trace: Couldn't activate tracepoint" | ||
| 115 | " probe to trace_power_end\n"); | ||
| 116 | goto fail_start; | ||
| 117 | } | ||
| 118 | ret = register_trace_power_mark(probe_power_mark); | ||
| 119 | if (ret) { | ||
| 120 | pr_info("power trace: Couldn't activate tracepoint" | ||
| 121 | " probe to trace_power_mark\n"); | ||
| 122 | goto fail_end; | ||
| 123 | } | ||
| 124 | return ret; | ||
| 125 | fail_end: | ||
| 126 | unregister_trace_power_end(probe_power_end); | ||
| 127 | fail_start: | ||
| 128 | unregister_trace_power_start(probe_power_start); | ||
| 129 | return ret; | ||
| 130 | } | ||
| 131 | |||
| 132 | static void start_power_trace(struct trace_array *tr) | ||
| 133 | { | ||
| 134 | trace_power_enabled = 1; | ||
| 135 | } | ||
| 136 | |||
| 137 | static void stop_power_trace(struct trace_array *tr) | ||
| 138 | { | ||
| 139 | trace_power_enabled = 0; | ||
| 140 | } | ||
| 141 | |||
| 142 | static void power_trace_reset(struct trace_array *tr) | ||
| 143 | { | ||
| 144 | trace_power_enabled = 0; | ||
| 145 | unregister_trace_power_start(probe_power_start); | ||
| 146 | unregister_trace_power_end(probe_power_end); | ||
| 147 | unregister_trace_power_mark(probe_power_mark); | ||
| 148 | } | ||
| 149 | |||
| 150 | |||
| 151 | static int power_trace_init(struct trace_array *tr) | ||
| 152 | { | ||
| 153 | power_trace = tr; | ||
| 154 | |||
| 155 | trace_power_enabled = 1; | ||
| 156 | tracing_power_register(); | ||
| 157 | |||
| 158 | tracing_reset_online_cpus(tr); | ||
| 159 | return 0; | ||
| 160 | } | ||
| 161 | |||
| 162 | static enum print_line_t power_print_line(struct trace_iterator *iter) | ||
| 163 | { | ||
| 164 | int ret = 0; | ||
| 165 | struct trace_entry *entry = iter->ent; | ||
| 166 | struct trace_power *field ; | ||
| 167 | struct power_trace *it; | ||
| 168 | struct trace_seq *s = &iter->seq; | ||
| 169 | struct timespec stamp; | ||
| 170 | struct timespec duration; | ||
| 171 | |||
| 172 | trace_assign_type(field, entry); | ||
| 173 | it = &field->state_data; | ||
| 174 | stamp = ktime_to_timespec(it->stamp); | ||
| 175 | duration = ktime_to_timespec(ktime_sub(it->end, it->stamp)); | ||
| 176 | |||
| 177 | if (entry->type == TRACE_POWER) { | ||
| 178 | if (it->type == POWER_CSTATE) | ||
| 179 | ret = trace_seq_printf(s, "[%5ld.%09ld] CSTATE: Going to C%i on cpu %i for %ld.%09ld\n", | ||
| 180 | stamp.tv_sec, | ||
| 181 | stamp.tv_nsec, | ||
| 182 | it->state, iter->cpu, | ||
| 183 | duration.tv_sec, | ||
| 184 | duration.tv_nsec); | ||
| 185 | if (it->type == POWER_PSTATE) | ||
| 186 | ret = trace_seq_printf(s, "[%5ld.%09ld] PSTATE: Going to P%i on cpu %i\n", | ||
| 187 | stamp.tv_sec, | ||
| 188 | stamp.tv_nsec, | ||
| 189 | it->state, iter->cpu); | ||
| 190 | if (!ret) | ||
| 191 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 192 | return TRACE_TYPE_HANDLED; | ||
| 193 | } | ||
| 194 | return TRACE_TYPE_UNHANDLED; | ||
| 195 | } | ||
| 196 | |||
| 197 | static void power_print_header(struct seq_file *s) | ||
| 198 | { | ||
| 199 | seq_puts(s, "# TIMESTAMP STATE EVENT\n"); | ||
| 200 | seq_puts(s, "# | | |\n"); | ||
| 201 | } | ||
| 202 | |||
| 203 | static struct tracer power_tracer __read_mostly = | ||
| 204 | { | ||
| 205 | .name = "power", | ||
| 206 | .init = power_trace_init, | ||
| 207 | .start = start_power_trace, | ||
| 208 | .stop = stop_power_trace, | ||
| 209 | .reset = power_trace_reset, | ||
| 210 | .print_line = power_print_line, | ||
| 211 | .print_header = power_print_header, | ||
| 212 | }; | ||
| 213 | |||
| 214 | static int init_power_trace(void) | ||
| 215 | { | ||
| 216 | return register_tracer(&power_tracer); | ||
| 217 | } | ||
| 218 | device_initcall(init_power_trace); | ||
diff --git a/kernel/trace/trace_printk.c b/kernel/trace/trace_printk.c index 687699d365ae..2547d8813cf0 100644 --- a/kernel/trace/trace_printk.c +++ b/kernel/trace/trace_printk.c | |||
| @@ -11,7 +11,6 @@ | |||
| 11 | #include <linux/ftrace.h> | 11 | #include <linux/ftrace.h> |
| 12 | #include <linux/string.h> | 12 | #include <linux/string.h> |
| 13 | #include <linux/module.h> | 13 | #include <linux/module.h> |
| 14 | #include <linux/marker.h> | ||
| 15 | #include <linux/mutex.h> | 14 | #include <linux/mutex.h> |
| 16 | #include <linux/ctype.h> | 15 | #include <linux/ctype.h> |
| 17 | #include <linux/list.h> | 16 | #include <linux/list.h> |
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index ad69f105a7c6..26185d727676 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c | |||
| @@ -24,6 +24,7 @@ static int __read_mostly tracer_enabled; | |||
| 24 | 24 | ||
| 25 | static struct task_struct *wakeup_task; | 25 | static struct task_struct *wakeup_task; |
| 26 | static int wakeup_cpu; | 26 | static int wakeup_cpu; |
| 27 | static int wakeup_current_cpu; | ||
| 27 | static unsigned wakeup_prio = -1; | 28 | static unsigned wakeup_prio = -1; |
| 28 | static int wakeup_rt; | 29 | static int wakeup_rt; |
| 29 | 30 | ||
| @@ -56,33 +57,23 @@ wakeup_tracer_call(unsigned long ip, unsigned long parent_ip) | |||
| 56 | resched = ftrace_preempt_disable(); | 57 | resched = ftrace_preempt_disable(); |
| 57 | 58 | ||
| 58 | cpu = raw_smp_processor_id(); | 59 | cpu = raw_smp_processor_id(); |
| 60 | if (cpu != wakeup_current_cpu) | ||
| 61 | goto out_enable; | ||
| 62 | |||
| 59 | data = tr->data[cpu]; | 63 | data = tr->data[cpu]; |
| 60 | disabled = atomic_inc_return(&data->disabled); | 64 | disabled = atomic_inc_return(&data->disabled); |
| 61 | if (unlikely(disabled != 1)) | 65 | if (unlikely(disabled != 1)) |
| 62 | goto out; | 66 | goto out; |
| 63 | 67 | ||
| 64 | local_irq_save(flags); | 68 | local_irq_save(flags); |
| 65 | __raw_spin_lock(&wakeup_lock); | ||
| 66 | |||
| 67 | if (unlikely(!wakeup_task)) | ||
| 68 | goto unlock; | ||
| 69 | |||
| 70 | /* | ||
| 71 | * The task can't disappear because it needs to | ||
| 72 | * wake up first, and we have the wakeup_lock. | ||
| 73 | */ | ||
| 74 | if (task_cpu(wakeup_task) != cpu) | ||
| 75 | goto unlock; | ||
| 76 | 69 | ||
| 77 | trace_function(tr, ip, parent_ip, flags, pc); | 70 | trace_function(tr, ip, parent_ip, flags, pc); |
| 78 | 71 | ||
| 79 | unlock: | ||
| 80 | __raw_spin_unlock(&wakeup_lock); | ||
| 81 | local_irq_restore(flags); | 72 | local_irq_restore(flags); |
| 82 | 73 | ||
| 83 | out: | 74 | out: |
| 84 | atomic_dec(&data->disabled); | 75 | atomic_dec(&data->disabled); |
| 85 | 76 | out_enable: | |
| 86 | ftrace_preempt_enable(resched); | 77 | ftrace_preempt_enable(resched); |
| 87 | } | 78 | } |
| 88 | 79 | ||
| @@ -107,11 +98,18 @@ static int report_latency(cycle_t delta) | |||
| 107 | return 1; | 98 | return 1; |
| 108 | } | 99 | } |
| 109 | 100 | ||
| 101 | static void probe_wakeup_migrate_task(struct task_struct *task, int cpu) | ||
| 102 | { | ||
| 103 | if (task != wakeup_task) | ||
| 104 | return; | ||
| 105 | |||
| 106 | wakeup_current_cpu = cpu; | ||
| 107 | } | ||
| 108 | |||
| 110 | static void notrace | 109 | static void notrace |
| 111 | probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, | 110 | probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, |
| 112 | struct task_struct *next) | 111 | struct task_struct *next) |
| 113 | { | 112 | { |
| 114 | unsigned long latency = 0, t0 = 0, t1 = 0; | ||
| 115 | struct trace_array_cpu *data; | 113 | struct trace_array_cpu *data; |
| 116 | cycle_t T0, T1, delta; | 114 | cycle_t T0, T1, delta; |
| 117 | unsigned long flags; | 115 | unsigned long flags; |
| @@ -157,10 +155,6 @@ probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, | |||
| 157 | trace_function(wakeup_trace, CALLER_ADDR0, CALLER_ADDR1, flags, pc); | 155 | trace_function(wakeup_trace, CALLER_ADDR0, CALLER_ADDR1, flags, pc); |
| 158 | tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); | 156 | tracing_sched_switch_trace(wakeup_trace, prev, next, flags, pc); |
| 159 | 157 | ||
| 160 | /* | ||
| 161 | * usecs conversion is slow so we try to delay the conversion | ||
| 162 | * as long as possible: | ||
| 163 | */ | ||
| 164 | T0 = data->preempt_timestamp; | 158 | T0 = data->preempt_timestamp; |
| 165 | T1 = ftrace_now(cpu); | 159 | T1 = ftrace_now(cpu); |
| 166 | delta = T1-T0; | 160 | delta = T1-T0; |
| @@ -168,13 +162,10 @@ probe_wakeup_sched_switch(struct rq *rq, struct task_struct *prev, | |||
| 168 | if (!report_latency(delta)) | 162 | if (!report_latency(delta)) |
| 169 | goto out_unlock; | 163 | goto out_unlock; |
| 170 | 164 | ||
| 171 | latency = nsecs_to_usecs(delta); | 165 | if (likely(!is_tracing_stopped())) { |
| 172 | 166 | tracing_max_latency = delta; | |
| 173 | tracing_max_latency = delta; | 167 | update_max_tr(wakeup_trace, wakeup_task, wakeup_cpu); |
| 174 | t0 = nsecs_to_usecs(T0); | 168 | } |
| 175 | t1 = nsecs_to_usecs(T1); | ||
| 176 | |||
| 177 | update_max_tr(wakeup_trace, wakeup_task, wakeup_cpu); | ||
| 178 | 169 | ||
| 179 | out_unlock: | 170 | out_unlock: |
| 180 | __wakeup_reset(wakeup_trace); | 171 | __wakeup_reset(wakeup_trace); |
| @@ -244,6 +235,7 @@ probe_wakeup(struct rq *rq, struct task_struct *p, int success) | |||
| 244 | __wakeup_reset(wakeup_trace); | 235 | __wakeup_reset(wakeup_trace); |
| 245 | 236 | ||
| 246 | wakeup_cpu = task_cpu(p); | 237 | wakeup_cpu = task_cpu(p); |
| 238 | wakeup_current_cpu = wakeup_cpu; | ||
| 247 | wakeup_prio = p->prio; | 239 | wakeup_prio = p->prio; |
| 248 | 240 | ||
| 249 | wakeup_task = p; | 241 | wakeup_task = p; |
| @@ -293,6 +285,13 @@ static void start_wakeup_tracer(struct trace_array *tr) | |||
| 293 | goto fail_deprobe_wake_new; | 285 | goto fail_deprobe_wake_new; |
| 294 | } | 286 | } |
| 295 | 287 | ||
| 288 | ret = register_trace_sched_migrate_task(probe_wakeup_migrate_task); | ||
| 289 | if (ret) { | ||
| 290 | pr_info("wakeup trace: Couldn't activate tracepoint" | ||
| 291 | " probe to kernel_sched_migrate_task\n"); | ||
| 292 | return; | ||
| 293 | } | ||
| 294 | |||
| 296 | wakeup_reset(tr); | 295 | wakeup_reset(tr); |
| 297 | 296 | ||
| 298 | /* | 297 | /* |
| @@ -325,6 +324,7 @@ static void stop_wakeup_tracer(struct trace_array *tr) | |||
| 325 | unregister_trace_sched_switch(probe_wakeup_sched_switch); | 324 | unregister_trace_sched_switch(probe_wakeup_sched_switch); |
| 326 | unregister_trace_sched_wakeup_new(probe_wakeup); | 325 | unregister_trace_sched_wakeup_new(probe_wakeup); |
| 327 | unregister_trace_sched_wakeup(probe_wakeup); | 326 | unregister_trace_sched_wakeup(probe_wakeup); |
| 327 | unregister_trace_sched_migrate_task(probe_wakeup_migrate_task); | ||
| 328 | } | 328 | } |
| 329 | 329 | ||
| 330 | static int __wakeup_tracer_init(struct trace_array *tr) | 330 | static int __wakeup_tracer_init(struct trace_array *tr) |
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index dfc55fed2099..1b050ab47120 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
| @@ -2,7 +2,7 @@ | |||
| 2 | #include <trace/events/syscalls.h> | 2 | #include <trace/events/syscalls.h> |
| 3 | #include <linux/kernel.h> | 3 | #include <linux/kernel.h> |
| 4 | #include <linux/ftrace.h> | 4 | #include <linux/ftrace.h> |
| 5 | #include <linux/perf_counter.h> | 5 | #include <linux/perf_event.h> |
| 6 | #include <asm/syscall.h> | 6 | #include <asm/syscall.h> |
| 7 | 7 | ||
| 8 | #include "trace_output.h" | 8 | #include "trace_output.h" |
| @@ -384,10 +384,13 @@ static int sys_prof_refcount_exit; | |||
| 384 | 384 | ||
| 385 | static void prof_syscall_enter(struct pt_regs *regs, long id) | 385 | static void prof_syscall_enter(struct pt_regs *regs, long id) |
| 386 | { | 386 | { |
| 387 | struct syscall_trace_enter *rec; | ||
| 388 | struct syscall_metadata *sys_data; | 387 | struct syscall_metadata *sys_data; |
| 388 | struct syscall_trace_enter *rec; | ||
| 389 | unsigned long flags; | ||
| 390 | char *raw_data; | ||
| 389 | int syscall_nr; | 391 | int syscall_nr; |
| 390 | int size; | 392 | int size; |
| 393 | int cpu; | ||
| 391 | 394 | ||
| 392 | syscall_nr = syscall_get_nr(current, regs); | 395 | syscall_nr = syscall_get_nr(current, regs); |
| 393 | if (!test_bit(syscall_nr, enabled_prof_enter_syscalls)) | 396 | if (!test_bit(syscall_nr, enabled_prof_enter_syscalls)) |
| @@ -402,20 +405,38 @@ static void prof_syscall_enter(struct pt_regs *regs, long id) | |||
| 402 | size = ALIGN(size + sizeof(u32), sizeof(u64)); | 405 | size = ALIGN(size + sizeof(u32), sizeof(u64)); |
| 403 | size -= sizeof(u32); | 406 | size -= sizeof(u32); |
| 404 | 407 | ||
| 405 | do { | 408 | if (WARN_ONCE(size > FTRACE_MAX_PROFILE_SIZE, |
| 406 | char raw_data[size]; | 409 | "profile buffer not large enough")) |
| 410 | return; | ||
| 411 | |||
| 412 | /* Protect the per cpu buffer, begin the rcu read side */ | ||
| 413 | local_irq_save(flags); | ||
| 407 | 414 | ||
| 408 | /* zero the dead bytes from align to not leak stack to user */ | 415 | cpu = smp_processor_id(); |
| 409 | *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL; | 416 | |
| 417 | if (in_nmi()) | ||
| 418 | raw_data = rcu_dereference(trace_profile_buf_nmi); | ||
| 419 | else | ||
| 420 | raw_data = rcu_dereference(trace_profile_buf); | ||
| 421 | |||
| 422 | if (!raw_data) | ||
| 423 | goto end; | ||
| 410 | 424 | ||
| 411 | rec = (struct syscall_trace_enter *) raw_data; | 425 | raw_data = per_cpu_ptr(raw_data, cpu); |
| 412 | tracing_generic_entry_update(&rec->ent, 0, 0); | 426 | |
| 413 | rec->ent.type = sys_data->enter_id; | 427 | /* zero the dead bytes from align to not leak stack to user */ |
| 414 | rec->nr = syscall_nr; | 428 | *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL; |
| 415 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, | 429 | |
| 416 | (unsigned long *)&rec->args); | 430 | rec = (struct syscall_trace_enter *) raw_data; |
| 417 | perf_tpcounter_event(sys_data->enter_id, 0, 1, rec, size); | 431 | tracing_generic_entry_update(&rec->ent, 0, 0); |
| 418 | } while(0); | 432 | rec->ent.type = sys_data->enter_id; |
| 433 | rec->nr = syscall_nr; | ||
| 434 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, | ||
| 435 | (unsigned long *)&rec->args); | ||
| 436 | perf_tp_event(sys_data->enter_id, 0, 1, rec, size); | ||
| 437 | |||
| 438 | end: | ||
| 439 | local_irq_restore(flags); | ||
| 419 | } | 440 | } |
| 420 | 441 | ||
| 421 | int reg_prof_syscall_enter(char *name) | 442 | int reg_prof_syscall_enter(char *name) |
| @@ -460,8 +481,12 @@ void unreg_prof_syscall_enter(char *name) | |||
| 460 | static void prof_syscall_exit(struct pt_regs *regs, long ret) | 481 | static void prof_syscall_exit(struct pt_regs *regs, long ret) |
| 461 | { | 482 | { |
| 462 | struct syscall_metadata *sys_data; | 483 | struct syscall_metadata *sys_data; |
| 463 | struct syscall_trace_exit rec; | 484 | struct syscall_trace_exit *rec; |
| 485 | unsigned long flags; | ||
| 464 | int syscall_nr; | 486 | int syscall_nr; |
| 487 | char *raw_data; | ||
| 488 | int size; | ||
| 489 | int cpu; | ||
| 465 | 490 | ||
| 466 | syscall_nr = syscall_get_nr(current, regs); | 491 | syscall_nr = syscall_get_nr(current, regs); |
| 467 | if (!test_bit(syscall_nr, enabled_prof_exit_syscalls)) | 492 | if (!test_bit(syscall_nr, enabled_prof_exit_syscalls)) |
| @@ -471,12 +496,46 @@ static void prof_syscall_exit(struct pt_regs *regs, long ret) | |||
| 471 | if (!sys_data) | 496 | if (!sys_data) |
| 472 | return; | 497 | return; |
| 473 | 498 | ||
| 474 | tracing_generic_entry_update(&rec.ent, 0, 0); | 499 | /* We can probably do that at build time */ |
| 475 | rec.ent.type = sys_data->exit_id; | 500 | size = ALIGN(sizeof(*rec) + sizeof(u32), sizeof(u64)); |
| 476 | rec.nr = syscall_nr; | 501 | size -= sizeof(u32); |
| 477 | rec.ret = syscall_get_return_value(current, regs); | ||
| 478 | 502 | ||
| 479 | perf_tpcounter_event(sys_data->exit_id, 0, 1, &rec, sizeof(rec)); | 503 | /* |
| 504 | * Impossible, but be paranoid with the future | ||
| 505 | * How to put this check outside runtime? | ||
| 506 | */ | ||
| 507 | if (WARN_ONCE(size > FTRACE_MAX_PROFILE_SIZE, | ||
| 508 | "exit event has grown above profile buffer size")) | ||
| 509 | return; | ||
| 510 | |||
| 511 | /* Protect the per cpu buffer, begin the rcu read side */ | ||
| 512 | local_irq_save(flags); | ||
| 513 | cpu = smp_processor_id(); | ||
| 514 | |||
| 515 | if (in_nmi()) | ||
| 516 | raw_data = rcu_dereference(trace_profile_buf_nmi); | ||
| 517 | else | ||
| 518 | raw_data = rcu_dereference(trace_profile_buf); | ||
| 519 | |||
| 520 | if (!raw_data) | ||
| 521 | goto end; | ||
| 522 | |||
| 523 | raw_data = per_cpu_ptr(raw_data, cpu); | ||
| 524 | |||
| 525 | /* zero the dead bytes from align to not leak stack to user */ | ||
| 526 | *(u64 *)(&raw_data[size - sizeof(u64)]) = 0ULL; | ||
| 527 | |||
| 528 | rec = (struct syscall_trace_exit *)raw_data; | ||
| 529 | |||
| 530 | tracing_generic_entry_update(&rec->ent, 0, 0); | ||
| 531 | rec->ent.type = sys_data->exit_id; | ||
| 532 | rec->nr = syscall_nr; | ||
| 533 | rec->ret = syscall_get_return_value(current, regs); | ||
| 534 | |||
| 535 | perf_tp_event(sys_data->exit_id, 0, 1, rec, size); | ||
| 536 | |||
| 537 | end: | ||
| 538 | local_irq_restore(flags); | ||
| 480 | } | 539 | } |
| 481 | 540 | ||
| 482 | int reg_prof_syscall_exit(char *name) | 541 | int reg_prof_syscall_exit(char *name) |
