diff options
Diffstat (limited to 'kernel/trace')
| -rw-r--r-- | kernel/trace/blktrace.c | 39 | ||||
| -rw-r--r-- | kernel/trace/ftrace.c | 58 | ||||
| -rw-r--r-- | kernel/trace/kmemtrace.c | 2 | ||||
| -rw-r--r-- | kernel/trace/ring_buffer.c | 14 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 12 | ||||
| -rw-r--r-- | kernel/trace/trace_branch.c | 8 | ||||
| -rw-r--r-- | kernel/trace/trace_event_profile.c | 15 | ||||
| -rw-r--r-- | kernel/trace/trace_events.c | 7 | ||||
| -rw-r--r-- | kernel/trace/trace_events_filter.c | 3 | ||||
| -rw-r--r-- | kernel/trace/trace_hw_branches.c | 8 | ||||
| -rw-r--r-- | kernel/trace/trace_output.c | 23 | ||||
| -rw-r--r-- | kernel/trace/trace_syscalls.c | 4 |
12 files changed, 123 insertions, 70 deletions
diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c index 3eb159c277c8..d9d6206e0b14 100644 --- a/kernel/trace/blktrace.c +++ b/kernel/trace/blktrace.c | |||
| @@ -856,6 +856,37 @@ static void blk_add_trace_remap(struct request_queue *q, struct bio *bio, | |||
| 856 | } | 856 | } |
| 857 | 857 | ||
| 858 | /** | 858 | /** |
| 859 | * blk_add_trace_rq_remap - Add a trace for a request-remap operation | ||
| 860 | * @q: queue the io is for | ||
| 861 | * @rq: the source request | ||
| 862 | * @dev: target device | ||
| 863 | * @from: source sector | ||
| 864 | * | ||
| 865 | * Description: | ||
| 866 | * Device mapper remaps request to other devices. | ||
| 867 | * Add a trace for that action. | ||
| 868 | * | ||
| 869 | **/ | ||
| 870 | static void blk_add_trace_rq_remap(struct request_queue *q, | ||
| 871 | struct request *rq, dev_t dev, | ||
| 872 | sector_t from) | ||
| 873 | { | ||
| 874 | struct blk_trace *bt = q->blk_trace; | ||
| 875 | struct blk_io_trace_remap r; | ||
| 876 | |||
| 877 | if (likely(!bt)) | ||
| 878 | return; | ||
| 879 | |||
| 880 | r.device_from = cpu_to_be32(dev); | ||
| 881 | r.device_to = cpu_to_be32(disk_devt(rq->rq_disk)); | ||
| 882 | r.sector_from = cpu_to_be64(from); | ||
| 883 | |||
| 884 | __blk_add_trace(bt, blk_rq_pos(rq), blk_rq_bytes(rq), | ||
| 885 | rq_data_dir(rq), BLK_TA_REMAP, !!rq->errors, | ||
| 886 | sizeof(r), &r); | ||
| 887 | } | ||
| 888 | |||
| 889 | /** | ||
| 859 | * blk_add_driver_data - Add binary message with driver-specific data | 890 | * blk_add_driver_data - Add binary message with driver-specific data |
| 860 | * @q: queue the io is for | 891 | * @q: queue the io is for |
| 861 | * @rq: io request | 892 | * @rq: io request |
| @@ -922,10 +953,13 @@ static void blk_register_tracepoints(void) | |||
| 922 | WARN_ON(ret); | 953 | WARN_ON(ret); |
| 923 | ret = register_trace_block_remap(blk_add_trace_remap); | 954 | ret = register_trace_block_remap(blk_add_trace_remap); |
| 924 | WARN_ON(ret); | 955 | WARN_ON(ret); |
| 956 | ret = register_trace_block_rq_remap(blk_add_trace_rq_remap); | ||
| 957 | WARN_ON(ret); | ||
| 925 | } | 958 | } |
| 926 | 959 | ||
| 927 | static void blk_unregister_tracepoints(void) | 960 | static void blk_unregister_tracepoints(void) |
| 928 | { | 961 | { |
| 962 | unregister_trace_block_rq_remap(blk_add_trace_rq_remap); | ||
| 929 | unregister_trace_block_remap(blk_add_trace_remap); | 963 | unregister_trace_block_remap(blk_add_trace_remap); |
| 930 | unregister_trace_block_split(blk_add_trace_split); | 964 | unregister_trace_block_split(blk_add_trace_split); |
| 931 | unregister_trace_block_unplug_io(blk_add_trace_unplug_io); | 965 | unregister_trace_block_unplug_io(blk_add_trace_unplug_io); |
| @@ -1657,6 +1691,11 @@ int blk_trace_init_sysfs(struct device *dev) | |||
| 1657 | return sysfs_create_group(&dev->kobj, &blk_trace_attr_group); | 1691 | return sysfs_create_group(&dev->kobj, &blk_trace_attr_group); |
| 1658 | } | 1692 | } |
| 1659 | 1693 | ||
| 1694 | void blk_trace_remove_sysfs(struct device *dev) | ||
| 1695 | { | ||
| 1696 | sysfs_remove_group(&dev->kobj, &blk_trace_attr_group); | ||
| 1697 | } | ||
| 1698 | |||
| 1660 | #endif /* CONFIG_BLK_DEV_IO_TRACE */ | 1699 | #endif /* CONFIG_BLK_DEV_IO_TRACE */ |
| 1661 | 1700 | ||
| 1662 | #ifdef CONFIG_EVENT_TRACING | 1701 | #ifdef CONFIG_EVENT_TRACING |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index a142579765bf..6dc4e5ef7a01 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -225,7 +225,11 @@ static void ftrace_update_pid_func(void) | |||
| 225 | if (ftrace_trace_function == ftrace_stub) | 225 | if (ftrace_trace_function == ftrace_stub) |
| 226 | return; | 226 | return; |
| 227 | 227 | ||
| 228 | #ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | ||
| 228 | func = ftrace_trace_function; | 229 | func = ftrace_trace_function; |
| 230 | #else | ||
| 231 | func = __ftrace_trace_function; | ||
| 232 | #endif | ||
| 229 | 233 | ||
| 230 | if (ftrace_pid_trace) { | 234 | if (ftrace_pid_trace) { |
| 231 | set_ftrace_pid_function(func); | 235 | set_ftrace_pid_function(func); |
| @@ -736,7 +740,7 @@ ftrace_profile_write(struct file *filp, const char __user *ubuf, | |||
| 736 | out: | 740 | out: |
| 737 | mutex_unlock(&ftrace_profile_lock); | 741 | mutex_unlock(&ftrace_profile_lock); |
| 738 | 742 | ||
| 739 | filp->f_pos += cnt; | 743 | *ppos += cnt; |
| 740 | 744 | ||
| 741 | return cnt; | 745 | return cnt; |
| 742 | } | 746 | } |
| @@ -1074,14 +1078,9 @@ static void ftrace_replace_code(int enable) | |||
| 1074 | failed = __ftrace_replace_code(rec, enable); | 1078 | failed = __ftrace_replace_code(rec, enable); |
| 1075 | if (failed) { | 1079 | if (failed) { |
| 1076 | rec->flags |= FTRACE_FL_FAILED; | 1080 | rec->flags |= FTRACE_FL_FAILED; |
| 1077 | if ((system_state == SYSTEM_BOOTING) || | 1081 | ftrace_bug(failed, rec->ip); |
| 1078 | !core_kernel_text(rec->ip)) { | 1082 | /* Stop processing */ |
| 1079 | ftrace_free_rec(rec); | 1083 | return; |
| 1080 | } else { | ||
| 1081 | ftrace_bug(failed, rec->ip); | ||
| 1082 | /* Stop processing */ | ||
| 1083 | return; | ||
| 1084 | } | ||
| 1085 | } | 1084 | } |
| 1086 | } while_for_each_ftrace_rec(); | 1085 | } while_for_each_ftrace_rec(); |
| 1087 | } | 1086 | } |
| @@ -1621,8 +1620,10 @@ ftrace_regex_open(struct inode *inode, struct file *file, int enable) | |||
| 1621 | if (!ret) { | 1620 | if (!ret) { |
| 1622 | struct seq_file *m = file->private_data; | 1621 | struct seq_file *m = file->private_data; |
| 1623 | m->private = iter; | 1622 | m->private = iter; |
| 1624 | } else | 1623 | } else { |
| 1624 | trace_parser_put(&iter->parser); | ||
| 1625 | kfree(iter); | 1625 | kfree(iter); |
| 1626 | } | ||
| 1626 | } else | 1627 | } else |
| 1627 | file->private_data = iter; | 1628 | file->private_data = iter; |
| 1628 | mutex_unlock(&ftrace_regex_lock); | 1629 | mutex_unlock(&ftrace_regex_lock); |
| @@ -2202,7 +2203,7 @@ ftrace_regex_write(struct file *file, const char __user *ubuf, | |||
| 2202 | struct trace_parser *parser; | 2203 | struct trace_parser *parser; |
| 2203 | ssize_t ret, read; | 2204 | ssize_t ret, read; |
| 2204 | 2205 | ||
| 2205 | if (!cnt || cnt < 0) | 2206 | if (!cnt) |
| 2206 | return 0; | 2207 | return 0; |
| 2207 | 2208 | ||
| 2208 | mutex_lock(&ftrace_regex_lock); | 2209 | mutex_lock(&ftrace_regex_lock); |
| @@ -2216,20 +2217,20 @@ ftrace_regex_write(struct file *file, const char __user *ubuf, | |||
| 2216 | parser = &iter->parser; | 2217 | parser = &iter->parser; |
| 2217 | read = trace_get_user(parser, ubuf, cnt, ppos); | 2218 | read = trace_get_user(parser, ubuf, cnt, ppos); |
| 2218 | 2219 | ||
| 2219 | if (trace_parser_loaded(parser) && | 2220 | if (read >= 0 && trace_parser_loaded(parser) && |
| 2220 | !trace_parser_cont(parser)) { | 2221 | !trace_parser_cont(parser)) { |
| 2221 | ret = ftrace_process_regex(parser->buffer, | 2222 | ret = ftrace_process_regex(parser->buffer, |
| 2222 | parser->idx, enable); | 2223 | parser->idx, enable); |
| 2223 | if (ret) | 2224 | if (ret) |
| 2224 | goto out; | 2225 | goto out_unlock; |
| 2225 | 2226 | ||
| 2226 | trace_parser_clear(parser); | 2227 | trace_parser_clear(parser); |
| 2227 | } | 2228 | } |
| 2228 | 2229 | ||
| 2229 | ret = read; | 2230 | ret = read; |
| 2230 | 2231 | out_unlock: | |
| 2231 | mutex_unlock(&ftrace_regex_lock); | 2232 | mutex_unlock(&ftrace_regex_lock); |
| 2232 | out: | 2233 | |
| 2233 | return ret; | 2234 | return ret; |
| 2234 | } | 2235 | } |
| 2235 | 2236 | ||
| @@ -2552,8 +2553,7 @@ ftrace_graph_write(struct file *file, const char __user *ubuf, | |||
| 2552 | size_t cnt, loff_t *ppos) | 2553 | size_t cnt, loff_t *ppos) |
| 2553 | { | 2554 | { |
| 2554 | struct trace_parser parser; | 2555 | struct trace_parser parser; |
| 2555 | size_t read = 0; | 2556 | ssize_t read, ret; |
| 2556 | ssize_t ret; | ||
| 2557 | 2557 | ||
| 2558 | if (!cnt || cnt < 0) | 2558 | if (!cnt || cnt < 0) |
| 2559 | return 0; | 2559 | return 0; |
| @@ -2562,29 +2562,31 @@ ftrace_graph_write(struct file *file, const char __user *ubuf, | |||
| 2562 | 2562 | ||
| 2563 | if (ftrace_graph_count >= FTRACE_GRAPH_MAX_FUNCS) { | 2563 | if (ftrace_graph_count >= FTRACE_GRAPH_MAX_FUNCS) { |
| 2564 | ret = -EBUSY; | 2564 | ret = -EBUSY; |
| 2565 | goto out; | 2565 | goto out_unlock; |
| 2566 | } | 2566 | } |
| 2567 | 2567 | ||
| 2568 | if (trace_parser_get_init(&parser, FTRACE_BUFF_MAX)) { | 2568 | if (trace_parser_get_init(&parser, FTRACE_BUFF_MAX)) { |
| 2569 | ret = -ENOMEM; | 2569 | ret = -ENOMEM; |
| 2570 | goto out; | 2570 | goto out_unlock; |
| 2571 | } | 2571 | } |
| 2572 | 2572 | ||
| 2573 | read = trace_get_user(&parser, ubuf, cnt, ppos); | 2573 | read = trace_get_user(&parser, ubuf, cnt, ppos); |
| 2574 | 2574 | ||
| 2575 | if (trace_parser_loaded((&parser))) { | 2575 | if (read >= 0 && trace_parser_loaded((&parser))) { |
| 2576 | parser.buffer[parser.idx] = 0; | 2576 | parser.buffer[parser.idx] = 0; |
| 2577 | 2577 | ||
| 2578 | /* we allow only one expression at a time */ | 2578 | /* we allow only one expression at a time */ |
| 2579 | ret = ftrace_set_func(ftrace_graph_funcs, &ftrace_graph_count, | 2579 | ret = ftrace_set_func(ftrace_graph_funcs, &ftrace_graph_count, |
| 2580 | parser.buffer); | 2580 | parser.buffer); |
| 2581 | if (ret) | 2581 | if (ret) |
| 2582 | goto out; | 2582 | goto out_free; |
| 2583 | } | 2583 | } |
| 2584 | 2584 | ||
| 2585 | ret = read; | 2585 | ret = read; |
| 2586 | out: | 2586 | |
| 2587 | out_free: | ||
| 2587 | trace_parser_put(&parser); | 2588 | trace_parser_put(&parser); |
| 2589 | out_unlock: | ||
| 2588 | mutex_unlock(&graph_lock); | 2590 | mutex_unlock(&graph_lock); |
| 2589 | 2591 | ||
| 2590 | return ret; | 2592 | return ret; |
| @@ -2655,19 +2657,17 @@ static int ftrace_convert_nops(struct module *mod, | |||
| 2655 | } | 2657 | } |
| 2656 | 2658 | ||
| 2657 | #ifdef CONFIG_MODULES | 2659 | #ifdef CONFIG_MODULES |
| 2658 | void ftrace_release(void *start, void *end) | 2660 | void ftrace_release_mod(struct module *mod) |
| 2659 | { | 2661 | { |
| 2660 | struct dyn_ftrace *rec; | 2662 | struct dyn_ftrace *rec; |
| 2661 | struct ftrace_page *pg; | 2663 | struct ftrace_page *pg; |
| 2662 | unsigned long s = (unsigned long)start; | ||
| 2663 | unsigned long e = (unsigned long)end; | ||
| 2664 | 2664 | ||
| 2665 | if (ftrace_disabled || !start || start == end) | 2665 | if (ftrace_disabled) |
| 2666 | return; | 2666 | return; |
| 2667 | 2667 | ||
| 2668 | mutex_lock(&ftrace_lock); | 2668 | mutex_lock(&ftrace_lock); |
| 2669 | do_for_each_ftrace_rec(pg, rec) { | 2669 | do_for_each_ftrace_rec(pg, rec) { |
| 2670 | if ((rec->ip >= s) && (rec->ip < e)) { | 2670 | if (within_module_core(rec->ip, mod)) { |
| 2671 | /* | 2671 | /* |
| 2672 | * rec->ip is changed in ftrace_free_rec() | 2672 | * rec->ip is changed in ftrace_free_rec() |
| 2673 | * It should not between s and e if record was freed. | 2673 | * It should not between s and e if record was freed. |
| @@ -2699,9 +2699,7 @@ static int ftrace_module_notify(struct notifier_block *self, | |||
| 2699 | mod->num_ftrace_callsites); | 2699 | mod->num_ftrace_callsites); |
| 2700 | break; | 2700 | break; |
| 2701 | case MODULE_STATE_GOING: | 2701 | case MODULE_STATE_GOING: |
| 2702 | ftrace_release(mod->ftrace_callsites, | 2702 | ftrace_release_mod(mod); |
| 2703 | mod->ftrace_callsites + | ||
| 2704 | mod->num_ftrace_callsites); | ||
| 2705 | break; | 2703 | break; |
| 2706 | } | 2704 | } |
| 2707 | 2705 | ||
diff --git a/kernel/trace/kmemtrace.c b/kernel/trace/kmemtrace.c index 81b1645c8549..a91da69f153a 100644 --- a/kernel/trace/kmemtrace.c +++ b/kernel/trace/kmemtrace.c | |||
| @@ -501,7 +501,7 @@ static int __init init_kmem_tracer(void) | |||
| 501 | return 1; | 501 | return 1; |
| 502 | } | 502 | } |
| 503 | 503 | ||
| 504 | if (!register_tracer(&kmem_tracer)) { | 504 | if (register_tracer(&kmem_tracer) != 0) { |
| 505 | pr_warning("Warning: could not register the kmem tracer\n"); | 505 | pr_warning("Warning: could not register the kmem tracer\n"); |
| 506 | return 1; | 506 | return 1; |
| 507 | } | 507 | } |
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index d4ff01970547..5dd017fea6f5 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -483,7 +483,7 @@ struct ring_buffer_iter { | |||
| 483 | /* Up this if you want to test the TIME_EXTENTS and normalization */ | 483 | /* Up this if you want to test the TIME_EXTENTS and normalization */ |
| 484 | #define DEBUG_SHIFT 0 | 484 | #define DEBUG_SHIFT 0 |
| 485 | 485 | ||
| 486 | static inline u64 rb_time_stamp(struct ring_buffer *buffer, int cpu) | 486 | static inline u64 rb_time_stamp(struct ring_buffer *buffer) |
| 487 | { | 487 | { |
| 488 | /* shift to debug/test normalization and TIME_EXTENTS */ | 488 | /* shift to debug/test normalization and TIME_EXTENTS */ |
| 489 | return buffer->clock() << DEBUG_SHIFT; | 489 | return buffer->clock() << DEBUG_SHIFT; |
| @@ -494,7 +494,7 @@ u64 ring_buffer_time_stamp(struct ring_buffer *buffer, int cpu) | |||
| 494 | u64 time; | 494 | u64 time; |
| 495 | 495 | ||
| 496 | preempt_disable_notrace(); | 496 | preempt_disable_notrace(); |
| 497 | time = rb_time_stamp(buffer, cpu); | 497 | time = rb_time_stamp(buffer); |
| 498 | preempt_enable_no_resched_notrace(); | 498 | preempt_enable_no_resched_notrace(); |
| 499 | 499 | ||
| 500 | return time; | 500 | return time; |
| @@ -599,7 +599,7 @@ static struct list_head *rb_list_head(struct list_head *list) | |||
| 599 | } | 599 | } |
| 600 | 600 | ||
| 601 | /* | 601 | /* |
| 602 | * rb_is_head_page - test if the give page is the head page | 602 | * rb_is_head_page - test if the given page is the head page |
| 603 | * | 603 | * |
| 604 | * Because the reader may move the head_page pointer, we can | 604 | * Because the reader may move the head_page pointer, we can |
| 605 | * not trust what the head page is (it may be pointing to | 605 | * not trust what the head page is (it may be pointing to |
| @@ -1193,6 +1193,7 @@ rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned nr_pages) | |||
| 1193 | atomic_inc(&cpu_buffer->record_disabled); | 1193 | atomic_inc(&cpu_buffer->record_disabled); |
| 1194 | synchronize_sched(); | 1194 | synchronize_sched(); |
| 1195 | 1195 | ||
| 1196 | spin_lock_irq(&cpu_buffer->reader_lock); | ||
| 1196 | rb_head_page_deactivate(cpu_buffer); | 1197 | rb_head_page_deactivate(cpu_buffer); |
| 1197 | 1198 | ||
| 1198 | for (i = 0; i < nr_pages; i++) { | 1199 | for (i = 0; i < nr_pages; i++) { |
| @@ -1207,6 +1208,7 @@ rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned nr_pages) | |||
| 1207 | return; | 1208 | return; |
| 1208 | 1209 | ||
| 1209 | rb_reset_cpu(cpu_buffer); | 1210 | rb_reset_cpu(cpu_buffer); |
| 1211 | spin_unlock_irq(&cpu_buffer->reader_lock); | ||
| 1210 | 1212 | ||
| 1211 | rb_check_pages(cpu_buffer); | 1213 | rb_check_pages(cpu_buffer); |
| 1212 | 1214 | ||
| @@ -1868,7 +1870,7 @@ rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1868 | * Nested commits always have zero deltas, so | 1870 | * Nested commits always have zero deltas, so |
| 1869 | * just reread the time stamp | 1871 | * just reread the time stamp |
| 1870 | */ | 1872 | */ |
| 1871 | *ts = rb_time_stamp(buffer, cpu_buffer->cpu); | 1873 | *ts = rb_time_stamp(buffer); |
| 1872 | next_page->page->time_stamp = *ts; | 1874 | next_page->page->time_stamp = *ts; |
| 1873 | } | 1875 | } |
| 1874 | 1876 | ||
| @@ -2111,7 +2113,7 @@ rb_reserve_next_event(struct ring_buffer *buffer, | |||
| 2111 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > 1000)) | 2113 | if (RB_WARN_ON(cpu_buffer, ++nr_loops > 1000)) |
| 2112 | goto out_fail; | 2114 | goto out_fail; |
| 2113 | 2115 | ||
| 2114 | ts = rb_time_stamp(cpu_buffer->buffer, cpu_buffer->cpu); | 2116 | ts = rb_time_stamp(cpu_buffer->buffer); |
| 2115 | 2117 | ||
| 2116 | /* | 2118 | /* |
| 2117 | * Only the first commit can update the timestamp. | 2119 | * Only the first commit can update the timestamp. |
| @@ -2681,7 +2683,7 @@ unsigned long ring_buffer_entries(struct ring_buffer *buffer) | |||
| 2681 | EXPORT_SYMBOL_GPL(ring_buffer_entries); | 2683 | EXPORT_SYMBOL_GPL(ring_buffer_entries); |
| 2682 | 2684 | ||
| 2683 | /** | 2685 | /** |
| 2684 | * ring_buffer_overrun_cpu - get the number of overruns in buffer | 2686 | * ring_buffer_overruns - get the number of overruns in buffer |
| 2685 | * @buffer: The ring buffer | 2687 | * @buffer: The ring buffer |
| 2686 | * | 2688 | * |
| 2687 | * Returns the total number of overruns in the ring buffer | 2689 | * Returns the total number of overruns in the ring buffer |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 411af37f4be4..b20d3ec75de9 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -415,7 +415,7 @@ int trace_get_user(struct trace_parser *parser, const char __user *ubuf, | |||
| 415 | 415 | ||
| 416 | /* read the non-space input */ | 416 | /* read the non-space input */ |
| 417 | while (cnt && !isspace(ch)) { | 417 | while (cnt && !isspace(ch)) { |
| 418 | if (parser->idx < parser->size) | 418 | if (parser->idx < parser->size - 1) |
| 419 | parser->buffer[parser->idx++] = ch; | 419 | parser->buffer[parser->idx++] = ch; |
| 420 | else { | 420 | else { |
| 421 | ret = -EINVAL; | 421 | ret = -EINVAL; |
| @@ -1393,7 +1393,7 @@ int trace_array_vprintk(struct trace_array *tr, | |||
| 1393 | 1393 | ||
| 1394 | int trace_vprintk(unsigned long ip, const char *fmt, va_list args) | 1394 | int trace_vprintk(unsigned long ip, const char *fmt, va_list args) |
| 1395 | { | 1395 | { |
| 1396 | return trace_array_printk(&global_trace, ip, fmt, args); | 1396 | return trace_array_vprintk(&global_trace, ip, fmt, args); |
| 1397 | } | 1397 | } |
| 1398 | EXPORT_SYMBOL_GPL(trace_vprintk); | 1398 | EXPORT_SYMBOL_GPL(trace_vprintk); |
| 1399 | 1399 | ||
| @@ -2440,7 +2440,7 @@ tracing_trace_options_write(struct file *filp, const char __user *ubuf, | |||
| 2440 | return ret; | 2440 | return ret; |
| 2441 | } | 2441 | } |
| 2442 | 2442 | ||
| 2443 | filp->f_pos += cnt; | 2443 | *ppos += cnt; |
| 2444 | 2444 | ||
| 2445 | return cnt; | 2445 | return cnt; |
| 2446 | } | 2446 | } |
| @@ -2582,7 +2582,7 @@ tracing_ctrl_write(struct file *filp, const char __user *ubuf, | |||
| 2582 | } | 2582 | } |
| 2583 | mutex_unlock(&trace_types_lock); | 2583 | mutex_unlock(&trace_types_lock); |
| 2584 | 2584 | ||
| 2585 | filp->f_pos += cnt; | 2585 | *ppos += cnt; |
| 2586 | 2586 | ||
| 2587 | return cnt; | 2587 | return cnt; |
| 2588 | } | 2588 | } |
| @@ -2764,7 +2764,7 @@ tracing_set_trace_write(struct file *filp, const char __user *ubuf, | |||
| 2764 | if (err) | 2764 | if (err) |
| 2765 | return err; | 2765 | return err; |
| 2766 | 2766 | ||
| 2767 | filp->f_pos += ret; | 2767 | *ppos += ret; |
| 2768 | 2768 | ||
| 2769 | return ret; | 2769 | return ret; |
| 2770 | } | 2770 | } |
| @@ -3299,7 +3299,7 @@ tracing_entries_write(struct file *filp, const char __user *ubuf, | |||
| 3299 | } | 3299 | } |
| 3300 | } | 3300 | } |
| 3301 | 3301 | ||
| 3302 | filp->f_pos += cnt; | 3302 | *ppos += cnt; |
| 3303 | 3303 | ||
| 3304 | /* If check pages failed, return ENOMEM */ | 3304 | /* If check pages failed, return ENOMEM */ |
| 3305 | if (tracing_disabled) | 3305 | if (tracing_disabled) |
diff --git a/kernel/trace/trace_branch.c b/kernel/trace/trace_branch.c index 7a7a9fd249a9..4a194f08f88c 100644 --- a/kernel/trace/trace_branch.c +++ b/kernel/trace/trace_branch.c | |||
| @@ -34,6 +34,7 @@ probe_likely_condition(struct ftrace_branch_data *f, int val, int expect) | |||
| 34 | struct trace_array *tr = branch_tracer; | 34 | struct trace_array *tr = branch_tracer; |
| 35 | struct ring_buffer_event *event; | 35 | struct ring_buffer_event *event; |
| 36 | struct trace_branch *entry; | 36 | struct trace_branch *entry; |
| 37 | struct ring_buffer *buffer; | ||
| 37 | unsigned long flags; | 38 | unsigned long flags; |
| 38 | int cpu, pc; | 39 | int cpu, pc; |
| 39 | const char *p; | 40 | const char *p; |
| @@ -54,7 +55,8 @@ probe_likely_condition(struct ftrace_branch_data *f, int val, int expect) | |||
| 54 | goto out; | 55 | goto out; |
| 55 | 56 | ||
| 56 | pc = preempt_count(); | 57 | pc = preempt_count(); |
| 57 | event = trace_buffer_lock_reserve(tr, TRACE_BRANCH, | 58 | buffer = tr->buffer; |
| 59 | event = trace_buffer_lock_reserve(buffer, TRACE_BRANCH, | ||
| 58 | sizeof(*entry), flags, pc); | 60 | sizeof(*entry), flags, pc); |
| 59 | if (!event) | 61 | if (!event) |
| 60 | goto out; | 62 | goto out; |
| @@ -74,8 +76,8 @@ probe_likely_condition(struct ftrace_branch_data *f, int val, int expect) | |||
| 74 | entry->line = f->line; | 76 | entry->line = f->line; |
| 75 | entry->correct = val == expect; | 77 | entry->correct = val == expect; |
| 76 | 78 | ||
| 77 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 79 | if (!filter_check_discard(call, entry, buffer, event)) |
| 78 | ring_buffer_unlock_commit(tr->buffer, event); | 80 | ring_buffer_unlock_commit(buffer, event); |
| 79 | 81 | ||
| 80 | out: | 82 | out: |
| 81 | atomic_dec(&tr->data[cpu]->disabled); | 83 | atomic_dec(&tr->data[cpu]->disabled); |
diff --git a/kernel/trace/trace_event_profile.c b/kernel/trace/trace_event_profile.c index dd44b8768867..8d5c171cc998 100644 --- a/kernel/trace/trace_event_profile.c +++ b/kernel/trace/trace_event_profile.c | |||
| @@ -31,7 +31,7 @@ static int ftrace_profile_enable_event(struct ftrace_event_call *event) | |||
| 31 | if (atomic_inc_return(&event->profile_count)) | 31 | if (atomic_inc_return(&event->profile_count)) |
| 32 | return 0; | 32 | return 0; |
| 33 | 33 | ||
| 34 | if (!total_profile_count++) { | 34 | if (!total_profile_count) { |
| 35 | buf = (char *)alloc_percpu(profile_buf_t); | 35 | buf = (char *)alloc_percpu(profile_buf_t); |
| 36 | if (!buf) | 36 | if (!buf) |
| 37 | goto fail_buf; | 37 | goto fail_buf; |
| @@ -46,14 +46,19 @@ static int ftrace_profile_enable_event(struct ftrace_event_call *event) | |||
| 46 | } | 46 | } |
| 47 | 47 | ||
| 48 | ret = event->profile_enable(); | 48 | ret = event->profile_enable(); |
| 49 | if (!ret) | 49 | if (!ret) { |
| 50 | total_profile_count++; | ||
| 50 | return 0; | 51 | return 0; |
| 52 | } | ||
| 51 | 53 | ||
| 52 | kfree(trace_profile_buf_nmi); | ||
| 53 | fail_buf_nmi: | 54 | fail_buf_nmi: |
| 54 | kfree(trace_profile_buf); | 55 | if (!total_profile_count) { |
| 56 | free_percpu(trace_profile_buf_nmi); | ||
| 57 | free_percpu(trace_profile_buf); | ||
| 58 | trace_profile_buf_nmi = NULL; | ||
| 59 | trace_profile_buf = NULL; | ||
| 60 | } | ||
| 55 | fail_buf: | 61 | fail_buf: |
| 56 | total_profile_count--; | ||
| 57 | atomic_dec(&event->profile_count); | 62 | atomic_dec(&event->profile_count); |
| 58 | 63 | ||
| 59 | return ret; | 64 | return ret; |
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index 6f03c8a1105e..d128f65778e6 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
| @@ -232,10 +232,9 @@ ftrace_event_write(struct file *file, const char __user *ubuf, | |||
| 232 | size_t cnt, loff_t *ppos) | 232 | size_t cnt, loff_t *ppos) |
| 233 | { | 233 | { |
| 234 | struct trace_parser parser; | 234 | struct trace_parser parser; |
| 235 | size_t read = 0; | 235 | ssize_t read, ret; |
| 236 | ssize_t ret; | ||
| 237 | 236 | ||
| 238 | if (!cnt || cnt < 0) | 237 | if (!cnt) |
| 239 | return 0; | 238 | return 0; |
| 240 | 239 | ||
| 241 | ret = tracing_update_buffers(); | 240 | ret = tracing_update_buffers(); |
| @@ -247,7 +246,7 @@ ftrace_event_write(struct file *file, const char __user *ubuf, | |||
| 247 | 246 | ||
| 248 | read = trace_get_user(&parser, ubuf, cnt, ppos); | 247 | read = trace_get_user(&parser, ubuf, cnt, ppos); |
| 249 | 248 | ||
| 250 | if (trace_parser_loaded((&parser))) { | 249 | if (read >= 0 && trace_parser_loaded((&parser))) { |
| 251 | int set = 1; | 250 | int set = 1; |
| 252 | 251 | ||
| 253 | if (*parser.buffer == '!') | 252 | if (*parser.buffer == '!') |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 23245785927f..98a6cc5c64ed 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
| @@ -933,8 +933,9 @@ static void postfix_clear(struct filter_parse_state *ps) | |||
| 933 | 933 | ||
| 934 | while (!list_empty(&ps->postfix)) { | 934 | while (!list_empty(&ps->postfix)) { |
| 935 | elt = list_first_entry(&ps->postfix, struct postfix_elt, list); | 935 | elt = list_first_entry(&ps->postfix, struct postfix_elt, list); |
| 936 | kfree(elt->operand); | ||
| 937 | list_del(&elt->list); | 936 | list_del(&elt->list); |
| 937 | kfree(elt->operand); | ||
| 938 | kfree(elt); | ||
| 938 | } | 939 | } |
| 939 | } | 940 | } |
| 940 | 941 | ||
diff --git a/kernel/trace/trace_hw_branches.c b/kernel/trace/trace_hw_branches.c index 23b63859130e..69543a905cd5 100644 --- a/kernel/trace/trace_hw_branches.c +++ b/kernel/trace/trace_hw_branches.c | |||
| @@ -165,6 +165,7 @@ void trace_hw_branch(u64 from, u64 to) | |||
| 165 | struct ftrace_event_call *call = &event_hw_branch; | 165 | struct ftrace_event_call *call = &event_hw_branch; |
| 166 | struct trace_array *tr = hw_branch_trace; | 166 | struct trace_array *tr = hw_branch_trace; |
| 167 | struct ring_buffer_event *event; | 167 | struct ring_buffer_event *event; |
| 168 | struct ring_buffer *buf; | ||
| 168 | struct hw_branch_entry *entry; | 169 | struct hw_branch_entry *entry; |
| 169 | unsigned long irq1; | 170 | unsigned long irq1; |
| 170 | int cpu; | 171 | int cpu; |
| @@ -180,7 +181,8 @@ void trace_hw_branch(u64 from, u64 to) | |||
| 180 | if (atomic_inc_return(&tr->data[cpu]->disabled) != 1) | 181 | if (atomic_inc_return(&tr->data[cpu]->disabled) != 1) |
| 181 | goto out; | 182 | goto out; |
| 182 | 183 | ||
| 183 | event = trace_buffer_lock_reserve(tr, TRACE_HW_BRANCHES, | 184 | buf = tr->buffer; |
| 185 | event = trace_buffer_lock_reserve(buf, TRACE_HW_BRANCHES, | ||
| 184 | sizeof(*entry), 0, 0); | 186 | sizeof(*entry), 0, 0); |
| 185 | if (!event) | 187 | if (!event) |
| 186 | goto out; | 188 | goto out; |
| @@ -189,8 +191,8 @@ void trace_hw_branch(u64 from, u64 to) | |||
| 189 | entry->ent.type = TRACE_HW_BRANCHES; | 191 | entry->ent.type = TRACE_HW_BRANCHES; |
| 190 | entry->from = from; | 192 | entry->from = from; |
| 191 | entry->to = to; | 193 | entry->to = to; |
| 192 | if (!filter_check_discard(call, entry, tr->buffer, event)) | 194 | if (!filter_check_discard(call, entry, buf, event)) |
| 193 | trace_buffer_unlock_commit(tr, event, 0, 0); | 195 | trace_buffer_unlock_commit(buf, event, 0, 0); |
| 194 | 196 | ||
| 195 | out: | 197 | out: |
| 196 | atomic_dec(&tr->data[cpu]->disabled); | 198 | atomic_dec(&tr->data[cpu]->disabled); |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index f572f44c6e1e..b6c12c6a1bcd 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
| @@ -69,6 +69,9 @@ enum print_line_t trace_print_printk_msg_only(struct trace_iterator *iter) | |||
| 69 | * @s: trace sequence descriptor | 69 | * @s: trace sequence descriptor |
| 70 | * @fmt: printf format string | 70 | * @fmt: printf format string |
| 71 | * | 71 | * |
| 72 | * It returns 0 if the trace oversizes the buffer's free | ||
| 73 | * space, 1 otherwise. | ||
| 74 | * | ||
| 72 | * The tracer may use either sequence operations or its own | 75 | * The tracer may use either sequence operations or its own |
| 73 | * copy to user routines. To simplify formating of a trace | 76 | * copy to user routines. To simplify formating of a trace |
| 74 | * trace_seq_printf is used to store strings into a special | 77 | * trace_seq_printf is used to store strings into a special |
| @@ -95,7 +98,7 @@ trace_seq_printf(struct trace_seq *s, const char *fmt, ...) | |||
| 95 | 98 | ||
| 96 | s->len += ret; | 99 | s->len += ret; |
| 97 | 100 | ||
| 98 | return len; | 101 | return 1; |
| 99 | } | 102 | } |
| 100 | EXPORT_SYMBOL_GPL(trace_seq_printf); | 103 | EXPORT_SYMBOL_GPL(trace_seq_printf); |
| 101 | 104 | ||
| @@ -486,16 +489,18 @@ int trace_print_lat_fmt(struct trace_seq *s, struct trace_entry *entry) | |||
| 486 | hardirq ? 'h' : softirq ? 's' : '.')) | 489 | hardirq ? 'h' : softirq ? 's' : '.')) |
| 487 | return 0; | 490 | return 0; |
| 488 | 491 | ||
| 489 | if (entry->lock_depth < 0) | 492 | if (entry->preempt_count) |
| 490 | ret = trace_seq_putc(s, '.'); | 493 | ret = trace_seq_printf(s, "%x", entry->preempt_count); |
| 491 | else | 494 | else |
| 492 | ret = trace_seq_printf(s, "%d", entry->lock_depth); | 495 | ret = trace_seq_putc(s, '.'); |
| 496 | |||
| 493 | if (!ret) | 497 | if (!ret) |
| 494 | return 0; | 498 | return 0; |
| 495 | 499 | ||
| 496 | if (entry->preempt_count) | 500 | if (entry->lock_depth < 0) |
| 497 | return trace_seq_printf(s, "%x", entry->preempt_count); | 501 | return trace_seq_putc(s, '.'); |
| 498 | return trace_seq_putc(s, '.'); | 502 | |
| 503 | return trace_seq_printf(s, "%d", entry->lock_depth); | ||
| 499 | } | 504 | } |
| 500 | 505 | ||
| 501 | static int | 506 | static int |
| @@ -883,7 +888,7 @@ static int trace_ctxwake_raw(struct trace_iterator *iter, char S) | |||
| 883 | trace_assign_type(field, iter->ent); | 888 | trace_assign_type(field, iter->ent); |
| 884 | 889 | ||
| 885 | if (!S) | 890 | if (!S) |
| 886 | task_state_char(field->prev_state); | 891 | S = task_state_char(field->prev_state); |
| 887 | T = task_state_char(field->next_state); | 892 | T = task_state_char(field->next_state); |
| 888 | if (!trace_seq_printf(&iter->seq, "%d %d %c %d %d %d %c\n", | 893 | if (!trace_seq_printf(&iter->seq, "%d %d %c %d %d %d %c\n", |
| 889 | field->prev_pid, | 894 | field->prev_pid, |
| @@ -918,7 +923,7 @@ static int trace_ctxwake_hex(struct trace_iterator *iter, char S) | |||
| 918 | trace_assign_type(field, iter->ent); | 923 | trace_assign_type(field, iter->ent); |
| 919 | 924 | ||
| 920 | if (!S) | 925 | if (!S) |
| 921 | task_state_char(field->prev_state); | 926 | S = task_state_char(field->prev_state); |
| 922 | T = task_state_char(field->next_state); | 927 | T = task_state_char(field->next_state); |
| 923 | 928 | ||
| 924 | SEQ_PUT_HEX_FIELD_RET(s, field->prev_pid); | 929 | SEQ_PUT_HEX_FIELD_RET(s, field->prev_pid); |
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index 9fbce6c9d2e1..527e17eae575 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
| @@ -166,7 +166,7 @@ int syscall_exit_format(struct ftrace_event_call *call, struct trace_seq *s) | |||
| 166 | "\tfield:%s %s;\toffset:%zu;\tsize:%zu;\n" | 166 | "\tfield:%s %s;\toffset:%zu;\tsize:%zu;\n" |
| 167 | "\tfield:%s %s;\toffset:%zu;\tsize:%zu;\n", | 167 | "\tfield:%s %s;\toffset:%zu;\tsize:%zu;\n", |
| 168 | SYSCALL_FIELD(int, nr), | 168 | SYSCALL_FIELD(int, nr), |
| 169 | SYSCALL_FIELD(unsigned long, ret)); | 169 | SYSCALL_FIELD(long, ret)); |
| 170 | if (!ret) | 170 | if (!ret) |
| 171 | return 0; | 171 | return 0; |
| 172 | 172 | ||
| @@ -212,7 +212,7 @@ int syscall_exit_define_fields(struct ftrace_event_call *call) | |||
| 212 | if (ret) | 212 | if (ret) |
| 213 | return ret; | 213 | return ret; |
| 214 | 214 | ||
| 215 | ret = trace_define_field(call, SYSCALL_FIELD(unsigned long, ret), 0, | 215 | ret = trace_define_field(call, SYSCALL_FIELD(long, ret), 0, |
| 216 | FILTER_OTHER); | 216 | FILTER_OTHER); |
| 217 | 217 | ||
| 218 | return ret; | 218 | return ret; |
