diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/rcupdate.c | 2 | ||||
| -rw-r--r-- | kernel/trace/ftrace.c | 17 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 37 | ||||
| -rw-r--r-- | kernel/trace/trace.h | 1 | ||||
| -rw-r--r-- | kernel/trace/trace_events.c | 207 | ||||
| -rw-r--r-- | kernel/trace/trace_syscalls.c | 10 |
6 files changed, 65 insertions, 209 deletions
diff --git a/kernel/rcupdate.c b/kernel/rcupdate.c index 33eb4620aa17..b02a339836b4 100644 --- a/kernel/rcupdate.c +++ b/kernel/rcupdate.c | |||
| @@ -122,7 +122,7 @@ struct lockdep_map rcu_sched_lock_map = | |||
| 122 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_sched", &rcu_sched_lock_key); | 122 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_sched", &rcu_sched_lock_key); |
| 123 | EXPORT_SYMBOL_GPL(rcu_sched_lock_map); | 123 | EXPORT_SYMBOL_GPL(rcu_sched_lock_map); |
| 124 | 124 | ||
| 125 | int debug_lockdep_rcu_enabled(void) | 125 | int notrace debug_lockdep_rcu_enabled(void) |
| 126 | { | 126 | { |
| 127 | return rcu_scheduler_active && debug_locks && | 127 | return rcu_scheduler_active && debug_locks && |
| 128 | current->lockdep_recursion == 0; | 128 | current->lockdep_recursion == 0; |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index a6d098c6df3f..03cf44ac54d3 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -1978,12 +1978,27 @@ int __weak ftrace_arch_code_modify_post_process(void) | |||
| 1978 | 1978 | ||
| 1979 | void ftrace_modify_all_code(int command) | 1979 | void ftrace_modify_all_code(int command) |
| 1980 | { | 1980 | { |
| 1981 | int update = command & FTRACE_UPDATE_TRACE_FUNC; | ||
| 1982 | |||
| 1983 | /* | ||
| 1984 | * If the ftrace_caller calls a ftrace_ops func directly, | ||
| 1985 | * we need to make sure that it only traces functions it | ||
| 1986 | * expects to trace. When doing the switch of functions, | ||
| 1987 | * we need to update to the ftrace_ops_list_func first | ||
| 1988 | * before the transition between old and new calls are set, | ||
| 1989 | * as the ftrace_ops_list_func will check the ops hashes | ||
| 1990 | * to make sure the ops are having the right functions | ||
| 1991 | * traced. | ||
| 1992 | */ | ||
| 1993 | if (update) | ||
| 1994 | ftrace_update_ftrace_func(ftrace_ops_list_func); | ||
| 1995 | |||
| 1981 | if (command & FTRACE_UPDATE_CALLS) | 1996 | if (command & FTRACE_UPDATE_CALLS) |
| 1982 | ftrace_replace_code(1); | 1997 | ftrace_replace_code(1); |
| 1983 | else if (command & FTRACE_DISABLE_CALLS) | 1998 | else if (command & FTRACE_DISABLE_CALLS) |
| 1984 | ftrace_replace_code(0); | 1999 | ftrace_replace_code(0); |
| 1985 | 2000 | ||
| 1986 | if (command & FTRACE_UPDATE_TRACE_FUNC) | 2001 | if (update && ftrace_trace_function != ftrace_ops_list_func) |
| 1987 | ftrace_update_ftrace_func(ftrace_trace_function); | 2002 | ftrace_update_ftrace_func(ftrace_trace_function); |
| 1988 | 2003 | ||
| 1989 | if (command & FTRACE_START_FUNC_RET) | 2004 | if (command & FTRACE_START_FUNC_RET) |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 496f94d57698..7974ba20557d 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -3166,11 +3166,6 @@ static const struct file_operations show_traces_fops = { | |||
| 3166 | }; | 3166 | }; |
| 3167 | 3167 | ||
| 3168 | /* | 3168 | /* |
| 3169 | * Only trace on a CPU if the bitmask is set: | ||
| 3170 | */ | ||
| 3171 | static cpumask_var_t tracing_cpumask; | ||
| 3172 | |||
| 3173 | /* | ||
| 3174 | * The tracer itself will not take this lock, but still we want | 3169 | * The tracer itself will not take this lock, but still we want |
| 3175 | * to provide a consistent cpumask to user-space: | 3170 | * to provide a consistent cpumask to user-space: |
| 3176 | */ | 3171 | */ |
| @@ -3186,11 +3181,12 @@ static ssize_t | |||
| 3186 | tracing_cpumask_read(struct file *filp, char __user *ubuf, | 3181 | tracing_cpumask_read(struct file *filp, char __user *ubuf, |
| 3187 | size_t count, loff_t *ppos) | 3182 | size_t count, loff_t *ppos) |
| 3188 | { | 3183 | { |
| 3184 | struct trace_array *tr = file_inode(filp)->i_private; | ||
| 3189 | int len; | 3185 | int len; |
| 3190 | 3186 | ||
| 3191 | mutex_lock(&tracing_cpumask_update_lock); | 3187 | mutex_lock(&tracing_cpumask_update_lock); |
| 3192 | 3188 | ||
| 3193 | len = cpumask_scnprintf(mask_str, count, tracing_cpumask); | 3189 | len = cpumask_scnprintf(mask_str, count, tr->tracing_cpumask); |
| 3194 | if (count - len < 2) { | 3190 | if (count - len < 2) { |
| 3195 | count = -EINVAL; | 3191 | count = -EINVAL; |
| 3196 | goto out_err; | 3192 | goto out_err; |
| @@ -3208,7 +3204,7 @@ static ssize_t | |||
| 3208 | tracing_cpumask_write(struct file *filp, const char __user *ubuf, | 3204 | tracing_cpumask_write(struct file *filp, const char __user *ubuf, |
| 3209 | size_t count, loff_t *ppos) | 3205 | size_t count, loff_t *ppos) |
| 3210 | { | 3206 | { |
| 3211 | struct trace_array *tr = filp->private_data; | 3207 | struct trace_array *tr = file_inode(filp)->i_private; |
| 3212 | cpumask_var_t tracing_cpumask_new; | 3208 | cpumask_var_t tracing_cpumask_new; |
| 3213 | int err, cpu; | 3209 | int err, cpu; |
| 3214 | 3210 | ||
| @@ -3228,12 +3224,12 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf, | |||
| 3228 | * Increase/decrease the disabled counter if we are | 3224 | * Increase/decrease the disabled counter if we are |
| 3229 | * about to flip a bit in the cpumask: | 3225 | * about to flip a bit in the cpumask: |
| 3230 | */ | 3226 | */ |
| 3231 | if (cpumask_test_cpu(cpu, tracing_cpumask) && | 3227 | if (cpumask_test_cpu(cpu, tr->tracing_cpumask) && |
| 3232 | !cpumask_test_cpu(cpu, tracing_cpumask_new)) { | 3228 | !cpumask_test_cpu(cpu, tracing_cpumask_new)) { |
| 3233 | atomic_inc(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); | 3229 | atomic_inc(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); |
| 3234 | ring_buffer_record_disable_cpu(tr->trace_buffer.buffer, cpu); | 3230 | ring_buffer_record_disable_cpu(tr->trace_buffer.buffer, cpu); |
| 3235 | } | 3231 | } |
| 3236 | if (!cpumask_test_cpu(cpu, tracing_cpumask) && | 3232 | if (!cpumask_test_cpu(cpu, tr->tracing_cpumask) && |
| 3237 | cpumask_test_cpu(cpu, tracing_cpumask_new)) { | 3233 | cpumask_test_cpu(cpu, tracing_cpumask_new)) { |
| 3238 | atomic_dec(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); | 3234 | atomic_dec(&per_cpu_ptr(tr->trace_buffer.data, cpu)->disabled); |
| 3239 | ring_buffer_record_enable_cpu(tr->trace_buffer.buffer, cpu); | 3235 | ring_buffer_record_enable_cpu(tr->trace_buffer.buffer, cpu); |
| @@ -3242,7 +3238,7 @@ tracing_cpumask_write(struct file *filp, const char __user *ubuf, | |||
| 3242 | arch_spin_unlock(&ftrace_max_lock); | 3238 | arch_spin_unlock(&ftrace_max_lock); |
| 3243 | local_irq_enable(); | 3239 | local_irq_enable(); |
| 3244 | 3240 | ||
| 3245 | cpumask_copy(tracing_cpumask, tracing_cpumask_new); | 3241 | cpumask_copy(tr->tracing_cpumask, tracing_cpumask_new); |
| 3246 | 3242 | ||
| 3247 | mutex_unlock(&tracing_cpumask_update_lock); | 3243 | mutex_unlock(&tracing_cpumask_update_lock); |
| 3248 | free_cpumask_var(tracing_cpumask_new); | 3244 | free_cpumask_var(tracing_cpumask_new); |
| @@ -3256,9 +3252,10 @@ err_unlock: | |||
| 3256 | } | 3252 | } |
| 3257 | 3253 | ||
| 3258 | static const struct file_operations tracing_cpumask_fops = { | 3254 | static const struct file_operations tracing_cpumask_fops = { |
| 3259 | .open = tracing_open_generic, | 3255 | .open = tracing_open_generic_tr, |
| 3260 | .read = tracing_cpumask_read, | 3256 | .read = tracing_cpumask_read, |
| 3261 | .write = tracing_cpumask_write, | 3257 | .write = tracing_cpumask_write, |
| 3258 | .release = tracing_release_generic_tr, | ||
| 3262 | .llseek = generic_file_llseek, | 3259 | .llseek = generic_file_llseek, |
| 3263 | }; | 3260 | }; |
| 3264 | 3261 | ||
| @@ -5938,6 +5935,11 @@ static int new_instance_create(const char *name) | |||
| 5938 | if (!tr->name) | 5935 | if (!tr->name) |
| 5939 | goto out_free_tr; | 5936 | goto out_free_tr; |
| 5940 | 5937 | ||
| 5938 | if (!alloc_cpumask_var(&tr->tracing_cpumask, GFP_KERNEL)) | ||
| 5939 | goto out_free_tr; | ||
| 5940 | |||
| 5941 | cpumask_copy(tr->tracing_cpumask, cpu_all_mask); | ||
| 5942 | |||
| 5941 | raw_spin_lock_init(&tr->start_lock); | 5943 | raw_spin_lock_init(&tr->start_lock); |
| 5942 | 5944 | ||
| 5943 | tr->current_trace = &nop_trace; | 5945 | tr->current_trace = &nop_trace; |
| @@ -5969,6 +5971,7 @@ static int new_instance_create(const char *name) | |||
| 5969 | out_free_tr: | 5971 | out_free_tr: |
| 5970 | if (tr->trace_buffer.buffer) | 5972 | if (tr->trace_buffer.buffer) |
| 5971 | ring_buffer_free(tr->trace_buffer.buffer); | 5973 | ring_buffer_free(tr->trace_buffer.buffer); |
| 5974 | free_cpumask_var(tr->tracing_cpumask); | ||
| 5972 | kfree(tr->name); | 5975 | kfree(tr->name); |
| 5973 | kfree(tr); | 5976 | kfree(tr); |
| 5974 | 5977 | ||
| @@ -6098,6 +6101,9 @@ init_tracer_debugfs(struct trace_array *tr, struct dentry *d_tracer) | |||
| 6098 | { | 6101 | { |
| 6099 | int cpu; | 6102 | int cpu; |
| 6100 | 6103 | ||
| 6104 | trace_create_file("tracing_cpumask", 0644, d_tracer, | ||
| 6105 | tr, &tracing_cpumask_fops); | ||
| 6106 | |||
| 6101 | trace_create_file("trace_options", 0644, d_tracer, | 6107 | trace_create_file("trace_options", 0644, d_tracer, |
| 6102 | tr, &tracing_iter_fops); | 6108 | tr, &tracing_iter_fops); |
| 6103 | 6109 | ||
| @@ -6147,9 +6153,6 @@ static __init int tracer_init_debugfs(void) | |||
| 6147 | 6153 | ||
| 6148 | init_tracer_debugfs(&global_trace, d_tracer); | 6154 | init_tracer_debugfs(&global_trace, d_tracer); |
| 6149 | 6155 | ||
| 6150 | trace_create_file("tracing_cpumask", 0644, d_tracer, | ||
| 6151 | &global_trace, &tracing_cpumask_fops); | ||
| 6152 | |||
| 6153 | trace_create_file("available_tracers", 0444, d_tracer, | 6156 | trace_create_file("available_tracers", 0444, d_tracer, |
| 6154 | &global_trace, &show_traces_fops); | 6157 | &global_trace, &show_traces_fops); |
| 6155 | 6158 | ||
| @@ -6371,7 +6374,7 @@ __init static int tracer_alloc_buffers(void) | |||
| 6371 | if (!alloc_cpumask_var(&tracing_buffer_mask, GFP_KERNEL)) | 6374 | if (!alloc_cpumask_var(&tracing_buffer_mask, GFP_KERNEL)) |
| 6372 | goto out; | 6375 | goto out; |
| 6373 | 6376 | ||
| 6374 | if (!alloc_cpumask_var(&tracing_cpumask, GFP_KERNEL)) | 6377 | if (!alloc_cpumask_var(&global_trace.tracing_cpumask, GFP_KERNEL)) |
| 6375 | goto out_free_buffer_mask; | 6378 | goto out_free_buffer_mask; |
| 6376 | 6379 | ||
| 6377 | /* Only allocate trace_printk buffers if a trace_printk exists */ | 6380 | /* Only allocate trace_printk buffers if a trace_printk exists */ |
| @@ -6386,7 +6389,7 @@ __init static int tracer_alloc_buffers(void) | |||
| 6386 | ring_buf_size = 1; | 6389 | ring_buf_size = 1; |
| 6387 | 6390 | ||
| 6388 | cpumask_copy(tracing_buffer_mask, cpu_possible_mask); | 6391 | cpumask_copy(tracing_buffer_mask, cpu_possible_mask); |
| 6389 | cpumask_copy(tracing_cpumask, cpu_all_mask); | 6392 | cpumask_copy(global_trace.tracing_cpumask, cpu_all_mask); |
| 6390 | 6393 | ||
| 6391 | raw_spin_lock_init(&global_trace.start_lock); | 6394 | raw_spin_lock_init(&global_trace.start_lock); |
| 6392 | 6395 | ||
| @@ -6441,7 +6444,7 @@ out_free_cpumask: | |||
| 6441 | #ifdef CONFIG_TRACER_MAX_TRACE | 6444 | #ifdef CONFIG_TRACER_MAX_TRACE |
| 6442 | free_percpu(global_trace.max_buffer.data); | 6445 | free_percpu(global_trace.max_buffer.data); |
| 6443 | #endif | 6446 | #endif |
| 6444 | free_cpumask_var(tracing_cpumask); | 6447 | free_cpumask_var(global_trace.tracing_cpumask); |
| 6445 | out_free_buffer_mask: | 6448 | out_free_buffer_mask: |
| 6446 | free_cpumask_var(tracing_buffer_mask); | 6449 | free_cpumask_var(tracing_buffer_mask); |
| 6447 | out: | 6450 | out: |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index fe39acd4c1aa..10c86fb7a2b4 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
| @@ -206,6 +206,7 @@ struct trace_array { | |||
| 206 | struct dentry *event_dir; | 206 | struct dentry *event_dir; |
| 207 | struct list_head systems; | 207 | struct list_head systems; |
| 208 | struct list_head events; | 208 | struct list_head events; |
| 209 | cpumask_var_t tracing_cpumask; /* only trace on set CPUs */ | ||
| 209 | int ref; | 210 | int ref; |
| 210 | }; | 211 | }; |
| 211 | 212 | ||
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index 29a7ebcfb426..368a4d50cc30 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
| @@ -1489,12 +1489,7 @@ event_subsystem_dir(struct trace_array *tr, const char *name, | |||
| 1489 | } | 1489 | } |
| 1490 | 1490 | ||
| 1491 | static int | 1491 | static int |
| 1492 | event_create_dir(struct dentry *parent, | 1492 | event_create_dir(struct dentry *parent, struct ftrace_event_file *file) |
| 1493 | struct ftrace_event_file *file, | ||
| 1494 | const struct file_operations *id, | ||
| 1495 | const struct file_operations *enable, | ||
| 1496 | const struct file_operations *filter, | ||
| 1497 | const struct file_operations *format) | ||
| 1498 | { | 1493 | { |
| 1499 | struct ftrace_event_call *call = file->event_call; | 1494 | struct ftrace_event_call *call = file->event_call; |
| 1500 | struct trace_array *tr = file->tr; | 1495 | struct trace_array *tr = file->tr; |
| @@ -1522,12 +1517,13 @@ event_create_dir(struct dentry *parent, | |||
| 1522 | 1517 | ||
| 1523 | if (call->class->reg && !(call->flags & TRACE_EVENT_FL_IGNORE_ENABLE)) | 1518 | if (call->class->reg && !(call->flags & TRACE_EVENT_FL_IGNORE_ENABLE)) |
| 1524 | trace_create_file("enable", 0644, file->dir, file, | 1519 | trace_create_file("enable", 0644, file->dir, file, |
| 1525 | enable); | 1520 | &ftrace_enable_fops); |
| 1526 | 1521 | ||
| 1527 | #ifdef CONFIG_PERF_EVENTS | 1522 | #ifdef CONFIG_PERF_EVENTS |
| 1528 | if (call->event.type && call->class->reg) | 1523 | if (call->event.type && call->class->reg) |
| 1529 | trace_create_file("id", 0444, file->dir, | 1524 | trace_create_file("id", 0444, file->dir, |
| 1530 | (void *)(long)call->event.type, id); | 1525 | (void *)(long)call->event.type, |
| 1526 | &ftrace_event_id_fops); | ||
| 1531 | #endif | 1527 | #endif |
| 1532 | 1528 | ||
| 1533 | /* | 1529 | /* |
| @@ -1544,10 +1540,10 @@ event_create_dir(struct dentry *parent, | |||
| 1544 | } | 1540 | } |
| 1545 | } | 1541 | } |
| 1546 | trace_create_file("filter", 0644, file->dir, call, | 1542 | trace_create_file("filter", 0644, file->dir, call, |
| 1547 | filter); | 1543 | &ftrace_event_filter_fops); |
| 1548 | 1544 | ||
| 1549 | trace_create_file("format", 0444, file->dir, call, | 1545 | trace_create_file("format", 0444, file->dir, call, |
| 1550 | format); | 1546 | &ftrace_event_format_fops); |
| 1551 | 1547 | ||
| 1552 | return 0; | 1548 | return 0; |
| 1553 | } | 1549 | } |
| @@ -1648,12 +1644,7 @@ trace_create_new_event(struct ftrace_event_call *call, | |||
| 1648 | 1644 | ||
| 1649 | /* Add an event to a trace directory */ | 1645 | /* Add an event to a trace directory */ |
| 1650 | static int | 1646 | static int |
| 1651 | __trace_add_new_event(struct ftrace_event_call *call, | 1647 | __trace_add_new_event(struct ftrace_event_call *call, struct trace_array *tr) |
| 1652 | struct trace_array *tr, | ||
| 1653 | const struct file_operations *id, | ||
| 1654 | const struct file_operations *enable, | ||
| 1655 | const struct file_operations *filter, | ||
| 1656 | const struct file_operations *format) | ||
| 1657 | { | 1648 | { |
| 1658 | struct ftrace_event_file *file; | 1649 | struct ftrace_event_file *file; |
| 1659 | 1650 | ||
| @@ -1661,7 +1652,7 @@ __trace_add_new_event(struct ftrace_event_call *call, | |||
| 1661 | if (!file) | 1652 | if (!file) |
| 1662 | return -ENOMEM; | 1653 | return -ENOMEM; |
| 1663 | 1654 | ||
| 1664 | return event_create_dir(tr->event_dir, file, id, enable, filter, format); | 1655 | return event_create_dir(tr->event_dir, file); |
| 1665 | } | 1656 | } |
| 1666 | 1657 | ||
| 1667 | /* | 1658 | /* |
| @@ -1683,8 +1674,7 @@ __trace_early_add_new_event(struct ftrace_event_call *call, | |||
| 1683 | } | 1674 | } |
| 1684 | 1675 | ||
| 1685 | struct ftrace_module_file_ops; | 1676 | struct ftrace_module_file_ops; |
| 1686 | static void __add_event_to_tracers(struct ftrace_event_call *call, | 1677 | static void __add_event_to_tracers(struct ftrace_event_call *call); |
| 1687 | struct ftrace_module_file_ops *file_ops); | ||
| 1688 | 1678 | ||
| 1689 | /* Add an additional event_call dynamically */ | 1679 | /* Add an additional event_call dynamically */ |
| 1690 | int trace_add_event_call(struct ftrace_event_call *call) | 1680 | int trace_add_event_call(struct ftrace_event_call *call) |
| @@ -1695,7 +1685,7 @@ int trace_add_event_call(struct ftrace_event_call *call) | |||
| 1695 | 1685 | ||
| 1696 | ret = __register_event(call, NULL); | 1686 | ret = __register_event(call, NULL); |
| 1697 | if (ret >= 0) | 1687 | if (ret >= 0) |
| 1698 | __add_event_to_tracers(call, NULL); | 1688 | __add_event_to_tracers(call); |
| 1699 | 1689 | ||
| 1700 | mutex_unlock(&event_mutex); | 1690 | mutex_unlock(&event_mutex); |
| 1701 | mutex_unlock(&trace_types_lock); | 1691 | mutex_unlock(&trace_types_lock); |
| @@ -1769,100 +1759,21 @@ int trace_remove_event_call(struct ftrace_event_call *call) | |||
| 1769 | 1759 | ||
| 1770 | #ifdef CONFIG_MODULES | 1760 | #ifdef CONFIG_MODULES |
| 1771 | 1761 | ||
| 1772 | static LIST_HEAD(ftrace_module_file_list); | ||
| 1773 | |||
| 1774 | /* | ||
| 1775 | * Modules must own their file_operations to keep up with | ||
| 1776 | * reference counting. | ||
| 1777 | */ | ||
| 1778 | struct ftrace_module_file_ops { | ||
| 1779 | struct list_head list; | ||
| 1780 | struct module *mod; | ||
| 1781 | struct file_operations id; | ||
| 1782 | struct file_operations enable; | ||
| 1783 | struct file_operations format; | ||
| 1784 | struct file_operations filter; | ||
| 1785 | }; | ||
| 1786 | |||
| 1787 | static struct ftrace_module_file_ops * | ||
| 1788 | find_ftrace_file_ops(struct ftrace_module_file_ops *file_ops, struct module *mod) | ||
| 1789 | { | ||
| 1790 | /* | ||
| 1791 | * As event_calls are added in groups by module, | ||
| 1792 | * when we find one file_ops, we don't need to search for | ||
| 1793 | * each call in that module, as the rest should be the | ||
| 1794 | * same. Only search for a new one if the last one did | ||
| 1795 | * not match. | ||
| 1796 | */ | ||
| 1797 | if (file_ops && mod == file_ops->mod) | ||
| 1798 | return file_ops; | ||
| 1799 | |||
| 1800 | list_for_each_entry(file_ops, &ftrace_module_file_list, list) { | ||
| 1801 | if (file_ops->mod == mod) | ||
| 1802 | return file_ops; | ||
| 1803 | } | ||
| 1804 | return NULL; | ||
| 1805 | } | ||
| 1806 | |||
| 1807 | static struct ftrace_module_file_ops * | ||
| 1808 | trace_create_file_ops(struct module *mod) | ||
| 1809 | { | ||
| 1810 | struct ftrace_module_file_ops *file_ops; | ||
| 1811 | |||
| 1812 | /* | ||
| 1813 | * This is a bit of a PITA. To allow for correct reference | ||
| 1814 | * counting, modules must "own" their file_operations. | ||
| 1815 | * To do this, we allocate the file operations that will be | ||
| 1816 | * used in the event directory. | ||
| 1817 | */ | ||
| 1818 | |||
| 1819 | file_ops = kmalloc(sizeof(*file_ops), GFP_KERNEL); | ||
| 1820 | if (!file_ops) | ||
| 1821 | return NULL; | ||
| 1822 | |||
| 1823 | file_ops->mod = mod; | ||
| 1824 | |||
| 1825 | file_ops->id = ftrace_event_id_fops; | ||
| 1826 | file_ops->id.owner = mod; | ||
| 1827 | |||
| 1828 | file_ops->enable = ftrace_enable_fops; | ||
| 1829 | file_ops->enable.owner = mod; | ||
| 1830 | |||
| 1831 | file_ops->filter = ftrace_event_filter_fops; | ||
| 1832 | file_ops->filter.owner = mod; | ||
| 1833 | |||
| 1834 | file_ops->format = ftrace_event_format_fops; | ||
| 1835 | file_ops->format.owner = mod; | ||
| 1836 | |||
| 1837 | list_add(&file_ops->list, &ftrace_module_file_list); | ||
| 1838 | |||
| 1839 | return file_ops; | ||
| 1840 | } | ||
| 1841 | |||
| 1842 | static void trace_module_add_events(struct module *mod) | 1762 | static void trace_module_add_events(struct module *mod) |
| 1843 | { | 1763 | { |
| 1844 | struct ftrace_module_file_ops *file_ops = NULL; | ||
| 1845 | struct ftrace_event_call **call, **start, **end; | 1764 | struct ftrace_event_call **call, **start, **end; |
| 1846 | 1765 | ||
| 1847 | start = mod->trace_events; | 1766 | start = mod->trace_events; |
| 1848 | end = mod->trace_events + mod->num_trace_events; | 1767 | end = mod->trace_events + mod->num_trace_events; |
| 1849 | 1768 | ||
| 1850 | if (start == end) | ||
| 1851 | return; | ||
| 1852 | |||
| 1853 | file_ops = trace_create_file_ops(mod); | ||
| 1854 | if (!file_ops) | ||
| 1855 | return; | ||
| 1856 | |||
| 1857 | for_each_event(call, start, end) { | 1769 | for_each_event(call, start, end) { |
| 1858 | __register_event(*call, mod); | 1770 | __register_event(*call, mod); |
| 1859 | __add_event_to_tracers(*call, file_ops); | 1771 | __add_event_to_tracers(*call); |
| 1860 | } | 1772 | } |
| 1861 | } | 1773 | } |
| 1862 | 1774 | ||
| 1863 | static void trace_module_remove_events(struct module *mod) | 1775 | static void trace_module_remove_events(struct module *mod) |
| 1864 | { | 1776 | { |
| 1865 | struct ftrace_module_file_ops *file_ops; | ||
| 1866 | struct ftrace_event_call *call, *p; | 1777 | struct ftrace_event_call *call, *p; |
| 1867 | bool clear_trace = false; | 1778 | bool clear_trace = false; |
| 1868 | 1779 | ||
| @@ -1874,16 +1785,6 @@ static void trace_module_remove_events(struct module *mod) | |||
| 1874 | __trace_remove_event_call(call); | 1785 | __trace_remove_event_call(call); |
| 1875 | } | 1786 | } |
| 1876 | } | 1787 | } |
| 1877 | |||
| 1878 | /* Now free the file_operations */ | ||
| 1879 | list_for_each_entry(file_ops, &ftrace_module_file_list, list) { | ||
| 1880 | if (file_ops->mod == mod) | ||
| 1881 | break; | ||
| 1882 | } | ||
| 1883 | if (&file_ops->list != &ftrace_module_file_list) { | ||
| 1884 | list_del(&file_ops->list); | ||
| 1885 | kfree(file_ops); | ||
| 1886 | } | ||
| 1887 | up_write(&trace_event_sem); | 1788 | up_write(&trace_event_sem); |
| 1888 | 1789 | ||
| 1889 | /* | 1790 | /* |
| @@ -1919,67 +1820,21 @@ static int trace_module_notify(struct notifier_block *self, | |||
| 1919 | return 0; | 1820 | return 0; |
| 1920 | } | 1821 | } |
| 1921 | 1822 | ||
| 1922 | static int | 1823 | static struct notifier_block trace_module_nb = { |
| 1923 | __trace_add_new_mod_event(struct ftrace_event_call *call, | 1824 | .notifier_call = trace_module_notify, |
| 1924 | struct trace_array *tr, | 1825 | .priority = 0, |
| 1925 | struct ftrace_module_file_ops *file_ops) | 1826 | }; |
| 1926 | { | ||
| 1927 | return __trace_add_new_event(call, tr, | ||
| 1928 | &file_ops->id, &file_ops->enable, | ||
| 1929 | &file_ops->filter, &file_ops->format); | ||
| 1930 | } | ||
| 1931 | |||
| 1932 | #else | ||
| 1933 | static inline struct ftrace_module_file_ops * | ||
| 1934 | find_ftrace_file_ops(struct ftrace_module_file_ops *file_ops, struct module *mod) | ||
| 1935 | { | ||
| 1936 | return NULL; | ||
| 1937 | } | ||
| 1938 | static inline int trace_module_notify(struct notifier_block *self, | ||
| 1939 | unsigned long val, void *data) | ||
| 1940 | { | ||
| 1941 | return 0; | ||
| 1942 | } | ||
| 1943 | static inline int | ||
| 1944 | __trace_add_new_mod_event(struct ftrace_event_call *call, | ||
| 1945 | struct trace_array *tr, | ||
| 1946 | struct ftrace_module_file_ops *file_ops) | ||
| 1947 | { | ||
| 1948 | return -ENODEV; | ||
| 1949 | } | ||
| 1950 | #endif /* CONFIG_MODULES */ | 1827 | #endif /* CONFIG_MODULES */ |
| 1951 | 1828 | ||
| 1952 | /* Create a new event directory structure for a trace directory. */ | 1829 | /* Create a new event directory structure for a trace directory. */ |
| 1953 | static void | 1830 | static void |
| 1954 | __trace_add_event_dirs(struct trace_array *tr) | 1831 | __trace_add_event_dirs(struct trace_array *tr) |
| 1955 | { | 1832 | { |
| 1956 | struct ftrace_module_file_ops *file_ops = NULL; | ||
| 1957 | struct ftrace_event_call *call; | 1833 | struct ftrace_event_call *call; |
| 1958 | int ret; | 1834 | int ret; |
| 1959 | 1835 | ||
| 1960 | list_for_each_entry(call, &ftrace_events, list) { | 1836 | list_for_each_entry(call, &ftrace_events, list) { |
| 1961 | if (call->mod) { | 1837 | ret = __trace_add_new_event(call, tr); |
| 1962 | /* | ||
| 1963 | * Directories for events by modules need to | ||
| 1964 | * keep module ref counts when opened (as we don't | ||
| 1965 | * want the module to disappear when reading one | ||
| 1966 | * of these files). The file_ops keep account of | ||
| 1967 | * the module ref count. | ||
| 1968 | */ | ||
| 1969 | file_ops = find_ftrace_file_ops(file_ops, call->mod); | ||
| 1970 | if (!file_ops) | ||
| 1971 | continue; /* Warn? */ | ||
| 1972 | ret = __trace_add_new_mod_event(call, tr, file_ops); | ||
| 1973 | if (ret < 0) | ||
| 1974 | pr_warning("Could not create directory for event %s\n", | ||
| 1975 | call->name); | ||
| 1976 | continue; | ||
| 1977 | } | ||
| 1978 | ret = __trace_add_new_event(call, tr, | ||
| 1979 | &ftrace_event_id_fops, | ||
| 1980 | &ftrace_enable_fops, | ||
| 1981 | &ftrace_event_filter_fops, | ||
| 1982 | &ftrace_event_format_fops); | ||
| 1983 | if (ret < 0) | 1838 | if (ret < 0) |
| 1984 | pr_warning("Could not create directory for event %s\n", | 1839 | pr_warning("Could not create directory for event %s\n", |
| 1985 | call->name); | 1840 | call->name); |
| @@ -2287,11 +2142,7 @@ __trace_early_add_event_dirs(struct trace_array *tr) | |||
| 2287 | 2142 | ||
| 2288 | 2143 | ||
| 2289 | list_for_each_entry(file, &tr->events, list) { | 2144 | list_for_each_entry(file, &tr->events, list) { |
| 2290 | ret = event_create_dir(tr->event_dir, file, | 2145 | ret = event_create_dir(tr->event_dir, file); |
| 2291 | &ftrace_event_id_fops, | ||
| 2292 | &ftrace_enable_fops, | ||
| 2293 | &ftrace_event_filter_fops, | ||
| 2294 | &ftrace_event_format_fops); | ||
| 2295 | if (ret < 0) | 2146 | if (ret < 0) |
| 2296 | pr_warning("Could not create directory for event %s\n", | 2147 | pr_warning("Could not create directory for event %s\n", |
| 2297 | file->event_call->name); | 2148 | file->event_call->name); |
| @@ -2332,29 +2183,14 @@ __trace_remove_event_dirs(struct trace_array *tr) | |||
| 2332 | remove_event_file_dir(file); | 2183 | remove_event_file_dir(file); |
| 2333 | } | 2184 | } |
| 2334 | 2185 | ||
| 2335 | static void | 2186 | static void __add_event_to_tracers(struct ftrace_event_call *call) |
| 2336 | __add_event_to_tracers(struct ftrace_event_call *call, | ||
| 2337 | struct ftrace_module_file_ops *file_ops) | ||
| 2338 | { | 2187 | { |
| 2339 | struct trace_array *tr; | 2188 | struct trace_array *tr; |
| 2340 | 2189 | ||
| 2341 | list_for_each_entry(tr, &ftrace_trace_arrays, list) { | 2190 | list_for_each_entry(tr, &ftrace_trace_arrays, list) |
| 2342 | if (file_ops) | 2191 | __trace_add_new_event(call, tr); |
| 2343 | __trace_add_new_mod_event(call, tr, file_ops); | ||
| 2344 | else | ||
| 2345 | __trace_add_new_event(call, tr, | ||
| 2346 | &ftrace_event_id_fops, | ||
| 2347 | &ftrace_enable_fops, | ||
| 2348 | &ftrace_event_filter_fops, | ||
| 2349 | &ftrace_event_format_fops); | ||
| 2350 | } | ||
| 2351 | } | 2192 | } |
| 2352 | 2193 | ||
| 2353 | static struct notifier_block trace_module_nb = { | ||
| 2354 | .notifier_call = trace_module_notify, | ||
| 2355 | .priority = 0, | ||
| 2356 | }; | ||
| 2357 | |||
| 2358 | extern struct ftrace_event_call *__start_ftrace_events[]; | 2194 | extern struct ftrace_event_call *__start_ftrace_events[]; |
| 2359 | extern struct ftrace_event_call *__stop_ftrace_events[]; | 2195 | extern struct ftrace_event_call *__stop_ftrace_events[]; |
| 2360 | 2196 | ||
| @@ -2559,10 +2395,11 @@ static __init int event_trace_init(void) | |||
| 2559 | if (ret) | 2395 | if (ret) |
| 2560 | return ret; | 2396 | return ret; |
| 2561 | 2397 | ||
| 2398 | #ifdef CONFIG_MODULES | ||
| 2562 | ret = register_module_notifier(&trace_module_nb); | 2399 | ret = register_module_notifier(&trace_module_nb); |
| 2563 | if (ret) | 2400 | if (ret) |
| 2564 | pr_warning("Failed to register trace events module notifier\n"); | 2401 | pr_warning("Failed to register trace events module notifier\n"); |
| 2565 | 2402 | #endif | |
| 2566 | return 0; | 2403 | return 0; |
| 2567 | } | 2404 | } |
| 2568 | early_initcall(event_trace_memsetup); | 2405 | early_initcall(event_trace_memsetup); |
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index 8fd03657bc7d..559329d9bd2f 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
| @@ -200,8 +200,8 @@ extern char *__bad_type_size(void); | |||
| 200 | #type, #name, offsetof(typeof(trace), name), \ | 200 | #type, #name, offsetof(typeof(trace), name), \ |
| 201 | sizeof(trace.name), is_signed_type(type) | 201 | sizeof(trace.name), is_signed_type(type) |
| 202 | 202 | ||
| 203 | static | 203 | static int __init |
| 204 | int __set_enter_print_fmt(struct syscall_metadata *entry, char *buf, int len) | 204 | __set_enter_print_fmt(struct syscall_metadata *entry, char *buf, int len) |
| 205 | { | 205 | { |
| 206 | int i; | 206 | int i; |
| 207 | int pos = 0; | 207 | int pos = 0; |
| @@ -228,7 +228,7 @@ int __set_enter_print_fmt(struct syscall_metadata *entry, char *buf, int len) | |||
| 228 | return pos; | 228 | return pos; |
| 229 | } | 229 | } |
| 230 | 230 | ||
| 231 | static int set_syscall_print_fmt(struct ftrace_event_call *call) | 231 | static int __init set_syscall_print_fmt(struct ftrace_event_call *call) |
| 232 | { | 232 | { |
| 233 | char *print_fmt; | 233 | char *print_fmt; |
| 234 | int len; | 234 | int len; |
| @@ -253,7 +253,7 @@ static int set_syscall_print_fmt(struct ftrace_event_call *call) | |||
| 253 | return 0; | 253 | return 0; |
| 254 | } | 254 | } |
| 255 | 255 | ||
| 256 | static void free_syscall_print_fmt(struct ftrace_event_call *call) | 256 | static void __init free_syscall_print_fmt(struct ftrace_event_call *call) |
| 257 | { | 257 | { |
| 258 | struct syscall_metadata *entry = call->data; | 258 | struct syscall_metadata *entry = call->data; |
| 259 | 259 | ||
| @@ -459,7 +459,7 @@ static void unreg_event_syscall_exit(struct ftrace_event_file *file, | |||
| 459 | mutex_unlock(&syscall_trace_lock); | 459 | mutex_unlock(&syscall_trace_lock); |
| 460 | } | 460 | } |
| 461 | 461 | ||
| 462 | static int init_syscall_trace(struct ftrace_event_call *call) | 462 | static int __init init_syscall_trace(struct ftrace_event_call *call) |
| 463 | { | 463 | { |
| 464 | int id; | 464 | int id; |
| 465 | int num; | 465 | int num; |
