diff options
Diffstat (limited to 'kernel/trace')
| -rw-r--r-- | kernel/trace/ftrace.c | 156 | ||||
| -rw-r--r-- | kernel/trace/ring_buffer.c | 66 | ||||
| -rw-r--r-- | kernel/trace/ring_buffer_benchmark.c | 2 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 326 | ||||
| -rw-r--r-- | kernel/trace/trace.h | 63 | ||||
| -rw-r--r-- | kernel/trace/trace_entries.h | 3 | ||||
| -rw-r--r-- | kernel/trace/trace_events.c | 139 | ||||
| -rw-r--r-- | kernel/trace/trace_events_filter.c | 6 | ||||
| -rw-r--r-- | kernel/trace/trace_functions.c | 3 | ||||
| -rw-r--r-- | kernel/trace/trace_functions_graph.c | 225 | ||||
| -rw-r--r-- | kernel/trace/trace_irqsoff.c | 4 | ||||
| -rw-r--r-- | kernel/trace/trace_kprobe.c | 332 | ||||
| -rw-r--r-- | kernel/trace/trace_mmiotrace.c | 2 | ||||
| -rw-r--r-- | kernel/trace/trace_output.c | 11 | ||||
| -rw-r--r-- | kernel/trace/trace_printk.c | 5 | ||||
| -rw-r--r-- | kernel/trace/trace_sched_wakeup.c | 4 | ||||
| -rw-r--r-- | kernel/trace/trace_stack.c | 13 |
17 files changed, 870 insertions, 490 deletions
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 1ee417fcbfa5..c3e4575e7829 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -32,7 +32,6 @@ | |||
| 32 | 32 | ||
| 33 | #include <trace/events/sched.h> | 33 | #include <trace/events/sched.h> |
| 34 | 34 | ||
| 35 | #include <asm/ftrace.h> | ||
| 36 | #include <asm/setup.h> | 35 | #include <asm/setup.h> |
| 37 | 36 | ||
| 38 | #include "trace_output.h" | 37 | #include "trace_output.h" |
| @@ -82,14 +81,14 @@ static int ftrace_disabled __read_mostly; | |||
| 82 | 81 | ||
| 83 | static DEFINE_MUTEX(ftrace_lock); | 82 | static DEFINE_MUTEX(ftrace_lock); |
| 84 | 83 | ||
| 85 | static struct ftrace_ops ftrace_list_end __read_mostly = | 84 | static struct ftrace_ops ftrace_list_end __read_mostly = { |
| 86 | { | ||
| 87 | .func = ftrace_stub, | 85 | .func = ftrace_stub, |
| 88 | }; | 86 | }; |
| 89 | 87 | ||
| 90 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; | 88 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; |
| 91 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; | 89 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; |
| 92 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; | 90 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; |
| 91 | static ftrace_func_t __ftrace_trace_function_delay __read_mostly = ftrace_stub; | ||
| 93 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; | 92 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; |
| 94 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; | 93 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; |
| 95 | static struct ftrace_ops global_ops; | 94 | static struct ftrace_ops global_ops; |
| @@ -148,9 +147,11 @@ void clear_ftrace_function(void) | |||
| 148 | { | 147 | { |
| 149 | ftrace_trace_function = ftrace_stub; | 148 | ftrace_trace_function = ftrace_stub; |
| 150 | __ftrace_trace_function = ftrace_stub; | 149 | __ftrace_trace_function = ftrace_stub; |
| 150 | __ftrace_trace_function_delay = ftrace_stub; | ||
| 151 | ftrace_pid_function = ftrace_stub; | 151 | ftrace_pid_function = ftrace_stub; |
| 152 | } | 152 | } |
| 153 | 153 | ||
| 154 | #undef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | ||
| 154 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 155 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 155 | /* | 156 | /* |
| 156 | * For those archs that do not test ftrace_trace_stop in their | 157 | * For those archs that do not test ftrace_trace_stop in their |
| @@ -210,7 +211,12 @@ static void update_ftrace_function(void) | |||
| 210 | #ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | 211 | #ifdef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 211 | ftrace_trace_function = func; | 212 | ftrace_trace_function = func; |
| 212 | #else | 213 | #else |
| 214 | #ifdef CONFIG_DYNAMIC_FTRACE | ||
| 215 | /* do not update till all functions have been modified */ | ||
| 216 | __ftrace_trace_function_delay = func; | ||
| 217 | #else | ||
| 213 | __ftrace_trace_function = func; | 218 | __ftrace_trace_function = func; |
| 219 | #endif | ||
| 214 | ftrace_trace_function = ftrace_test_stop_func; | 220 | ftrace_trace_function = ftrace_test_stop_func; |
| 215 | #endif | 221 | #endif |
| 216 | } | 222 | } |
| @@ -785,8 +791,7 @@ static void unregister_ftrace_profiler(void) | |||
| 785 | unregister_ftrace_graph(); | 791 | unregister_ftrace_graph(); |
| 786 | } | 792 | } |
| 787 | #else | 793 | #else |
| 788 | static struct ftrace_ops ftrace_profile_ops __read_mostly = | 794 | static struct ftrace_ops ftrace_profile_ops __read_mostly = { |
| 789 | { | ||
| 790 | .func = function_profile_call, | 795 | .func = function_profile_call, |
| 791 | }; | 796 | }; |
| 792 | 797 | ||
| @@ -806,19 +811,10 @@ ftrace_profile_write(struct file *filp, const char __user *ubuf, | |||
| 806 | size_t cnt, loff_t *ppos) | 811 | size_t cnt, loff_t *ppos) |
| 807 | { | 812 | { |
| 808 | unsigned long val; | 813 | unsigned long val; |
| 809 | char buf[64]; /* big enough to hold a number */ | ||
| 810 | int ret; | 814 | int ret; |
| 811 | 815 | ||
| 812 | if (cnt >= sizeof(buf)) | 816 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 813 | return -EINVAL; | 817 | if (ret) |
| 814 | |||
| 815 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 816 | return -EFAULT; | ||
| 817 | |||
| 818 | buf[cnt] = 0; | ||
| 819 | |||
| 820 | ret = strict_strtoul(buf, 10, &val); | ||
| 821 | if (ret < 0) | ||
| 822 | return ret; | 818 | return ret; |
| 823 | 819 | ||
| 824 | val = !!val; | 820 | val = !!val; |
| @@ -1182,8 +1178,14 @@ alloc_and_copy_ftrace_hash(int size_bits, struct ftrace_hash *hash) | |||
| 1182 | return NULL; | 1178 | return NULL; |
| 1183 | } | 1179 | } |
| 1184 | 1180 | ||
| 1181 | static void | ||
| 1182 | ftrace_hash_rec_disable(struct ftrace_ops *ops, int filter_hash); | ||
| 1183 | static void | ||
| 1184 | ftrace_hash_rec_enable(struct ftrace_ops *ops, int filter_hash); | ||
| 1185 | |||
| 1185 | static int | 1186 | static int |
| 1186 | ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | 1187 | ftrace_hash_move(struct ftrace_ops *ops, int enable, |
| 1188 | struct ftrace_hash **dst, struct ftrace_hash *src) | ||
| 1187 | { | 1189 | { |
| 1188 | struct ftrace_func_entry *entry; | 1190 | struct ftrace_func_entry *entry; |
| 1189 | struct hlist_node *tp, *tn; | 1191 | struct hlist_node *tp, *tn; |
| @@ -1193,9 +1195,16 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1193 | unsigned long key; | 1195 | unsigned long key; |
| 1194 | int size = src->count; | 1196 | int size = src->count; |
| 1195 | int bits = 0; | 1197 | int bits = 0; |
| 1198 | int ret; | ||
| 1196 | int i; | 1199 | int i; |
| 1197 | 1200 | ||
| 1198 | /* | 1201 | /* |
| 1202 | * Remove the current set, update the hash and add | ||
| 1203 | * them back. | ||
| 1204 | */ | ||
| 1205 | ftrace_hash_rec_disable(ops, enable); | ||
| 1206 | |||
| 1207 | /* | ||
| 1199 | * If the new source is empty, just free dst and assign it | 1208 | * If the new source is empty, just free dst and assign it |
| 1200 | * the empty_hash. | 1209 | * the empty_hash. |
| 1201 | */ | 1210 | */ |
| @@ -1215,9 +1224,10 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1215 | if (bits > FTRACE_HASH_MAX_BITS) | 1224 | if (bits > FTRACE_HASH_MAX_BITS) |
| 1216 | bits = FTRACE_HASH_MAX_BITS; | 1225 | bits = FTRACE_HASH_MAX_BITS; |
| 1217 | 1226 | ||
| 1227 | ret = -ENOMEM; | ||
| 1218 | new_hash = alloc_ftrace_hash(bits); | 1228 | new_hash = alloc_ftrace_hash(bits); |
| 1219 | if (!new_hash) | 1229 | if (!new_hash) |
| 1220 | return -ENOMEM; | 1230 | goto out; |
| 1221 | 1231 | ||
| 1222 | size = 1 << src->size_bits; | 1232 | size = 1 << src->size_bits; |
| 1223 | for (i = 0; i < size; i++) { | 1233 | for (i = 0; i < size; i++) { |
| @@ -1236,7 +1246,16 @@ ftrace_hash_move(struct ftrace_hash **dst, struct ftrace_hash *src) | |||
| 1236 | rcu_assign_pointer(*dst, new_hash); | 1246 | rcu_assign_pointer(*dst, new_hash); |
| 1237 | free_ftrace_hash_rcu(old_hash); | 1247 | free_ftrace_hash_rcu(old_hash); |
| 1238 | 1248 | ||
| 1239 | return 0; | 1249 | ret = 0; |
| 1250 | out: | ||
| 1251 | /* | ||
| 1252 | * Enable regardless of ret: | ||
| 1253 | * On success, we enable the new hash. | ||
| 1254 | * On failure, we re-enable the original hash. | ||
| 1255 | */ | ||
| 1256 | ftrace_hash_rec_enable(ops, enable); | ||
| 1257 | |||
| 1258 | return ret; | ||
| 1240 | } | 1259 | } |
| 1241 | 1260 | ||
| 1242 | /* | 1261 | /* |
| @@ -1596,6 +1615,12 @@ static int __ftrace_modify_code(void *data) | |||
| 1596 | { | 1615 | { |
| 1597 | int *command = data; | 1616 | int *command = data; |
| 1598 | 1617 | ||
| 1618 | /* | ||
| 1619 | * Do not call function tracer while we update the code. | ||
| 1620 | * We are in stop machine, no worrying about races. | ||
| 1621 | */ | ||
| 1622 | function_trace_stop++; | ||
| 1623 | |||
| 1599 | if (*command & FTRACE_ENABLE_CALLS) | 1624 | if (*command & FTRACE_ENABLE_CALLS) |
| 1600 | ftrace_replace_code(1); | 1625 | ftrace_replace_code(1); |
| 1601 | else if (*command & FTRACE_DISABLE_CALLS) | 1626 | else if (*command & FTRACE_DISABLE_CALLS) |
| @@ -1609,6 +1634,18 @@ static int __ftrace_modify_code(void *data) | |||
| 1609 | else if (*command & FTRACE_STOP_FUNC_RET) | 1634 | else if (*command & FTRACE_STOP_FUNC_RET) |
| 1610 | ftrace_disable_ftrace_graph_caller(); | 1635 | ftrace_disable_ftrace_graph_caller(); |
| 1611 | 1636 | ||
| 1637 | #ifndef CONFIG_HAVE_FUNCTION_TRACE_MCOUNT_TEST | ||
| 1638 | /* | ||
| 1639 | * For archs that call ftrace_test_stop_func(), we must | ||
| 1640 | * wait till after we update all the function callers | ||
| 1641 | * before we update the callback. This keeps different | ||
| 1642 | * ops that record different functions from corrupting | ||
| 1643 | * each other. | ||
| 1644 | */ | ||
| 1645 | __ftrace_trace_function = __ftrace_trace_function_delay; | ||
| 1646 | #endif | ||
| 1647 | function_trace_stop--; | ||
| 1648 | |||
| 1612 | return 0; | 1649 | return 0; |
| 1613 | } | 1650 | } |
| 1614 | 1651 | ||
| @@ -1744,10 +1781,36 @@ static cycle_t ftrace_update_time; | |||
| 1744 | static unsigned long ftrace_update_cnt; | 1781 | static unsigned long ftrace_update_cnt; |
| 1745 | unsigned long ftrace_update_tot_cnt; | 1782 | unsigned long ftrace_update_tot_cnt; |
| 1746 | 1783 | ||
| 1784 | static int ops_traces_mod(struct ftrace_ops *ops) | ||
| 1785 | { | ||
| 1786 | struct ftrace_hash *hash; | ||
| 1787 | |||
| 1788 | hash = ops->filter_hash; | ||
| 1789 | return !!(!hash || !hash->count); | ||
| 1790 | } | ||
| 1791 | |||
| 1747 | static int ftrace_update_code(struct module *mod) | 1792 | static int ftrace_update_code(struct module *mod) |
| 1748 | { | 1793 | { |
| 1749 | struct dyn_ftrace *p; | 1794 | struct dyn_ftrace *p; |
| 1750 | cycle_t start, stop; | 1795 | cycle_t start, stop; |
| 1796 | unsigned long ref = 0; | ||
| 1797 | |||
| 1798 | /* | ||
| 1799 | * When adding a module, we need to check if tracers are | ||
| 1800 | * currently enabled and if they are set to trace all functions. | ||
| 1801 | * If they are, we need to enable the module functions as well | ||
| 1802 | * as update the reference counts for those function records. | ||
| 1803 | */ | ||
| 1804 | if (mod) { | ||
| 1805 | struct ftrace_ops *ops; | ||
| 1806 | |||
| 1807 | for (ops = ftrace_ops_list; | ||
| 1808 | ops != &ftrace_list_end; ops = ops->next) { | ||
| 1809 | if (ops->flags & FTRACE_OPS_FL_ENABLED && | ||
| 1810 | ops_traces_mod(ops)) | ||
| 1811 | ref++; | ||
| 1812 | } | ||
| 1813 | } | ||
| 1751 | 1814 | ||
| 1752 | start = ftrace_now(raw_smp_processor_id()); | 1815 | start = ftrace_now(raw_smp_processor_id()); |
| 1753 | ftrace_update_cnt = 0; | 1816 | ftrace_update_cnt = 0; |
| @@ -1760,7 +1823,7 @@ static int ftrace_update_code(struct module *mod) | |||
| 1760 | 1823 | ||
| 1761 | p = ftrace_new_addrs; | 1824 | p = ftrace_new_addrs; |
| 1762 | ftrace_new_addrs = p->newlist; | 1825 | ftrace_new_addrs = p->newlist; |
| 1763 | p->flags = 0L; | 1826 | p->flags = ref; |
| 1764 | 1827 | ||
| 1765 | /* | 1828 | /* |
| 1766 | * Do the initial record conversion from mcount jump | 1829 | * Do the initial record conversion from mcount jump |
| @@ -1783,7 +1846,7 @@ static int ftrace_update_code(struct module *mod) | |||
| 1783 | * conversion puts the module to the correct state, thus | 1846 | * conversion puts the module to the correct state, thus |
| 1784 | * passing the ftrace_make_call check. | 1847 | * passing the ftrace_make_call check. |
| 1785 | */ | 1848 | */ |
| 1786 | if (ftrace_start_up) { | 1849 | if (ftrace_start_up && ref) { |
| 1787 | int failed = __ftrace_replace_code(p, 1); | 1850 | int failed = __ftrace_replace_code(p, 1); |
| 1788 | if (failed) { | 1851 | if (failed) { |
| 1789 | ftrace_bug(failed, p->ip); | 1852 | ftrace_bug(failed, p->ip); |
| @@ -2407,10 +2470,9 @@ ftrace_match_module_records(struct ftrace_hash *hash, char *buff, char *mod) | |||
| 2407 | */ | 2470 | */ |
| 2408 | 2471 | ||
| 2409 | static int | 2472 | static int |
| 2410 | ftrace_mod_callback(char *func, char *cmd, char *param, int enable) | 2473 | ftrace_mod_callback(struct ftrace_hash *hash, |
| 2474 | char *func, char *cmd, char *param, int enable) | ||
| 2411 | { | 2475 | { |
| 2412 | struct ftrace_ops *ops = &global_ops; | ||
| 2413 | struct ftrace_hash *hash; | ||
| 2414 | char *mod; | 2476 | char *mod; |
| 2415 | int ret = -EINVAL; | 2477 | int ret = -EINVAL; |
| 2416 | 2478 | ||
| @@ -2430,11 +2492,6 @@ ftrace_mod_callback(char *func, char *cmd, char *param, int enable) | |||
| 2430 | if (!strlen(mod)) | 2492 | if (!strlen(mod)) |
| 2431 | return ret; | 2493 | return ret; |
| 2432 | 2494 | ||
| 2433 | if (enable) | ||
| 2434 | hash = ops->filter_hash; | ||
| 2435 | else | ||
| 2436 | hash = ops->notrace_hash; | ||
| 2437 | |||
| 2438 | ret = ftrace_match_module_records(hash, func, mod); | 2495 | ret = ftrace_match_module_records(hash, func, mod); |
| 2439 | if (!ret) | 2496 | if (!ret) |
| 2440 | ret = -EINVAL; | 2497 | ret = -EINVAL; |
| @@ -2740,7 +2797,7 @@ static int ftrace_process_regex(struct ftrace_hash *hash, | |||
| 2740 | { | 2797 | { |
| 2741 | char *func, *command, *next = buff; | 2798 | char *func, *command, *next = buff; |
| 2742 | struct ftrace_func_command *p; | 2799 | struct ftrace_func_command *p; |
| 2743 | int ret; | 2800 | int ret = -EINVAL; |
| 2744 | 2801 | ||
| 2745 | func = strsep(&next, ":"); | 2802 | func = strsep(&next, ":"); |
| 2746 | 2803 | ||
| @@ -2760,7 +2817,7 @@ static int ftrace_process_regex(struct ftrace_hash *hash, | |||
| 2760 | mutex_lock(&ftrace_cmd_mutex); | 2817 | mutex_lock(&ftrace_cmd_mutex); |
| 2761 | list_for_each_entry(p, &ftrace_commands, list) { | 2818 | list_for_each_entry(p, &ftrace_commands, list) { |
| 2762 | if (strcmp(p->name, command) == 0) { | 2819 | if (strcmp(p->name, command) == 0) { |
| 2763 | ret = p->func(func, command, next, enable); | 2820 | ret = p->func(hash, func, command, next, enable); |
| 2764 | goto out_unlock; | 2821 | goto out_unlock; |
| 2765 | } | 2822 | } |
| 2766 | } | 2823 | } |
| @@ -2857,7 +2914,11 @@ ftrace_set_regex(struct ftrace_ops *ops, unsigned char *buf, int len, | |||
| 2857 | ftrace_match_records(hash, buf, len); | 2914 | ftrace_match_records(hash, buf, len); |
| 2858 | 2915 | ||
| 2859 | mutex_lock(&ftrace_lock); | 2916 | mutex_lock(&ftrace_lock); |
| 2860 | ret = ftrace_hash_move(orig_hash, hash); | 2917 | ret = ftrace_hash_move(ops, enable, orig_hash, hash); |
| 2918 | if (!ret && ops->flags & FTRACE_OPS_FL_ENABLED | ||
| 2919 | && ftrace_enabled) | ||
| 2920 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); | ||
| 2921 | |||
| 2861 | mutex_unlock(&ftrace_lock); | 2922 | mutex_unlock(&ftrace_lock); |
| 2862 | 2923 | ||
| 2863 | mutex_unlock(&ftrace_regex_lock); | 2924 | mutex_unlock(&ftrace_regex_lock); |
| @@ -3040,18 +3101,12 @@ ftrace_regex_release(struct inode *inode, struct file *file) | |||
| 3040 | orig_hash = &iter->ops->notrace_hash; | 3101 | orig_hash = &iter->ops->notrace_hash; |
| 3041 | 3102 | ||
| 3042 | mutex_lock(&ftrace_lock); | 3103 | mutex_lock(&ftrace_lock); |
| 3043 | /* | 3104 | ret = ftrace_hash_move(iter->ops, filter_hash, |
| 3044 | * Remove the current set, update the hash and add | 3105 | orig_hash, iter->hash); |
| 3045 | * them back. | 3106 | if (!ret && (iter->ops->flags & FTRACE_OPS_FL_ENABLED) |
| 3046 | */ | 3107 | && ftrace_enabled) |
| 3047 | ftrace_hash_rec_disable(iter->ops, filter_hash); | 3108 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); |
| 3048 | ret = ftrace_hash_move(orig_hash, iter->hash); | 3109 | |
| 3049 | if (!ret) { | ||
| 3050 | ftrace_hash_rec_enable(iter->ops, filter_hash); | ||
| 3051 | if (iter->ops->flags & FTRACE_OPS_FL_ENABLED | ||
| 3052 | && ftrace_enabled) | ||
| 3053 | ftrace_run_update_code(FTRACE_ENABLE_CALLS); | ||
| 3054 | } | ||
| 3055 | mutex_unlock(&ftrace_lock); | 3110 | mutex_unlock(&ftrace_lock); |
| 3056 | } | 3111 | } |
| 3057 | free_ftrace_hash(iter->hash); | 3112 | free_ftrace_hash(iter->hash); |
| @@ -3330,6 +3385,7 @@ static int ftrace_process_locs(struct module *mod, | |||
| 3330 | { | 3385 | { |
| 3331 | unsigned long *p; | 3386 | unsigned long *p; |
| 3332 | unsigned long addr; | 3387 | unsigned long addr; |
| 3388 | unsigned long flags = 0; /* Shut up gcc */ | ||
| 3333 | 3389 | ||
| 3334 | mutex_lock(&ftrace_lock); | 3390 | mutex_lock(&ftrace_lock); |
| 3335 | p = start; | 3391 | p = start; |
| @@ -3346,7 +3402,19 @@ static int ftrace_process_locs(struct module *mod, | |||
| 3346 | ftrace_record_ip(addr); | 3402 | ftrace_record_ip(addr); |
| 3347 | } | 3403 | } |
| 3348 | 3404 | ||
| 3405 | /* | ||
| 3406 | * We only need to disable interrupts on start up | ||
| 3407 | * because we are modifying code that an interrupt | ||
| 3408 | * may execute, and the modification is not atomic. | ||
| 3409 | * But for modules, nothing runs the code we modify | ||
| 3410 | * until we are finished with it, and there's no | ||
| 3411 | * reason to cause large interrupt latencies while we do it. | ||
| 3412 | */ | ||
| 3413 | if (!mod) | ||
| 3414 | local_irq_save(flags); | ||
| 3349 | ftrace_update_code(mod); | 3415 | ftrace_update_code(mod); |
| 3416 | if (!mod) | ||
| 3417 | local_irq_restore(flags); | ||
| 3350 | mutex_unlock(&ftrace_lock); | 3418 | mutex_unlock(&ftrace_lock); |
| 3351 | 3419 | ||
| 3352 | return 0; | 3420 | return 0; |
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index b0c7aa407943..731201bf4acc 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -997,15 +997,21 @@ static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 997 | unsigned nr_pages) | 997 | unsigned nr_pages) |
| 998 | { | 998 | { |
| 999 | struct buffer_page *bpage, *tmp; | 999 | struct buffer_page *bpage, *tmp; |
| 1000 | unsigned long addr; | ||
| 1001 | LIST_HEAD(pages); | 1000 | LIST_HEAD(pages); |
| 1002 | unsigned i; | 1001 | unsigned i; |
| 1003 | 1002 | ||
| 1004 | WARN_ON(!nr_pages); | 1003 | WARN_ON(!nr_pages); |
| 1005 | 1004 | ||
| 1006 | for (i = 0; i < nr_pages; i++) { | 1005 | for (i = 0; i < nr_pages; i++) { |
| 1006 | struct page *page; | ||
| 1007 | /* | ||
| 1008 | * __GFP_NORETRY flag makes sure that the allocation fails | ||
| 1009 | * gracefully without invoking oom-killer and the system is | ||
| 1010 | * not destabilized. | ||
| 1011 | */ | ||
| 1007 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), | 1012 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), |
| 1008 | GFP_KERNEL, cpu_to_node(cpu_buffer->cpu)); | 1013 | GFP_KERNEL | __GFP_NORETRY, |
| 1014 | cpu_to_node(cpu_buffer->cpu)); | ||
| 1009 | if (!bpage) | 1015 | if (!bpage) |
| 1010 | goto free_pages; | 1016 | goto free_pages; |
| 1011 | 1017 | ||
| @@ -1013,10 +1019,11 @@ static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, | |||
| 1013 | 1019 | ||
| 1014 | list_add(&bpage->list, &pages); | 1020 | list_add(&bpage->list, &pages); |
| 1015 | 1021 | ||
| 1016 | addr = __get_free_page(GFP_KERNEL); | 1022 | page = alloc_pages_node(cpu_to_node(cpu_buffer->cpu), |
| 1017 | if (!addr) | 1023 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 1024 | if (!page) | ||
| 1018 | goto free_pages; | 1025 | goto free_pages; |
| 1019 | bpage->page = (void *)addr; | 1026 | bpage->page = page_address(page); |
| 1020 | rb_init_page(bpage->page); | 1027 | rb_init_page(bpage->page); |
| 1021 | } | 1028 | } |
| 1022 | 1029 | ||
| @@ -1045,7 +1052,7 @@ rb_allocate_cpu_buffer(struct ring_buffer *buffer, int cpu) | |||
| 1045 | { | 1052 | { |
| 1046 | struct ring_buffer_per_cpu *cpu_buffer; | 1053 | struct ring_buffer_per_cpu *cpu_buffer; |
| 1047 | struct buffer_page *bpage; | 1054 | struct buffer_page *bpage; |
| 1048 | unsigned long addr; | 1055 | struct page *page; |
| 1049 | int ret; | 1056 | int ret; |
| 1050 | 1057 | ||
| 1051 | cpu_buffer = kzalloc_node(ALIGN(sizeof(*cpu_buffer), cache_line_size()), | 1058 | cpu_buffer = kzalloc_node(ALIGN(sizeof(*cpu_buffer), cache_line_size()), |
| @@ -1067,10 +1074,10 @@ rb_allocate_cpu_buffer(struct ring_buffer *buffer, int cpu) | |||
| 1067 | rb_check_bpage(cpu_buffer, bpage); | 1074 | rb_check_bpage(cpu_buffer, bpage); |
| 1068 | 1075 | ||
| 1069 | cpu_buffer->reader_page = bpage; | 1076 | cpu_buffer->reader_page = bpage; |
| 1070 | addr = __get_free_page(GFP_KERNEL); | 1077 | page = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL, 0); |
| 1071 | if (!addr) | 1078 | if (!page) |
| 1072 | goto fail_free_reader; | 1079 | goto fail_free_reader; |
| 1073 | bpage->page = (void *)addr; | 1080 | bpage->page = page_address(page); |
| 1074 | rb_init_page(bpage->page); | 1081 | rb_init_page(bpage->page); |
| 1075 | 1082 | ||
| 1076 | INIT_LIST_HEAD(&cpu_buffer->reader_page->list); | 1083 | INIT_LIST_HEAD(&cpu_buffer->reader_page->list); |
| @@ -1314,7 +1321,6 @@ int ring_buffer_resize(struct ring_buffer *buffer, unsigned long size) | |||
| 1314 | unsigned nr_pages, rm_pages, new_pages; | 1321 | unsigned nr_pages, rm_pages, new_pages; |
| 1315 | struct buffer_page *bpage, *tmp; | 1322 | struct buffer_page *bpage, *tmp; |
| 1316 | unsigned long buffer_size; | 1323 | unsigned long buffer_size; |
| 1317 | unsigned long addr; | ||
| 1318 | LIST_HEAD(pages); | 1324 | LIST_HEAD(pages); |
| 1319 | int i, cpu; | 1325 | int i, cpu; |
| 1320 | 1326 | ||
| @@ -1375,16 +1381,24 @@ int ring_buffer_resize(struct ring_buffer *buffer, unsigned long size) | |||
| 1375 | 1381 | ||
| 1376 | for_each_buffer_cpu(buffer, cpu) { | 1382 | for_each_buffer_cpu(buffer, cpu) { |
| 1377 | for (i = 0; i < new_pages; i++) { | 1383 | for (i = 0; i < new_pages; i++) { |
| 1384 | struct page *page; | ||
| 1385 | /* | ||
| 1386 | * __GFP_NORETRY flag makes sure that the allocation | ||
| 1387 | * fails gracefully without invoking oom-killer and | ||
| 1388 | * the system is not destabilized. | ||
| 1389 | */ | ||
| 1378 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), | 1390 | bpage = kzalloc_node(ALIGN(sizeof(*bpage), |
| 1379 | cache_line_size()), | 1391 | cache_line_size()), |
| 1380 | GFP_KERNEL, cpu_to_node(cpu)); | 1392 | GFP_KERNEL | __GFP_NORETRY, |
| 1393 | cpu_to_node(cpu)); | ||
| 1381 | if (!bpage) | 1394 | if (!bpage) |
| 1382 | goto free_pages; | 1395 | goto free_pages; |
| 1383 | list_add(&bpage->list, &pages); | 1396 | list_add(&bpage->list, &pages); |
| 1384 | addr = __get_free_page(GFP_KERNEL); | 1397 | page = alloc_pages_node(cpu_to_node(cpu), |
| 1385 | if (!addr) | 1398 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 1399 | if (!page) | ||
| 1386 | goto free_pages; | 1400 | goto free_pages; |
| 1387 | bpage->page = (void *)addr; | 1401 | bpage->page = page_address(page); |
| 1388 | rb_init_page(bpage->page); | 1402 | rb_init_page(bpage->page); |
| 1389 | } | 1403 | } |
| 1390 | } | 1404 | } |
| @@ -3730,16 +3744,17 @@ EXPORT_SYMBOL_GPL(ring_buffer_swap_cpu); | |||
| 3730 | * Returns: | 3744 | * Returns: |
| 3731 | * The page allocated, or NULL on error. | 3745 | * The page allocated, or NULL on error. |
| 3732 | */ | 3746 | */ |
| 3733 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer) | 3747 | void *ring_buffer_alloc_read_page(struct ring_buffer *buffer, int cpu) |
| 3734 | { | 3748 | { |
| 3735 | struct buffer_data_page *bpage; | 3749 | struct buffer_data_page *bpage; |
| 3736 | unsigned long addr; | 3750 | struct page *page; |
| 3737 | 3751 | ||
| 3738 | addr = __get_free_page(GFP_KERNEL); | 3752 | page = alloc_pages_node(cpu_to_node(cpu), |
| 3739 | if (!addr) | 3753 | GFP_KERNEL | __GFP_NORETRY, 0); |
| 3754 | if (!page) | ||
| 3740 | return NULL; | 3755 | return NULL; |
| 3741 | 3756 | ||
| 3742 | bpage = (void *)addr; | 3757 | bpage = page_address(page); |
| 3743 | 3758 | ||
| 3744 | rb_init_page(bpage); | 3759 | rb_init_page(bpage); |
| 3745 | 3760 | ||
| @@ -3978,20 +3993,11 @@ rb_simple_write(struct file *filp, const char __user *ubuf, | |||
| 3978 | size_t cnt, loff_t *ppos) | 3993 | size_t cnt, loff_t *ppos) |
| 3979 | { | 3994 | { |
| 3980 | unsigned long *p = filp->private_data; | 3995 | unsigned long *p = filp->private_data; |
| 3981 | char buf[64]; | ||
| 3982 | unsigned long val; | 3996 | unsigned long val; |
| 3983 | int ret; | 3997 | int ret; |
| 3984 | 3998 | ||
| 3985 | if (cnt >= sizeof(buf)) | 3999 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 3986 | return -EINVAL; | 4000 | if (ret) |
| 3987 | |||
| 3988 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 3989 | return -EFAULT; | ||
| 3990 | |||
| 3991 | buf[cnt] = 0; | ||
| 3992 | |||
| 3993 | ret = strict_strtoul(buf, 10, &val); | ||
| 3994 | if (ret < 0) | ||
| 3995 | return ret; | 4001 | return ret; |
| 3996 | 4002 | ||
| 3997 | if (val) | 4003 | if (val) |
diff --git a/kernel/trace/ring_buffer_benchmark.c b/kernel/trace/ring_buffer_benchmark.c index 302f8a614635..a5457d577b98 100644 --- a/kernel/trace/ring_buffer_benchmark.c +++ b/kernel/trace/ring_buffer_benchmark.c | |||
| @@ -106,7 +106,7 @@ static enum event_status read_page(int cpu) | |||
| 106 | int inc; | 106 | int inc; |
| 107 | int i; | 107 | int i; |
| 108 | 108 | ||
| 109 | bpage = ring_buffer_alloc_read_page(buffer); | 109 | bpage = ring_buffer_alloc_read_page(buffer, cpu); |
| 110 | if (!bpage) | 110 | if (!bpage) |
| 111 | return EVENT_DROPPED; | 111 | return EVENT_DROPPED; |
| 112 | 112 | ||
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index ee9c921d7f21..e5df02c69b1d 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -343,26 +343,27 @@ unsigned long trace_flags = TRACE_ITER_PRINT_PARENT | TRACE_ITER_PRINTK | | |||
| 343 | static int trace_stop_count; | 343 | static int trace_stop_count; |
| 344 | static DEFINE_SPINLOCK(tracing_start_lock); | 344 | static DEFINE_SPINLOCK(tracing_start_lock); |
| 345 | 345 | ||
| 346 | static void wakeup_work_handler(struct work_struct *work) | ||
| 347 | { | ||
| 348 | wake_up(&trace_wait); | ||
| 349 | } | ||
| 350 | |||
| 351 | static DECLARE_DELAYED_WORK(wakeup_work, wakeup_work_handler); | ||
| 352 | |||
| 346 | /** | 353 | /** |
| 347 | * trace_wake_up - wake up tasks waiting for trace input | 354 | * trace_wake_up - wake up tasks waiting for trace input |
| 348 | * | 355 | * |
| 349 | * Simply wakes up any task that is blocked on the trace_wait | 356 | * Schedules a delayed work to wake up any task that is blocked on the |
| 350 | * queue. These is used with trace_poll for tasks polling the trace. | 357 | * trace_wait queue. These is used with trace_poll for tasks polling the |
| 358 | * trace. | ||
| 351 | */ | 359 | */ |
| 352 | void trace_wake_up(void) | 360 | void trace_wake_up(void) |
| 353 | { | 361 | { |
| 354 | int cpu; | 362 | const unsigned long delay = msecs_to_jiffies(2); |
| 355 | 363 | ||
| 356 | if (trace_flags & TRACE_ITER_BLOCK) | 364 | if (trace_flags & TRACE_ITER_BLOCK) |
| 357 | return; | 365 | return; |
| 358 | /* | 366 | schedule_delayed_work(&wakeup_work, delay); |
| 359 | * The runqueue_is_locked() can fail, but this is the best we | ||
| 360 | * have for now: | ||
| 361 | */ | ||
| 362 | cpu = get_cpu(); | ||
| 363 | if (!runqueue_is_locked(cpu)) | ||
| 364 | wake_up(&trace_wait); | ||
| 365 | put_cpu(); | ||
| 366 | } | 367 | } |
| 367 | 368 | ||
| 368 | static int __init set_buf_size(char *str) | 369 | static int __init set_buf_size(char *str) |
| @@ -424,6 +425,7 @@ static const char *trace_options[] = { | |||
| 424 | "graph-time", | 425 | "graph-time", |
| 425 | "record-cmd", | 426 | "record-cmd", |
| 426 | "overwrite", | 427 | "overwrite", |
| 428 | "disable_on_free", | ||
| 427 | NULL | 429 | NULL |
| 428 | }; | 430 | }; |
| 429 | 431 | ||
| @@ -1191,6 +1193,18 @@ void trace_nowake_buffer_unlock_commit(struct ring_buffer *buffer, | |||
| 1191 | } | 1193 | } |
| 1192 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); | 1194 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit); |
| 1193 | 1195 | ||
| 1196 | void trace_nowake_buffer_unlock_commit_regs(struct ring_buffer *buffer, | ||
| 1197 | struct ring_buffer_event *event, | ||
| 1198 | unsigned long flags, int pc, | ||
| 1199 | struct pt_regs *regs) | ||
| 1200 | { | ||
| 1201 | ring_buffer_unlock_commit(buffer, event); | ||
| 1202 | |||
| 1203 | ftrace_trace_stack_regs(buffer, flags, 0, pc, regs); | ||
| 1204 | ftrace_trace_userstack(buffer, flags, pc); | ||
| 1205 | } | ||
| 1206 | EXPORT_SYMBOL_GPL(trace_nowake_buffer_unlock_commit_regs); | ||
| 1207 | |||
| 1194 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, | 1208 | void trace_current_buffer_discard_commit(struct ring_buffer *buffer, |
| 1195 | struct ring_buffer_event *event) | 1209 | struct ring_buffer_event *event) |
| 1196 | { | 1210 | { |
| @@ -1234,30 +1248,103 @@ ftrace(struct trace_array *tr, struct trace_array_cpu *data, | |||
| 1234 | } | 1248 | } |
| 1235 | 1249 | ||
| 1236 | #ifdef CONFIG_STACKTRACE | 1250 | #ifdef CONFIG_STACKTRACE |
| 1251 | |||
| 1252 | #define FTRACE_STACK_MAX_ENTRIES (PAGE_SIZE / sizeof(unsigned long)) | ||
| 1253 | struct ftrace_stack { | ||
| 1254 | unsigned long calls[FTRACE_STACK_MAX_ENTRIES]; | ||
| 1255 | }; | ||
| 1256 | |||
| 1257 | static DEFINE_PER_CPU(struct ftrace_stack, ftrace_stack); | ||
| 1258 | static DEFINE_PER_CPU(int, ftrace_stack_reserve); | ||
| 1259 | |||
| 1237 | static void __ftrace_trace_stack(struct ring_buffer *buffer, | 1260 | static void __ftrace_trace_stack(struct ring_buffer *buffer, |
| 1238 | unsigned long flags, | 1261 | unsigned long flags, |
| 1239 | int skip, int pc) | 1262 | int skip, int pc, struct pt_regs *regs) |
| 1240 | { | 1263 | { |
| 1241 | struct ftrace_event_call *call = &event_kernel_stack; | 1264 | struct ftrace_event_call *call = &event_kernel_stack; |
| 1242 | struct ring_buffer_event *event; | 1265 | struct ring_buffer_event *event; |
| 1243 | struct stack_entry *entry; | 1266 | struct stack_entry *entry; |
| 1244 | struct stack_trace trace; | 1267 | struct stack_trace trace; |
| 1268 | int use_stack; | ||
| 1269 | int size = FTRACE_STACK_ENTRIES; | ||
| 1270 | |||
| 1271 | trace.nr_entries = 0; | ||
| 1272 | trace.skip = skip; | ||
| 1273 | |||
| 1274 | /* | ||
| 1275 | * Since events can happen in NMIs there's no safe way to | ||
| 1276 | * use the per cpu ftrace_stacks. We reserve it and if an interrupt | ||
| 1277 | * or NMI comes in, it will just have to use the default | ||
| 1278 | * FTRACE_STACK_SIZE. | ||
| 1279 | */ | ||
| 1280 | preempt_disable_notrace(); | ||
| 1281 | |||
| 1282 | use_stack = ++__get_cpu_var(ftrace_stack_reserve); | ||
| 1283 | /* | ||
| 1284 | * We don't need any atomic variables, just a barrier. | ||
| 1285 | * If an interrupt comes in, we don't care, because it would | ||
| 1286 | * have exited and put the counter back to what we want. | ||
| 1287 | * We just need a barrier to keep gcc from moving things | ||
| 1288 | * around. | ||
| 1289 | */ | ||
| 1290 | barrier(); | ||
| 1291 | if (use_stack == 1) { | ||
| 1292 | trace.entries = &__get_cpu_var(ftrace_stack).calls[0]; | ||
| 1293 | trace.max_entries = FTRACE_STACK_MAX_ENTRIES; | ||
| 1294 | |||
| 1295 | if (regs) | ||
| 1296 | save_stack_trace_regs(regs, &trace); | ||
| 1297 | else | ||
| 1298 | save_stack_trace(&trace); | ||
| 1299 | |||
| 1300 | if (trace.nr_entries > size) | ||
| 1301 | size = trace.nr_entries; | ||
| 1302 | } else | ||
| 1303 | /* From now on, use_stack is a boolean */ | ||
| 1304 | use_stack = 0; | ||
| 1305 | |||
| 1306 | size *= sizeof(unsigned long); | ||
| 1245 | 1307 | ||
| 1246 | event = trace_buffer_lock_reserve(buffer, TRACE_STACK, | 1308 | event = trace_buffer_lock_reserve(buffer, TRACE_STACK, |
| 1247 | sizeof(*entry), flags, pc); | 1309 | sizeof(*entry) + size, flags, pc); |
| 1248 | if (!event) | 1310 | if (!event) |
| 1249 | return; | 1311 | goto out; |
| 1250 | entry = ring_buffer_event_data(event); | 1312 | entry = ring_buffer_event_data(event); |
| 1251 | memset(&entry->caller, 0, sizeof(entry->caller)); | ||
| 1252 | 1313 | ||
| 1253 | trace.nr_entries = 0; | 1314 | memset(&entry->caller, 0, size); |
| 1254 | trace.max_entries = FTRACE_STACK_ENTRIES; | 1315 | |
| 1255 | trace.skip = skip; | 1316 | if (use_stack) |
| 1256 | trace.entries = entry->caller; | 1317 | memcpy(&entry->caller, trace.entries, |
| 1318 | trace.nr_entries * sizeof(unsigned long)); | ||
| 1319 | else { | ||
| 1320 | trace.max_entries = FTRACE_STACK_ENTRIES; | ||
| 1321 | trace.entries = entry->caller; | ||
| 1322 | if (regs) | ||
| 1323 | save_stack_trace_regs(regs, &trace); | ||
| 1324 | else | ||
| 1325 | save_stack_trace(&trace); | ||
| 1326 | } | ||
| 1327 | |||
| 1328 | entry->size = trace.nr_entries; | ||
| 1257 | 1329 | ||
| 1258 | save_stack_trace(&trace); | ||
| 1259 | if (!filter_check_discard(call, entry, buffer, event)) | 1330 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1260 | ring_buffer_unlock_commit(buffer, event); | 1331 | ring_buffer_unlock_commit(buffer, event); |
| 1332 | |||
| 1333 | out: | ||
| 1334 | /* Again, don't let gcc optimize things here */ | ||
| 1335 | barrier(); | ||
| 1336 | __get_cpu_var(ftrace_stack_reserve)--; | ||
| 1337 | preempt_enable_notrace(); | ||
| 1338 | |||
| 1339 | } | ||
| 1340 | |||
| 1341 | void ftrace_trace_stack_regs(struct ring_buffer *buffer, unsigned long flags, | ||
| 1342 | int skip, int pc, struct pt_regs *regs) | ||
| 1343 | { | ||
| 1344 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) | ||
| 1345 | return; | ||
| 1346 | |||
| 1347 | __ftrace_trace_stack(buffer, flags, skip, pc, regs); | ||
| 1261 | } | 1348 | } |
| 1262 | 1349 | ||
| 1263 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | 1350 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, |
| @@ -1266,13 +1353,13 @@ void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | |||
| 1266 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) | 1353 | if (!(trace_flags & TRACE_ITER_STACKTRACE)) |
| 1267 | return; | 1354 | return; |
| 1268 | 1355 | ||
| 1269 | __ftrace_trace_stack(buffer, flags, skip, pc); | 1356 | __ftrace_trace_stack(buffer, flags, skip, pc, NULL); |
| 1270 | } | 1357 | } |
| 1271 | 1358 | ||
| 1272 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, | 1359 | void __trace_stack(struct trace_array *tr, unsigned long flags, int skip, |
| 1273 | int pc) | 1360 | int pc) |
| 1274 | { | 1361 | { |
| 1275 | __ftrace_trace_stack(tr->buffer, flags, skip, pc); | 1362 | __ftrace_trace_stack(tr->buffer, flags, skip, pc, NULL); |
| 1276 | } | 1363 | } |
| 1277 | 1364 | ||
| 1278 | /** | 1365 | /** |
| @@ -1288,7 +1375,7 @@ void trace_dump_stack(void) | |||
| 1288 | local_save_flags(flags); | 1375 | local_save_flags(flags); |
| 1289 | 1376 | ||
| 1290 | /* skipping 3 traces, seems to get us at the caller of this function */ | 1377 | /* skipping 3 traces, seems to get us at the caller of this function */ |
| 1291 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count()); | 1378 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count(), NULL); |
| 1292 | } | 1379 | } |
| 1293 | 1380 | ||
| 1294 | static DEFINE_PER_CPU(int, user_stack_count); | 1381 | static DEFINE_PER_CPU(int, user_stack_count); |
| @@ -1536,7 +1623,12 @@ peek_next_entry(struct trace_iterator *iter, int cpu, u64 *ts, | |||
| 1536 | 1623 | ||
| 1537 | ftrace_enable_cpu(); | 1624 | ftrace_enable_cpu(); |
| 1538 | 1625 | ||
| 1539 | return event ? ring_buffer_event_data(event) : NULL; | 1626 | if (event) { |
| 1627 | iter->ent_size = ring_buffer_event_length(event); | ||
| 1628 | return ring_buffer_event_data(event); | ||
| 1629 | } | ||
| 1630 | iter->ent_size = 0; | ||
| 1631 | return NULL; | ||
| 1540 | } | 1632 | } |
| 1541 | 1633 | ||
| 1542 | static struct trace_entry * | 1634 | static struct trace_entry * |
| @@ -2051,6 +2143,9 @@ void trace_default_header(struct seq_file *m) | |||
| 2051 | { | 2143 | { |
| 2052 | struct trace_iterator *iter = m->private; | 2144 | struct trace_iterator *iter = m->private; |
| 2053 | 2145 | ||
| 2146 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 2147 | return; | ||
| 2148 | |||
| 2054 | if (iter->iter_flags & TRACE_FILE_LAT_FMT) { | 2149 | if (iter->iter_flags & TRACE_FILE_LAT_FMT) { |
| 2055 | /* print nothing if the buffers are empty */ | 2150 | /* print nothing if the buffers are empty */ |
| 2056 | if (trace_empty(iter)) | 2151 | if (trace_empty(iter)) |
| @@ -2701,20 +2796,11 @@ tracing_ctrl_write(struct file *filp, const char __user *ubuf, | |||
| 2701 | size_t cnt, loff_t *ppos) | 2796 | size_t cnt, loff_t *ppos) |
| 2702 | { | 2797 | { |
| 2703 | struct trace_array *tr = filp->private_data; | 2798 | struct trace_array *tr = filp->private_data; |
| 2704 | char buf[64]; | ||
| 2705 | unsigned long val; | 2799 | unsigned long val; |
| 2706 | int ret; | 2800 | int ret; |
| 2707 | 2801 | ||
| 2708 | if (cnt >= sizeof(buf)) | 2802 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 2709 | return -EINVAL; | 2803 | if (ret) |
| 2710 | |||
| 2711 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 2712 | return -EFAULT; | ||
| 2713 | |||
| 2714 | buf[cnt] = 0; | ||
| 2715 | |||
| 2716 | ret = strict_strtoul(buf, 10, &val); | ||
| 2717 | if (ret < 0) | ||
| 2718 | return ret; | 2804 | return ret; |
| 2719 | 2805 | ||
| 2720 | val = !!val; | 2806 | val = !!val; |
| @@ -2767,7 +2853,7 @@ int tracer_init(struct tracer *t, struct trace_array *tr) | |||
| 2767 | return t->init(tr); | 2853 | return t->init(tr); |
| 2768 | } | 2854 | } |
| 2769 | 2855 | ||
| 2770 | static int tracing_resize_ring_buffer(unsigned long size) | 2856 | static int __tracing_resize_ring_buffer(unsigned long size) |
| 2771 | { | 2857 | { |
| 2772 | int ret; | 2858 | int ret; |
| 2773 | 2859 | ||
| @@ -2819,6 +2905,41 @@ static int tracing_resize_ring_buffer(unsigned long size) | |||
| 2819 | return ret; | 2905 | return ret; |
| 2820 | } | 2906 | } |
| 2821 | 2907 | ||
| 2908 | static ssize_t tracing_resize_ring_buffer(unsigned long size) | ||
| 2909 | { | ||
| 2910 | int cpu, ret = size; | ||
| 2911 | |||
| 2912 | mutex_lock(&trace_types_lock); | ||
| 2913 | |||
| 2914 | tracing_stop(); | ||
| 2915 | |||
| 2916 | /* disable all cpu buffers */ | ||
| 2917 | for_each_tracing_cpu(cpu) { | ||
| 2918 | if (global_trace.data[cpu]) | ||
| 2919 | atomic_inc(&global_trace.data[cpu]->disabled); | ||
| 2920 | if (max_tr.data[cpu]) | ||
| 2921 | atomic_inc(&max_tr.data[cpu]->disabled); | ||
| 2922 | } | ||
| 2923 | |||
| 2924 | if (size != global_trace.entries) | ||
| 2925 | ret = __tracing_resize_ring_buffer(size); | ||
| 2926 | |||
| 2927 | if (ret < 0) | ||
| 2928 | ret = -ENOMEM; | ||
| 2929 | |||
| 2930 | for_each_tracing_cpu(cpu) { | ||
| 2931 | if (global_trace.data[cpu]) | ||
| 2932 | atomic_dec(&global_trace.data[cpu]->disabled); | ||
| 2933 | if (max_tr.data[cpu]) | ||
| 2934 | atomic_dec(&max_tr.data[cpu]->disabled); | ||
| 2935 | } | ||
| 2936 | |||
| 2937 | tracing_start(); | ||
| 2938 | mutex_unlock(&trace_types_lock); | ||
| 2939 | |||
| 2940 | return ret; | ||
| 2941 | } | ||
| 2942 | |||
| 2822 | 2943 | ||
| 2823 | /** | 2944 | /** |
| 2824 | * tracing_update_buffers - used by tracing facility to expand ring buffers | 2945 | * tracing_update_buffers - used by tracing facility to expand ring buffers |
| @@ -2836,7 +2957,7 @@ int tracing_update_buffers(void) | |||
| 2836 | 2957 | ||
| 2837 | mutex_lock(&trace_types_lock); | 2958 | mutex_lock(&trace_types_lock); |
| 2838 | if (!ring_buffer_expanded) | 2959 | if (!ring_buffer_expanded) |
| 2839 | ret = tracing_resize_ring_buffer(trace_buf_size); | 2960 | ret = __tracing_resize_ring_buffer(trace_buf_size); |
| 2840 | mutex_unlock(&trace_types_lock); | 2961 | mutex_unlock(&trace_types_lock); |
| 2841 | 2962 | ||
| 2842 | return ret; | 2963 | return ret; |
| @@ -2860,7 +2981,7 @@ static int tracing_set_tracer(const char *buf) | |||
| 2860 | mutex_lock(&trace_types_lock); | 2981 | mutex_lock(&trace_types_lock); |
| 2861 | 2982 | ||
| 2862 | if (!ring_buffer_expanded) { | 2983 | if (!ring_buffer_expanded) { |
| 2863 | ret = tracing_resize_ring_buffer(trace_buf_size); | 2984 | ret = __tracing_resize_ring_buffer(trace_buf_size); |
| 2864 | if (ret < 0) | 2985 | if (ret < 0) |
| 2865 | goto out; | 2986 | goto out; |
| 2866 | ret = 0; | 2987 | ret = 0; |
| @@ -2966,20 +3087,11 @@ tracing_max_lat_write(struct file *filp, const char __user *ubuf, | |||
| 2966 | size_t cnt, loff_t *ppos) | 3087 | size_t cnt, loff_t *ppos) |
| 2967 | { | 3088 | { |
| 2968 | unsigned long *ptr = filp->private_data; | 3089 | unsigned long *ptr = filp->private_data; |
| 2969 | char buf[64]; | ||
| 2970 | unsigned long val; | 3090 | unsigned long val; |
| 2971 | int ret; | 3091 | int ret; |
| 2972 | 3092 | ||
| 2973 | if (cnt >= sizeof(buf)) | 3093 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 2974 | return -EINVAL; | 3094 | if (ret) |
| 2975 | |||
| 2976 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 2977 | return -EFAULT; | ||
| 2978 | |||
| 2979 | buf[cnt] = 0; | ||
| 2980 | |||
| 2981 | ret = strict_strtoul(buf, 10, &val); | ||
| 2982 | if (ret < 0) | ||
| 2983 | return ret; | 3095 | return ret; |
| 2984 | 3096 | ||
| 2985 | *ptr = val * 1000; | 3097 | *ptr = val * 1000; |
| @@ -3434,67 +3546,54 @@ tracing_entries_write(struct file *filp, const char __user *ubuf, | |||
| 3434 | size_t cnt, loff_t *ppos) | 3546 | size_t cnt, loff_t *ppos) |
| 3435 | { | 3547 | { |
| 3436 | unsigned long val; | 3548 | unsigned long val; |
| 3437 | char buf[64]; | 3549 | int ret; |
| 3438 | int ret, cpu; | ||
| 3439 | |||
| 3440 | if (cnt >= sizeof(buf)) | ||
| 3441 | return -EINVAL; | ||
| 3442 | |||
| 3443 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 3444 | return -EFAULT; | ||
| 3445 | |||
| 3446 | buf[cnt] = 0; | ||
| 3447 | 3550 | ||
| 3448 | ret = strict_strtoul(buf, 10, &val); | 3551 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 3449 | if (ret < 0) | 3552 | if (ret) |
| 3450 | return ret; | 3553 | return ret; |
| 3451 | 3554 | ||
| 3452 | /* must have at least 1 entry */ | 3555 | /* must have at least 1 entry */ |
| 3453 | if (!val) | 3556 | if (!val) |
| 3454 | return -EINVAL; | 3557 | return -EINVAL; |
| 3455 | 3558 | ||
| 3456 | mutex_lock(&trace_types_lock); | ||
| 3457 | |||
| 3458 | tracing_stop(); | ||
| 3459 | |||
| 3460 | /* disable all cpu buffers */ | ||
| 3461 | for_each_tracing_cpu(cpu) { | ||
| 3462 | if (global_trace.data[cpu]) | ||
| 3463 | atomic_inc(&global_trace.data[cpu]->disabled); | ||
| 3464 | if (max_tr.data[cpu]) | ||
| 3465 | atomic_inc(&max_tr.data[cpu]->disabled); | ||
| 3466 | } | ||
| 3467 | |||
| 3468 | /* value is in KB */ | 3559 | /* value is in KB */ |
| 3469 | val <<= 10; | 3560 | val <<= 10; |
| 3470 | 3561 | ||
| 3471 | if (val != global_trace.entries) { | 3562 | ret = tracing_resize_ring_buffer(val); |
| 3472 | ret = tracing_resize_ring_buffer(val); | 3563 | if (ret < 0) |
| 3473 | if (ret < 0) { | 3564 | return ret; |
| 3474 | cnt = ret; | ||
| 3475 | goto out; | ||
| 3476 | } | ||
| 3477 | } | ||
| 3478 | 3565 | ||
| 3479 | *ppos += cnt; | 3566 | *ppos += cnt; |
| 3480 | 3567 | ||
| 3481 | /* If check pages failed, return ENOMEM */ | 3568 | return cnt; |
| 3482 | if (tracing_disabled) | 3569 | } |
| 3483 | cnt = -ENOMEM; | ||
| 3484 | out: | ||
| 3485 | for_each_tracing_cpu(cpu) { | ||
| 3486 | if (global_trace.data[cpu]) | ||
| 3487 | atomic_dec(&global_trace.data[cpu]->disabled); | ||
| 3488 | if (max_tr.data[cpu]) | ||
| 3489 | atomic_dec(&max_tr.data[cpu]->disabled); | ||
| 3490 | } | ||
| 3491 | 3570 | ||
| 3492 | tracing_start(); | 3571 | static ssize_t |
| 3493 | mutex_unlock(&trace_types_lock); | 3572 | tracing_free_buffer_write(struct file *filp, const char __user *ubuf, |
| 3573 | size_t cnt, loff_t *ppos) | ||
| 3574 | { | ||
| 3575 | /* | ||
| 3576 | * There is no need to read what the user has written, this function | ||
| 3577 | * is just to make sure that there is no error when "echo" is used | ||
| 3578 | */ | ||
| 3579 | |||
| 3580 | *ppos += cnt; | ||
| 3494 | 3581 | ||
| 3495 | return cnt; | 3582 | return cnt; |
| 3496 | } | 3583 | } |
| 3497 | 3584 | ||
| 3585 | static int | ||
| 3586 | tracing_free_buffer_release(struct inode *inode, struct file *filp) | ||
| 3587 | { | ||
| 3588 | /* disable tracing ? */ | ||
| 3589 | if (trace_flags & TRACE_ITER_STOP_ON_FREE) | ||
| 3590 | tracing_off(); | ||
| 3591 | /* resize the ring buffer to 0 */ | ||
| 3592 | tracing_resize_ring_buffer(0); | ||
| 3593 | |||
| 3594 | return 0; | ||
| 3595 | } | ||
| 3596 | |||
| 3498 | static int mark_printk(const char *fmt, ...) | 3597 | static int mark_printk(const char *fmt, ...) |
| 3499 | { | 3598 | { |
| 3500 | int ret; | 3599 | int ret; |
| @@ -3640,6 +3739,11 @@ static const struct file_operations tracing_entries_fops = { | |||
| 3640 | .llseek = generic_file_llseek, | 3739 | .llseek = generic_file_llseek, |
| 3641 | }; | 3740 | }; |
| 3642 | 3741 | ||
| 3742 | static const struct file_operations tracing_free_buffer_fops = { | ||
| 3743 | .write = tracing_free_buffer_write, | ||
| 3744 | .release = tracing_free_buffer_release, | ||
| 3745 | }; | ||
| 3746 | |||
| 3643 | static const struct file_operations tracing_mark_fops = { | 3747 | static const struct file_operations tracing_mark_fops = { |
| 3644 | .open = tracing_open_generic, | 3748 | .open = tracing_open_generic, |
| 3645 | .write = tracing_mark_write, | 3749 | .write = tracing_mark_write, |
| @@ -3696,7 +3800,7 @@ tracing_buffers_read(struct file *filp, char __user *ubuf, | |||
| 3696 | return 0; | 3800 | return 0; |
| 3697 | 3801 | ||
| 3698 | if (!info->spare) | 3802 | if (!info->spare) |
| 3699 | info->spare = ring_buffer_alloc_read_page(info->tr->buffer); | 3803 | info->spare = ring_buffer_alloc_read_page(info->tr->buffer, info->cpu); |
| 3700 | if (!info->spare) | 3804 | if (!info->spare) |
| 3701 | return -ENOMEM; | 3805 | return -ENOMEM; |
| 3702 | 3806 | ||
| @@ -3853,7 +3957,7 @@ tracing_buffers_splice_read(struct file *file, loff_t *ppos, | |||
| 3853 | 3957 | ||
| 3854 | ref->ref = 1; | 3958 | ref->ref = 1; |
| 3855 | ref->buffer = info->tr->buffer; | 3959 | ref->buffer = info->tr->buffer; |
| 3856 | ref->page = ring_buffer_alloc_read_page(ref->buffer); | 3960 | ref->page = ring_buffer_alloc_read_page(ref->buffer, info->cpu); |
| 3857 | if (!ref->page) { | 3961 | if (!ref->page) { |
| 3858 | kfree(ref); | 3962 | kfree(ref); |
| 3859 | break; | 3963 | break; |
| @@ -3862,8 +3966,7 @@ tracing_buffers_splice_read(struct file *file, loff_t *ppos, | |||
| 3862 | r = ring_buffer_read_page(ref->buffer, &ref->page, | 3966 | r = ring_buffer_read_page(ref->buffer, &ref->page, |
| 3863 | len, info->cpu, 1); | 3967 | len, info->cpu, 1); |
| 3864 | if (r < 0) { | 3968 | if (r < 0) { |
| 3865 | ring_buffer_free_read_page(ref->buffer, | 3969 | ring_buffer_free_read_page(ref->buffer, ref->page); |
| 3866 | ref->page); | ||
| 3867 | kfree(ref); | 3970 | kfree(ref); |
| 3868 | break; | 3971 | break; |
| 3869 | } | 3972 | } |
| @@ -4099,19 +4202,10 @@ trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 4099 | { | 4202 | { |
| 4100 | struct trace_option_dentry *topt = filp->private_data; | 4203 | struct trace_option_dentry *topt = filp->private_data; |
| 4101 | unsigned long val; | 4204 | unsigned long val; |
| 4102 | char buf[64]; | ||
| 4103 | int ret; | 4205 | int ret; |
| 4104 | 4206 | ||
| 4105 | if (cnt >= sizeof(buf)) | 4207 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 4106 | return -EINVAL; | 4208 | if (ret) |
| 4107 | |||
| 4108 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 4109 | return -EFAULT; | ||
| 4110 | |||
| 4111 | buf[cnt] = 0; | ||
| 4112 | |||
| 4113 | ret = strict_strtoul(buf, 10, &val); | ||
| 4114 | if (ret < 0) | ||
| 4115 | return ret; | 4209 | return ret; |
| 4116 | 4210 | ||
| 4117 | if (val != 0 && val != 1) | 4211 | if (val != 0 && val != 1) |
| @@ -4159,20 +4253,11 @@ trace_options_core_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 4159 | loff_t *ppos) | 4253 | loff_t *ppos) |
| 4160 | { | 4254 | { |
| 4161 | long index = (long)filp->private_data; | 4255 | long index = (long)filp->private_data; |
| 4162 | char buf[64]; | ||
| 4163 | unsigned long val; | 4256 | unsigned long val; |
| 4164 | int ret; | 4257 | int ret; |
| 4165 | 4258 | ||
| 4166 | if (cnt >= sizeof(buf)) | 4259 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 4167 | return -EINVAL; | 4260 | if (ret) |
| 4168 | |||
| 4169 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 4170 | return -EFAULT; | ||
| 4171 | |||
| 4172 | buf[cnt] = 0; | ||
| 4173 | |||
| 4174 | ret = strict_strtoul(buf, 10, &val); | ||
| 4175 | if (ret < 0) | ||
| 4176 | return ret; | 4261 | return ret; |
| 4177 | 4262 | ||
| 4178 | if (val != 0 && val != 1) | 4263 | if (val != 0 && val != 1) |
| @@ -4365,6 +4450,9 @@ static __init int tracer_init_debugfs(void) | |||
| 4365 | trace_create_file("buffer_size_kb", 0644, d_tracer, | 4450 | trace_create_file("buffer_size_kb", 0644, d_tracer, |
| 4366 | &global_trace, &tracing_entries_fops); | 4451 | &global_trace, &tracing_entries_fops); |
| 4367 | 4452 | ||
| 4453 | trace_create_file("free_buffer", 0644, d_tracer, | ||
| 4454 | &global_trace, &tracing_free_buffer_fops); | ||
| 4455 | |||
| 4368 | trace_create_file("trace_marker", 0220, d_tracer, | 4456 | trace_create_file("trace_marker", 0220, d_tracer, |
| 4369 | NULL, &tracing_mark_fops); | 4457 | NULL, &tracing_mark_fops); |
| 4370 | 4458 | ||
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index 229f8591f61d..616846bcfee5 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
| @@ -2,7 +2,7 @@ | |||
| 2 | #define _LINUX_KERNEL_TRACE_H | 2 | #define _LINUX_KERNEL_TRACE_H |
| 3 | 3 | ||
| 4 | #include <linux/fs.h> | 4 | #include <linux/fs.h> |
| 5 | #include <asm/atomic.h> | 5 | #include <linux/atomic.h> |
| 6 | #include <linux/sched.h> | 6 | #include <linux/sched.h> |
| 7 | #include <linux/clocksource.h> | 7 | #include <linux/clocksource.h> |
| 8 | #include <linux/ring_buffer.h> | 8 | #include <linux/ring_buffer.h> |
| @@ -278,6 +278,29 @@ struct tracer { | |||
| 278 | }; | 278 | }; |
| 279 | 279 | ||
| 280 | 280 | ||
| 281 | /* Only current can touch trace_recursion */ | ||
| 282 | #define trace_recursion_inc() do { (current)->trace_recursion++; } while (0) | ||
| 283 | #define trace_recursion_dec() do { (current)->trace_recursion--; } while (0) | ||
| 284 | |||
| 285 | /* Ring buffer has the 10 LSB bits to count */ | ||
| 286 | #define trace_recursion_buffer() ((current)->trace_recursion & 0x3ff) | ||
| 287 | |||
| 288 | /* for function tracing recursion */ | ||
| 289 | #define TRACE_INTERNAL_BIT (1<<11) | ||
| 290 | #define TRACE_GLOBAL_BIT (1<<12) | ||
| 291 | /* | ||
| 292 | * Abuse of the trace_recursion. | ||
| 293 | * As we need a way to maintain state if we are tracing the function | ||
| 294 | * graph in irq because we want to trace a particular function that | ||
| 295 | * was called in irq context but we have irq tracing off. Since this | ||
| 296 | * can only be modified by current, we can reuse trace_recursion. | ||
| 297 | */ | ||
| 298 | #define TRACE_IRQ_BIT (1<<13) | ||
| 299 | |||
| 300 | #define trace_recursion_set(bit) do { (current)->trace_recursion |= (bit); } while (0) | ||
| 301 | #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(bit); } while (0) | ||
| 302 | #define trace_recursion_test(bit) ((current)->trace_recursion & (bit)) | ||
| 303 | |||
| 281 | #define TRACE_PIPE_ALL_CPU -1 | 304 | #define TRACE_PIPE_ALL_CPU -1 |
| 282 | 305 | ||
| 283 | int tracer_init(struct tracer *t, struct trace_array *tr); | 306 | int tracer_init(struct tracer *t, struct trace_array *tr); |
| @@ -389,6 +412,9 @@ void update_max_tr_single(struct trace_array *tr, | |||
| 389 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, | 412 | void ftrace_trace_stack(struct ring_buffer *buffer, unsigned long flags, |
| 390 | int skip, int pc); | 413 | int skip, int pc); |
| 391 | 414 | ||
| 415 | void ftrace_trace_stack_regs(struct ring_buffer *buffer, unsigned long flags, | ||
| 416 | int skip, int pc, struct pt_regs *regs); | ||
| 417 | |||
| 392 | void ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, | 418 | void ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, |
| 393 | int pc); | 419 | int pc); |
| 394 | 420 | ||
| @@ -400,6 +426,12 @@ static inline void ftrace_trace_stack(struct ring_buffer *buffer, | |||
| 400 | { | 426 | { |
| 401 | } | 427 | } |
| 402 | 428 | ||
| 429 | static inline void ftrace_trace_stack_regs(struct ring_buffer *buffer, | ||
| 430 | unsigned long flags, int skip, | ||
| 431 | int pc, struct pt_regs *regs) | ||
| 432 | { | ||
| 433 | } | ||
| 434 | |||
| 403 | static inline void ftrace_trace_userstack(struct ring_buffer *buffer, | 435 | static inline void ftrace_trace_userstack(struct ring_buffer *buffer, |
| 404 | unsigned long flags, int pc) | 436 | unsigned long flags, int pc) |
| 405 | { | 437 | { |
| @@ -507,8 +539,18 @@ static inline int ftrace_graph_addr(unsigned long addr) | |||
| 507 | return 1; | 539 | return 1; |
| 508 | 540 | ||
| 509 | for (i = 0; i < ftrace_graph_count; i++) { | 541 | for (i = 0; i < ftrace_graph_count; i++) { |
| 510 | if (addr == ftrace_graph_funcs[i]) | 542 | if (addr == ftrace_graph_funcs[i]) { |
| 543 | /* | ||
| 544 | * If no irqs are to be traced, but a set_graph_function | ||
| 545 | * is set, and called by an interrupt handler, we still | ||
| 546 | * want to trace it. | ||
| 547 | */ | ||
| 548 | if (in_irq()) | ||
| 549 | trace_recursion_set(TRACE_IRQ_BIT); | ||
| 550 | else | ||
| 551 | trace_recursion_clear(TRACE_IRQ_BIT); | ||
| 511 | return 1; | 552 | return 1; |
| 553 | } | ||
| 512 | } | 554 | } |
| 513 | 555 | ||
| 514 | return 0; | 556 | return 0; |
| @@ -609,6 +651,7 @@ enum trace_iterator_flags { | |||
| 609 | TRACE_ITER_GRAPH_TIME = 0x80000, | 651 | TRACE_ITER_GRAPH_TIME = 0x80000, |
| 610 | TRACE_ITER_RECORD_CMD = 0x100000, | 652 | TRACE_ITER_RECORD_CMD = 0x100000, |
| 611 | TRACE_ITER_OVERWRITE = 0x200000, | 653 | TRACE_ITER_OVERWRITE = 0x200000, |
| 654 | TRACE_ITER_STOP_ON_FREE = 0x400000, | ||
| 612 | }; | 655 | }; |
| 613 | 656 | ||
| 614 | /* | 657 | /* |
| @@ -677,6 +720,7 @@ struct event_subsystem { | |||
| 677 | struct dentry *entry; | 720 | struct dentry *entry; |
| 678 | struct event_filter *filter; | 721 | struct event_filter *filter; |
| 679 | int nr_events; | 722 | int nr_events; |
| 723 | int ref_count; | ||
| 680 | }; | 724 | }; |
| 681 | 725 | ||
| 682 | #define FILTER_PRED_INVALID ((unsigned short)-1) | 726 | #define FILTER_PRED_INVALID ((unsigned short)-1) |
| @@ -784,19 +828,4 @@ extern const char *__stop___trace_bprintk_fmt[]; | |||
| 784 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) | 828 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) |
| 785 | #include "trace_entries.h" | 829 | #include "trace_entries.h" |
| 786 | 830 | ||
| 787 | /* Only current can touch trace_recursion */ | ||
| 788 | #define trace_recursion_inc() do { (current)->trace_recursion++; } while (0) | ||
| 789 | #define trace_recursion_dec() do { (current)->trace_recursion--; } while (0) | ||
| 790 | |||
| 791 | /* Ring buffer has the 10 LSB bits to count */ | ||
| 792 | #define trace_recursion_buffer() ((current)->trace_recursion & 0x3ff) | ||
| 793 | |||
| 794 | /* for function tracing recursion */ | ||
| 795 | #define TRACE_INTERNAL_BIT (1<<11) | ||
| 796 | #define TRACE_GLOBAL_BIT (1<<12) | ||
| 797 | |||
| 798 | #define trace_recursion_set(bit) do { (current)->trace_recursion |= (bit); } while (0) | ||
| 799 | #define trace_recursion_clear(bit) do { (current)->trace_recursion &= ~(bit); } while (0) | ||
| 800 | #define trace_recursion_test(bit) ((current)->trace_recursion & (bit)) | ||
| 801 | |||
| 802 | #endif /* _LINUX_KERNEL_TRACE_H */ | 831 | #endif /* _LINUX_KERNEL_TRACE_H */ |
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h index e32744c84d94..93365907f219 100644 --- a/kernel/trace/trace_entries.h +++ b/kernel/trace/trace_entries.h | |||
| @@ -161,7 +161,8 @@ FTRACE_ENTRY(kernel_stack, stack_entry, | |||
| 161 | TRACE_STACK, | 161 | TRACE_STACK, |
| 162 | 162 | ||
| 163 | F_STRUCT( | 163 | F_STRUCT( |
| 164 | __array( unsigned long, caller, FTRACE_STACK_ENTRIES ) | 164 | __field( int, size ) |
| 165 | __dynamic_array(unsigned long, caller ) | ||
| 165 | ), | 166 | ), |
| 166 | 167 | ||
| 167 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" | 168 | F_printk("\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n" |
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index 686ec399f2a8..581876f9f387 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
| @@ -244,6 +244,35 @@ static void ftrace_clear_events(void) | |||
| 244 | mutex_unlock(&event_mutex); | 244 | mutex_unlock(&event_mutex); |
| 245 | } | 245 | } |
| 246 | 246 | ||
| 247 | static void __put_system(struct event_subsystem *system) | ||
| 248 | { | ||
| 249 | struct event_filter *filter = system->filter; | ||
| 250 | |||
| 251 | WARN_ON_ONCE(system->ref_count == 0); | ||
| 252 | if (--system->ref_count) | ||
| 253 | return; | ||
| 254 | |||
| 255 | if (filter) { | ||
| 256 | kfree(filter->filter_string); | ||
| 257 | kfree(filter); | ||
| 258 | } | ||
| 259 | kfree(system->name); | ||
| 260 | kfree(system); | ||
| 261 | } | ||
| 262 | |||
| 263 | static void __get_system(struct event_subsystem *system) | ||
| 264 | { | ||
| 265 | WARN_ON_ONCE(system->ref_count == 0); | ||
| 266 | system->ref_count++; | ||
| 267 | } | ||
| 268 | |||
| 269 | static void put_system(struct event_subsystem *system) | ||
| 270 | { | ||
| 271 | mutex_lock(&event_mutex); | ||
| 272 | __put_system(system); | ||
| 273 | mutex_unlock(&event_mutex); | ||
| 274 | } | ||
| 275 | |||
| 247 | /* | 276 | /* |
| 248 | * __ftrace_set_clr_event(NULL, NULL, NULL, set) will set/unset all events. | 277 | * __ftrace_set_clr_event(NULL, NULL, NULL, set) will set/unset all events. |
| 249 | */ | 278 | */ |
| @@ -486,20 +515,11 @@ event_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 486 | loff_t *ppos) | 515 | loff_t *ppos) |
| 487 | { | 516 | { |
| 488 | struct ftrace_event_call *call = filp->private_data; | 517 | struct ftrace_event_call *call = filp->private_data; |
| 489 | char buf[64]; | ||
| 490 | unsigned long val; | 518 | unsigned long val; |
| 491 | int ret; | 519 | int ret; |
| 492 | 520 | ||
| 493 | if (cnt >= sizeof(buf)) | 521 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 494 | return -EINVAL; | 522 | if (ret) |
| 495 | |||
| 496 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 497 | return -EFAULT; | ||
| 498 | |||
| 499 | buf[cnt] = 0; | ||
| 500 | |||
| 501 | ret = strict_strtoul(buf, 10, &val); | ||
| 502 | if (ret < 0) | ||
| 503 | return ret; | 523 | return ret; |
| 504 | 524 | ||
| 505 | ret = tracing_update_buffers(); | 525 | ret = tracing_update_buffers(); |
| @@ -528,7 +548,7 @@ system_enable_read(struct file *filp, char __user *ubuf, size_t cnt, | |||
| 528 | loff_t *ppos) | 548 | loff_t *ppos) |
| 529 | { | 549 | { |
| 530 | const char set_to_char[4] = { '?', '0', '1', 'X' }; | 550 | const char set_to_char[4] = { '?', '0', '1', 'X' }; |
| 531 | const char *system = filp->private_data; | 551 | struct event_subsystem *system = filp->private_data; |
| 532 | struct ftrace_event_call *call; | 552 | struct ftrace_event_call *call; |
| 533 | char buf[2]; | 553 | char buf[2]; |
| 534 | int set = 0; | 554 | int set = 0; |
| @@ -539,7 +559,7 @@ system_enable_read(struct file *filp, char __user *ubuf, size_t cnt, | |||
| 539 | if (!call->name || !call->class || !call->class->reg) | 559 | if (!call->name || !call->class || !call->class->reg) |
| 540 | continue; | 560 | continue; |
| 541 | 561 | ||
| 542 | if (system && strcmp(call->class->system, system) != 0) | 562 | if (system && strcmp(call->class->system, system->name) != 0) |
| 543 | continue; | 563 | continue; |
| 544 | 564 | ||
| 545 | /* | 565 | /* |
| @@ -569,21 +589,13 @@ static ssize_t | |||
| 569 | system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | 589 | system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, |
| 570 | loff_t *ppos) | 590 | loff_t *ppos) |
| 571 | { | 591 | { |
| 572 | const char *system = filp->private_data; | 592 | struct event_subsystem *system = filp->private_data; |
| 593 | const char *name = NULL; | ||
| 573 | unsigned long val; | 594 | unsigned long val; |
| 574 | char buf[64]; | ||
| 575 | ssize_t ret; | 595 | ssize_t ret; |
| 576 | 596 | ||
| 577 | if (cnt >= sizeof(buf)) | 597 | ret = kstrtoul_from_user(ubuf, cnt, 10, &val); |
| 578 | return -EINVAL; | 598 | if (ret) |
| 579 | |||
| 580 | if (copy_from_user(&buf, ubuf, cnt)) | ||
| 581 | return -EFAULT; | ||
| 582 | |||
| 583 | buf[cnt] = 0; | ||
| 584 | |||
| 585 | ret = strict_strtoul(buf, 10, &val); | ||
| 586 | if (ret < 0) | ||
| 587 | return ret; | 599 | return ret; |
| 588 | 600 | ||
| 589 | ret = tracing_update_buffers(); | 601 | ret = tracing_update_buffers(); |
| @@ -593,7 +605,14 @@ system_enable_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 593 | if (val != 0 && val != 1) | 605 | if (val != 0 && val != 1) |
| 594 | return -EINVAL; | 606 | return -EINVAL; |
| 595 | 607 | ||
| 596 | ret = __ftrace_set_clr_event(NULL, system, NULL, val); | 608 | /* |
| 609 | * Opening of "enable" adds a ref count to system, | ||
| 610 | * so the name is safe to use. | ||
| 611 | */ | ||
| 612 | if (system) | ||
| 613 | name = system->name; | ||
| 614 | |||
| 615 | ret = __ftrace_set_clr_event(NULL, name, NULL, val); | ||
| 597 | if (ret) | 616 | if (ret) |
| 598 | goto out; | 617 | goto out; |
| 599 | 618 | ||
| @@ -826,6 +845,52 @@ event_filter_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 826 | return cnt; | 845 | return cnt; |
| 827 | } | 846 | } |
| 828 | 847 | ||
| 848 | static LIST_HEAD(event_subsystems); | ||
| 849 | |||
| 850 | static int subsystem_open(struct inode *inode, struct file *filp) | ||
| 851 | { | ||
| 852 | struct event_subsystem *system = NULL; | ||
| 853 | int ret; | ||
| 854 | |||
| 855 | if (!inode->i_private) | ||
| 856 | goto skip_search; | ||
| 857 | |||
| 858 | /* Make sure the system still exists */ | ||
| 859 | mutex_lock(&event_mutex); | ||
| 860 | list_for_each_entry(system, &event_subsystems, list) { | ||
| 861 | if (system == inode->i_private) { | ||
| 862 | /* Don't open systems with no events */ | ||
| 863 | if (!system->nr_events) { | ||
| 864 | system = NULL; | ||
| 865 | break; | ||
| 866 | } | ||
| 867 | __get_system(system); | ||
| 868 | break; | ||
| 869 | } | ||
| 870 | } | ||
| 871 | mutex_unlock(&event_mutex); | ||
| 872 | |||
| 873 | if (system != inode->i_private) | ||
| 874 | return -ENODEV; | ||
| 875 | |||
| 876 | skip_search: | ||
| 877 | ret = tracing_open_generic(inode, filp); | ||
| 878 | if (ret < 0 && system) | ||
| 879 | put_system(system); | ||
| 880 | |||
| 881 | return ret; | ||
| 882 | } | ||
| 883 | |||
| 884 | static int subsystem_release(struct inode *inode, struct file *file) | ||
| 885 | { | ||
| 886 | struct event_subsystem *system = inode->i_private; | ||
| 887 | |||
| 888 | if (system) | ||
| 889 | put_system(system); | ||
| 890 | |||
| 891 | return 0; | ||
| 892 | } | ||
| 893 | |||
| 829 | static ssize_t | 894 | static ssize_t |
| 830 | subsystem_filter_read(struct file *filp, char __user *ubuf, size_t cnt, | 895 | subsystem_filter_read(struct file *filp, char __user *ubuf, size_t cnt, |
| 831 | loff_t *ppos) | 896 | loff_t *ppos) |
| @@ -963,17 +1028,19 @@ static const struct file_operations ftrace_event_filter_fops = { | |||
| 963 | }; | 1028 | }; |
| 964 | 1029 | ||
| 965 | static const struct file_operations ftrace_subsystem_filter_fops = { | 1030 | static const struct file_operations ftrace_subsystem_filter_fops = { |
| 966 | .open = tracing_open_generic, | 1031 | .open = subsystem_open, |
| 967 | .read = subsystem_filter_read, | 1032 | .read = subsystem_filter_read, |
| 968 | .write = subsystem_filter_write, | 1033 | .write = subsystem_filter_write, |
| 969 | .llseek = default_llseek, | 1034 | .llseek = default_llseek, |
| 1035 | .release = subsystem_release, | ||
| 970 | }; | 1036 | }; |
| 971 | 1037 | ||
| 972 | static const struct file_operations ftrace_system_enable_fops = { | 1038 | static const struct file_operations ftrace_system_enable_fops = { |
| 973 | .open = tracing_open_generic, | 1039 | .open = subsystem_open, |
| 974 | .read = system_enable_read, | 1040 | .read = system_enable_read, |
| 975 | .write = system_enable_write, | 1041 | .write = system_enable_write, |
| 976 | .llseek = default_llseek, | 1042 | .llseek = default_llseek, |
| 1043 | .release = subsystem_release, | ||
| 977 | }; | 1044 | }; |
| 978 | 1045 | ||
| 979 | static const struct file_operations ftrace_show_header_fops = { | 1046 | static const struct file_operations ftrace_show_header_fops = { |
| @@ -1002,8 +1069,6 @@ static struct dentry *event_trace_events_dir(void) | |||
| 1002 | return d_events; | 1069 | return d_events; |
| 1003 | } | 1070 | } |
| 1004 | 1071 | ||
| 1005 | static LIST_HEAD(event_subsystems); | ||
| 1006 | |||
| 1007 | static struct dentry * | 1072 | static struct dentry * |
| 1008 | event_subsystem_dir(const char *name, struct dentry *d_events) | 1073 | event_subsystem_dir(const char *name, struct dentry *d_events) |
| 1009 | { | 1074 | { |
| @@ -1013,6 +1078,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1013 | /* First see if we did not already create this dir */ | 1078 | /* First see if we did not already create this dir */ |
| 1014 | list_for_each_entry(system, &event_subsystems, list) { | 1079 | list_for_each_entry(system, &event_subsystems, list) { |
| 1015 | if (strcmp(system->name, name) == 0) { | 1080 | if (strcmp(system->name, name) == 0) { |
| 1081 | __get_system(system); | ||
| 1016 | system->nr_events++; | 1082 | system->nr_events++; |
| 1017 | return system->entry; | 1083 | return system->entry; |
| 1018 | } | 1084 | } |
| @@ -1035,6 +1101,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1035 | } | 1101 | } |
| 1036 | 1102 | ||
| 1037 | system->nr_events = 1; | 1103 | system->nr_events = 1; |
| 1104 | system->ref_count = 1; | ||
| 1038 | system->name = kstrdup(name, GFP_KERNEL); | 1105 | system->name = kstrdup(name, GFP_KERNEL); |
| 1039 | if (!system->name) { | 1106 | if (!system->name) { |
| 1040 | debugfs_remove(system->entry); | 1107 | debugfs_remove(system->entry); |
| @@ -1062,8 +1129,7 @@ event_subsystem_dir(const char *name, struct dentry *d_events) | |||
| 1062 | "'%s/filter' entry\n", name); | 1129 | "'%s/filter' entry\n", name); |
| 1063 | } | 1130 | } |
| 1064 | 1131 | ||
| 1065 | trace_create_file("enable", 0644, system->entry, | 1132 | trace_create_file("enable", 0644, system->entry, system, |
| 1066 | (void *)system->name, | ||
| 1067 | &ftrace_system_enable_fops); | 1133 | &ftrace_system_enable_fops); |
| 1068 | 1134 | ||
| 1069 | return system->entry; | 1135 | return system->entry; |
| @@ -1184,16 +1250,9 @@ static void remove_subsystem_dir(const char *name) | |||
| 1184 | list_for_each_entry(system, &event_subsystems, list) { | 1250 | list_for_each_entry(system, &event_subsystems, list) { |
| 1185 | if (strcmp(system->name, name) == 0) { | 1251 | if (strcmp(system->name, name) == 0) { |
| 1186 | if (!--system->nr_events) { | 1252 | if (!--system->nr_events) { |
| 1187 | struct event_filter *filter = system->filter; | ||
| 1188 | |||
| 1189 | debugfs_remove_recursive(system->entry); | 1253 | debugfs_remove_recursive(system->entry); |
| 1190 | list_del(&system->list); | 1254 | list_del(&system->list); |
| 1191 | if (filter) { | 1255 | __put_system(system); |
| 1192 | kfree(filter->filter_string); | ||
| 1193 | kfree(filter); | ||
| 1194 | } | ||
| 1195 | kfree(system->name); | ||
| 1196 | kfree(system); | ||
| 1197 | } | 1256 | } |
| 1198 | break; | 1257 | break; |
| 1199 | } | 1258 | } |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 8008ddcfbf20..256764ecccd6 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
| @@ -1886,6 +1886,12 @@ int apply_subsystem_event_filter(struct event_subsystem *system, | |||
| 1886 | 1886 | ||
| 1887 | mutex_lock(&event_mutex); | 1887 | mutex_lock(&event_mutex); |
| 1888 | 1888 | ||
| 1889 | /* Make sure the system still has events */ | ||
| 1890 | if (!system->nr_events) { | ||
| 1891 | err = -ENODEV; | ||
| 1892 | goto out_unlock; | ||
| 1893 | } | ||
| 1894 | |||
| 1889 | if (!strcmp(strstrip(filter_string), "0")) { | 1895 | if (!strcmp(strstrip(filter_string), "0")) { |
| 1890 | filter_free_subsystem_preds(system); | 1896 | filter_free_subsystem_preds(system); |
| 1891 | remove_filter_string(system->filter); | 1897 | remove_filter_string(system->filter); |
diff --git a/kernel/trace/trace_functions.c b/kernel/trace/trace_functions.c index 8d0e1cc4e974..c7b0c6a7db09 100644 --- a/kernel/trace/trace_functions.c +++ b/kernel/trace/trace_functions.c | |||
| @@ -324,7 +324,8 @@ ftrace_trace_onoff_unreg(char *glob, char *cmd, char *param) | |||
| 324 | } | 324 | } |
| 325 | 325 | ||
| 326 | static int | 326 | static int |
| 327 | ftrace_trace_onoff_callback(char *glob, char *cmd, char *param, int enable) | 327 | ftrace_trace_onoff_callback(struct ftrace_hash *hash, |
| 328 | char *glob, char *cmd, char *param, int enable) | ||
| 328 | { | 329 | { |
| 329 | struct ftrace_probe_ops *ops; | 330 | struct ftrace_probe_ops *ops; |
| 330 | void *count = (void *)-1; | 331 | void *count = (void *)-1; |
diff --git a/kernel/trace/trace_functions_graph.c b/kernel/trace/trace_functions_graph.c index 962cdb24ed81..a7d2a4c653d8 100644 --- a/kernel/trace/trace_functions_graph.c +++ b/kernel/trace/trace_functions_graph.c | |||
| @@ -74,6 +74,20 @@ static struct tracer_flags tracer_flags = { | |||
| 74 | 74 | ||
| 75 | static struct trace_array *graph_array; | 75 | static struct trace_array *graph_array; |
| 76 | 76 | ||
| 77 | /* | ||
| 78 | * DURATION column is being also used to display IRQ signs, | ||
| 79 | * following values are used by print_graph_irq and others | ||
| 80 | * to fill in space into DURATION column. | ||
| 81 | */ | ||
| 82 | enum { | ||
| 83 | DURATION_FILL_FULL = -1, | ||
| 84 | DURATION_FILL_START = -2, | ||
| 85 | DURATION_FILL_END = -3, | ||
| 86 | }; | ||
| 87 | |||
| 88 | static enum print_line_t | ||
| 89 | print_graph_duration(unsigned long long duration, struct trace_seq *s, | ||
| 90 | u32 flags); | ||
| 77 | 91 | ||
| 78 | /* Add a function return address to the trace stack on thread info.*/ | 92 | /* Add a function return address to the trace stack on thread info.*/ |
| 79 | int | 93 | int |
| @@ -213,7 +227,7 @@ int __trace_graph_entry(struct trace_array *tr, | |||
| 213 | 227 | ||
| 214 | static inline int ftrace_graph_ignore_irqs(void) | 228 | static inline int ftrace_graph_ignore_irqs(void) |
| 215 | { | 229 | { |
| 216 | if (!ftrace_graph_skip_irqs) | 230 | if (!ftrace_graph_skip_irqs || trace_recursion_test(TRACE_IRQ_BIT)) |
| 217 | return 0; | 231 | return 0; |
| 218 | 232 | ||
| 219 | return in_irq(); | 233 | return in_irq(); |
| @@ -577,32 +591,6 @@ get_return_for_leaf(struct trace_iterator *iter, | |||
| 577 | return next; | 591 | return next; |
| 578 | } | 592 | } |
| 579 | 593 | ||
| 580 | /* Signal a overhead of time execution to the output */ | ||
| 581 | static int | ||
| 582 | print_graph_overhead(unsigned long long duration, struct trace_seq *s, | ||
| 583 | u32 flags) | ||
| 584 | { | ||
| 585 | /* If duration disappear, we don't need anything */ | ||
| 586 | if (!(flags & TRACE_GRAPH_PRINT_DURATION)) | ||
| 587 | return 1; | ||
| 588 | |||
| 589 | /* Non nested entry or return */ | ||
| 590 | if (duration == -1) | ||
| 591 | return trace_seq_printf(s, " "); | ||
| 592 | |||
| 593 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { | ||
| 594 | /* Duration exceeded 100 msecs */ | ||
| 595 | if (duration > 100000ULL) | ||
| 596 | return trace_seq_printf(s, "! "); | ||
| 597 | |||
| 598 | /* Duration exceeded 10 msecs */ | ||
| 599 | if (duration > 10000ULL) | ||
| 600 | return trace_seq_printf(s, "+ "); | ||
| 601 | } | ||
| 602 | |||
| 603 | return trace_seq_printf(s, " "); | ||
| 604 | } | ||
| 605 | |||
| 606 | static int print_graph_abs_time(u64 t, struct trace_seq *s) | 594 | static int print_graph_abs_time(u64 t, struct trace_seq *s) |
| 607 | { | 595 | { |
| 608 | unsigned long usecs_rem; | 596 | unsigned long usecs_rem; |
| @@ -625,34 +613,36 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
| 625 | addr >= (unsigned long)__irqentry_text_end) | 613 | addr >= (unsigned long)__irqentry_text_end) |
| 626 | return TRACE_TYPE_UNHANDLED; | 614 | return TRACE_TYPE_UNHANDLED; |
| 627 | 615 | ||
| 628 | /* Absolute time */ | 616 | if (trace_flags & TRACE_ITER_CONTEXT_INFO) { |
| 629 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { | 617 | /* Absolute time */ |
| 630 | ret = print_graph_abs_time(iter->ts, s); | 618 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { |
| 631 | if (!ret) | 619 | ret = print_graph_abs_time(iter->ts, s); |
| 632 | return TRACE_TYPE_PARTIAL_LINE; | 620 | if (!ret) |
| 633 | } | 621 | return TRACE_TYPE_PARTIAL_LINE; |
| 622 | } | ||
| 634 | 623 | ||
| 635 | /* Cpu */ | 624 | /* Cpu */ |
| 636 | if (flags & TRACE_GRAPH_PRINT_CPU) { | 625 | if (flags & TRACE_GRAPH_PRINT_CPU) { |
| 637 | ret = print_graph_cpu(s, cpu); | 626 | ret = print_graph_cpu(s, cpu); |
| 638 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 627 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 639 | return TRACE_TYPE_PARTIAL_LINE; | 628 | return TRACE_TYPE_PARTIAL_LINE; |
| 640 | } | 629 | } |
| 641 | 630 | ||
| 642 | /* Proc */ | 631 | /* Proc */ |
| 643 | if (flags & TRACE_GRAPH_PRINT_PROC) { | 632 | if (flags & TRACE_GRAPH_PRINT_PROC) { |
| 644 | ret = print_graph_proc(s, pid); | 633 | ret = print_graph_proc(s, pid); |
| 645 | if (ret == TRACE_TYPE_PARTIAL_LINE) | 634 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 646 | return TRACE_TYPE_PARTIAL_LINE; | 635 | return TRACE_TYPE_PARTIAL_LINE; |
| 647 | ret = trace_seq_printf(s, " | "); | 636 | ret = trace_seq_printf(s, " | "); |
| 648 | if (!ret) | 637 | if (!ret) |
| 649 | return TRACE_TYPE_PARTIAL_LINE; | 638 | return TRACE_TYPE_PARTIAL_LINE; |
| 639 | } | ||
| 650 | } | 640 | } |
| 651 | 641 | ||
| 652 | /* No overhead */ | 642 | /* No overhead */ |
| 653 | ret = print_graph_overhead(-1, s, flags); | 643 | ret = print_graph_duration(DURATION_FILL_START, s, flags); |
| 654 | if (!ret) | 644 | if (ret != TRACE_TYPE_HANDLED) |
| 655 | return TRACE_TYPE_PARTIAL_LINE; | 645 | return ret; |
| 656 | 646 | ||
| 657 | if (type == TRACE_GRAPH_ENT) | 647 | if (type == TRACE_GRAPH_ENT) |
| 658 | ret = trace_seq_printf(s, "==========>"); | 648 | ret = trace_seq_printf(s, "==========>"); |
| @@ -662,9 +652,10 @@ print_graph_irq(struct trace_iterator *iter, unsigned long addr, | |||
| 662 | if (!ret) | 652 | if (!ret) |
| 663 | return TRACE_TYPE_PARTIAL_LINE; | 653 | return TRACE_TYPE_PARTIAL_LINE; |
| 664 | 654 | ||
| 665 | /* Don't close the duration column if haven't one */ | 655 | ret = print_graph_duration(DURATION_FILL_END, s, flags); |
| 666 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 656 | if (ret != TRACE_TYPE_HANDLED) |
| 667 | trace_seq_printf(s, " |"); | 657 | return ret; |
| 658 | |||
| 668 | ret = trace_seq_printf(s, "\n"); | 659 | ret = trace_seq_printf(s, "\n"); |
| 669 | 660 | ||
| 670 | if (!ret) | 661 | if (!ret) |
| @@ -716,9 +707,49 @@ trace_print_graph_duration(unsigned long long duration, struct trace_seq *s) | |||
| 716 | } | 707 | } |
| 717 | 708 | ||
| 718 | static enum print_line_t | 709 | static enum print_line_t |
| 719 | print_graph_duration(unsigned long long duration, struct trace_seq *s) | 710 | print_graph_duration(unsigned long long duration, struct trace_seq *s, |
| 711 | u32 flags) | ||
| 720 | { | 712 | { |
| 721 | int ret; | 713 | int ret = -1; |
| 714 | |||
| 715 | if (!(flags & TRACE_GRAPH_PRINT_DURATION) || | ||
| 716 | !(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 717 | return TRACE_TYPE_HANDLED; | ||
| 718 | |||
| 719 | /* No real adata, just filling the column with spaces */ | ||
| 720 | switch (duration) { | ||
| 721 | case DURATION_FILL_FULL: | ||
| 722 | ret = trace_seq_printf(s, " | "); | ||
| 723 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 724 | case DURATION_FILL_START: | ||
| 725 | ret = trace_seq_printf(s, " "); | ||
| 726 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 727 | case DURATION_FILL_END: | ||
| 728 | ret = trace_seq_printf(s, " |"); | ||
| 729 | return ret ? TRACE_TYPE_HANDLED : TRACE_TYPE_PARTIAL_LINE; | ||
| 730 | } | ||
| 731 | |||
| 732 | /* Signal a overhead of time execution to the output */ | ||
| 733 | if (flags & TRACE_GRAPH_PRINT_OVERHEAD) { | ||
| 734 | /* Duration exceeded 100 msecs */ | ||
| 735 | if (duration > 100000ULL) | ||
| 736 | ret = trace_seq_printf(s, "! "); | ||
| 737 | /* Duration exceeded 10 msecs */ | ||
| 738 | else if (duration > 10000ULL) | ||
| 739 | ret = trace_seq_printf(s, "+ "); | ||
| 740 | } | ||
| 741 | |||
| 742 | /* | ||
| 743 | * The -1 means we either did not exceed the duration tresholds | ||
| 744 | * or we dont want to print out the overhead. Either way we need | ||
| 745 | * to fill out the space. | ||
| 746 | */ | ||
| 747 | if (ret == -1) | ||
| 748 | ret = trace_seq_printf(s, " "); | ||
| 749 | |||
| 750 | /* Catching here any failure happenned above */ | ||
| 751 | if (!ret) | ||
| 752 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 722 | 753 | ||
| 723 | ret = trace_print_graph_duration(duration, s); | 754 | ret = trace_print_graph_duration(duration, s); |
| 724 | if (ret != TRACE_TYPE_HANDLED) | 755 | if (ret != TRACE_TYPE_HANDLED) |
| @@ -767,18 +798,11 @@ print_graph_entry_leaf(struct trace_iterator *iter, | |||
| 767 | cpu_data->enter_funcs[call->depth] = 0; | 798 | cpu_data->enter_funcs[call->depth] = 0; |
| 768 | } | 799 | } |
| 769 | 800 | ||
| 770 | /* Overhead */ | 801 | /* Overhead and duration */ |
| 771 | ret = print_graph_overhead(duration, s, flags); | 802 | ret = print_graph_duration(duration, s, flags); |
| 772 | if (!ret) | 803 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 773 | return TRACE_TYPE_PARTIAL_LINE; | 804 | return TRACE_TYPE_PARTIAL_LINE; |
| 774 | 805 | ||
| 775 | /* Duration */ | ||
| 776 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | ||
| 777 | ret = print_graph_duration(duration, s); | ||
| 778 | if (ret == TRACE_TYPE_PARTIAL_LINE) | ||
| 779 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 780 | } | ||
| 781 | |||
| 782 | /* Function */ | 806 | /* Function */ |
| 783 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 807 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
| 784 | ret = trace_seq_printf(s, " "); | 808 | ret = trace_seq_printf(s, " "); |
| @@ -815,17 +839,10 @@ print_graph_entry_nested(struct trace_iterator *iter, | |||
| 815 | cpu_data->enter_funcs[call->depth] = call->func; | 839 | cpu_data->enter_funcs[call->depth] = call->func; |
| 816 | } | 840 | } |
| 817 | 841 | ||
| 818 | /* No overhead */ | ||
| 819 | ret = print_graph_overhead(-1, s, flags); | ||
| 820 | if (!ret) | ||
| 821 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 822 | |||
| 823 | /* No time */ | 842 | /* No time */ |
| 824 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | 843 | ret = print_graph_duration(DURATION_FILL_FULL, s, flags); |
| 825 | ret = trace_seq_printf(s, " | "); | 844 | if (ret != TRACE_TYPE_HANDLED) |
| 826 | if (!ret) | 845 | return ret; |
| 827 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 828 | } | ||
| 829 | 846 | ||
| 830 | /* Function */ | 847 | /* Function */ |
| 831 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { | 848 | for (i = 0; i < call->depth * TRACE_GRAPH_INDENT; i++) { |
| @@ -865,6 +882,9 @@ print_graph_prologue(struct trace_iterator *iter, struct trace_seq *s, | |||
| 865 | return TRACE_TYPE_PARTIAL_LINE; | 882 | return TRACE_TYPE_PARTIAL_LINE; |
| 866 | } | 883 | } |
| 867 | 884 | ||
| 885 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 886 | return 0; | ||
| 887 | |||
| 868 | /* Absolute time */ | 888 | /* Absolute time */ |
| 869 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { | 889 | if (flags & TRACE_GRAPH_PRINT_ABS_TIME) { |
| 870 | ret = print_graph_abs_time(iter->ts, s); | 890 | ret = print_graph_abs_time(iter->ts, s); |
| @@ -1078,18 +1098,11 @@ print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s, | |||
| 1078 | if (print_graph_prologue(iter, s, 0, 0, flags)) | 1098 | if (print_graph_prologue(iter, s, 0, 0, flags)) |
| 1079 | return TRACE_TYPE_PARTIAL_LINE; | 1099 | return TRACE_TYPE_PARTIAL_LINE; |
| 1080 | 1100 | ||
| 1081 | /* Overhead */ | 1101 | /* Overhead and duration */ |
| 1082 | ret = print_graph_overhead(duration, s, flags); | 1102 | ret = print_graph_duration(duration, s, flags); |
| 1083 | if (!ret) | 1103 | if (ret == TRACE_TYPE_PARTIAL_LINE) |
| 1084 | return TRACE_TYPE_PARTIAL_LINE; | 1104 | return TRACE_TYPE_PARTIAL_LINE; |
| 1085 | 1105 | ||
| 1086 | /* Duration */ | ||
| 1087 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | ||
| 1088 | ret = print_graph_duration(duration, s); | ||
| 1089 | if (ret == TRACE_TYPE_PARTIAL_LINE) | ||
| 1090 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1091 | } | ||
| 1092 | |||
| 1093 | /* Closing brace */ | 1106 | /* Closing brace */ |
| 1094 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { | 1107 | for (i = 0; i < trace->depth * TRACE_GRAPH_INDENT; i++) { |
| 1095 | ret = trace_seq_printf(s, " "); | 1108 | ret = trace_seq_printf(s, " "); |
| @@ -1146,17 +1159,10 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
| 1146 | if (print_graph_prologue(iter, s, 0, 0, flags)) | 1159 | if (print_graph_prologue(iter, s, 0, 0, flags)) |
| 1147 | return TRACE_TYPE_PARTIAL_LINE; | 1160 | return TRACE_TYPE_PARTIAL_LINE; |
| 1148 | 1161 | ||
| 1149 | /* No overhead */ | ||
| 1150 | ret = print_graph_overhead(-1, s, flags); | ||
| 1151 | if (!ret) | ||
| 1152 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1153 | |||
| 1154 | /* No time */ | 1162 | /* No time */ |
| 1155 | if (flags & TRACE_GRAPH_PRINT_DURATION) { | 1163 | ret = print_graph_duration(DURATION_FILL_FULL, s, flags); |
| 1156 | ret = trace_seq_printf(s, " | "); | 1164 | if (ret != TRACE_TYPE_HANDLED) |
| 1157 | if (!ret) | 1165 | return ret; |
| 1158 | return TRACE_TYPE_PARTIAL_LINE; | ||
| 1159 | } | ||
| 1160 | 1166 | ||
| 1161 | /* Indentation */ | 1167 | /* Indentation */ |
| 1162 | if (depth > 0) | 1168 | if (depth > 0) |
| @@ -1207,7 +1213,7 @@ print_graph_comment(struct trace_seq *s, struct trace_entry *ent, | |||
| 1207 | 1213 | ||
| 1208 | 1214 | ||
| 1209 | enum print_line_t | 1215 | enum print_line_t |
| 1210 | __print_graph_function_flags(struct trace_iterator *iter, u32 flags) | 1216 | print_graph_function_flags(struct trace_iterator *iter, u32 flags) |
| 1211 | { | 1217 | { |
| 1212 | struct ftrace_graph_ent_entry *field; | 1218 | struct ftrace_graph_ent_entry *field; |
| 1213 | struct fgraph_data *data = iter->private; | 1219 | struct fgraph_data *data = iter->private; |
| @@ -1270,18 +1276,7 @@ __print_graph_function_flags(struct trace_iterator *iter, u32 flags) | |||
| 1270 | static enum print_line_t | 1276 | static enum print_line_t |
| 1271 | print_graph_function(struct trace_iterator *iter) | 1277 | print_graph_function(struct trace_iterator *iter) |
| 1272 | { | 1278 | { |
| 1273 | return __print_graph_function_flags(iter, tracer_flags.val); | 1279 | return print_graph_function_flags(iter, tracer_flags.val); |
| 1274 | } | ||
| 1275 | |||
| 1276 | enum print_line_t print_graph_function_flags(struct trace_iterator *iter, | ||
| 1277 | u32 flags) | ||
| 1278 | { | ||
| 1279 | if (trace_flags & TRACE_ITER_LATENCY_FMT) | ||
| 1280 | flags |= TRACE_GRAPH_PRINT_DURATION; | ||
| 1281 | else | ||
| 1282 | flags |= TRACE_GRAPH_PRINT_ABS_TIME; | ||
| 1283 | |||
| 1284 | return __print_graph_function_flags(iter, flags); | ||
| 1285 | } | 1280 | } |
| 1286 | 1281 | ||
| 1287 | static enum print_line_t | 1282 | static enum print_line_t |
| @@ -1309,8 +1304,7 @@ static void print_lat_header(struct seq_file *s, u32 flags) | |||
| 1309 | seq_printf(s, "#%.*s / _----=> need-resched \n", size, spaces); | 1304 | seq_printf(s, "#%.*s / _----=> need-resched \n", size, spaces); |
| 1310 | seq_printf(s, "#%.*s| / _---=> hardirq/softirq \n", size, spaces); | 1305 | seq_printf(s, "#%.*s| / _---=> hardirq/softirq \n", size, spaces); |
| 1311 | seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); | 1306 | seq_printf(s, "#%.*s|| / _--=> preempt-depth \n", size, spaces); |
| 1312 | seq_printf(s, "#%.*s||| / _-=> lock-depth \n", size, spaces); | 1307 | seq_printf(s, "#%.*s||| / \n", size, spaces); |
| 1313 | seq_printf(s, "#%.*s|||| / \n", size, spaces); | ||
| 1314 | } | 1308 | } |
| 1315 | 1309 | ||
| 1316 | static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | 1310 | static void __print_graph_headers_flags(struct seq_file *s, u32 flags) |
| @@ -1329,7 +1323,7 @@ static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1329 | if (flags & TRACE_GRAPH_PRINT_PROC) | 1323 | if (flags & TRACE_GRAPH_PRINT_PROC) |
| 1330 | seq_printf(s, " TASK/PID "); | 1324 | seq_printf(s, " TASK/PID "); |
| 1331 | if (lat) | 1325 | if (lat) |
| 1332 | seq_printf(s, "|||||"); | 1326 | seq_printf(s, "||||"); |
| 1333 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 1327 | if (flags & TRACE_GRAPH_PRINT_DURATION) |
| 1334 | seq_printf(s, " DURATION "); | 1328 | seq_printf(s, " DURATION "); |
| 1335 | seq_printf(s, " FUNCTION CALLS\n"); | 1329 | seq_printf(s, " FUNCTION CALLS\n"); |
| @@ -1343,7 +1337,7 @@ static void __print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1343 | if (flags & TRACE_GRAPH_PRINT_PROC) | 1337 | if (flags & TRACE_GRAPH_PRINT_PROC) |
| 1344 | seq_printf(s, " | | "); | 1338 | seq_printf(s, " | | "); |
| 1345 | if (lat) | 1339 | if (lat) |
| 1346 | seq_printf(s, "|||||"); | 1340 | seq_printf(s, "||||"); |
| 1347 | if (flags & TRACE_GRAPH_PRINT_DURATION) | 1341 | if (flags & TRACE_GRAPH_PRINT_DURATION) |
| 1348 | seq_printf(s, " | | "); | 1342 | seq_printf(s, " | | "); |
| 1349 | seq_printf(s, " | | | |\n"); | 1343 | seq_printf(s, " | | | |\n"); |
| @@ -1358,15 +1352,16 @@ void print_graph_headers_flags(struct seq_file *s, u32 flags) | |||
| 1358 | { | 1352 | { |
| 1359 | struct trace_iterator *iter = s->private; | 1353 | struct trace_iterator *iter = s->private; |
| 1360 | 1354 | ||
| 1355 | if (!(trace_flags & TRACE_ITER_CONTEXT_INFO)) | ||
| 1356 | return; | ||
| 1357 | |||
| 1361 | if (trace_flags & TRACE_ITER_LATENCY_FMT) { | 1358 | if (trace_flags & TRACE_ITER_LATENCY_FMT) { |
| 1362 | /* print nothing if the buffers are empty */ | 1359 | /* print nothing if the buffers are empty */ |
| 1363 | if (trace_empty(iter)) | 1360 | if (trace_empty(iter)) |
| 1364 | return; | 1361 | return; |
| 1365 | 1362 | ||
| 1366 | print_trace_header(s, iter); | 1363 | print_trace_header(s, iter); |
| 1367 | flags |= TRACE_GRAPH_PRINT_DURATION; | 1364 | } |
| 1368 | } else | ||
| 1369 | flags |= TRACE_GRAPH_PRINT_ABS_TIME; | ||
| 1370 | 1365 | ||
| 1371 | __print_graph_headers_flags(s, flags); | 1366 | __print_graph_headers_flags(s, flags); |
| 1372 | } | 1367 | } |
diff --git a/kernel/trace/trace_irqsoff.c b/kernel/trace/trace_irqsoff.c index c77424be284d..667aa8cc0cfc 100644 --- a/kernel/trace/trace_irqsoff.c +++ b/kernel/trace/trace_irqsoff.c | |||
| @@ -226,7 +226,9 @@ static void irqsoff_trace_close(struct trace_iterator *iter) | |||
| 226 | } | 226 | } |
| 227 | 227 | ||
| 228 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_CPU | \ | 228 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_CPU | \ |
| 229 | TRACE_GRAPH_PRINT_PROC) | 229 | TRACE_GRAPH_PRINT_PROC | \ |
| 230 | TRACE_GRAPH_PRINT_ABS_TIME | \ | ||
| 231 | TRACE_GRAPH_PRINT_DURATION) | ||
| 230 | 232 | ||
| 231 | static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) | 233 | static enum print_line_t irqsoff_print_line(struct trace_iterator *iter) |
| 232 | { | 234 | { |
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index f925c45f0afa..5fb3697bf0e5 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
| @@ -343,6 +343,14 @@ DEFINE_BASIC_FETCH_FUNCS(deref) | |||
| 343 | DEFINE_FETCH_deref(string) | 343 | DEFINE_FETCH_deref(string) |
| 344 | DEFINE_FETCH_deref(string_size) | 344 | DEFINE_FETCH_deref(string_size) |
| 345 | 345 | ||
| 346 | static __kprobes void update_deref_fetch_param(struct deref_fetch_param *data) | ||
| 347 | { | ||
| 348 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | ||
| 349 | update_deref_fetch_param(data->orig.data); | ||
| 350 | else if (CHECK_FETCH_FUNCS(symbol, data->orig.fn)) | ||
| 351 | update_symbol_cache(data->orig.data); | ||
| 352 | } | ||
| 353 | |||
| 346 | static __kprobes void free_deref_fetch_param(struct deref_fetch_param *data) | 354 | static __kprobes void free_deref_fetch_param(struct deref_fetch_param *data) |
| 347 | { | 355 | { |
| 348 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | 356 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) |
| @@ -377,6 +385,19 @@ DEFINE_BASIC_FETCH_FUNCS(bitfield) | |||
| 377 | #define fetch_bitfield_string_size NULL | 385 | #define fetch_bitfield_string_size NULL |
| 378 | 386 | ||
| 379 | static __kprobes void | 387 | static __kprobes void |
| 388 | update_bitfield_fetch_param(struct bitfield_fetch_param *data) | ||
| 389 | { | ||
| 390 | /* | ||
| 391 | * Don't check the bitfield itself, because this must be the | ||
| 392 | * last fetch function. | ||
| 393 | */ | ||
| 394 | if (CHECK_FETCH_FUNCS(deref, data->orig.fn)) | ||
| 395 | update_deref_fetch_param(data->orig.data); | ||
| 396 | else if (CHECK_FETCH_FUNCS(symbol, data->orig.fn)) | ||
| 397 | update_symbol_cache(data->orig.data); | ||
| 398 | } | ||
| 399 | |||
| 400 | static __kprobes void | ||
| 380 | free_bitfield_fetch_param(struct bitfield_fetch_param *data) | 401 | free_bitfield_fetch_param(struct bitfield_fetch_param *data) |
| 381 | { | 402 | { |
| 382 | /* | 403 | /* |
| @@ -389,6 +410,7 @@ free_bitfield_fetch_param(struct bitfield_fetch_param *data) | |||
| 389 | free_symbol_cache(data->orig.data); | 410 | free_symbol_cache(data->orig.data); |
| 390 | kfree(data); | 411 | kfree(data); |
| 391 | } | 412 | } |
| 413 | |||
| 392 | /* Default (unsigned long) fetch type */ | 414 | /* Default (unsigned long) fetch type */ |
| 393 | #define __DEFAULT_FETCH_TYPE(t) u##t | 415 | #define __DEFAULT_FETCH_TYPE(t) u##t |
| 394 | #define _DEFAULT_FETCH_TYPE(t) __DEFAULT_FETCH_TYPE(t) | 416 | #define _DEFAULT_FETCH_TYPE(t) __DEFAULT_FETCH_TYPE(t) |
| @@ -536,6 +558,7 @@ struct probe_arg { | |||
| 536 | /* Flags for trace_probe */ | 558 | /* Flags for trace_probe */ |
| 537 | #define TP_FLAG_TRACE 1 | 559 | #define TP_FLAG_TRACE 1 |
| 538 | #define TP_FLAG_PROFILE 2 | 560 | #define TP_FLAG_PROFILE 2 |
| 561 | #define TP_FLAG_REGISTERED 4 | ||
| 539 | 562 | ||
| 540 | struct trace_probe { | 563 | struct trace_probe { |
| 541 | struct list_head list; | 564 | struct list_head list; |
| @@ -555,16 +578,49 @@ struct trace_probe { | |||
| 555 | (sizeof(struct probe_arg) * (n))) | 578 | (sizeof(struct probe_arg) * (n))) |
| 556 | 579 | ||
| 557 | 580 | ||
| 558 | static __kprobes int probe_is_return(struct trace_probe *tp) | 581 | static __kprobes int trace_probe_is_return(struct trace_probe *tp) |
| 559 | { | 582 | { |
| 560 | return tp->rp.handler != NULL; | 583 | return tp->rp.handler != NULL; |
| 561 | } | 584 | } |
| 562 | 585 | ||
| 563 | static __kprobes const char *probe_symbol(struct trace_probe *tp) | 586 | static __kprobes const char *trace_probe_symbol(struct trace_probe *tp) |
| 564 | { | 587 | { |
| 565 | return tp->symbol ? tp->symbol : "unknown"; | 588 | return tp->symbol ? tp->symbol : "unknown"; |
| 566 | } | 589 | } |
| 567 | 590 | ||
| 591 | static __kprobes unsigned long trace_probe_offset(struct trace_probe *tp) | ||
| 592 | { | ||
| 593 | return tp->rp.kp.offset; | ||
| 594 | } | ||
| 595 | |||
| 596 | static __kprobes bool trace_probe_is_enabled(struct trace_probe *tp) | ||
| 597 | { | ||
| 598 | return !!(tp->flags & (TP_FLAG_TRACE | TP_FLAG_PROFILE)); | ||
| 599 | } | ||
| 600 | |||
| 601 | static __kprobes bool trace_probe_is_registered(struct trace_probe *tp) | ||
| 602 | { | ||
| 603 | return !!(tp->flags & TP_FLAG_REGISTERED); | ||
| 604 | } | ||
| 605 | |||
| 606 | static __kprobes bool trace_probe_has_gone(struct trace_probe *tp) | ||
| 607 | { | ||
| 608 | return !!(kprobe_gone(&tp->rp.kp)); | ||
| 609 | } | ||
| 610 | |||
| 611 | static __kprobes bool trace_probe_within_module(struct trace_probe *tp, | ||
| 612 | struct module *mod) | ||
| 613 | { | ||
| 614 | int len = strlen(mod->name); | ||
| 615 | const char *name = trace_probe_symbol(tp); | ||
| 616 | return strncmp(mod->name, name, len) == 0 && name[len] == ':'; | ||
| 617 | } | ||
| 618 | |||
| 619 | static __kprobes bool trace_probe_is_on_module(struct trace_probe *tp) | ||
| 620 | { | ||
| 621 | return !!strchr(trace_probe_symbol(tp), ':'); | ||
| 622 | } | ||
| 623 | |||
| 568 | static int register_probe_event(struct trace_probe *tp); | 624 | static int register_probe_event(struct trace_probe *tp); |
| 569 | static void unregister_probe_event(struct trace_probe *tp); | 625 | static void unregister_probe_event(struct trace_probe *tp); |
| 570 | 626 | ||
| @@ -646,6 +702,16 @@ error: | |||
| 646 | return ERR_PTR(ret); | 702 | return ERR_PTR(ret); |
| 647 | } | 703 | } |
| 648 | 704 | ||
| 705 | static void update_probe_arg(struct probe_arg *arg) | ||
| 706 | { | ||
| 707 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) | ||
| 708 | update_bitfield_fetch_param(arg->fetch.data); | ||
| 709 | else if (CHECK_FETCH_FUNCS(deref, arg->fetch.fn)) | ||
| 710 | update_deref_fetch_param(arg->fetch.data); | ||
| 711 | else if (CHECK_FETCH_FUNCS(symbol, arg->fetch.fn)) | ||
| 712 | update_symbol_cache(arg->fetch.data); | ||
| 713 | } | ||
| 714 | |||
| 649 | static void free_probe_arg(struct probe_arg *arg) | 715 | static void free_probe_arg(struct probe_arg *arg) |
| 650 | { | 716 | { |
| 651 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) | 717 | if (CHECK_FETCH_FUNCS(bitfield, arg->fetch.fn)) |
| @@ -671,7 +737,7 @@ static void free_trace_probe(struct trace_probe *tp) | |||
| 671 | kfree(tp); | 737 | kfree(tp); |
| 672 | } | 738 | } |
| 673 | 739 | ||
| 674 | static struct trace_probe *find_probe_event(const char *event, | 740 | static struct trace_probe *find_trace_probe(const char *event, |
| 675 | const char *group) | 741 | const char *group) |
| 676 | { | 742 | { |
| 677 | struct trace_probe *tp; | 743 | struct trace_probe *tp; |
| @@ -683,13 +749,96 @@ static struct trace_probe *find_probe_event(const char *event, | |||
| 683 | return NULL; | 749 | return NULL; |
| 684 | } | 750 | } |
| 685 | 751 | ||
| 752 | /* Enable trace_probe - @flag must be TP_FLAG_TRACE or TP_FLAG_PROFILE */ | ||
| 753 | static int enable_trace_probe(struct trace_probe *tp, int flag) | ||
| 754 | { | ||
| 755 | int ret = 0; | ||
| 756 | |||
| 757 | tp->flags |= flag; | ||
| 758 | if (trace_probe_is_enabled(tp) && trace_probe_is_registered(tp) && | ||
| 759 | !trace_probe_has_gone(tp)) { | ||
| 760 | if (trace_probe_is_return(tp)) | ||
| 761 | ret = enable_kretprobe(&tp->rp); | ||
| 762 | else | ||
| 763 | ret = enable_kprobe(&tp->rp.kp); | ||
| 764 | } | ||
| 765 | |||
| 766 | return ret; | ||
| 767 | } | ||
| 768 | |||
| 769 | /* Disable trace_probe - @flag must be TP_FLAG_TRACE or TP_FLAG_PROFILE */ | ||
| 770 | static void disable_trace_probe(struct trace_probe *tp, int flag) | ||
| 771 | { | ||
| 772 | tp->flags &= ~flag; | ||
| 773 | if (!trace_probe_is_enabled(tp) && trace_probe_is_registered(tp)) { | ||
| 774 | if (trace_probe_is_return(tp)) | ||
| 775 | disable_kretprobe(&tp->rp); | ||
| 776 | else | ||
| 777 | disable_kprobe(&tp->rp.kp); | ||
| 778 | } | ||
| 779 | } | ||
| 780 | |||
| 781 | /* Internal register function - just handle k*probes and flags */ | ||
| 782 | static int __register_trace_probe(struct trace_probe *tp) | ||
| 783 | { | ||
| 784 | int i, ret; | ||
| 785 | |||
| 786 | if (trace_probe_is_registered(tp)) | ||
| 787 | return -EINVAL; | ||
| 788 | |||
| 789 | for (i = 0; i < tp->nr_args; i++) | ||
| 790 | update_probe_arg(&tp->args[i]); | ||
| 791 | |||
| 792 | /* Set/clear disabled flag according to tp->flag */ | ||
| 793 | if (trace_probe_is_enabled(tp)) | ||
| 794 | tp->rp.kp.flags &= ~KPROBE_FLAG_DISABLED; | ||
| 795 | else | ||
| 796 | tp->rp.kp.flags |= KPROBE_FLAG_DISABLED; | ||
| 797 | |||
| 798 | if (trace_probe_is_return(tp)) | ||
| 799 | ret = register_kretprobe(&tp->rp); | ||
| 800 | else | ||
| 801 | ret = register_kprobe(&tp->rp.kp); | ||
| 802 | |||
| 803 | if (ret == 0) | ||
| 804 | tp->flags |= TP_FLAG_REGISTERED; | ||
| 805 | else { | ||
| 806 | pr_warning("Could not insert probe at %s+%lu: %d\n", | ||
| 807 | trace_probe_symbol(tp), trace_probe_offset(tp), ret); | ||
| 808 | if (ret == -ENOENT && trace_probe_is_on_module(tp)) { | ||
| 809 | pr_warning("This probe might be able to register after" | ||
| 810 | "target module is loaded. Continue.\n"); | ||
| 811 | ret = 0; | ||
| 812 | } else if (ret == -EILSEQ) { | ||
| 813 | pr_warning("Probing address(0x%p) is not an " | ||
| 814 | "instruction boundary.\n", | ||
| 815 | tp->rp.kp.addr); | ||
| 816 | ret = -EINVAL; | ||
| 817 | } | ||
| 818 | } | ||
| 819 | |||
| 820 | return ret; | ||
| 821 | } | ||
| 822 | |||
| 823 | /* Internal unregister function - just handle k*probes and flags */ | ||
| 824 | static void __unregister_trace_probe(struct trace_probe *tp) | ||
| 825 | { | ||
| 826 | if (trace_probe_is_registered(tp)) { | ||
| 827 | if (trace_probe_is_return(tp)) | ||
| 828 | unregister_kretprobe(&tp->rp); | ||
| 829 | else | ||
| 830 | unregister_kprobe(&tp->rp.kp); | ||
| 831 | tp->flags &= ~TP_FLAG_REGISTERED; | ||
| 832 | /* Cleanup kprobe for reuse */ | ||
| 833 | if (tp->rp.kp.symbol_name) | ||
| 834 | tp->rp.kp.addr = NULL; | ||
| 835 | } | ||
| 836 | } | ||
| 837 | |||
| 686 | /* Unregister a trace_probe and probe_event: call with locking probe_lock */ | 838 | /* Unregister a trace_probe and probe_event: call with locking probe_lock */ |
| 687 | static void unregister_trace_probe(struct trace_probe *tp) | 839 | static void unregister_trace_probe(struct trace_probe *tp) |
| 688 | { | 840 | { |
| 689 | if (probe_is_return(tp)) | 841 | __unregister_trace_probe(tp); |
| 690 | unregister_kretprobe(&tp->rp); | ||
| 691 | else | ||
| 692 | unregister_kprobe(&tp->rp.kp); | ||
| 693 | list_del(&tp->list); | 842 | list_del(&tp->list); |
| 694 | unregister_probe_event(tp); | 843 | unregister_probe_event(tp); |
| 695 | } | 844 | } |
| @@ -702,41 +851,65 @@ static int register_trace_probe(struct trace_probe *tp) | |||
| 702 | 851 | ||
| 703 | mutex_lock(&probe_lock); | 852 | mutex_lock(&probe_lock); |
| 704 | 853 | ||
| 705 | /* register as an event */ | 854 | /* Delete old (same name) event if exist */ |
| 706 | old_tp = find_probe_event(tp->call.name, tp->call.class->system); | 855 | old_tp = find_trace_probe(tp->call.name, tp->call.class->system); |
| 707 | if (old_tp) { | 856 | if (old_tp) { |
| 708 | /* delete old event */ | ||
| 709 | unregister_trace_probe(old_tp); | 857 | unregister_trace_probe(old_tp); |
| 710 | free_trace_probe(old_tp); | 858 | free_trace_probe(old_tp); |
| 711 | } | 859 | } |
| 860 | |||
| 861 | /* Register new event */ | ||
| 712 | ret = register_probe_event(tp); | 862 | ret = register_probe_event(tp); |
| 713 | if (ret) { | 863 | if (ret) { |
| 714 | pr_warning("Failed to register probe event(%d)\n", ret); | 864 | pr_warning("Failed to register probe event(%d)\n", ret); |
| 715 | goto end; | 865 | goto end; |
| 716 | } | 866 | } |
| 717 | 867 | ||
| 718 | tp->rp.kp.flags |= KPROBE_FLAG_DISABLED; | 868 | /* Register k*probe */ |
| 719 | if (probe_is_return(tp)) | 869 | ret = __register_trace_probe(tp); |
| 720 | ret = register_kretprobe(&tp->rp); | 870 | if (ret < 0) |
| 721 | else | ||
| 722 | ret = register_kprobe(&tp->rp.kp); | ||
| 723 | |||
| 724 | if (ret) { | ||
| 725 | pr_warning("Could not insert probe(%d)\n", ret); | ||
| 726 | if (ret == -EILSEQ) { | ||
| 727 | pr_warning("Probing address(0x%p) is not an " | ||
| 728 | "instruction boundary.\n", | ||
| 729 | tp->rp.kp.addr); | ||
| 730 | ret = -EINVAL; | ||
| 731 | } | ||
| 732 | unregister_probe_event(tp); | 871 | unregister_probe_event(tp); |
| 733 | } else | 872 | else |
| 734 | list_add_tail(&tp->list, &probe_list); | 873 | list_add_tail(&tp->list, &probe_list); |
| 874 | |||
| 735 | end: | 875 | end: |
| 736 | mutex_unlock(&probe_lock); | 876 | mutex_unlock(&probe_lock); |
| 737 | return ret; | 877 | return ret; |
| 738 | } | 878 | } |
| 739 | 879 | ||
| 880 | /* Module notifier call back, checking event on the module */ | ||
| 881 | static int trace_probe_module_callback(struct notifier_block *nb, | ||
| 882 | unsigned long val, void *data) | ||
| 883 | { | ||
| 884 | struct module *mod = data; | ||
| 885 | struct trace_probe *tp; | ||
| 886 | int ret; | ||
| 887 | |||
| 888 | if (val != MODULE_STATE_COMING) | ||
| 889 | return NOTIFY_DONE; | ||
| 890 | |||
| 891 | /* Update probes on coming module */ | ||
| 892 | mutex_lock(&probe_lock); | ||
| 893 | list_for_each_entry(tp, &probe_list, list) { | ||
| 894 | if (trace_probe_within_module(tp, mod)) { | ||
| 895 | __unregister_trace_probe(tp); | ||
| 896 | ret = __register_trace_probe(tp); | ||
| 897 | if (ret) | ||
| 898 | pr_warning("Failed to re-register probe %s on" | ||
| 899 | "%s: %d\n", | ||
| 900 | tp->call.name, mod->name, ret); | ||
| 901 | } | ||
| 902 | } | ||
| 903 | mutex_unlock(&probe_lock); | ||
| 904 | |||
| 905 | return NOTIFY_DONE; | ||
| 906 | } | ||
| 907 | |||
| 908 | static struct notifier_block trace_probe_module_nb = { | ||
| 909 | .notifier_call = trace_probe_module_callback, | ||
| 910 | .priority = 1 /* Invoked after kprobe module callback */ | ||
| 911 | }; | ||
| 912 | |||
| 740 | /* Split symbol and offset. */ | 913 | /* Split symbol and offset. */ |
| 741 | static int split_symbol_offset(char *symbol, unsigned long *offset) | 914 | static int split_symbol_offset(char *symbol, unsigned long *offset) |
| 742 | { | 915 | { |
| @@ -962,8 +1135,8 @@ static int create_trace_probe(int argc, char **argv) | |||
| 962 | { | 1135 | { |
| 963 | /* | 1136 | /* |
| 964 | * Argument syntax: | 1137 | * Argument syntax: |
| 965 | * - Add kprobe: p[:[GRP/]EVENT] KSYM[+OFFS]|KADDR [FETCHARGS] | 1138 | * - Add kprobe: p[:[GRP/]EVENT] [MOD:]KSYM[+OFFS]|KADDR [FETCHARGS] |
| 966 | * - Add kretprobe: r[:[GRP/]EVENT] KSYM[+0] [FETCHARGS] | 1139 | * - Add kretprobe: r[:[GRP/]EVENT] [MOD:]KSYM[+0] [FETCHARGS] |
| 967 | * Fetch args: | 1140 | * Fetch args: |
| 968 | * $retval : fetch return value | 1141 | * $retval : fetch return value |
| 969 | * $stack : fetch stack address | 1142 | * $stack : fetch stack address |
| @@ -1025,7 +1198,7 @@ static int create_trace_probe(int argc, char **argv) | |||
| 1025 | return -EINVAL; | 1198 | return -EINVAL; |
| 1026 | } | 1199 | } |
| 1027 | mutex_lock(&probe_lock); | 1200 | mutex_lock(&probe_lock); |
| 1028 | tp = find_probe_event(event, group); | 1201 | tp = find_trace_probe(event, group); |
| 1029 | if (!tp) { | 1202 | if (!tp) { |
| 1030 | mutex_unlock(&probe_lock); | 1203 | mutex_unlock(&probe_lock); |
| 1031 | pr_info("Event %s/%s doesn't exist.\n", group, event); | 1204 | pr_info("Event %s/%s doesn't exist.\n", group, event); |
| @@ -1144,7 +1317,7 @@ error: | |||
| 1144 | return ret; | 1317 | return ret; |
| 1145 | } | 1318 | } |
| 1146 | 1319 | ||
| 1147 | static void cleanup_all_probes(void) | 1320 | static void release_all_trace_probes(void) |
| 1148 | { | 1321 | { |
| 1149 | struct trace_probe *tp; | 1322 | struct trace_probe *tp; |
| 1150 | 1323 | ||
| @@ -1158,7 +1331,6 @@ static void cleanup_all_probes(void) | |||
| 1158 | mutex_unlock(&probe_lock); | 1331 | mutex_unlock(&probe_lock); |
| 1159 | } | 1332 | } |
| 1160 | 1333 | ||
| 1161 | |||
| 1162 | /* Probes listing interfaces */ | 1334 | /* Probes listing interfaces */ |
| 1163 | static void *probes_seq_start(struct seq_file *m, loff_t *pos) | 1335 | static void *probes_seq_start(struct seq_file *m, loff_t *pos) |
| 1164 | { | 1336 | { |
| @@ -1181,15 +1353,16 @@ static int probes_seq_show(struct seq_file *m, void *v) | |||
| 1181 | struct trace_probe *tp = v; | 1353 | struct trace_probe *tp = v; |
| 1182 | int i; | 1354 | int i; |
| 1183 | 1355 | ||
| 1184 | seq_printf(m, "%c", probe_is_return(tp) ? 'r' : 'p'); | 1356 | seq_printf(m, "%c", trace_probe_is_return(tp) ? 'r' : 'p'); |
| 1185 | seq_printf(m, ":%s/%s", tp->call.class->system, tp->call.name); | 1357 | seq_printf(m, ":%s/%s", tp->call.class->system, tp->call.name); |
| 1186 | 1358 | ||
| 1187 | if (!tp->symbol) | 1359 | if (!tp->symbol) |
| 1188 | seq_printf(m, " 0x%p", tp->rp.kp.addr); | 1360 | seq_printf(m, " 0x%p", tp->rp.kp.addr); |
| 1189 | else if (tp->rp.kp.offset) | 1361 | else if (tp->rp.kp.offset) |
| 1190 | seq_printf(m, " %s+%u", probe_symbol(tp), tp->rp.kp.offset); | 1362 | seq_printf(m, " %s+%u", trace_probe_symbol(tp), |
| 1363 | tp->rp.kp.offset); | ||
| 1191 | else | 1364 | else |
| 1192 | seq_printf(m, " %s", probe_symbol(tp)); | 1365 | seq_printf(m, " %s", trace_probe_symbol(tp)); |
| 1193 | 1366 | ||
| 1194 | for (i = 0; i < tp->nr_args; i++) | 1367 | for (i = 0; i < tp->nr_args; i++) |
| 1195 | seq_printf(m, " %s=%s", tp->args[i].name, tp->args[i].comm); | 1368 | seq_printf(m, " %s=%s", tp->args[i].name, tp->args[i].comm); |
| @@ -1209,7 +1382,7 @@ static int probes_open(struct inode *inode, struct file *file) | |||
| 1209 | { | 1382 | { |
| 1210 | if ((file->f_mode & FMODE_WRITE) && | 1383 | if ((file->f_mode & FMODE_WRITE) && |
| 1211 | (file->f_flags & O_TRUNC)) | 1384 | (file->f_flags & O_TRUNC)) |
| 1212 | cleanup_all_probes(); | 1385 | release_all_trace_probes(); |
| 1213 | 1386 | ||
| 1214 | return seq_open(file, &probes_seq_op); | 1387 | return seq_open(file, &probes_seq_op); |
| 1215 | } | 1388 | } |
| @@ -1397,7 +1570,8 @@ static __kprobes void kprobe_trace_func(struct kprobe *kp, struct pt_regs *regs) | |||
| 1397 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); | 1570 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); |
| 1398 | 1571 | ||
| 1399 | if (!filter_current_check_discard(buffer, call, entry, event)) | 1572 | if (!filter_current_check_discard(buffer, call, entry, event)) |
| 1400 | trace_nowake_buffer_unlock_commit(buffer, event, irq_flags, pc); | 1573 | trace_nowake_buffer_unlock_commit_regs(buffer, event, |
| 1574 | irq_flags, pc, regs); | ||
| 1401 | } | 1575 | } |
| 1402 | 1576 | ||
| 1403 | /* Kretprobe handler */ | 1577 | /* Kretprobe handler */ |
| @@ -1429,7 +1603,8 @@ static __kprobes void kretprobe_trace_func(struct kretprobe_instance *ri, | |||
| 1429 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); | 1603 | store_trace_args(sizeof(*entry), tp, regs, (u8 *)&entry[1], dsize); |
| 1430 | 1604 | ||
| 1431 | if (!filter_current_check_discard(buffer, call, entry, event)) | 1605 | if (!filter_current_check_discard(buffer, call, entry, event)) |
| 1432 | trace_nowake_buffer_unlock_commit(buffer, event, irq_flags, pc); | 1606 | trace_nowake_buffer_unlock_commit_regs(buffer, event, |
| 1607 | irq_flags, pc, regs); | ||
| 1433 | } | 1608 | } |
| 1434 | 1609 | ||
| 1435 | /* Event entry printers */ | 1610 | /* Event entry printers */ |
| @@ -1511,30 +1686,6 @@ partial: | |||
| 1511 | return TRACE_TYPE_PARTIAL_LINE; | 1686 | return TRACE_TYPE_PARTIAL_LINE; |
| 1512 | } | 1687 | } |
| 1513 | 1688 | ||
| 1514 | static int probe_event_enable(struct ftrace_event_call *call) | ||
| 1515 | { | ||
| 1516 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1517 | |||
| 1518 | tp->flags |= TP_FLAG_TRACE; | ||
| 1519 | if (probe_is_return(tp)) | ||
| 1520 | return enable_kretprobe(&tp->rp); | ||
| 1521 | else | ||
| 1522 | return enable_kprobe(&tp->rp.kp); | ||
| 1523 | } | ||
| 1524 | |||
| 1525 | static void probe_event_disable(struct ftrace_event_call *call) | ||
| 1526 | { | ||
| 1527 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1528 | |||
| 1529 | tp->flags &= ~TP_FLAG_TRACE; | ||
| 1530 | if (!(tp->flags & (TP_FLAG_TRACE | TP_FLAG_PROFILE))) { | ||
| 1531 | if (probe_is_return(tp)) | ||
| 1532 | disable_kretprobe(&tp->rp); | ||
| 1533 | else | ||
| 1534 | disable_kprobe(&tp->rp.kp); | ||
| 1535 | } | ||
| 1536 | } | ||
| 1537 | |||
| 1538 | #undef DEFINE_FIELD | 1689 | #undef DEFINE_FIELD |
| 1539 | #define DEFINE_FIELD(type, item, name, is_signed) \ | 1690 | #define DEFINE_FIELD(type, item, name, is_signed) \ |
| 1540 | do { \ | 1691 | do { \ |
| @@ -1596,7 +1747,7 @@ static int __set_print_fmt(struct trace_probe *tp, char *buf, int len) | |||
| 1596 | 1747 | ||
| 1597 | const char *fmt, *arg; | 1748 | const char *fmt, *arg; |
| 1598 | 1749 | ||
| 1599 | if (!probe_is_return(tp)) { | 1750 | if (!trace_probe_is_return(tp)) { |
| 1600 | fmt = "(%lx)"; | 1751 | fmt = "(%lx)"; |
| 1601 | arg = "REC->" FIELD_STRING_IP; | 1752 | arg = "REC->" FIELD_STRING_IP; |
| 1602 | } else { | 1753 | } else { |
| @@ -1713,49 +1864,25 @@ static __kprobes void kretprobe_perf_func(struct kretprobe_instance *ri, | |||
| 1713 | head = this_cpu_ptr(call->perf_events); | 1864 | head = this_cpu_ptr(call->perf_events); |
| 1714 | perf_trace_buf_submit(entry, size, rctx, entry->ret_ip, 1, regs, head); | 1865 | perf_trace_buf_submit(entry, size, rctx, entry->ret_ip, 1, regs, head); |
| 1715 | } | 1866 | } |
| 1716 | |||
| 1717 | static int probe_perf_enable(struct ftrace_event_call *call) | ||
| 1718 | { | ||
| 1719 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1720 | |||
| 1721 | tp->flags |= TP_FLAG_PROFILE; | ||
| 1722 | |||
| 1723 | if (probe_is_return(tp)) | ||
| 1724 | return enable_kretprobe(&tp->rp); | ||
| 1725 | else | ||
| 1726 | return enable_kprobe(&tp->rp.kp); | ||
| 1727 | } | ||
| 1728 | |||
| 1729 | static void probe_perf_disable(struct ftrace_event_call *call) | ||
| 1730 | { | ||
| 1731 | struct trace_probe *tp = (struct trace_probe *)call->data; | ||
| 1732 | |||
| 1733 | tp->flags &= ~TP_FLAG_PROFILE; | ||
| 1734 | |||
| 1735 | if (!(tp->flags & TP_FLAG_TRACE)) { | ||
| 1736 | if (probe_is_return(tp)) | ||
| 1737 | disable_kretprobe(&tp->rp); | ||
| 1738 | else | ||
| 1739 | disable_kprobe(&tp->rp.kp); | ||
| 1740 | } | ||
| 1741 | } | ||
| 1742 | #endif /* CONFIG_PERF_EVENTS */ | 1867 | #endif /* CONFIG_PERF_EVENTS */ |
| 1743 | 1868 | ||
| 1744 | static __kprobes | 1869 | static __kprobes |
| 1745 | int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) | 1870 | int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) |
| 1746 | { | 1871 | { |
| 1872 | struct trace_probe *tp = (struct trace_probe *)event->data; | ||
| 1873 | |||
| 1747 | switch (type) { | 1874 | switch (type) { |
| 1748 | case TRACE_REG_REGISTER: | 1875 | case TRACE_REG_REGISTER: |
| 1749 | return probe_event_enable(event); | 1876 | return enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1750 | case TRACE_REG_UNREGISTER: | 1877 | case TRACE_REG_UNREGISTER: |
| 1751 | probe_event_disable(event); | 1878 | disable_trace_probe(tp, TP_FLAG_TRACE); |
| 1752 | return 0; | 1879 | return 0; |
| 1753 | 1880 | ||
| 1754 | #ifdef CONFIG_PERF_EVENTS | 1881 | #ifdef CONFIG_PERF_EVENTS |
| 1755 | case TRACE_REG_PERF_REGISTER: | 1882 | case TRACE_REG_PERF_REGISTER: |
| 1756 | return probe_perf_enable(event); | 1883 | return enable_trace_probe(tp, TP_FLAG_PROFILE); |
| 1757 | case TRACE_REG_PERF_UNREGISTER: | 1884 | case TRACE_REG_PERF_UNREGISTER: |
| 1758 | probe_perf_disable(event); | 1885 | disable_trace_probe(tp, TP_FLAG_PROFILE); |
| 1759 | return 0; | 1886 | return 0; |
| 1760 | #endif | 1887 | #endif |
| 1761 | } | 1888 | } |
| @@ -1805,7 +1932,7 @@ static int register_probe_event(struct trace_probe *tp) | |||
| 1805 | 1932 | ||
| 1806 | /* Initialize ftrace_event_call */ | 1933 | /* Initialize ftrace_event_call */ |
| 1807 | INIT_LIST_HEAD(&call->class->fields); | 1934 | INIT_LIST_HEAD(&call->class->fields); |
| 1808 | if (probe_is_return(tp)) { | 1935 | if (trace_probe_is_return(tp)) { |
| 1809 | call->event.funcs = &kretprobe_funcs; | 1936 | call->event.funcs = &kretprobe_funcs; |
| 1810 | call->class->define_fields = kretprobe_event_define_fields; | 1937 | call->class->define_fields = kretprobe_event_define_fields; |
| 1811 | } else { | 1938 | } else { |
| @@ -1844,6 +1971,9 @@ static __init int init_kprobe_trace(void) | |||
| 1844 | struct dentry *d_tracer; | 1971 | struct dentry *d_tracer; |
| 1845 | struct dentry *entry; | 1972 | struct dentry *entry; |
| 1846 | 1973 | ||
| 1974 | if (register_module_notifier(&trace_probe_module_nb)) | ||
| 1975 | return -EINVAL; | ||
| 1976 | |||
| 1847 | d_tracer = tracing_init_dentry(); | 1977 | d_tracer = tracing_init_dentry(); |
| 1848 | if (!d_tracer) | 1978 | if (!d_tracer) |
| 1849 | return 0; | 1979 | return 0; |
| @@ -1870,8 +2000,12 @@ fs_initcall(init_kprobe_trace); | |||
| 1870 | 2000 | ||
| 1871 | #ifdef CONFIG_FTRACE_STARTUP_TEST | 2001 | #ifdef CONFIG_FTRACE_STARTUP_TEST |
| 1872 | 2002 | ||
| 1873 | static int kprobe_trace_selftest_target(int a1, int a2, int a3, | 2003 | /* |
| 1874 | int a4, int a5, int a6) | 2004 | * The "__used" keeps gcc from removing the function symbol |
| 2005 | * from the kallsyms table. | ||
| 2006 | */ | ||
| 2007 | static __used int kprobe_trace_selftest_target(int a1, int a2, int a3, | ||
| 2008 | int a4, int a5, int a6) | ||
| 1875 | { | 2009 | { |
| 1876 | return a1 + a2 + a3 + a4 + a5 + a6; | 2010 | return a1 + a2 + a3 + a4 + a5 + a6; |
| 1877 | } | 2011 | } |
| @@ -1893,12 +2027,12 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1893 | warn++; | 2027 | warn++; |
| 1894 | } else { | 2028 | } else { |
| 1895 | /* Enable trace point */ | 2029 | /* Enable trace point */ |
| 1896 | tp = find_probe_event("testprobe", KPROBE_EVENT_SYSTEM); | 2030 | tp = find_trace_probe("testprobe", KPROBE_EVENT_SYSTEM); |
| 1897 | if (WARN_ON_ONCE(tp == NULL)) { | 2031 | if (WARN_ON_ONCE(tp == NULL)) { |
| 1898 | pr_warning("error on getting new probe.\n"); | 2032 | pr_warning("error on getting new probe.\n"); |
| 1899 | warn++; | 2033 | warn++; |
| 1900 | } else | 2034 | } else |
| 1901 | probe_event_enable(&tp->call); | 2035 | enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1902 | } | 2036 | } |
| 1903 | 2037 | ||
| 1904 | ret = command_trace_probe("r:testprobe2 kprobe_trace_selftest_target " | 2038 | ret = command_trace_probe("r:testprobe2 kprobe_trace_selftest_target " |
| @@ -1908,12 +2042,12 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1908 | warn++; | 2042 | warn++; |
| 1909 | } else { | 2043 | } else { |
| 1910 | /* Enable trace point */ | 2044 | /* Enable trace point */ |
| 1911 | tp = find_probe_event("testprobe2", KPROBE_EVENT_SYSTEM); | 2045 | tp = find_trace_probe("testprobe2", KPROBE_EVENT_SYSTEM); |
| 1912 | if (WARN_ON_ONCE(tp == NULL)) { | 2046 | if (WARN_ON_ONCE(tp == NULL)) { |
| 1913 | pr_warning("error on getting new probe.\n"); | 2047 | pr_warning("error on getting new probe.\n"); |
| 1914 | warn++; | 2048 | warn++; |
| 1915 | } else | 2049 | } else |
| 1916 | probe_event_enable(&tp->call); | 2050 | enable_trace_probe(tp, TP_FLAG_TRACE); |
| 1917 | } | 2051 | } |
| 1918 | 2052 | ||
| 1919 | if (warn) | 2053 | if (warn) |
| @@ -1934,7 +2068,7 @@ static __init int kprobe_trace_self_tests_init(void) | |||
| 1934 | } | 2068 | } |
| 1935 | 2069 | ||
| 1936 | end: | 2070 | end: |
| 1937 | cleanup_all_probes(); | 2071 | release_all_trace_probes(); |
| 1938 | if (warn) | 2072 | if (warn) |
| 1939 | pr_cont("NG: Some tests are failed. Please check them.\n"); | 2073 | pr_cont("NG: Some tests are failed. Please check them.\n"); |
| 1940 | else | 2074 | else |
diff --git a/kernel/trace/trace_mmiotrace.c b/kernel/trace/trace_mmiotrace.c index 017fa376505d..fd3c8aae55e5 100644 --- a/kernel/trace/trace_mmiotrace.c +++ b/kernel/trace/trace_mmiotrace.c | |||
| @@ -12,7 +12,7 @@ | |||
| 12 | #include <linux/slab.h> | 12 | #include <linux/slab.h> |
| 13 | #include <linux/time.h> | 13 | #include <linux/time.h> |
| 14 | 14 | ||
| 15 | #include <asm/atomic.h> | 15 | #include <linux/atomic.h> |
| 16 | 16 | ||
| 17 | #include "trace.h" | 17 | #include "trace.h" |
| 18 | #include "trace_output.h" | 18 | #include "trace_output.h" |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index e37de492a9e1..51999309a6cf 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
| @@ -1107,19 +1107,20 @@ static enum print_line_t trace_stack_print(struct trace_iterator *iter, | |||
| 1107 | { | 1107 | { |
| 1108 | struct stack_entry *field; | 1108 | struct stack_entry *field; |
| 1109 | struct trace_seq *s = &iter->seq; | 1109 | struct trace_seq *s = &iter->seq; |
| 1110 | int i; | 1110 | unsigned long *p; |
| 1111 | unsigned long *end; | ||
| 1111 | 1112 | ||
| 1112 | trace_assign_type(field, iter->ent); | 1113 | trace_assign_type(field, iter->ent); |
| 1114 | end = (unsigned long *)((long)iter->ent + iter->ent_size); | ||
| 1113 | 1115 | ||
| 1114 | if (!trace_seq_puts(s, "<stack trace>\n")) | 1116 | if (!trace_seq_puts(s, "<stack trace>\n")) |
| 1115 | goto partial; | 1117 | goto partial; |
| 1116 | for (i = 0; i < FTRACE_STACK_ENTRIES; i++) { | 1118 | |
| 1117 | if (!field->caller[i] || (field->caller[i] == ULONG_MAX)) | 1119 | for (p = field->caller; p && *p != ULONG_MAX && p < end; p++) { |
| 1118 | break; | ||
| 1119 | if (!trace_seq_puts(s, " => ")) | 1120 | if (!trace_seq_puts(s, " => ")) |
| 1120 | goto partial; | 1121 | goto partial; |
| 1121 | 1122 | ||
| 1122 | if (!seq_print_ip_sym(s, field->caller[i], flags)) | 1123 | if (!seq_print_ip_sym(s, *p, flags)) |
| 1123 | goto partial; | 1124 | goto partial; |
| 1124 | if (!trace_seq_puts(s, "\n")) | 1125 | if (!trace_seq_puts(s, "\n")) |
| 1125 | goto partial; | 1126 | goto partial; |
diff --git a/kernel/trace/trace_printk.c b/kernel/trace/trace_printk.c index dff763b7baf1..1f06468a10d7 100644 --- a/kernel/trace/trace_printk.c +++ b/kernel/trace/trace_printk.c | |||
| @@ -240,13 +240,10 @@ static const char **find_next(void *v, loff_t *pos) | |||
| 240 | const char **fmt = v; | 240 | const char **fmt = v; |
| 241 | int start_index; | 241 | int start_index; |
| 242 | 242 | ||
| 243 | if (!fmt) | ||
| 244 | fmt = __start___trace_bprintk_fmt + *pos; | ||
| 245 | |||
| 246 | start_index = __stop___trace_bprintk_fmt - __start___trace_bprintk_fmt; | 243 | start_index = __stop___trace_bprintk_fmt - __start___trace_bprintk_fmt; |
| 247 | 244 | ||
| 248 | if (*pos < start_index) | 245 | if (*pos < start_index) |
| 249 | return fmt; | 246 | return __start___trace_bprintk_fmt + *pos; |
| 250 | 247 | ||
| 251 | return find_next_mod_format(start_index, v, fmt, pos); | 248 | return find_next_mod_format(start_index, v, fmt, pos); |
| 252 | } | 249 | } |
diff --git a/kernel/trace/trace_sched_wakeup.c b/kernel/trace/trace_sched_wakeup.c index f029dd4fd2ca..e4a70c0c71b6 100644 --- a/kernel/trace/trace_sched_wakeup.c +++ b/kernel/trace/trace_sched_wakeup.c | |||
| @@ -227,7 +227,9 @@ static void wakeup_trace_close(struct trace_iterator *iter) | |||
| 227 | graph_trace_close(iter); | 227 | graph_trace_close(iter); |
| 228 | } | 228 | } |
| 229 | 229 | ||
| 230 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_PROC) | 230 | #define GRAPH_TRACER_FLAGS (TRACE_GRAPH_PRINT_PROC | \ |
| 231 | TRACE_GRAPH_PRINT_ABS_TIME | \ | ||
| 232 | TRACE_GRAPH_PRINT_DURATION) | ||
| 231 | 233 | ||
| 232 | static enum print_line_t wakeup_print_line(struct trace_iterator *iter) | 234 | static enum print_line_t wakeup_print_line(struct trace_iterator *iter) |
| 233 | { | 235 | { |
diff --git a/kernel/trace/trace_stack.c b/kernel/trace/trace_stack.c index b0b53b8e4c25..77575b386d97 100644 --- a/kernel/trace/trace_stack.c +++ b/kernel/trace/trace_stack.c | |||
| @@ -156,20 +156,11 @@ stack_max_size_write(struct file *filp, const char __user *ubuf, | |||
| 156 | { | 156 | { |
| 157 | long *ptr = filp->private_data; | 157 | long *ptr = filp->private_data; |
| 158 | unsigned long val, flags; | 158 | unsigned long val, flags; |
| 159 | char buf[64]; | ||
| 160 | int ret; | 159 | int ret; |
| 161 | int cpu; | 160 | int cpu; |
| 162 | 161 | ||
| 163 | if (count >= sizeof(buf)) | 162 | ret = kstrtoul_from_user(ubuf, count, 10, &val); |
| 164 | return -EINVAL; | 163 | if (ret) |
| 165 | |||
| 166 | if (copy_from_user(&buf, ubuf, count)) | ||
| 167 | return -EFAULT; | ||
| 168 | |||
| 169 | buf[count] = 0; | ||
| 170 | |||
| 171 | ret = strict_strtoul(buf, 10, &val); | ||
| 172 | if (ret < 0) | ||
| 173 | return ret; | 164 | return ret; |
| 174 | 165 | ||
| 175 | local_irq_save(flags); | 166 | local_irq_save(flags); |
