diff options
Diffstat (limited to 'kernel/trace')
| -rw-r--r-- | kernel/trace/Kconfig | 116 | ||||
| -rw-r--r-- | kernel/trace/ftrace.c | 6 | ||||
| -rw-r--r-- | kernel/trace/ring_buffer.c | 28 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 7 | ||||
| -rw-r--r-- | kernel/trace/trace_events_filter.c | 29 | ||||
| -rw-r--r-- | kernel/trace/trace_export.c | 7 | ||||
| -rw-r--r-- | kernel/trace/trace_kprobe.c | 7 | ||||
| -rw-r--r-- | kernel/trace/trace_ksym.c | 140 |
8 files changed, 174 insertions, 166 deletions
diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig index d006554888dc..60e2ce0181ee 100644 --- a/kernel/trace/Kconfig +++ b/kernel/trace/Kconfig | |||
| @@ -12,39 +12,37 @@ config NOP_TRACER | |||
| 12 | config HAVE_FTRACE_NMI_ENTER | 12 | config HAVE_FTRACE_NMI_ENTER |
| 13 | bool | 13 | bool |
| 14 | help | 14 | help |
| 15 | See Documentation/trace/ftrace-implementation.txt | 15 | See Documentation/trace/ftrace-design.txt |
| 16 | 16 | ||
| 17 | config HAVE_FUNCTION_TRACER | 17 | config HAVE_FUNCTION_TRACER |
| 18 | bool | 18 | bool |
| 19 | help | 19 | help |
| 20 | See Documentation/trace/ftrace-implementation.txt | 20 | See Documentation/trace/ftrace-design.txt |
| 21 | 21 | ||
| 22 | config HAVE_FUNCTION_GRAPH_TRACER | 22 | config HAVE_FUNCTION_GRAPH_TRACER |
| 23 | bool | 23 | bool |
| 24 | help | 24 | help |
| 25 | See Documentation/trace/ftrace-implementation.txt | 25 | See Documentation/trace/ftrace-design.txt |
| 26 | 26 | ||
| 27 | config HAVE_FUNCTION_GRAPH_FP_TEST | 27 | config HAVE_FUNCTION_GRAPH_FP_TEST |
| 28 | bool | 28 | bool |
| 29 | help | 29 | help |
| 30 | An arch may pass in a unique value (frame pointer) to both the | 30 | See Documentation/trace/ftrace-design.txt |
| 31 | entering and exiting of a function. On exit, the value is compared | ||
| 32 | and if it does not match, then it will panic the kernel. | ||
| 33 | 31 | ||
| 34 | config HAVE_FUNCTION_TRACE_MCOUNT_TEST | 32 | config HAVE_FUNCTION_TRACE_MCOUNT_TEST |
| 35 | bool | 33 | bool |
| 36 | help | 34 | help |
| 37 | See Documentation/trace/ftrace-implementation.txt | 35 | See Documentation/trace/ftrace-design.txt |
| 38 | 36 | ||
| 39 | config HAVE_DYNAMIC_FTRACE | 37 | config HAVE_DYNAMIC_FTRACE |
| 40 | bool | 38 | bool |
| 41 | help | 39 | help |
| 42 | See Documentation/trace/ftrace-implementation.txt | 40 | See Documentation/trace/ftrace-design.txt |
| 43 | 41 | ||
| 44 | config HAVE_FTRACE_MCOUNT_RECORD | 42 | config HAVE_FTRACE_MCOUNT_RECORD |
| 45 | bool | 43 | bool |
| 46 | help | 44 | help |
| 47 | See Documentation/trace/ftrace-implementation.txt | 45 | See Documentation/trace/ftrace-design.txt |
| 48 | 46 | ||
| 49 | config HAVE_HW_BRANCH_TRACER | 47 | config HAVE_HW_BRANCH_TRACER |
| 50 | bool | 48 | bool |
| @@ -52,7 +50,7 @@ config HAVE_HW_BRANCH_TRACER | |||
| 52 | config HAVE_SYSCALL_TRACEPOINTS | 50 | config HAVE_SYSCALL_TRACEPOINTS |
| 53 | bool | 51 | bool |
| 54 | help | 52 | help |
| 55 | See Documentation/trace/ftrace-implementation.txt | 53 | See Documentation/trace/ftrace-design.txt |
| 56 | 54 | ||
| 57 | config TRACER_MAX_TRACE | 55 | config TRACER_MAX_TRACE |
| 58 | bool | 56 | bool |
| @@ -83,7 +81,7 @@ config RING_BUFFER_ALLOW_SWAP | |||
| 83 | # This allows those options to appear when no other tracer is selected. But the | 81 | # This allows those options to appear when no other tracer is selected. But the |
| 84 | # options do not appear when something else selects it. We need the two options | 82 | # options do not appear when something else selects it. We need the two options |
| 85 | # GENERIC_TRACER and TRACING to avoid circular dependencies to accomplish the | 83 | # GENERIC_TRACER and TRACING to avoid circular dependencies to accomplish the |
| 86 | # hidding of the automatic options. | 84 | # hiding of the automatic options. |
| 87 | 85 | ||
| 88 | config TRACING | 86 | config TRACING |
| 89 | bool | 87 | bool |
| @@ -119,7 +117,7 @@ menuconfig FTRACE | |||
| 119 | bool "Tracers" | 117 | bool "Tracers" |
| 120 | default y if DEBUG_KERNEL | 118 | default y if DEBUG_KERNEL |
| 121 | help | 119 | help |
| 122 | Enable the kernel tracing infrastructure. | 120 | Enable the kernel tracing infrastructure. |
| 123 | 121 | ||
| 124 | if FTRACE | 122 | if FTRACE |
| 125 | 123 | ||
| @@ -133,7 +131,7 @@ config FUNCTION_TRACER | |||
| 133 | help | 131 | help |
| 134 | Enable the kernel to trace every kernel function. This is done | 132 | Enable the kernel to trace every kernel function. This is done |
| 135 | by using a compiler feature to insert a small, 5-byte No-Operation | 133 | by using a compiler feature to insert a small, 5-byte No-Operation |
| 136 | instruction to the beginning of every kernel function, which NOP | 134 | instruction at the beginning of every kernel function, which NOP |
| 137 | sequence is then dynamically patched into a tracer call when | 135 | sequence is then dynamically patched into a tracer call when |
| 138 | tracing is enabled by the administrator. If it's runtime disabled | 136 | tracing is enabled by the administrator. If it's runtime disabled |
| 139 | (the bootup default), then the overhead of the instructions is very | 137 | (the bootup default), then the overhead of the instructions is very |
| @@ -150,7 +148,7 @@ config FUNCTION_GRAPH_TRACER | |||
| 150 | and its entry. | 148 | and its entry. |
| 151 | Its first purpose is to trace the duration of functions and | 149 | Its first purpose is to trace the duration of functions and |
| 152 | draw a call graph for each thread with some information like | 150 | draw a call graph for each thread with some information like |
| 153 | the return value. This is done by setting the current return | 151 | the return value. This is done by setting the current return |
| 154 | address on the current task structure into a stack of calls. | 152 | address on the current task structure into a stack of calls. |
| 155 | 153 | ||
| 156 | 154 | ||
| @@ -173,7 +171,7 @@ config IRQSOFF_TRACER | |||
| 173 | 171 | ||
| 174 | echo 0 > /sys/kernel/debug/tracing/tracing_max_latency | 172 | echo 0 > /sys/kernel/debug/tracing/tracing_max_latency |
| 175 | 173 | ||
| 176 | (Note that kernel size and overhead increases with this option | 174 | (Note that kernel size and overhead increase with this option |
| 177 | enabled. This option and the preempt-off timing option can be | 175 | enabled. This option and the preempt-off timing option can be |
| 178 | used together or separately.) | 176 | used together or separately.) |
| 179 | 177 | ||
| @@ -186,7 +184,7 @@ config PREEMPT_TRACER | |||
| 186 | select TRACER_MAX_TRACE | 184 | select TRACER_MAX_TRACE |
| 187 | select RING_BUFFER_ALLOW_SWAP | 185 | select RING_BUFFER_ALLOW_SWAP |
| 188 | help | 186 | help |
| 189 | This option measures the time spent in preemption off critical | 187 | This option measures the time spent in preemption-off critical |
| 190 | sections, with microsecond accuracy. | 188 | sections, with microsecond accuracy. |
| 191 | 189 | ||
| 192 | The default measurement method is a maximum search, which is | 190 | The default measurement method is a maximum search, which is |
| @@ -195,7 +193,7 @@ config PREEMPT_TRACER | |||
| 195 | 193 | ||
| 196 | echo 0 > /sys/kernel/debug/tracing/tracing_max_latency | 194 | echo 0 > /sys/kernel/debug/tracing/tracing_max_latency |
| 197 | 195 | ||
| 198 | (Note that kernel size and overhead increases with this option | 196 | (Note that kernel size and overhead increase with this option |
| 199 | enabled. This option and the irqs-off timing option can be | 197 | enabled. This option and the irqs-off timing option can be |
| 200 | used together or separately.) | 198 | used together or separately.) |
| 201 | 199 | ||
| @@ -222,7 +220,7 @@ config ENABLE_DEFAULT_TRACERS | |||
| 222 | depends on !GENERIC_TRACER | 220 | depends on !GENERIC_TRACER |
| 223 | select TRACING | 221 | select TRACING |
| 224 | help | 222 | help |
| 225 | This tracer hooks to various trace points in the kernel | 223 | This tracer hooks to various trace points in the kernel, |
| 226 | allowing the user to pick and choose which trace point they | 224 | allowing the user to pick and choose which trace point they |
| 227 | want to trace. It also includes the sched_switch tracer plugin. | 225 | want to trace. It also includes the sched_switch tracer plugin. |
| 228 | 226 | ||
| @@ -265,19 +263,19 @@ choice | |||
| 265 | The likely/unlikely profiler only looks at the conditions that | 263 | The likely/unlikely profiler only looks at the conditions that |
| 266 | are annotated with a likely or unlikely macro. | 264 | are annotated with a likely or unlikely macro. |
| 267 | 265 | ||
| 268 | The "all branch" profiler will profile every if statement in the | 266 | The "all branch" profiler will profile every if-statement in the |
| 269 | kernel. This profiler will also enable the likely/unlikely | 267 | kernel. This profiler will also enable the likely/unlikely |
| 270 | profiler as well. | 268 | profiler. |
| 271 | 269 | ||
| 272 | Either of the above profilers add a bit of overhead to the system. | 270 | Either of the above profilers adds a bit of overhead to the system. |
| 273 | If unsure choose "No branch profiling". | 271 | If unsure, choose "No branch profiling". |
| 274 | 272 | ||
| 275 | config BRANCH_PROFILE_NONE | 273 | config BRANCH_PROFILE_NONE |
| 276 | bool "No branch profiling" | 274 | bool "No branch profiling" |
| 277 | help | 275 | help |
| 278 | No branch profiling. Branch profiling adds a bit of overhead. | 276 | No branch profiling. Branch profiling adds a bit of overhead. |
| 279 | Only enable it if you want to analyse the branching behavior. | 277 | Only enable it if you want to analyse the branching behavior. |
| 280 | Otherwise keep it disabled. | 278 | Otherwise keep it disabled. |
| 281 | 279 | ||
| 282 | config PROFILE_ANNOTATED_BRANCHES | 280 | config PROFILE_ANNOTATED_BRANCHES |
| 283 | bool "Trace likely/unlikely profiler" | 281 | bool "Trace likely/unlikely profiler" |
| @@ -288,7 +286,7 @@ config PROFILE_ANNOTATED_BRANCHES | |||
| 288 | 286 | ||
| 289 | /sys/kernel/debug/tracing/profile_annotated_branch | 287 | /sys/kernel/debug/tracing/profile_annotated_branch |
| 290 | 288 | ||
| 291 | Note: this will add a significant overhead, only turn this | 289 | Note: this will add a significant overhead; only turn this |
| 292 | on if you need to profile the system's use of these macros. | 290 | on if you need to profile the system's use of these macros. |
| 293 | 291 | ||
| 294 | config PROFILE_ALL_BRANCHES | 292 | config PROFILE_ALL_BRANCHES |
| @@ -305,7 +303,7 @@ config PROFILE_ALL_BRANCHES | |||
| 305 | 303 | ||
| 306 | This configuration, when enabled, will impose a great overhead | 304 | This configuration, when enabled, will impose a great overhead |
| 307 | on the system. This should only be enabled when the system | 305 | on the system. This should only be enabled when the system |
| 308 | is to be analyzed | 306 | is to be analyzed in much detail. |
| 309 | endchoice | 307 | endchoice |
| 310 | 308 | ||
| 311 | config TRACING_BRANCHES | 309 | config TRACING_BRANCHES |
| @@ -335,7 +333,7 @@ config POWER_TRACER | |||
| 335 | depends on X86 | 333 | depends on X86 |
| 336 | select GENERIC_TRACER | 334 | select GENERIC_TRACER |
| 337 | help | 335 | help |
| 338 | This tracer helps developers to analyze and optimize the kernels | 336 | This tracer helps developers to analyze and optimize the kernel's |
| 339 | power management decisions, specifically the C-state and P-state | 337 | power management decisions, specifically the C-state and P-state |
| 340 | behavior. | 338 | behavior. |
| 341 | 339 | ||
| @@ -391,14 +389,14 @@ config HW_BRANCH_TRACER | |||
| 391 | select GENERIC_TRACER | 389 | select GENERIC_TRACER |
| 392 | help | 390 | help |
| 393 | This tracer records all branches on the system in a circular | 391 | This tracer records all branches on the system in a circular |
| 394 | buffer giving access to the last N branches for each cpu. | 392 | buffer, giving access to the last N branches for each cpu. |
| 395 | 393 | ||
| 396 | config KMEMTRACE | 394 | config KMEMTRACE |
| 397 | bool "Trace SLAB allocations" | 395 | bool "Trace SLAB allocations" |
| 398 | select GENERIC_TRACER | 396 | select GENERIC_TRACER |
| 399 | help | 397 | help |
| 400 | kmemtrace provides tracing for slab allocator functions, such as | 398 | kmemtrace provides tracing for slab allocator functions, such as |
| 401 | kmalloc, kfree, kmem_cache_alloc, kmem_cache_free etc.. Collected | 399 | kmalloc, kfree, kmem_cache_alloc, kmem_cache_free, etc. Collected |
| 402 | data is then fed to the userspace application in order to analyse | 400 | data is then fed to the userspace application in order to analyse |
| 403 | allocation hotspots, internal fragmentation and so on, making it | 401 | allocation hotspots, internal fragmentation and so on, making it |
| 404 | possible to see how well an allocator performs, as well as debug | 402 | possible to see how well an allocator performs, as well as debug |
| @@ -417,15 +415,15 @@ config WORKQUEUE_TRACER | |||
| 417 | bool "Trace workqueues" | 415 | bool "Trace workqueues" |
| 418 | select GENERIC_TRACER | 416 | select GENERIC_TRACER |
| 419 | help | 417 | help |
| 420 | The workqueue tracer provides some statistical informations | 418 | The workqueue tracer provides some statistical information |
| 421 | about each cpu workqueue thread such as the number of the | 419 | about each cpu workqueue thread such as the number of the |
| 422 | works inserted and executed since their creation. It can help | 420 | works inserted and executed since their creation. It can help |
| 423 | to evaluate the amount of work each of them have to perform. | 421 | to evaluate the amount of work each of them has to perform. |
| 424 | For example it can help a developer to decide whether he should | 422 | For example it can help a developer to decide whether he should |
| 425 | choose a per cpu workqueue instead of a singlethreaded one. | 423 | choose a per-cpu workqueue instead of a singlethreaded one. |
| 426 | 424 | ||
| 427 | config BLK_DEV_IO_TRACE | 425 | config BLK_DEV_IO_TRACE |
| 428 | bool "Support for tracing block io actions" | 426 | bool "Support for tracing block IO actions" |
| 429 | depends on SYSFS | 427 | depends on SYSFS |
| 430 | depends on BLOCK | 428 | depends on BLOCK |
| 431 | select RELAY | 429 | select RELAY |
| @@ -456,15 +454,15 @@ config KPROBE_EVENT | |||
| 456 | select TRACING | 454 | select TRACING |
| 457 | default y | 455 | default y |
| 458 | help | 456 | help |
| 459 | This allows the user to add tracing events (similar to tracepoints) on the fly | 457 | This allows the user to add tracing events (similar to tracepoints) |
| 460 | via the ftrace interface. See Documentation/trace/kprobetrace.txt | 458 | on the fly via the ftrace interface. See |
| 461 | for more details. | 459 | Documentation/trace/kprobetrace.txt for more details. |
| 462 | 460 | ||
| 463 | Those events can be inserted wherever kprobes can probe, and record | 461 | Those events can be inserted wherever kprobes can probe, and record |
| 464 | various register and memory values. | 462 | various register and memory values. |
| 465 | 463 | ||
| 466 | This option is also required by perf-probe subcommand of perf tools. If | 464 | This option is also required by perf-probe subcommand of perf tools. |
| 467 | you want to use perf tools, this option is strongly recommended. | 465 | If you want to use perf tools, this option is strongly recommended. |
| 468 | 466 | ||
| 469 | config DYNAMIC_FTRACE | 467 | config DYNAMIC_FTRACE |
| 470 | bool "enable/disable ftrace tracepoints dynamically" | 468 | bool "enable/disable ftrace tracepoints dynamically" |
| @@ -472,32 +470,32 @@ config DYNAMIC_FTRACE | |||
| 472 | depends on HAVE_DYNAMIC_FTRACE | 470 | depends on HAVE_DYNAMIC_FTRACE |
| 473 | default y | 471 | default y |
| 474 | help | 472 | help |
| 475 | This option will modify all the calls to ftrace dynamically | 473 | This option will modify all the calls to ftrace dynamically |
| 476 | (will patch them out of the binary image and replaces them | 474 | (will patch them out of the binary image and replace them |
| 477 | with a No-Op instruction) as they are called. A table is | 475 | with a No-Op instruction) as they are called. A table is |
| 478 | created to dynamically enable them again. | 476 | created to dynamically enable them again. |
| 479 | 477 | ||
| 480 | This way a CONFIG_FUNCTION_TRACER kernel is slightly larger, but otherwise | 478 | This way a CONFIG_FUNCTION_TRACER kernel is slightly larger, but |
| 481 | has native performance as long as no tracing is active. | 479 | otherwise has native performance as long as no tracing is active. |
| 482 | 480 | ||
| 483 | The changes to the code are done by a kernel thread that | 481 | The changes to the code are done by a kernel thread that |
| 484 | wakes up once a second and checks to see if any ftrace calls | 482 | wakes up once a second and checks to see if any ftrace calls |
| 485 | were made. If so, it runs stop_machine (stops all CPUS) | 483 | were made. If so, it runs stop_machine (stops all CPUS) |
| 486 | and modifies the code to jump over the call to ftrace. | 484 | and modifies the code to jump over the call to ftrace. |
| 487 | 485 | ||
| 488 | config FUNCTION_PROFILER | 486 | config FUNCTION_PROFILER |
| 489 | bool "Kernel function profiler" | 487 | bool "Kernel function profiler" |
| 490 | depends on FUNCTION_TRACER | 488 | depends on FUNCTION_TRACER |
| 491 | default n | 489 | default n |
| 492 | help | 490 | help |
| 493 | This option enables the kernel function profiler. A file is created | 491 | This option enables the kernel function profiler. A file is created |
| 494 | in debugfs called function_profile_enabled which defaults to zero. | 492 | in debugfs called function_profile_enabled which defaults to zero. |
| 495 | When a 1 is echoed into this file profiling begins, and when a | 493 | When a 1 is echoed into this file profiling begins, and when a |
| 496 | zero is entered, profiling stops. A file in the trace_stats | 494 | zero is entered, profiling stops. A "functions" file is created in |
| 497 | directory called functions, that show the list of functions that | 495 | the trace_stats directory; this file shows the list of functions that |
| 498 | have been hit and their counters. | 496 | have been hit and their counters. |
| 499 | 497 | ||
| 500 | If in doubt, say N | 498 | If in doubt, say N. |
| 501 | 499 | ||
| 502 | config FTRACE_MCOUNT_RECORD | 500 | config FTRACE_MCOUNT_RECORD |
| 503 | def_bool y | 501 | def_bool y |
| @@ -556,8 +554,8 @@ config RING_BUFFER_BENCHMARK | |||
| 556 | tristate "Ring buffer benchmark stress tester" | 554 | tristate "Ring buffer benchmark stress tester" |
| 557 | depends on RING_BUFFER | 555 | depends on RING_BUFFER |
| 558 | help | 556 | help |
| 559 | This option creates a test to stress the ring buffer and bench mark it. | 557 | This option creates a test to stress the ring buffer and benchmark it. |
| 560 | It creates its own ring buffer such that it will not interfer with | 558 | It creates its own ring buffer such that it will not interfere with |
| 561 | any other users of the ring buffer (such as ftrace). It then creates | 559 | any other users of the ring buffer (such as ftrace). It then creates |
| 562 | a producer and consumer that will run for 10 seconds and sleep for | 560 | a producer and consumer that will run for 10 seconds and sleep for |
| 563 | 10 seconds. Each interval it will print out the number of events | 561 | 10 seconds. Each interval it will print out the number of events |
| @@ -566,7 +564,7 @@ config RING_BUFFER_BENCHMARK | |||
| 566 | It does not disable interrupts or raise its priority, so it may be | 564 | It does not disable interrupts or raise its priority, so it may be |
| 567 | affected by processes that are running. | 565 | affected by processes that are running. |
| 568 | 566 | ||
| 569 | If unsure, say N | 567 | If unsure, say N. |
| 570 | 568 | ||
| 571 | endif # FTRACE | 569 | endif # FTRACE |
| 572 | 570 | ||
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 7968762c8167..1e6640f80454 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
| @@ -1690,7 +1690,7 @@ ftrace_regex_lseek(struct file *file, loff_t offset, int origin) | |||
| 1690 | static int ftrace_match(char *str, char *regex, int len, int type) | 1690 | static int ftrace_match(char *str, char *regex, int len, int type) |
| 1691 | { | 1691 | { |
| 1692 | int matched = 0; | 1692 | int matched = 0; |
| 1693 | char *ptr; | 1693 | int slen; |
| 1694 | 1694 | ||
| 1695 | switch (type) { | 1695 | switch (type) { |
| 1696 | case MATCH_FULL: | 1696 | case MATCH_FULL: |
| @@ -1706,8 +1706,8 @@ static int ftrace_match(char *str, char *regex, int len, int type) | |||
| 1706 | matched = 1; | 1706 | matched = 1; |
| 1707 | break; | 1707 | break; |
| 1708 | case MATCH_END_ONLY: | 1708 | case MATCH_END_ONLY: |
| 1709 | ptr = strstr(str, regex); | 1709 | slen = strlen(str); |
| 1710 | if (ptr && (ptr[len] == 0)) | 1710 | if (slen >= len && memcmp(str + slen - len, regex, len) == 0) |
| 1711 | matched = 1; | 1711 | matched = 1; |
| 1712 | break; | 1712 | break; |
| 1713 | } | 1713 | } |
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index 2326b04c95c4..8c1b2d290718 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
| @@ -464,6 +464,8 @@ struct ring_buffer_iter { | |||
| 464 | struct ring_buffer_per_cpu *cpu_buffer; | 464 | struct ring_buffer_per_cpu *cpu_buffer; |
| 465 | unsigned long head; | 465 | unsigned long head; |
| 466 | struct buffer_page *head_page; | 466 | struct buffer_page *head_page; |
| 467 | struct buffer_page *cache_reader_page; | ||
| 468 | unsigned long cache_read; | ||
| 467 | u64 read_stamp; | 469 | u64 read_stamp; |
| 468 | }; | 470 | }; |
| 469 | 471 | ||
| @@ -2716,6 +2718,8 @@ static void rb_iter_reset(struct ring_buffer_iter *iter) | |||
| 2716 | iter->read_stamp = cpu_buffer->read_stamp; | 2718 | iter->read_stamp = cpu_buffer->read_stamp; |
| 2717 | else | 2719 | else |
| 2718 | iter->read_stamp = iter->head_page->page->time_stamp; | 2720 | iter->read_stamp = iter->head_page->page->time_stamp; |
| 2721 | iter->cache_reader_page = cpu_buffer->reader_page; | ||
| 2722 | iter->cache_read = cpu_buffer->read; | ||
| 2719 | } | 2723 | } |
| 2720 | 2724 | ||
| 2721 | /** | 2725 | /** |
| @@ -2869,7 +2873,7 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) | |||
| 2869 | * Splice the empty reader page into the list around the head. | 2873 | * Splice the empty reader page into the list around the head. |
| 2870 | */ | 2874 | */ |
| 2871 | reader = rb_set_head_page(cpu_buffer); | 2875 | reader = rb_set_head_page(cpu_buffer); |
| 2872 | cpu_buffer->reader_page->list.next = reader->list.next; | 2876 | cpu_buffer->reader_page->list.next = rb_list_head(reader->list.next); |
| 2873 | cpu_buffer->reader_page->list.prev = reader->list.prev; | 2877 | cpu_buffer->reader_page->list.prev = reader->list.prev; |
| 2874 | 2878 | ||
| 2875 | /* | 2879 | /* |
| @@ -2906,7 +2910,7 @@ rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) | |||
| 2906 | * | 2910 | * |
| 2907 | * Now make the new head point back to the reader page. | 2911 | * Now make the new head point back to the reader page. |
| 2908 | */ | 2912 | */ |
| 2909 | reader->list.next->prev = &cpu_buffer->reader_page->list; | 2913 | rb_list_head(reader->list.next)->prev = &cpu_buffer->reader_page->list; |
| 2910 | rb_inc_page(cpu_buffer, &cpu_buffer->head_page); | 2914 | rb_inc_page(cpu_buffer, &cpu_buffer->head_page); |
| 2911 | 2915 | ||
| 2912 | /* Finally update the reader page to the new head */ | 2916 | /* Finally update the reader page to the new head */ |
| @@ -3060,13 +3064,22 @@ rb_iter_peek(struct ring_buffer_iter *iter, u64 *ts) | |||
| 3060 | struct ring_buffer_event *event; | 3064 | struct ring_buffer_event *event; |
| 3061 | int nr_loops = 0; | 3065 | int nr_loops = 0; |
| 3062 | 3066 | ||
| 3063 | if (ring_buffer_iter_empty(iter)) | ||
| 3064 | return NULL; | ||
| 3065 | |||
| 3066 | cpu_buffer = iter->cpu_buffer; | 3067 | cpu_buffer = iter->cpu_buffer; |
| 3067 | buffer = cpu_buffer->buffer; | 3068 | buffer = cpu_buffer->buffer; |
| 3068 | 3069 | ||
| 3070 | /* | ||
| 3071 | * Check if someone performed a consuming read to | ||
| 3072 | * the buffer. A consuming read invalidates the iterator | ||
| 3073 | * and we need to reset the iterator in this case. | ||
| 3074 | */ | ||
| 3075 | if (unlikely(iter->cache_read != cpu_buffer->read || | ||
| 3076 | iter->cache_reader_page != cpu_buffer->reader_page)) | ||
| 3077 | rb_iter_reset(iter); | ||
| 3078 | |||
| 3069 | again: | 3079 | again: |
| 3080 | if (ring_buffer_iter_empty(iter)) | ||
| 3081 | return NULL; | ||
| 3082 | |||
| 3070 | /* | 3083 | /* |
| 3071 | * We repeat when a timestamp is encountered. | 3084 | * We repeat when a timestamp is encountered. |
| 3072 | * We can get multiple timestamps by nested interrupts or also | 3085 | * We can get multiple timestamps by nested interrupts or also |
| @@ -3081,6 +3094,11 @@ rb_iter_peek(struct ring_buffer_iter *iter, u64 *ts) | |||
| 3081 | if (rb_per_cpu_empty(cpu_buffer)) | 3094 | if (rb_per_cpu_empty(cpu_buffer)) |
| 3082 | return NULL; | 3095 | return NULL; |
| 3083 | 3096 | ||
| 3097 | if (iter->head >= local_read(&iter->head_page->page->commit)) { | ||
| 3098 | rb_inc_iter(iter); | ||
| 3099 | goto again; | ||
| 3100 | } | ||
| 3101 | |||
| 3084 | event = rb_iter_head_event(iter); | 3102 | event = rb_iter_head_event(iter); |
| 3085 | 3103 | ||
| 3086 | switch (event->type_len) { | 3104 | switch (event->type_len) { |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 8b9f20ab8eed..eac6875cb990 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -951,6 +951,11 @@ void trace_find_cmdline(int pid, char comm[]) | |||
| 951 | return; | 951 | return; |
| 952 | } | 952 | } |
| 953 | 953 | ||
| 954 | if (WARN_ON_ONCE(pid < 0)) { | ||
| 955 | strcpy(comm, "<XXX>"); | ||
| 956 | return; | ||
| 957 | } | ||
| 958 | |||
| 954 | if (pid > PID_MAX_DEFAULT) { | 959 | if (pid > PID_MAX_DEFAULT) { |
| 955 | strcpy(comm, "<...>"); | 960 | strcpy(comm, "<...>"); |
| 956 | return; | 961 | return; |
| @@ -3949,7 +3954,7 @@ trace_options_write(struct file *filp, const char __user *ubuf, size_t cnt, | |||
| 3949 | if (!!(topt->flags->val & topt->opt->bit) != val) { | 3954 | if (!!(topt->flags->val & topt->opt->bit) != val) { |
| 3950 | mutex_lock(&trace_types_lock); | 3955 | mutex_lock(&trace_types_lock); |
| 3951 | ret = __set_tracer_option(current_trace, topt->flags, | 3956 | ret = __set_tracer_option(current_trace, topt->flags, |
| 3952 | topt->opt, val); | 3957 | topt->opt, !val); |
| 3953 | mutex_unlock(&trace_types_lock); | 3958 | mutex_unlock(&trace_types_lock); |
| 3954 | if (ret) | 3959 | if (ret) |
| 3955 | return ret; | 3960 | return ret; |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 50504cb228de..e42af9aad69f 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
| @@ -211,8 +211,9 @@ static int filter_pred_pchar(struct filter_pred *pred, void *event, | |||
| 211 | { | 211 | { |
| 212 | char **addr = (char **)(event + pred->offset); | 212 | char **addr = (char **)(event + pred->offset); |
| 213 | int cmp, match; | 213 | int cmp, match; |
| 214 | int len = strlen(*addr) + 1; /* including tailing '\0' */ | ||
| 214 | 215 | ||
| 215 | cmp = pred->regex.match(*addr, &pred->regex, pred->regex.field_len); | 216 | cmp = pred->regex.match(*addr, &pred->regex, len); |
| 216 | 217 | ||
| 217 | match = cmp ^ pred->not; | 218 | match = cmp ^ pred->not; |
| 218 | 219 | ||
| @@ -251,7 +252,18 @@ static int filter_pred_none(struct filter_pred *pred, void *event, | |||
| 251 | return 0; | 252 | return 0; |
| 252 | } | 253 | } |
| 253 | 254 | ||
| 254 | /* Basic regex callbacks */ | 255 | /* |
| 256 | * regex_match_foo - Basic regex callbacks | ||
| 257 | * | ||
| 258 | * @str: the string to be searched | ||
| 259 | * @r: the regex structure containing the pattern string | ||
| 260 | * @len: the length of the string to be searched (including '\0') | ||
| 261 | * | ||
| 262 | * Note: | ||
| 263 | * - @str might not be NULL-terminated if it's of type DYN_STRING | ||
| 264 | * or STATIC_STRING | ||
| 265 | */ | ||
| 266 | |||
| 255 | static int regex_match_full(char *str, struct regex *r, int len) | 267 | static int regex_match_full(char *str, struct regex *r, int len) |
| 256 | { | 268 | { |
| 257 | if (strncmp(str, r->pattern, len) == 0) | 269 | if (strncmp(str, r->pattern, len) == 0) |
| @@ -261,23 +273,24 @@ static int regex_match_full(char *str, struct regex *r, int len) | |||
| 261 | 273 | ||
| 262 | static int regex_match_front(char *str, struct regex *r, int len) | 274 | static int regex_match_front(char *str, struct regex *r, int len) |
| 263 | { | 275 | { |
| 264 | if (strncmp(str, r->pattern, len) == 0) | 276 | if (strncmp(str, r->pattern, r->len) == 0) |
| 265 | return 1; | 277 | return 1; |
| 266 | return 0; | 278 | return 0; |
| 267 | } | 279 | } |
| 268 | 280 | ||
| 269 | static int regex_match_middle(char *str, struct regex *r, int len) | 281 | static int regex_match_middle(char *str, struct regex *r, int len) |
| 270 | { | 282 | { |
| 271 | if (strstr(str, r->pattern)) | 283 | if (strnstr(str, r->pattern, len)) |
| 272 | return 1; | 284 | return 1; |
| 273 | return 0; | 285 | return 0; |
| 274 | } | 286 | } |
| 275 | 287 | ||
| 276 | static int regex_match_end(char *str, struct regex *r, int len) | 288 | static int regex_match_end(char *str, struct regex *r, int len) |
| 277 | { | 289 | { |
| 278 | char *ptr = strstr(str, r->pattern); | 290 | int strlen = len - 1; |
| 279 | 291 | ||
| 280 | if (ptr && (ptr[r->len] == 0)) | 292 | if (strlen >= r->len && |
| 293 | memcmp(str + strlen - r->len, r->pattern, r->len) == 0) | ||
| 281 | return 1; | 294 | return 1; |
| 282 | return 0; | 295 | return 0; |
| 283 | } | 296 | } |
| @@ -781,10 +794,8 @@ static int filter_add_pred(struct filter_parse_state *ps, | |||
| 781 | pred->regex.field_len = field->size; | 794 | pred->regex.field_len = field->size; |
| 782 | } else if (field->filter_type == FILTER_DYN_STRING) | 795 | } else if (field->filter_type == FILTER_DYN_STRING) |
| 783 | fn = filter_pred_strloc; | 796 | fn = filter_pred_strloc; |
| 784 | else { | 797 | else |
| 785 | fn = filter_pred_pchar; | 798 | fn = filter_pred_pchar; |
| 786 | pred->regex.field_len = strlen(pred->regex.pattern); | ||
| 787 | } | ||
| 788 | } else { | 799 | } else { |
| 789 | if (field->is_signed) | 800 | if (field->is_signed) |
| 790 | ret = strict_strtoll(pred->regex.pattern, 0, &val); | 801 | ret = strict_strtoll(pred->regex.pattern, 0, &val); |
diff --git a/kernel/trace/trace_export.c b/kernel/trace/trace_export.c index 458e5bfe26d0..d4fa5dc1ee4e 100644 --- a/kernel/trace/trace_export.c +++ b/kernel/trace/trace_export.c | |||
| @@ -158,7 +158,8 @@ ftrace_format_##name(struct ftrace_event_call *unused, \ | |||
| 158 | BUILD_BUG_ON(len > MAX_FILTER_STR_VAL); \ | 158 | BUILD_BUG_ON(len > MAX_FILTER_STR_VAL); \ |
| 159 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ | 159 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ |
| 160 | offsetof(typeof(field), item), \ | 160 | offsetof(typeof(field), item), \ |
| 161 | sizeof(field.item), 0, FILTER_OTHER); \ | 161 | sizeof(field.item), \ |
| 162 | is_signed_type(type), FILTER_OTHER); \ | ||
| 162 | if (ret) \ | 163 | if (ret) \ |
| 163 | return ret; | 164 | return ret; |
| 164 | 165 | ||
| @@ -168,8 +169,8 @@ ftrace_format_##name(struct ftrace_event_call *unused, \ | |||
| 168 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ | 169 | ret = trace_define_field(event_call, #type "[" #len "]", #item, \ |
| 169 | offsetof(typeof(field), \ | 170 | offsetof(typeof(field), \ |
| 170 | container.item), \ | 171 | container.item), \ |
| 171 | sizeof(field.container.item), 0, \ | 172 | sizeof(field.container.item), \ |
| 172 | FILTER_OTHER); \ | 173 | is_signed_type(type), FILTER_OTHER); \ |
| 173 | if (ret) \ | 174 | if (ret) \ |
| 174 | return ret; | 175 | return ret; |
| 175 | 176 | ||
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index 375f81a568dc..6ea90c0e2c96 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
| @@ -1201,10 +1201,11 @@ static int __probe_event_show_format(struct trace_seq *s, | |||
| 1201 | #undef SHOW_FIELD | 1201 | #undef SHOW_FIELD |
| 1202 | #define SHOW_FIELD(type, item, name) \ | 1202 | #define SHOW_FIELD(type, item, name) \ |
| 1203 | do { \ | 1203 | do { \ |
| 1204 | ret = trace_seq_printf(s, "\tfield: " #type " %s;\t" \ | 1204 | ret = trace_seq_printf(s, "\tfield:" #type " %s;\t" \ |
| 1205 | "offset:%u;\tsize:%u;\n", name, \ | 1205 | "offset:%u;\tsize:%u;\tsigned:%d;\n", name,\ |
| 1206 | (unsigned int)offsetof(typeof(field), item),\ | 1206 | (unsigned int)offsetof(typeof(field), item),\ |
| 1207 | (unsigned int)sizeof(type)); \ | 1207 | (unsigned int)sizeof(type), \ |
| 1208 | is_signed_type(type)); \ | ||
| 1208 | if (!ret) \ | 1209 | if (!ret) \ |
| 1209 | return 0; \ | 1210 | return 0; \ |
| 1210 | } while (0) | 1211 | } while (0) |
diff --git a/kernel/trace/trace_ksym.c b/kernel/trace/trace_ksym.c index faf37fa4408c..94103cdcf9d8 100644 --- a/kernel/trace/trace_ksym.c +++ b/kernel/trace/trace_ksym.c | |||
| @@ -26,12 +26,13 @@ | |||
| 26 | #include <linux/fs.h> | 26 | #include <linux/fs.h> |
| 27 | 27 | ||
| 28 | #include "trace_output.h" | 28 | #include "trace_output.h" |
| 29 | #include "trace_stat.h" | ||
| 30 | #include "trace.h" | 29 | #include "trace.h" |
| 31 | 30 | ||
| 32 | #include <linux/hw_breakpoint.h> | 31 | #include <linux/hw_breakpoint.h> |
| 33 | #include <asm/hw_breakpoint.h> | 32 | #include <asm/hw_breakpoint.h> |
| 34 | 33 | ||
| 34 | #include <asm/atomic.h> | ||
| 35 | |||
| 35 | /* | 36 | /* |
| 36 | * For now, let us restrict the no. of symbols traced simultaneously to number | 37 | * For now, let us restrict the no. of symbols traced simultaneously to number |
| 37 | * of available hardware breakpoint registers. | 38 | * of available hardware breakpoint registers. |
| @@ -44,7 +45,7 @@ struct trace_ksym { | |||
| 44 | struct perf_event **ksym_hbp; | 45 | struct perf_event **ksym_hbp; |
| 45 | struct perf_event_attr attr; | 46 | struct perf_event_attr attr; |
| 46 | #ifdef CONFIG_PROFILE_KSYM_TRACER | 47 | #ifdef CONFIG_PROFILE_KSYM_TRACER |
| 47 | unsigned long counter; | 48 | atomic64_t counter; |
| 48 | #endif | 49 | #endif |
| 49 | struct hlist_node ksym_hlist; | 50 | struct hlist_node ksym_hlist; |
| 50 | }; | 51 | }; |
| @@ -69,9 +70,8 @@ void ksym_collect_stats(unsigned long hbp_hit_addr) | |||
| 69 | 70 | ||
| 70 | rcu_read_lock(); | 71 | rcu_read_lock(); |
| 71 | hlist_for_each_entry_rcu(entry, node, &ksym_filter_head, ksym_hlist) { | 72 | hlist_for_each_entry_rcu(entry, node, &ksym_filter_head, ksym_hlist) { |
| 72 | if ((entry->attr.bp_addr == hbp_hit_addr) && | 73 | if (entry->attr.bp_addr == hbp_hit_addr) { |
| 73 | (entry->counter <= MAX_UL_INT)) { | 74 | atomic64_inc(&entry->counter); |
| 74 | entry->counter++; | ||
| 75 | break; | 75 | break; |
| 76 | } | 76 | } |
| 77 | } | 77 | } |
| @@ -197,7 +197,6 @@ int process_new_ksym_entry(char *ksymname, int op, unsigned long addr) | |||
| 197 | entry->attr.bp_addr = addr; | 197 | entry->attr.bp_addr = addr; |
| 198 | entry->attr.bp_len = HW_BREAKPOINT_LEN_4; | 198 | entry->attr.bp_len = HW_BREAKPOINT_LEN_4; |
| 199 | 199 | ||
| 200 | ret = -EAGAIN; | ||
| 201 | entry->ksym_hbp = register_wide_hw_breakpoint(&entry->attr, | 200 | entry->ksym_hbp = register_wide_hw_breakpoint(&entry->attr, |
| 202 | ksym_hbp_handler); | 201 | ksym_hbp_handler); |
| 203 | 202 | ||
| @@ -300,8 +299,8 @@ static ssize_t ksym_trace_filter_write(struct file *file, | |||
| 300 | * 2: echo 0 > ksym_trace_filter | 299 | * 2: echo 0 > ksym_trace_filter |
| 301 | * 3: echo "*:---" > ksym_trace_filter | 300 | * 3: echo "*:---" > ksym_trace_filter |
| 302 | */ | 301 | */ |
| 303 | if (!buf[0] || !strcmp(buf, "0") || | 302 | if (!input_string[0] || !strcmp(input_string, "0") || |
| 304 | !strcmp(buf, "*:---")) { | 303 | !strcmp(input_string, "*:---")) { |
| 305 | __ksym_trace_reset(); | 304 | __ksym_trace_reset(); |
| 306 | ret = 0; | 305 | ret = 0; |
| 307 | goto out; | 306 | goto out; |
| @@ -444,102 +443,77 @@ struct tracer ksym_tracer __read_mostly = | |||
| 444 | .print_line = ksym_trace_output | 443 | .print_line = ksym_trace_output |
| 445 | }; | 444 | }; |
| 446 | 445 | ||
| 447 | __init static int init_ksym_trace(void) | ||
| 448 | { | ||
| 449 | struct dentry *d_tracer; | ||
| 450 | struct dentry *entry; | ||
| 451 | |||
| 452 | d_tracer = tracing_init_dentry(); | ||
| 453 | ksym_filter_entry_count = 0; | ||
| 454 | |||
| 455 | entry = debugfs_create_file("ksym_trace_filter", 0644, d_tracer, | ||
| 456 | NULL, &ksym_tracing_fops); | ||
| 457 | if (!entry) | ||
| 458 | pr_warning("Could not create debugfs " | ||
| 459 | "'ksym_trace_filter' file\n"); | ||
| 460 | |||
| 461 | return register_tracer(&ksym_tracer); | ||
| 462 | } | ||
| 463 | device_initcall(init_ksym_trace); | ||
| 464 | |||
| 465 | |||
| 466 | #ifdef CONFIG_PROFILE_KSYM_TRACER | 446 | #ifdef CONFIG_PROFILE_KSYM_TRACER |
| 467 | static int ksym_tracer_stat_headers(struct seq_file *m) | 447 | static int ksym_profile_show(struct seq_file *m, void *v) |
| 468 | { | 448 | { |
| 449 | struct hlist_node *node; | ||
| 450 | struct trace_ksym *entry; | ||
| 451 | int access_type = 0; | ||
| 452 | char fn_name[KSYM_NAME_LEN]; | ||
| 453 | |||
| 469 | seq_puts(m, " Access Type "); | 454 | seq_puts(m, " Access Type "); |
| 470 | seq_puts(m, " Symbol Counter\n"); | 455 | seq_puts(m, " Symbol Counter\n"); |
| 471 | seq_puts(m, " ----------- "); | 456 | seq_puts(m, " ----------- "); |
| 472 | seq_puts(m, " ------ -------\n"); | 457 | seq_puts(m, " ------ -------\n"); |
| 473 | return 0; | ||
| 474 | } | ||
| 475 | 458 | ||
| 476 | static int ksym_tracer_stat_show(struct seq_file *m, void *v) | 459 | rcu_read_lock(); |
| 477 | { | 460 | hlist_for_each_entry_rcu(entry, node, &ksym_filter_head, ksym_hlist) { |
| 478 | struct hlist_node *stat = v; | ||
| 479 | struct trace_ksym *entry; | ||
| 480 | int access_type = 0; | ||
| 481 | char fn_name[KSYM_NAME_LEN]; | ||
| 482 | 461 | ||
| 483 | entry = hlist_entry(stat, struct trace_ksym, ksym_hlist); | 462 | access_type = entry->attr.bp_type; |
| 484 | 463 | ||
| 485 | access_type = entry->attr.bp_type; | 464 | switch (access_type) { |
| 465 | case HW_BREAKPOINT_R: | ||
| 466 | seq_puts(m, " R "); | ||
| 467 | break; | ||
| 468 | case HW_BREAKPOINT_W: | ||
| 469 | seq_puts(m, " W "); | ||
| 470 | break; | ||
| 471 | case HW_BREAKPOINT_R | HW_BREAKPOINT_W: | ||
| 472 | seq_puts(m, " RW "); | ||
| 473 | break; | ||
| 474 | default: | ||
| 475 | seq_puts(m, " NA "); | ||
| 476 | } | ||
| 486 | 477 | ||
| 487 | switch (access_type) { | 478 | if (lookup_symbol_name(entry->attr.bp_addr, fn_name) >= 0) |
| 488 | case HW_BREAKPOINT_R: | 479 | seq_printf(m, " %-36s", fn_name); |
| 489 | seq_puts(m, " R "); | 480 | else |
| 490 | break; | 481 | seq_printf(m, " %-36s", "<NA>"); |
| 491 | case HW_BREAKPOINT_W: | 482 | seq_printf(m, " %15llu\n", |
| 492 | seq_puts(m, " W "); | 483 | (unsigned long long)atomic64_read(&entry->counter)); |
| 493 | break; | ||
| 494 | case HW_BREAKPOINT_R | HW_BREAKPOINT_W: | ||
| 495 | seq_puts(m, " RW "); | ||
| 496 | break; | ||
| 497 | default: | ||
| 498 | seq_puts(m, " NA "); | ||
| 499 | } | 484 | } |
| 500 | 485 | rcu_read_unlock(); | |
| 501 | if (lookup_symbol_name(entry->attr.bp_addr, fn_name) >= 0) | ||
| 502 | seq_printf(m, " %-36s", fn_name); | ||
| 503 | else | ||
| 504 | seq_printf(m, " %-36s", "<NA>"); | ||
| 505 | seq_printf(m, " %15lu\n", entry->counter); | ||
| 506 | 486 | ||
| 507 | return 0; | 487 | return 0; |
| 508 | } | 488 | } |
| 509 | 489 | ||
| 510 | static void *ksym_tracer_stat_start(struct tracer_stat *trace) | 490 | static int ksym_profile_open(struct inode *node, struct file *file) |
| 511 | { | 491 | { |
| 512 | return ksym_filter_head.first; | 492 | return single_open(file, ksym_profile_show, NULL); |
| 513 | } | ||
| 514 | |||
| 515 | static void * | ||
| 516 | ksym_tracer_stat_next(void *v, int idx) | ||
| 517 | { | ||
| 518 | struct hlist_node *stat = v; | ||
| 519 | |||
| 520 | return stat->next; | ||
| 521 | } | 493 | } |
| 522 | 494 | ||
| 523 | static struct tracer_stat ksym_tracer_stats = { | 495 | static const struct file_operations ksym_profile_fops = { |
| 524 | .name = "ksym_tracer", | 496 | .open = ksym_profile_open, |
| 525 | .stat_start = ksym_tracer_stat_start, | 497 | .read = seq_read, |
| 526 | .stat_next = ksym_tracer_stat_next, | 498 | .llseek = seq_lseek, |
| 527 | .stat_headers = ksym_tracer_stat_headers, | 499 | .release = single_release, |
| 528 | .stat_show = ksym_tracer_stat_show | ||
| 529 | }; | 500 | }; |
| 501 | #endif /* CONFIG_PROFILE_KSYM_TRACER */ | ||
| 530 | 502 | ||
| 531 | __init static int ksym_tracer_stat_init(void) | 503 | __init static int init_ksym_trace(void) |
| 532 | { | 504 | { |
| 533 | int ret; | 505 | struct dentry *d_tracer; |
| 534 | 506 | ||
| 535 | ret = register_stat_tracer(&ksym_tracer_stats); | 507 | d_tracer = tracing_init_dentry(); |
| 536 | if (ret) { | ||
| 537 | printk(KERN_WARNING "Warning: could not register " | ||
| 538 | "ksym tracer stats\n"); | ||
| 539 | return 1; | ||
| 540 | } | ||
| 541 | 508 | ||
| 542 | return 0; | 509 | trace_create_file("ksym_trace_filter", 0644, d_tracer, |
| 510 | NULL, &ksym_tracing_fops); | ||
| 511 | |||
| 512 | #ifdef CONFIG_PROFILE_KSYM_TRACER | ||
| 513 | trace_create_file("ksym_profile", 0444, d_tracer, | ||
| 514 | NULL, &ksym_profile_fops); | ||
| 515 | #endif | ||
| 516 | |||
| 517 | return register_tracer(&ksym_tracer); | ||
| 543 | } | 518 | } |
| 544 | fs_initcall(ksym_tracer_stat_init); | 519 | device_initcall(init_ksym_trace); |
| 545 | #endif /* CONFIG_PROFILE_KSYM_TRACER */ | ||
