diff options
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/events/core.c | 69 | ||||
-rw-r--r-- | kernel/events/hw_breakpoint.c | 7 | ||||
-rw-r--r-- | kernel/irq/chip.c | 2 | ||||
-rw-r--r-- | kernel/jump_label.c | 135 | ||||
-rw-r--r-- | kernel/printk.c | 5 | ||||
-rw-r--r-- | kernel/sched/core.c | 18 | ||||
-rw-r--r-- | kernel/sched/fair.c | 8 | ||||
-rw-r--r-- | kernel/sched/sched.h | 14 | ||||
-rw-r--r-- | kernel/signal.c | 28 | ||||
-rw-r--r-- | kernel/softirq.c | 6 | ||||
-rw-r--r-- | kernel/trace/ftrace.c | 134 | ||||
-rw-r--r-- | kernel/trace/trace.c | 6 | ||||
-rw-r--r-- | kernel/trace/trace.h | 38 | ||||
-rw-r--r-- | kernel/trace/trace_entries.h | 54 | ||||
-rw-r--r-- | kernel/trace/trace_event_perf.c | 208 | ||||
-rw-r--r-- | kernel/trace/trace_events.c | 12 | ||||
-rw-r--r-- | kernel/trace/trace_events_filter.c | 175 | ||||
-rw-r--r-- | kernel/trace/trace_export.c | 64 | ||||
-rw-r--r-- | kernel/trace/trace_kprobe.c | 8 | ||||
-rw-r--r-- | kernel/trace/trace_output.c | 12 | ||||
-rw-r--r-- | kernel/trace/trace_syscalls.c | 22 | ||||
-rw-r--r-- | kernel/tracepoint.c | 20 | ||||
-rw-r--r-- | kernel/watchdog.c | 24 |
23 files changed, 810 insertions, 259 deletions
diff --git a/kernel/events/core.c b/kernel/events/core.c index 1b5c081d8b9f..e8b32ac75ce3 100644 --- a/kernel/events/core.c +++ b/kernel/events/core.c | |||
@@ -128,7 +128,7 @@ enum event_type_t { | |||
128 | * perf_sched_events : >0 events exist | 128 | * perf_sched_events : >0 events exist |
129 | * perf_cgroup_events: >0 per-cpu cgroup events exist on this cpu | 129 | * perf_cgroup_events: >0 per-cpu cgroup events exist on this cpu |
130 | */ | 130 | */ |
131 | struct jump_label_key_deferred perf_sched_events __read_mostly; | 131 | struct static_key_deferred perf_sched_events __read_mostly; |
132 | static DEFINE_PER_CPU(atomic_t, perf_cgroup_events); | 132 | static DEFINE_PER_CPU(atomic_t, perf_cgroup_events); |
133 | 133 | ||
134 | static atomic_t nr_mmap_events __read_mostly; | 134 | static atomic_t nr_mmap_events __read_mostly; |
@@ -2778,7 +2778,7 @@ static void free_event(struct perf_event *event) | |||
2778 | 2778 | ||
2779 | if (!event->parent) { | 2779 | if (!event->parent) { |
2780 | if (event->attach_state & PERF_ATTACH_TASK) | 2780 | if (event->attach_state & PERF_ATTACH_TASK) |
2781 | jump_label_dec_deferred(&perf_sched_events); | 2781 | static_key_slow_dec_deferred(&perf_sched_events); |
2782 | if (event->attr.mmap || event->attr.mmap_data) | 2782 | if (event->attr.mmap || event->attr.mmap_data) |
2783 | atomic_dec(&nr_mmap_events); | 2783 | atomic_dec(&nr_mmap_events); |
2784 | if (event->attr.comm) | 2784 | if (event->attr.comm) |
@@ -2789,7 +2789,7 @@ static void free_event(struct perf_event *event) | |||
2789 | put_callchain_buffers(); | 2789 | put_callchain_buffers(); |
2790 | if (is_cgroup_event(event)) { | 2790 | if (is_cgroup_event(event)) { |
2791 | atomic_dec(&per_cpu(perf_cgroup_events, event->cpu)); | 2791 | atomic_dec(&per_cpu(perf_cgroup_events, event->cpu)); |
2792 | jump_label_dec_deferred(&perf_sched_events); | 2792 | static_key_slow_dec_deferred(&perf_sched_events); |
2793 | } | 2793 | } |
2794 | } | 2794 | } |
2795 | 2795 | ||
@@ -3238,10 +3238,6 @@ int perf_event_task_disable(void) | |||
3238 | return 0; | 3238 | return 0; |
3239 | } | 3239 | } |
3240 | 3240 | ||
3241 | #ifndef PERF_EVENT_INDEX_OFFSET | ||
3242 | # define PERF_EVENT_INDEX_OFFSET 0 | ||
3243 | #endif | ||
3244 | |||
3245 | static int perf_event_index(struct perf_event *event) | 3241 | static int perf_event_index(struct perf_event *event) |
3246 | { | 3242 | { |
3247 | if (event->hw.state & PERF_HES_STOPPED) | 3243 | if (event->hw.state & PERF_HES_STOPPED) |
@@ -3250,21 +3246,26 @@ static int perf_event_index(struct perf_event *event) | |||
3250 | if (event->state != PERF_EVENT_STATE_ACTIVE) | 3246 | if (event->state != PERF_EVENT_STATE_ACTIVE) |
3251 | return 0; | 3247 | return 0; |
3252 | 3248 | ||
3253 | return event->hw.idx + 1 - PERF_EVENT_INDEX_OFFSET; | 3249 | return event->pmu->event_idx(event); |
3254 | } | 3250 | } |
3255 | 3251 | ||
3256 | static void calc_timer_values(struct perf_event *event, | 3252 | static void calc_timer_values(struct perf_event *event, |
3253 | u64 *now, | ||
3257 | u64 *enabled, | 3254 | u64 *enabled, |
3258 | u64 *running) | 3255 | u64 *running) |
3259 | { | 3256 | { |
3260 | u64 now, ctx_time; | 3257 | u64 ctx_time; |
3261 | 3258 | ||
3262 | now = perf_clock(); | 3259 | *now = perf_clock(); |
3263 | ctx_time = event->shadow_ctx_time + now; | 3260 | ctx_time = event->shadow_ctx_time + *now; |
3264 | *enabled = ctx_time - event->tstamp_enabled; | 3261 | *enabled = ctx_time - event->tstamp_enabled; |
3265 | *running = ctx_time - event->tstamp_running; | 3262 | *running = ctx_time - event->tstamp_running; |
3266 | } | 3263 | } |
3267 | 3264 | ||
3265 | void __weak perf_update_user_clock(struct perf_event_mmap_page *userpg, u64 now) | ||
3266 | { | ||
3267 | } | ||
3268 | |||
3268 | /* | 3269 | /* |
3269 | * Callers need to ensure there can be no nesting of this function, otherwise | 3270 | * Callers need to ensure there can be no nesting of this function, otherwise |
3270 | * the seqlock logic goes bad. We can not serialize this because the arch | 3271 | * the seqlock logic goes bad. We can not serialize this because the arch |
@@ -3274,7 +3275,7 @@ void perf_event_update_userpage(struct perf_event *event) | |||
3274 | { | 3275 | { |
3275 | struct perf_event_mmap_page *userpg; | 3276 | struct perf_event_mmap_page *userpg; |
3276 | struct ring_buffer *rb; | 3277 | struct ring_buffer *rb; |
3277 | u64 enabled, running; | 3278 | u64 enabled, running, now; |
3278 | 3279 | ||
3279 | rcu_read_lock(); | 3280 | rcu_read_lock(); |
3280 | /* | 3281 | /* |
@@ -3286,7 +3287,7 @@ void perf_event_update_userpage(struct perf_event *event) | |||
3286 | * because of locking issue as we can be called in | 3287 | * because of locking issue as we can be called in |
3287 | * NMI context | 3288 | * NMI context |
3288 | */ | 3289 | */ |
3289 | calc_timer_values(event, &enabled, &running); | 3290 | calc_timer_values(event, &now, &enabled, &running); |
3290 | rb = rcu_dereference(event->rb); | 3291 | rb = rcu_dereference(event->rb); |
3291 | if (!rb) | 3292 | if (!rb) |
3292 | goto unlock; | 3293 | goto unlock; |
@@ -3302,7 +3303,7 @@ void perf_event_update_userpage(struct perf_event *event) | |||
3302 | barrier(); | 3303 | barrier(); |
3303 | userpg->index = perf_event_index(event); | 3304 | userpg->index = perf_event_index(event); |
3304 | userpg->offset = perf_event_count(event); | 3305 | userpg->offset = perf_event_count(event); |
3305 | if (event->state == PERF_EVENT_STATE_ACTIVE) | 3306 | if (userpg->index) |
3306 | userpg->offset -= local64_read(&event->hw.prev_count); | 3307 | userpg->offset -= local64_read(&event->hw.prev_count); |
3307 | 3308 | ||
3308 | userpg->time_enabled = enabled + | 3309 | userpg->time_enabled = enabled + |
@@ -3311,6 +3312,8 @@ void perf_event_update_userpage(struct perf_event *event) | |||
3311 | userpg->time_running = running + | 3312 | userpg->time_running = running + |
3312 | atomic64_read(&event->child_total_time_running); | 3313 | atomic64_read(&event->child_total_time_running); |
3313 | 3314 | ||
3315 | perf_update_user_clock(userpg, now); | ||
3316 | |||
3314 | barrier(); | 3317 | barrier(); |
3315 | ++userpg->lock; | 3318 | ++userpg->lock; |
3316 | preempt_enable(); | 3319 | preempt_enable(); |
@@ -3568,6 +3571,8 @@ static int perf_mmap(struct file *file, struct vm_area_struct *vma) | |||
3568 | event->mmap_user = get_current_user(); | 3571 | event->mmap_user = get_current_user(); |
3569 | vma->vm_mm->pinned_vm += event->mmap_locked; | 3572 | vma->vm_mm->pinned_vm += event->mmap_locked; |
3570 | 3573 | ||
3574 | perf_event_update_userpage(event); | ||
3575 | |||
3571 | unlock: | 3576 | unlock: |
3572 | if (!ret) | 3577 | if (!ret) |
3573 | atomic_inc(&event->mmap_count); | 3578 | atomic_inc(&event->mmap_count); |
@@ -3799,7 +3804,7 @@ static void perf_output_read_group(struct perf_output_handle *handle, | |||
3799 | static void perf_output_read(struct perf_output_handle *handle, | 3804 | static void perf_output_read(struct perf_output_handle *handle, |
3800 | struct perf_event *event) | 3805 | struct perf_event *event) |
3801 | { | 3806 | { |
3802 | u64 enabled = 0, running = 0; | 3807 | u64 enabled = 0, running = 0, now; |
3803 | u64 read_format = event->attr.read_format; | 3808 | u64 read_format = event->attr.read_format; |
3804 | 3809 | ||
3805 | /* | 3810 | /* |
@@ -3812,7 +3817,7 @@ static void perf_output_read(struct perf_output_handle *handle, | |||
3812 | * NMI context | 3817 | * NMI context |
3813 | */ | 3818 | */ |
3814 | if (read_format & PERF_FORMAT_TOTAL_TIMES) | 3819 | if (read_format & PERF_FORMAT_TOTAL_TIMES) |
3815 | calc_timer_values(event, &enabled, &running); | 3820 | calc_timer_values(event, &now, &enabled, &running); |
3816 | 3821 | ||
3817 | if (event->attr.read_format & PERF_FORMAT_GROUP) | 3822 | if (event->attr.read_format & PERF_FORMAT_GROUP) |
3818 | perf_output_read_group(handle, event, enabled, running); | 3823 | perf_output_read_group(handle, event, enabled, running); |
@@ -4986,7 +4991,7 @@ fail: | |||
4986 | return err; | 4991 | return err; |
4987 | } | 4992 | } |
4988 | 4993 | ||
4989 | struct jump_label_key perf_swevent_enabled[PERF_COUNT_SW_MAX]; | 4994 | struct static_key perf_swevent_enabled[PERF_COUNT_SW_MAX]; |
4990 | 4995 | ||
4991 | static void sw_perf_event_destroy(struct perf_event *event) | 4996 | static void sw_perf_event_destroy(struct perf_event *event) |
4992 | { | 4997 | { |
@@ -4994,7 +4999,7 @@ static void sw_perf_event_destroy(struct perf_event *event) | |||
4994 | 4999 | ||
4995 | WARN_ON(event->parent); | 5000 | WARN_ON(event->parent); |
4996 | 5001 | ||
4997 | jump_label_dec(&perf_swevent_enabled[event_id]); | 5002 | static_key_slow_dec(&perf_swevent_enabled[event_id]); |
4998 | swevent_hlist_put(event); | 5003 | swevent_hlist_put(event); |
4999 | } | 5004 | } |
5000 | 5005 | ||
@@ -5024,13 +5029,18 @@ static int perf_swevent_init(struct perf_event *event) | |||
5024 | if (err) | 5029 | if (err) |
5025 | return err; | 5030 | return err; |
5026 | 5031 | ||
5027 | jump_label_inc(&perf_swevent_enabled[event_id]); | 5032 | static_key_slow_inc(&perf_swevent_enabled[event_id]); |
5028 | event->destroy = sw_perf_event_destroy; | 5033 | event->destroy = sw_perf_event_destroy; |
5029 | } | 5034 | } |
5030 | 5035 | ||
5031 | return 0; | 5036 | return 0; |
5032 | } | 5037 | } |
5033 | 5038 | ||
5039 | static int perf_swevent_event_idx(struct perf_event *event) | ||
5040 | { | ||
5041 | return 0; | ||
5042 | } | ||
5043 | |||
5034 | static struct pmu perf_swevent = { | 5044 | static struct pmu perf_swevent = { |
5035 | .task_ctx_nr = perf_sw_context, | 5045 | .task_ctx_nr = perf_sw_context, |
5036 | 5046 | ||
@@ -5040,6 +5050,8 @@ static struct pmu perf_swevent = { | |||
5040 | .start = perf_swevent_start, | 5050 | .start = perf_swevent_start, |
5041 | .stop = perf_swevent_stop, | 5051 | .stop = perf_swevent_stop, |
5042 | .read = perf_swevent_read, | 5052 | .read = perf_swevent_read, |
5053 | |||
5054 | .event_idx = perf_swevent_event_idx, | ||
5043 | }; | 5055 | }; |
5044 | 5056 | ||
5045 | #ifdef CONFIG_EVENT_TRACING | 5057 | #ifdef CONFIG_EVENT_TRACING |
@@ -5126,6 +5138,8 @@ static struct pmu perf_tracepoint = { | |||
5126 | .start = perf_swevent_start, | 5138 | .start = perf_swevent_start, |
5127 | .stop = perf_swevent_stop, | 5139 | .stop = perf_swevent_stop, |
5128 | .read = perf_swevent_read, | 5140 | .read = perf_swevent_read, |
5141 | |||
5142 | .event_idx = perf_swevent_event_idx, | ||
5129 | }; | 5143 | }; |
5130 | 5144 | ||
5131 | static inline void perf_tp_register(void) | 5145 | static inline void perf_tp_register(void) |
@@ -5345,6 +5359,8 @@ static struct pmu perf_cpu_clock = { | |||
5345 | .start = cpu_clock_event_start, | 5359 | .start = cpu_clock_event_start, |
5346 | .stop = cpu_clock_event_stop, | 5360 | .stop = cpu_clock_event_stop, |
5347 | .read = cpu_clock_event_read, | 5361 | .read = cpu_clock_event_read, |
5362 | |||
5363 | .event_idx = perf_swevent_event_idx, | ||
5348 | }; | 5364 | }; |
5349 | 5365 | ||
5350 | /* | 5366 | /* |
@@ -5417,6 +5433,8 @@ static struct pmu perf_task_clock = { | |||
5417 | .start = task_clock_event_start, | 5433 | .start = task_clock_event_start, |
5418 | .stop = task_clock_event_stop, | 5434 | .stop = task_clock_event_stop, |
5419 | .read = task_clock_event_read, | 5435 | .read = task_clock_event_read, |
5436 | |||
5437 | .event_idx = perf_swevent_event_idx, | ||
5420 | }; | 5438 | }; |
5421 | 5439 | ||
5422 | static void perf_pmu_nop_void(struct pmu *pmu) | 5440 | static void perf_pmu_nop_void(struct pmu *pmu) |
@@ -5444,6 +5462,11 @@ static void perf_pmu_cancel_txn(struct pmu *pmu) | |||
5444 | perf_pmu_enable(pmu); | 5462 | perf_pmu_enable(pmu); |
5445 | } | 5463 | } |
5446 | 5464 | ||
5465 | static int perf_event_idx_default(struct perf_event *event) | ||
5466 | { | ||
5467 | return event->hw.idx + 1; | ||
5468 | } | ||
5469 | |||
5447 | /* | 5470 | /* |
5448 | * Ensures all contexts with the same task_ctx_nr have the same | 5471 | * Ensures all contexts with the same task_ctx_nr have the same |
5449 | * pmu_cpu_context too. | 5472 | * pmu_cpu_context too. |
@@ -5530,6 +5553,7 @@ static int pmu_dev_alloc(struct pmu *pmu) | |||
5530 | if (!pmu->dev) | 5553 | if (!pmu->dev) |
5531 | goto out; | 5554 | goto out; |
5532 | 5555 | ||
5556 | pmu->dev->groups = pmu->attr_groups; | ||
5533 | device_initialize(pmu->dev); | 5557 | device_initialize(pmu->dev); |
5534 | ret = dev_set_name(pmu->dev, "%s", pmu->name); | 5558 | ret = dev_set_name(pmu->dev, "%s", pmu->name); |
5535 | if (ret) | 5559 | if (ret) |
@@ -5633,6 +5657,9 @@ got_cpu_context: | |||
5633 | pmu->pmu_disable = perf_pmu_nop_void; | 5657 | pmu->pmu_disable = perf_pmu_nop_void; |
5634 | } | 5658 | } |
5635 | 5659 | ||
5660 | if (!pmu->event_idx) | ||
5661 | pmu->event_idx = perf_event_idx_default; | ||
5662 | |||
5636 | list_add_rcu(&pmu->entry, &pmus); | 5663 | list_add_rcu(&pmu->entry, &pmus); |
5637 | ret = 0; | 5664 | ret = 0; |
5638 | unlock: | 5665 | unlock: |
@@ -5825,7 +5852,7 @@ done: | |||
5825 | 5852 | ||
5826 | if (!event->parent) { | 5853 | if (!event->parent) { |
5827 | if (event->attach_state & PERF_ATTACH_TASK) | 5854 | if (event->attach_state & PERF_ATTACH_TASK) |
5828 | jump_label_inc(&perf_sched_events.key); | 5855 | static_key_slow_inc(&perf_sched_events.key); |
5829 | if (event->attr.mmap || event->attr.mmap_data) | 5856 | if (event->attr.mmap || event->attr.mmap_data) |
5830 | atomic_inc(&nr_mmap_events); | 5857 | atomic_inc(&nr_mmap_events); |
5831 | if (event->attr.comm) | 5858 | if (event->attr.comm) |
@@ -6063,7 +6090,7 @@ SYSCALL_DEFINE5(perf_event_open, | |||
6063 | * - that may need work on context switch | 6090 | * - that may need work on context switch |
6064 | */ | 6091 | */ |
6065 | atomic_inc(&per_cpu(perf_cgroup_events, event->cpu)); | 6092 | atomic_inc(&per_cpu(perf_cgroup_events, event->cpu)); |
6066 | jump_label_inc(&perf_sched_events.key); | 6093 | static_key_slow_inc(&perf_sched_events.key); |
6067 | } | 6094 | } |
6068 | 6095 | ||
6069 | /* | 6096 | /* |
diff --git a/kernel/events/hw_breakpoint.c b/kernel/events/hw_breakpoint.c index ee706ce44aa0..3330022a7ac1 100644 --- a/kernel/events/hw_breakpoint.c +++ b/kernel/events/hw_breakpoint.c | |||
@@ -613,6 +613,11 @@ static void hw_breakpoint_stop(struct perf_event *bp, int flags) | |||
613 | bp->hw.state = PERF_HES_STOPPED; | 613 | bp->hw.state = PERF_HES_STOPPED; |
614 | } | 614 | } |
615 | 615 | ||
616 | static int hw_breakpoint_event_idx(struct perf_event *bp) | ||
617 | { | ||
618 | return 0; | ||
619 | } | ||
620 | |||
616 | static struct pmu perf_breakpoint = { | 621 | static struct pmu perf_breakpoint = { |
617 | .task_ctx_nr = perf_sw_context, /* could eventually get its own */ | 622 | .task_ctx_nr = perf_sw_context, /* could eventually get its own */ |
618 | 623 | ||
@@ -622,6 +627,8 @@ static struct pmu perf_breakpoint = { | |||
622 | .start = hw_breakpoint_start, | 627 | .start = hw_breakpoint_start, |
623 | .stop = hw_breakpoint_stop, | 628 | .stop = hw_breakpoint_stop, |
624 | .read = hw_breakpoint_pmu_read, | 629 | .read = hw_breakpoint_pmu_read, |
630 | |||
631 | .event_idx = hw_breakpoint_event_idx, | ||
625 | }; | 632 | }; |
626 | 633 | ||
627 | int __init init_hw_breakpoint(void) | 634 | int __init init_hw_breakpoint(void) |
diff --git a/kernel/irq/chip.c b/kernel/irq/chip.c index fb7db75ee0c8..838687ed802a 100644 --- a/kernel/irq/chip.c +++ b/kernel/irq/chip.c | |||
@@ -16,6 +16,8 @@ | |||
16 | #include <linux/interrupt.h> | 16 | #include <linux/interrupt.h> |
17 | #include <linux/kernel_stat.h> | 17 | #include <linux/kernel_stat.h> |
18 | 18 | ||
19 | #include <trace/events/irq.h> | ||
20 | |||
19 | #include "internals.h" | 21 | #include "internals.h" |
20 | 22 | ||
21 | /** | 23 | /** |
diff --git a/kernel/jump_label.c b/kernel/jump_label.c index 01d3b70fc98a..43049192b5ec 100644 --- a/kernel/jump_label.c +++ b/kernel/jump_label.c | |||
@@ -12,7 +12,7 @@ | |||
12 | #include <linux/slab.h> | 12 | #include <linux/slab.h> |
13 | #include <linux/sort.h> | 13 | #include <linux/sort.h> |
14 | #include <linux/err.h> | 14 | #include <linux/err.h> |
15 | #include <linux/jump_label.h> | 15 | #include <linux/static_key.h> |
16 | 16 | ||
17 | #ifdef HAVE_JUMP_LABEL | 17 | #ifdef HAVE_JUMP_LABEL |
18 | 18 | ||
@@ -29,11 +29,6 @@ void jump_label_unlock(void) | |||
29 | mutex_unlock(&jump_label_mutex); | 29 | mutex_unlock(&jump_label_mutex); |
30 | } | 30 | } |
31 | 31 | ||
32 | bool jump_label_enabled(struct jump_label_key *key) | ||
33 | { | ||
34 | return !!atomic_read(&key->enabled); | ||
35 | } | ||
36 | |||
37 | static int jump_label_cmp(const void *a, const void *b) | 32 | static int jump_label_cmp(const void *a, const void *b) |
38 | { | 33 | { |
39 | const struct jump_entry *jea = a; | 34 | const struct jump_entry *jea = a; |
@@ -58,56 +53,66 @@ jump_label_sort_entries(struct jump_entry *start, struct jump_entry *stop) | |||
58 | sort(start, size, sizeof(struct jump_entry), jump_label_cmp, NULL); | 53 | sort(start, size, sizeof(struct jump_entry), jump_label_cmp, NULL); |
59 | } | 54 | } |
60 | 55 | ||
61 | static void jump_label_update(struct jump_label_key *key, int enable); | 56 | static void jump_label_update(struct static_key *key, int enable); |
62 | 57 | ||
63 | void jump_label_inc(struct jump_label_key *key) | 58 | void static_key_slow_inc(struct static_key *key) |
64 | { | 59 | { |
65 | if (atomic_inc_not_zero(&key->enabled)) | 60 | if (atomic_inc_not_zero(&key->enabled)) |
66 | return; | 61 | return; |
67 | 62 | ||
68 | jump_label_lock(); | 63 | jump_label_lock(); |
69 | if (atomic_read(&key->enabled) == 0) | 64 | if (atomic_read(&key->enabled) == 0) { |
70 | jump_label_update(key, JUMP_LABEL_ENABLE); | 65 | if (!jump_label_get_branch_default(key)) |
66 | jump_label_update(key, JUMP_LABEL_ENABLE); | ||
67 | else | ||
68 | jump_label_update(key, JUMP_LABEL_DISABLE); | ||
69 | } | ||
71 | atomic_inc(&key->enabled); | 70 | atomic_inc(&key->enabled); |
72 | jump_label_unlock(); | 71 | jump_label_unlock(); |
73 | } | 72 | } |
74 | EXPORT_SYMBOL_GPL(jump_label_inc); | 73 | EXPORT_SYMBOL_GPL(static_key_slow_inc); |
75 | 74 | ||
76 | static void __jump_label_dec(struct jump_label_key *key, | 75 | static void __static_key_slow_dec(struct static_key *key, |
77 | unsigned long rate_limit, struct delayed_work *work) | 76 | unsigned long rate_limit, struct delayed_work *work) |
78 | { | 77 | { |
79 | if (!atomic_dec_and_mutex_lock(&key->enabled, &jump_label_mutex)) | 78 | if (!atomic_dec_and_mutex_lock(&key->enabled, &jump_label_mutex)) { |
79 | WARN(atomic_read(&key->enabled) < 0, | ||
80 | "jump label: negative count!\n"); | ||
80 | return; | 81 | return; |
82 | } | ||
81 | 83 | ||
82 | if (rate_limit) { | 84 | if (rate_limit) { |
83 | atomic_inc(&key->enabled); | 85 | atomic_inc(&key->enabled); |
84 | schedule_delayed_work(work, rate_limit); | 86 | schedule_delayed_work(work, rate_limit); |
85 | } else | 87 | } else { |
86 | jump_label_update(key, JUMP_LABEL_DISABLE); | 88 | if (!jump_label_get_branch_default(key)) |
87 | 89 | jump_label_update(key, JUMP_LABEL_DISABLE); | |
90 | else | ||
91 | jump_label_update(key, JUMP_LABEL_ENABLE); | ||
92 | } | ||
88 | jump_label_unlock(); | 93 | jump_label_unlock(); |
89 | } | 94 | } |
90 | EXPORT_SYMBOL_GPL(jump_label_dec); | ||
91 | 95 | ||
92 | static void jump_label_update_timeout(struct work_struct *work) | 96 | static void jump_label_update_timeout(struct work_struct *work) |
93 | { | 97 | { |
94 | struct jump_label_key_deferred *key = | 98 | struct static_key_deferred *key = |
95 | container_of(work, struct jump_label_key_deferred, work.work); | 99 | container_of(work, struct static_key_deferred, work.work); |
96 | __jump_label_dec(&key->key, 0, NULL); | 100 | __static_key_slow_dec(&key->key, 0, NULL); |
97 | } | 101 | } |
98 | 102 | ||
99 | void jump_label_dec(struct jump_label_key *key) | 103 | void static_key_slow_dec(struct static_key *key) |
100 | { | 104 | { |
101 | __jump_label_dec(key, 0, NULL); | 105 | __static_key_slow_dec(key, 0, NULL); |
102 | } | 106 | } |
107 | EXPORT_SYMBOL_GPL(static_key_slow_dec); | ||
103 | 108 | ||
104 | void jump_label_dec_deferred(struct jump_label_key_deferred *key) | 109 | void static_key_slow_dec_deferred(struct static_key_deferred *key) |
105 | { | 110 | { |
106 | __jump_label_dec(&key->key, key->timeout, &key->work); | 111 | __static_key_slow_dec(&key->key, key->timeout, &key->work); |
107 | } | 112 | } |
113 | EXPORT_SYMBOL_GPL(static_key_slow_dec_deferred); | ||
108 | 114 | ||
109 | 115 | void jump_label_rate_limit(struct static_key_deferred *key, | |
110 | void jump_label_rate_limit(struct jump_label_key_deferred *key, | ||
111 | unsigned long rl) | 116 | unsigned long rl) |
112 | { | 117 | { |
113 | key->timeout = rl; | 118 | key->timeout = rl; |
@@ -150,7 +155,7 @@ void __weak __init_or_module arch_jump_label_transform_static(struct jump_entry | |||
150 | arch_jump_label_transform(entry, type); | 155 | arch_jump_label_transform(entry, type); |
151 | } | 156 | } |
152 | 157 | ||
153 | static void __jump_label_update(struct jump_label_key *key, | 158 | static void __jump_label_update(struct static_key *key, |
154 | struct jump_entry *entry, | 159 | struct jump_entry *entry, |
155 | struct jump_entry *stop, int enable) | 160 | struct jump_entry *stop, int enable) |
156 | { | 161 | { |
@@ -167,27 +172,40 @@ static void __jump_label_update(struct jump_label_key *key, | |||
167 | } | 172 | } |
168 | } | 173 | } |
169 | 174 | ||
175 | static enum jump_label_type jump_label_type(struct static_key *key) | ||
176 | { | ||
177 | bool true_branch = jump_label_get_branch_default(key); | ||
178 | bool state = static_key_enabled(key); | ||
179 | |||
180 | if ((!true_branch && state) || (true_branch && !state)) | ||
181 | return JUMP_LABEL_ENABLE; | ||
182 | |||
183 | return JUMP_LABEL_DISABLE; | ||
184 | } | ||
185 | |||
170 | void __init jump_label_init(void) | 186 | void __init jump_label_init(void) |
171 | { | 187 | { |
172 | struct jump_entry *iter_start = __start___jump_table; | 188 | struct jump_entry *iter_start = __start___jump_table; |
173 | struct jump_entry *iter_stop = __stop___jump_table; | 189 | struct jump_entry *iter_stop = __stop___jump_table; |
174 | struct jump_label_key *key = NULL; | 190 | struct static_key *key = NULL; |
175 | struct jump_entry *iter; | 191 | struct jump_entry *iter; |
176 | 192 | ||
177 | jump_label_lock(); | 193 | jump_label_lock(); |
178 | jump_label_sort_entries(iter_start, iter_stop); | 194 | jump_label_sort_entries(iter_start, iter_stop); |
179 | 195 | ||
180 | for (iter = iter_start; iter < iter_stop; iter++) { | 196 | for (iter = iter_start; iter < iter_stop; iter++) { |
181 | struct jump_label_key *iterk; | 197 | struct static_key *iterk; |
182 | 198 | ||
183 | iterk = (struct jump_label_key *)(unsigned long)iter->key; | 199 | iterk = (struct static_key *)(unsigned long)iter->key; |
184 | arch_jump_label_transform_static(iter, jump_label_enabled(iterk) ? | 200 | arch_jump_label_transform_static(iter, jump_label_type(iterk)); |
185 | JUMP_LABEL_ENABLE : JUMP_LABEL_DISABLE); | ||
186 | if (iterk == key) | 201 | if (iterk == key) |
187 | continue; | 202 | continue; |
188 | 203 | ||
189 | key = iterk; | 204 | key = iterk; |
190 | key->entries = iter; | 205 | /* |
206 | * Set key->entries to iter, but preserve JUMP_LABEL_TRUE_BRANCH. | ||
207 | */ | ||
208 | *((unsigned long *)&key->entries) += (unsigned long)iter; | ||
191 | #ifdef CONFIG_MODULES | 209 | #ifdef CONFIG_MODULES |
192 | key->next = NULL; | 210 | key->next = NULL; |
193 | #endif | 211 | #endif |
@@ -197,8 +215,8 @@ void __init jump_label_init(void) | |||
197 | 215 | ||
198 | #ifdef CONFIG_MODULES | 216 | #ifdef CONFIG_MODULES |
199 | 217 | ||
200 | struct jump_label_mod { | 218 | struct static_key_mod { |
201 | struct jump_label_mod *next; | 219 | struct static_key_mod *next; |
202 | struct jump_entry *entries; | 220 | struct jump_entry *entries; |
203 | struct module *mod; | 221 | struct module *mod; |
204 | }; | 222 | }; |
@@ -218,9 +236,9 @@ static int __jump_label_mod_text_reserved(void *start, void *end) | |||
218 | start, end); | 236 | start, end); |
219 | } | 237 | } |
220 | 238 | ||
221 | static void __jump_label_mod_update(struct jump_label_key *key, int enable) | 239 | static void __jump_label_mod_update(struct static_key *key, int enable) |
222 | { | 240 | { |
223 | struct jump_label_mod *mod = key->next; | 241 | struct static_key_mod *mod = key->next; |
224 | 242 | ||
225 | while (mod) { | 243 | while (mod) { |
226 | struct module *m = mod->mod; | 244 | struct module *m = mod->mod; |
@@ -251,11 +269,7 @@ void jump_label_apply_nops(struct module *mod) | |||
251 | return; | 269 | return; |
252 | 270 | ||
253 | for (iter = iter_start; iter < iter_stop; iter++) { | 271 | for (iter = iter_start; iter < iter_stop; iter++) { |
254 | struct jump_label_key *iterk; | 272 | arch_jump_label_transform_static(iter, JUMP_LABEL_DISABLE); |
255 | |||
256 | iterk = (struct jump_label_key *)(unsigned long)iter->key; | ||
257 | arch_jump_label_transform_static(iter, jump_label_enabled(iterk) ? | ||
258 | JUMP_LABEL_ENABLE : JUMP_LABEL_DISABLE); | ||
259 | } | 273 | } |
260 | } | 274 | } |
261 | 275 | ||
@@ -264,8 +278,8 @@ static int jump_label_add_module(struct module *mod) | |||
264 | struct jump_entry *iter_start = mod->jump_entries; | 278 | struct jump_entry *iter_start = mod->jump_entries; |
265 | struct jump_entry *iter_stop = iter_start + mod->num_jump_entries; | 279 | struct jump_entry *iter_stop = iter_start + mod->num_jump_entries; |
266 | struct jump_entry *iter; | 280 | struct jump_entry *iter; |
267 | struct jump_label_key *key = NULL; | 281 | struct static_key *key = NULL; |
268 | struct jump_label_mod *jlm; | 282 | struct static_key_mod *jlm; |
269 | 283 | ||
270 | /* if the module doesn't have jump label entries, just return */ | 284 | /* if the module doesn't have jump label entries, just return */ |
271 | if (iter_start == iter_stop) | 285 | if (iter_start == iter_stop) |
@@ -274,28 +288,30 @@ static int jump_label_add_module(struct module *mod) | |||
274 | jump_label_sort_entries(iter_start, iter_stop); | 288 | jump_label_sort_entries(iter_start, iter_stop); |
275 | 289 | ||
276 | for (iter = iter_start; iter < iter_stop; iter++) { | 290 | for (iter = iter_start; iter < iter_stop; iter++) { |
277 | if (iter->key == (jump_label_t)(unsigned long)key) | 291 | struct static_key *iterk; |
278 | continue; | ||
279 | 292 | ||
280 | key = (struct jump_label_key *)(unsigned long)iter->key; | 293 | iterk = (struct static_key *)(unsigned long)iter->key; |
294 | if (iterk == key) | ||
295 | continue; | ||
281 | 296 | ||
297 | key = iterk; | ||
282 | if (__module_address(iter->key) == mod) { | 298 | if (__module_address(iter->key) == mod) { |
283 | atomic_set(&key->enabled, 0); | 299 | /* |
284 | key->entries = iter; | 300 | * Set key->entries to iter, but preserve JUMP_LABEL_TRUE_BRANCH. |
301 | */ | ||
302 | *((unsigned long *)&key->entries) += (unsigned long)iter; | ||
285 | key->next = NULL; | 303 | key->next = NULL; |
286 | continue; | 304 | continue; |
287 | } | 305 | } |
288 | 306 | jlm = kzalloc(sizeof(struct static_key_mod), GFP_KERNEL); | |
289 | jlm = kzalloc(sizeof(struct jump_label_mod), GFP_KERNEL); | ||
290 | if (!jlm) | 307 | if (!jlm) |
291 | return -ENOMEM; | 308 | return -ENOMEM; |
292 | |||
293 | jlm->mod = mod; | 309 | jlm->mod = mod; |
294 | jlm->entries = iter; | 310 | jlm->entries = iter; |
295 | jlm->next = key->next; | 311 | jlm->next = key->next; |
296 | key->next = jlm; | 312 | key->next = jlm; |
297 | 313 | ||
298 | if (jump_label_enabled(key)) | 314 | if (jump_label_type(key) == JUMP_LABEL_ENABLE) |
299 | __jump_label_update(key, iter, iter_stop, JUMP_LABEL_ENABLE); | 315 | __jump_label_update(key, iter, iter_stop, JUMP_LABEL_ENABLE); |
300 | } | 316 | } |
301 | 317 | ||
@@ -307,14 +323,14 @@ static void jump_label_del_module(struct module *mod) | |||
307 | struct jump_entry *iter_start = mod->jump_entries; | 323 | struct jump_entry *iter_start = mod->jump_entries; |
308 | struct jump_entry *iter_stop = iter_start + mod->num_jump_entries; | 324 | struct jump_entry *iter_stop = iter_start + mod->num_jump_entries; |
309 | struct jump_entry *iter; | 325 | struct jump_entry *iter; |
310 | struct jump_label_key *key = NULL; | 326 | struct static_key *key = NULL; |
311 | struct jump_label_mod *jlm, **prev; | 327 | struct static_key_mod *jlm, **prev; |
312 | 328 | ||
313 | for (iter = iter_start; iter < iter_stop; iter++) { | 329 | for (iter = iter_start; iter < iter_stop; iter++) { |
314 | if (iter->key == (jump_label_t)(unsigned long)key) | 330 | if (iter->key == (jump_label_t)(unsigned long)key) |
315 | continue; | 331 | continue; |
316 | 332 | ||
317 | key = (struct jump_label_key *)(unsigned long)iter->key; | 333 | key = (struct static_key *)(unsigned long)iter->key; |
318 | 334 | ||
319 | if (__module_address(iter->key) == mod) | 335 | if (__module_address(iter->key) == mod) |
320 | continue; | 336 | continue; |
@@ -416,12 +432,13 @@ int jump_label_text_reserved(void *start, void *end) | |||
416 | return ret; | 432 | return ret; |
417 | } | 433 | } |
418 | 434 | ||
419 | static void jump_label_update(struct jump_label_key *key, int enable) | 435 | static void jump_label_update(struct static_key *key, int enable) |
420 | { | 436 | { |
421 | struct jump_entry *entry = key->entries, *stop = __stop___jump_table; | 437 | struct jump_entry *stop = __stop___jump_table; |
438 | struct jump_entry *entry = jump_label_get_entries(key); | ||
422 | 439 | ||
423 | #ifdef CONFIG_MODULES | 440 | #ifdef CONFIG_MODULES |
424 | struct module *mod = __module_address((jump_label_t)key); | 441 | struct module *mod = __module_address((unsigned long)key); |
425 | 442 | ||
426 | __jump_label_mod_update(key, enable); | 443 | __jump_label_mod_update(key, enable); |
427 | 444 | ||
diff --git a/kernel/printk.c b/kernel/printk.c index 13c0a1143f49..cb8a6bd697c6 100644 --- a/kernel/printk.c +++ b/kernel/printk.c | |||
@@ -44,6 +44,9 @@ | |||
44 | 44 | ||
45 | #include <asm/uaccess.h> | 45 | #include <asm/uaccess.h> |
46 | 46 | ||
47 | #define CREATE_TRACE_POINTS | ||
48 | #include <trace/events/printk.h> | ||
49 | |||
47 | /* | 50 | /* |
48 | * Architectures can override it: | 51 | * Architectures can override it: |
49 | */ | 52 | */ |
@@ -542,6 +545,8 @@ MODULE_PARM_DESC(ignore_loglevel, "ignore loglevel setting, to" | |||
542 | static void _call_console_drivers(unsigned start, | 545 | static void _call_console_drivers(unsigned start, |
543 | unsigned end, int msg_log_level) | 546 | unsigned end, int msg_log_level) |
544 | { | 547 | { |
548 | trace_console(&LOG_BUF(0), start, end, log_buf_len); | ||
549 | |||
545 | if ((msg_log_level < console_loglevel || ignore_loglevel) && | 550 | if ((msg_log_level < console_loglevel || ignore_loglevel) && |
546 | console_drivers && start != end) { | 551 | console_drivers && start != end) { |
547 | if ((start & LOG_BUF_MASK) > (end & LOG_BUF_MASK)) { | 552 | if ((start & LOG_BUF_MASK) > (end & LOG_BUF_MASK)) { |
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index b342f57879e6..6c41ba49767a 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c | |||
@@ -162,13 +162,13 @@ static int sched_feat_show(struct seq_file *m, void *v) | |||
162 | 162 | ||
163 | #ifdef HAVE_JUMP_LABEL | 163 | #ifdef HAVE_JUMP_LABEL |
164 | 164 | ||
165 | #define jump_label_key__true jump_label_key_enabled | 165 | #define jump_label_key__true STATIC_KEY_INIT_TRUE |
166 | #define jump_label_key__false jump_label_key_disabled | 166 | #define jump_label_key__false STATIC_KEY_INIT_FALSE |
167 | 167 | ||
168 | #define SCHED_FEAT(name, enabled) \ | 168 | #define SCHED_FEAT(name, enabled) \ |
169 | jump_label_key__##enabled , | 169 | jump_label_key__##enabled , |
170 | 170 | ||
171 | struct jump_label_key sched_feat_keys[__SCHED_FEAT_NR] = { | 171 | struct static_key sched_feat_keys[__SCHED_FEAT_NR] = { |
172 | #include "features.h" | 172 | #include "features.h" |
173 | }; | 173 | }; |
174 | 174 | ||
@@ -176,14 +176,14 @@ struct jump_label_key sched_feat_keys[__SCHED_FEAT_NR] = { | |||
176 | 176 | ||
177 | static void sched_feat_disable(int i) | 177 | static void sched_feat_disable(int i) |
178 | { | 178 | { |
179 | if (jump_label_enabled(&sched_feat_keys[i])) | 179 | if (static_key_enabled(&sched_feat_keys[i])) |
180 | jump_label_dec(&sched_feat_keys[i]); | 180 | static_key_slow_dec(&sched_feat_keys[i]); |
181 | } | 181 | } |
182 | 182 | ||
183 | static void sched_feat_enable(int i) | 183 | static void sched_feat_enable(int i) |
184 | { | 184 | { |
185 | if (!jump_label_enabled(&sched_feat_keys[i])) | 185 | if (!static_key_enabled(&sched_feat_keys[i])) |
186 | jump_label_inc(&sched_feat_keys[i]); | 186 | static_key_slow_inc(&sched_feat_keys[i]); |
187 | } | 187 | } |
188 | #else | 188 | #else |
189 | static void sched_feat_disable(int i) { }; | 189 | static void sched_feat_disable(int i) { }; |
@@ -894,7 +894,7 @@ static void update_rq_clock_task(struct rq *rq, s64 delta) | |||
894 | delta -= irq_delta; | 894 | delta -= irq_delta; |
895 | #endif | 895 | #endif |
896 | #ifdef CONFIG_PARAVIRT_TIME_ACCOUNTING | 896 | #ifdef CONFIG_PARAVIRT_TIME_ACCOUNTING |
897 | if (static_branch((¶virt_steal_rq_enabled))) { | 897 | if (static_key_false((¶virt_steal_rq_enabled))) { |
898 | u64 st; | 898 | u64 st; |
899 | 899 | ||
900 | steal = paravirt_steal_clock(cpu_of(rq)); | 900 | steal = paravirt_steal_clock(cpu_of(rq)); |
@@ -2755,7 +2755,7 @@ void account_idle_time(cputime_t cputime) | |||
2755 | static __always_inline bool steal_account_process_tick(void) | 2755 | static __always_inline bool steal_account_process_tick(void) |
2756 | { | 2756 | { |
2757 | #ifdef CONFIG_PARAVIRT | 2757 | #ifdef CONFIG_PARAVIRT |
2758 | if (static_branch(¶virt_steal_enabled)) { | 2758 | if (static_key_false(¶virt_steal_enabled)) { |
2759 | u64 steal, st = 0; | 2759 | u64 steal, st = 0; |
2760 | 2760 | ||
2761 | steal = paravirt_steal_clock(smp_processor_id()); | 2761 | steal = paravirt_steal_clock(smp_processor_id()); |
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index aca16b843b7e..fd974faf467d 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -1401,20 +1401,20 @@ entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) | |||
1401 | #ifdef CONFIG_CFS_BANDWIDTH | 1401 | #ifdef CONFIG_CFS_BANDWIDTH |
1402 | 1402 | ||
1403 | #ifdef HAVE_JUMP_LABEL | 1403 | #ifdef HAVE_JUMP_LABEL |
1404 | static struct jump_label_key __cfs_bandwidth_used; | 1404 | static struct static_key __cfs_bandwidth_used; |
1405 | 1405 | ||
1406 | static inline bool cfs_bandwidth_used(void) | 1406 | static inline bool cfs_bandwidth_used(void) |
1407 | { | 1407 | { |
1408 | return static_branch(&__cfs_bandwidth_used); | 1408 | return static_key_false(&__cfs_bandwidth_used); |
1409 | } | 1409 | } |
1410 | 1410 | ||
1411 | void account_cfs_bandwidth_used(int enabled, int was_enabled) | 1411 | void account_cfs_bandwidth_used(int enabled, int was_enabled) |
1412 | { | 1412 | { |
1413 | /* only need to count groups transitioning between enabled/!enabled */ | 1413 | /* only need to count groups transitioning between enabled/!enabled */ |
1414 | if (enabled && !was_enabled) | 1414 | if (enabled && !was_enabled) |
1415 | jump_label_inc(&__cfs_bandwidth_used); | 1415 | static_key_slow_inc(&__cfs_bandwidth_used); |
1416 | else if (!enabled && was_enabled) | 1416 | else if (!enabled && was_enabled) |
1417 | jump_label_dec(&__cfs_bandwidth_used); | 1417 | static_key_slow_dec(&__cfs_bandwidth_used); |
1418 | } | 1418 | } |
1419 | #else /* HAVE_JUMP_LABEL */ | 1419 | #else /* HAVE_JUMP_LABEL */ |
1420 | static bool cfs_bandwidth_used(void) | 1420 | static bool cfs_bandwidth_used(void) |
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index 98c0c2623db8..b4cd6d8ea150 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h | |||
@@ -611,7 +611,7 @@ static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) | |||
611 | * Tunables that become constants when CONFIG_SCHED_DEBUG is off: | 611 | * Tunables that become constants when CONFIG_SCHED_DEBUG is off: |
612 | */ | 612 | */ |
613 | #ifdef CONFIG_SCHED_DEBUG | 613 | #ifdef CONFIG_SCHED_DEBUG |
614 | # include <linux/jump_label.h> | 614 | # include <linux/static_key.h> |
615 | # define const_debug __read_mostly | 615 | # define const_debug __read_mostly |
616 | #else | 616 | #else |
617 | # define const_debug const | 617 | # define const_debug const |
@@ -630,18 +630,18 @@ enum { | |||
630 | #undef SCHED_FEAT | 630 | #undef SCHED_FEAT |
631 | 631 | ||
632 | #if defined(CONFIG_SCHED_DEBUG) && defined(HAVE_JUMP_LABEL) | 632 | #if defined(CONFIG_SCHED_DEBUG) && defined(HAVE_JUMP_LABEL) |
633 | static __always_inline bool static_branch__true(struct jump_label_key *key) | 633 | static __always_inline bool static_branch__true(struct static_key *key) |
634 | { | 634 | { |
635 | return likely(static_branch(key)); /* Not out of line branch. */ | 635 | return static_key_true(key); /* Not out of line branch. */ |
636 | } | 636 | } |
637 | 637 | ||
638 | static __always_inline bool static_branch__false(struct jump_label_key *key) | 638 | static __always_inline bool static_branch__false(struct static_key *key) |
639 | { | 639 | { |
640 | return unlikely(static_branch(key)); /* Out of line branch. */ | 640 | return static_key_false(key); /* Out of line branch. */ |
641 | } | 641 | } |
642 | 642 | ||
643 | #define SCHED_FEAT(name, enabled) \ | 643 | #define SCHED_FEAT(name, enabled) \ |
644 | static __always_inline bool static_branch_##name(struct jump_label_key *key) \ | 644 | static __always_inline bool static_branch_##name(struct static_key *key) \ |
645 | { \ | 645 | { \ |
646 | return static_branch__##enabled(key); \ | 646 | return static_branch__##enabled(key); \ |
647 | } | 647 | } |
@@ -650,7 +650,7 @@ static __always_inline bool static_branch_##name(struct jump_label_key *key) \ | |||
650 | 650 | ||
651 | #undef SCHED_FEAT | 651 | #undef SCHED_FEAT |
652 | 652 | ||
653 | extern struct jump_label_key sched_feat_keys[__SCHED_FEAT_NR]; | 653 | extern struct static_key sched_feat_keys[__SCHED_FEAT_NR]; |
654 | #define sched_feat(x) (static_branch_##x(&sched_feat_keys[__SCHED_FEAT_##x])) | 654 | #define sched_feat(x) (static_branch_##x(&sched_feat_keys[__SCHED_FEAT_##x])) |
655 | #else /* !(SCHED_DEBUG && HAVE_JUMP_LABEL) */ | 655 | #else /* !(SCHED_DEBUG && HAVE_JUMP_LABEL) */ |
656 | #define sched_feat(x) (sysctl_sched_features & (1UL << __SCHED_FEAT_##x)) | 656 | #define sched_feat(x) (sysctl_sched_features & (1UL << __SCHED_FEAT_##x)) |
diff --git a/kernel/signal.c b/kernel/signal.c index c73c4284160e..8511e39813c7 100644 --- a/kernel/signal.c +++ b/kernel/signal.c | |||
@@ -1054,13 +1054,13 @@ static int __send_signal(int sig, struct siginfo *info, struct task_struct *t, | |||
1054 | struct sigpending *pending; | 1054 | struct sigpending *pending; |
1055 | struct sigqueue *q; | 1055 | struct sigqueue *q; |
1056 | int override_rlimit; | 1056 | int override_rlimit; |
1057 | 1057 | int ret = 0, result; | |
1058 | trace_signal_generate(sig, info, t); | ||
1059 | 1058 | ||
1060 | assert_spin_locked(&t->sighand->siglock); | 1059 | assert_spin_locked(&t->sighand->siglock); |
1061 | 1060 | ||
1061 | result = TRACE_SIGNAL_IGNORED; | ||
1062 | if (!prepare_signal(sig, t, from_ancestor_ns)) | 1062 | if (!prepare_signal(sig, t, from_ancestor_ns)) |
1063 | return 0; | 1063 | goto ret; |
1064 | 1064 | ||
1065 | pending = group ? &t->signal->shared_pending : &t->pending; | 1065 | pending = group ? &t->signal->shared_pending : &t->pending; |
1066 | /* | 1066 | /* |
@@ -1068,8 +1068,11 @@ static int __send_signal(int sig, struct siginfo *info, struct task_struct *t, | |||
1068 | * exactly one non-rt signal, so that we can get more | 1068 | * exactly one non-rt signal, so that we can get more |
1069 | * detailed information about the cause of the signal. | 1069 | * detailed information about the cause of the signal. |
1070 | */ | 1070 | */ |
1071 | result = TRACE_SIGNAL_ALREADY_PENDING; | ||
1071 | if (legacy_queue(pending, sig)) | 1072 | if (legacy_queue(pending, sig)) |
1072 | return 0; | 1073 | goto ret; |
1074 | |||
1075 | result = TRACE_SIGNAL_DELIVERED; | ||
1073 | /* | 1076 | /* |
1074 | * fast-pathed signals for kernel-internal things like SIGSTOP | 1077 | * fast-pathed signals for kernel-internal things like SIGSTOP |
1075 | * or SIGKILL. | 1078 | * or SIGKILL. |
@@ -1127,14 +1130,15 @@ static int __send_signal(int sig, struct siginfo *info, struct task_struct *t, | |||
1127 | * signal was rt and sent by user using something | 1130 | * signal was rt and sent by user using something |
1128 | * other than kill(). | 1131 | * other than kill(). |
1129 | */ | 1132 | */ |
1130 | trace_signal_overflow_fail(sig, group, info); | 1133 | result = TRACE_SIGNAL_OVERFLOW_FAIL; |
1131 | return -EAGAIN; | 1134 | ret = -EAGAIN; |
1135 | goto ret; | ||
1132 | } else { | 1136 | } else { |
1133 | /* | 1137 | /* |
1134 | * This is a silent loss of information. We still | 1138 | * This is a silent loss of information. We still |
1135 | * send the signal, but the *info bits are lost. | 1139 | * send the signal, but the *info bits are lost. |
1136 | */ | 1140 | */ |
1137 | trace_signal_lose_info(sig, group, info); | 1141 | result = TRACE_SIGNAL_LOSE_INFO; |
1138 | } | 1142 | } |
1139 | } | 1143 | } |
1140 | 1144 | ||
@@ -1142,7 +1146,9 @@ out_set: | |||
1142 | signalfd_notify(t, sig); | 1146 | signalfd_notify(t, sig); |
1143 | sigaddset(&pending->signal, sig); | 1147 | sigaddset(&pending->signal, sig); |
1144 | complete_signal(sig, t, group); | 1148 | complete_signal(sig, t, group); |
1145 | return 0; | 1149 | ret: |
1150 | trace_signal_generate(sig, info, t, group, result); | ||
1151 | return ret; | ||
1146 | } | 1152 | } |
1147 | 1153 | ||
1148 | static int send_signal(int sig, struct siginfo *info, struct task_struct *t, | 1154 | static int send_signal(int sig, struct siginfo *info, struct task_struct *t, |
@@ -1585,7 +1591,7 @@ int send_sigqueue(struct sigqueue *q, struct task_struct *t, int group) | |||
1585 | int sig = q->info.si_signo; | 1591 | int sig = q->info.si_signo; |
1586 | struct sigpending *pending; | 1592 | struct sigpending *pending; |
1587 | unsigned long flags; | 1593 | unsigned long flags; |
1588 | int ret; | 1594 | int ret, result; |
1589 | 1595 | ||
1590 | BUG_ON(!(q->flags & SIGQUEUE_PREALLOC)); | 1596 | BUG_ON(!(q->flags & SIGQUEUE_PREALLOC)); |
1591 | 1597 | ||
@@ -1594,6 +1600,7 @@ int send_sigqueue(struct sigqueue *q, struct task_struct *t, int group) | |||
1594 | goto ret; | 1600 | goto ret; |
1595 | 1601 | ||
1596 | ret = 1; /* the signal is ignored */ | 1602 | ret = 1; /* the signal is ignored */ |
1603 | result = TRACE_SIGNAL_IGNORED; | ||
1597 | if (!prepare_signal(sig, t, 0)) | 1604 | if (!prepare_signal(sig, t, 0)) |
1598 | goto out; | 1605 | goto out; |
1599 | 1606 | ||
@@ -1605,6 +1612,7 @@ int send_sigqueue(struct sigqueue *q, struct task_struct *t, int group) | |||
1605 | */ | 1612 | */ |
1606 | BUG_ON(q->info.si_code != SI_TIMER); | 1613 | BUG_ON(q->info.si_code != SI_TIMER); |
1607 | q->info.si_overrun++; | 1614 | q->info.si_overrun++; |
1615 | result = TRACE_SIGNAL_ALREADY_PENDING; | ||
1608 | goto out; | 1616 | goto out; |
1609 | } | 1617 | } |
1610 | q->info.si_overrun = 0; | 1618 | q->info.si_overrun = 0; |
@@ -1614,7 +1622,9 @@ int send_sigqueue(struct sigqueue *q, struct task_struct *t, int group) | |||
1614 | list_add_tail(&q->list, &pending->list); | 1622 | list_add_tail(&q->list, &pending->list); |
1615 | sigaddset(&pending->signal, sig); | 1623 | sigaddset(&pending->signal, sig); |
1616 | complete_signal(sig, t, group); | 1624 | complete_signal(sig, t, group); |
1625 | result = TRACE_SIGNAL_DELIVERED; | ||
1617 | out: | 1626 | out: |
1627 | trace_signal_generate(sig, &q->info, t, group, result); | ||
1618 | unlock_task_sighand(t, &flags); | 1628 | unlock_task_sighand(t, &flags); |
1619 | ret: | 1629 | ret: |
1620 | return ret; | 1630 | return ret; |
diff --git a/kernel/softirq.c b/kernel/softirq.c index 4eb3a0fa351e..06d40993594a 100644 --- a/kernel/softirq.c +++ b/kernel/softirq.c | |||
@@ -385,6 +385,12 @@ void raise_softirq(unsigned int nr) | |||
385 | local_irq_restore(flags); | 385 | local_irq_restore(flags); |
386 | } | 386 | } |
387 | 387 | ||
388 | void __raise_softirq_irqoff(unsigned int nr) | ||
389 | { | ||
390 | trace_softirq_raise(nr); | ||
391 | or_softirq_pending(1UL << nr); | ||
392 | } | ||
393 | |||
388 | void open_softirq(int nr, void (*action)(struct softirq_action *)) | 394 | void open_softirq(int nr, void (*action)(struct softirq_action *)) |
389 | { | 395 | { |
390 | softirq_vec[nr].action = action; | 396 | softirq_vec[nr].action = action; |
diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 683d559a0eef..867bd1dd2dd0 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c | |||
@@ -62,6 +62,8 @@ | |||
62 | #define FTRACE_HASH_DEFAULT_BITS 10 | 62 | #define FTRACE_HASH_DEFAULT_BITS 10 |
63 | #define FTRACE_HASH_MAX_BITS 12 | 63 | #define FTRACE_HASH_MAX_BITS 12 |
64 | 64 | ||
65 | #define FL_GLOBAL_CONTROL_MASK (FTRACE_OPS_FL_GLOBAL | FTRACE_OPS_FL_CONTROL) | ||
66 | |||
65 | /* ftrace_enabled is a method to turn ftrace on or off */ | 67 | /* ftrace_enabled is a method to turn ftrace on or off */ |
66 | int ftrace_enabled __read_mostly; | 68 | int ftrace_enabled __read_mostly; |
67 | static int last_ftrace_enabled; | 69 | static int last_ftrace_enabled; |
@@ -89,12 +91,14 @@ static struct ftrace_ops ftrace_list_end __read_mostly = { | |||
89 | }; | 91 | }; |
90 | 92 | ||
91 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; | 93 | static struct ftrace_ops *ftrace_global_list __read_mostly = &ftrace_list_end; |
94 | static struct ftrace_ops *ftrace_control_list __read_mostly = &ftrace_list_end; | ||
92 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; | 95 | static struct ftrace_ops *ftrace_ops_list __read_mostly = &ftrace_list_end; |
93 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; | 96 | ftrace_func_t ftrace_trace_function __read_mostly = ftrace_stub; |
94 | static ftrace_func_t __ftrace_trace_function_delay __read_mostly = ftrace_stub; | 97 | static ftrace_func_t __ftrace_trace_function_delay __read_mostly = ftrace_stub; |
95 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; | 98 | ftrace_func_t __ftrace_trace_function __read_mostly = ftrace_stub; |
96 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; | 99 | ftrace_func_t ftrace_pid_function __read_mostly = ftrace_stub; |
97 | static struct ftrace_ops global_ops; | 100 | static struct ftrace_ops global_ops; |
101 | static struct ftrace_ops control_ops; | ||
98 | 102 | ||
99 | static void | 103 | static void |
100 | ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip); | 104 | ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip); |
@@ -168,6 +172,32 @@ static void ftrace_test_stop_func(unsigned long ip, unsigned long parent_ip) | |||
168 | } | 172 | } |
169 | #endif | 173 | #endif |
170 | 174 | ||
175 | static void control_ops_disable_all(struct ftrace_ops *ops) | ||
176 | { | ||
177 | int cpu; | ||
178 | |||
179 | for_each_possible_cpu(cpu) | ||
180 | *per_cpu_ptr(ops->disabled, cpu) = 1; | ||
181 | } | ||
182 | |||
183 | static int control_ops_alloc(struct ftrace_ops *ops) | ||
184 | { | ||
185 | int __percpu *disabled; | ||
186 | |||
187 | disabled = alloc_percpu(int); | ||
188 | if (!disabled) | ||
189 | return -ENOMEM; | ||
190 | |||
191 | ops->disabled = disabled; | ||
192 | control_ops_disable_all(ops); | ||
193 | return 0; | ||
194 | } | ||
195 | |||
196 | static void control_ops_free(struct ftrace_ops *ops) | ||
197 | { | ||
198 | free_percpu(ops->disabled); | ||
199 | } | ||
200 | |||
171 | static void update_global_ops(void) | 201 | static void update_global_ops(void) |
172 | { | 202 | { |
173 | ftrace_func_t func; | 203 | ftrace_func_t func; |
@@ -259,6 +289,26 @@ static int remove_ftrace_ops(struct ftrace_ops **list, struct ftrace_ops *ops) | |||
259 | return 0; | 289 | return 0; |
260 | } | 290 | } |
261 | 291 | ||
292 | static void add_ftrace_list_ops(struct ftrace_ops **list, | ||
293 | struct ftrace_ops *main_ops, | ||
294 | struct ftrace_ops *ops) | ||
295 | { | ||
296 | int first = *list == &ftrace_list_end; | ||
297 | add_ftrace_ops(list, ops); | ||
298 | if (first) | ||
299 | add_ftrace_ops(&ftrace_ops_list, main_ops); | ||
300 | } | ||
301 | |||
302 | static int remove_ftrace_list_ops(struct ftrace_ops **list, | ||
303 | struct ftrace_ops *main_ops, | ||
304 | struct ftrace_ops *ops) | ||
305 | { | ||
306 | int ret = remove_ftrace_ops(list, ops); | ||
307 | if (!ret && *list == &ftrace_list_end) | ||
308 | ret = remove_ftrace_ops(&ftrace_ops_list, main_ops); | ||
309 | return ret; | ||
310 | } | ||
311 | |||
262 | static int __register_ftrace_function(struct ftrace_ops *ops) | 312 | static int __register_ftrace_function(struct ftrace_ops *ops) |
263 | { | 313 | { |
264 | if (ftrace_disabled) | 314 | if (ftrace_disabled) |
@@ -270,15 +320,20 @@ static int __register_ftrace_function(struct ftrace_ops *ops) | |||
270 | if (WARN_ON(ops->flags & FTRACE_OPS_FL_ENABLED)) | 320 | if (WARN_ON(ops->flags & FTRACE_OPS_FL_ENABLED)) |
271 | return -EBUSY; | 321 | return -EBUSY; |
272 | 322 | ||
323 | /* We don't support both control and global flags set. */ | ||
324 | if ((ops->flags & FL_GLOBAL_CONTROL_MASK) == FL_GLOBAL_CONTROL_MASK) | ||
325 | return -EINVAL; | ||
326 | |||
273 | if (!core_kernel_data((unsigned long)ops)) | 327 | if (!core_kernel_data((unsigned long)ops)) |
274 | ops->flags |= FTRACE_OPS_FL_DYNAMIC; | 328 | ops->flags |= FTRACE_OPS_FL_DYNAMIC; |
275 | 329 | ||
276 | if (ops->flags & FTRACE_OPS_FL_GLOBAL) { | 330 | if (ops->flags & FTRACE_OPS_FL_GLOBAL) { |
277 | int first = ftrace_global_list == &ftrace_list_end; | 331 | add_ftrace_list_ops(&ftrace_global_list, &global_ops, ops); |
278 | add_ftrace_ops(&ftrace_global_list, ops); | ||
279 | ops->flags |= FTRACE_OPS_FL_ENABLED; | 332 | ops->flags |= FTRACE_OPS_FL_ENABLED; |
280 | if (first) | 333 | } else if (ops->flags & FTRACE_OPS_FL_CONTROL) { |
281 | add_ftrace_ops(&ftrace_ops_list, &global_ops); | 334 | if (control_ops_alloc(ops)) |
335 | return -ENOMEM; | ||
336 | add_ftrace_list_ops(&ftrace_control_list, &control_ops, ops); | ||
282 | } else | 337 | } else |
283 | add_ftrace_ops(&ftrace_ops_list, ops); | 338 | add_ftrace_ops(&ftrace_ops_list, ops); |
284 | 339 | ||
@@ -302,11 +357,23 @@ static int __unregister_ftrace_function(struct ftrace_ops *ops) | |||
302 | return -EINVAL; | 357 | return -EINVAL; |
303 | 358 | ||
304 | if (ops->flags & FTRACE_OPS_FL_GLOBAL) { | 359 | if (ops->flags & FTRACE_OPS_FL_GLOBAL) { |
305 | ret = remove_ftrace_ops(&ftrace_global_list, ops); | 360 | ret = remove_ftrace_list_ops(&ftrace_global_list, |
306 | if (!ret && ftrace_global_list == &ftrace_list_end) | 361 | &global_ops, ops); |
307 | ret = remove_ftrace_ops(&ftrace_ops_list, &global_ops); | ||
308 | if (!ret) | 362 | if (!ret) |
309 | ops->flags &= ~FTRACE_OPS_FL_ENABLED; | 363 | ops->flags &= ~FTRACE_OPS_FL_ENABLED; |
364 | } else if (ops->flags & FTRACE_OPS_FL_CONTROL) { | ||
365 | ret = remove_ftrace_list_ops(&ftrace_control_list, | ||
366 | &control_ops, ops); | ||
367 | if (!ret) { | ||
368 | /* | ||
369 | * The ftrace_ops is now removed from the list, | ||
370 | * so there'll be no new users. We must ensure | ||
371 | * all current users are done before we free | ||
372 | * the control data. | ||
373 | */ | ||
374 | synchronize_sched(); | ||
375 | control_ops_free(ops); | ||
376 | } | ||
310 | } else | 377 | } else |
311 | ret = remove_ftrace_ops(&ftrace_ops_list, ops); | 378 | ret = remove_ftrace_ops(&ftrace_ops_list, ops); |
312 | 379 | ||
@@ -1119,6 +1186,12 @@ static void free_ftrace_hash_rcu(struct ftrace_hash *hash) | |||
1119 | call_rcu_sched(&hash->rcu, __free_ftrace_hash_rcu); | 1186 | call_rcu_sched(&hash->rcu, __free_ftrace_hash_rcu); |
1120 | } | 1187 | } |
1121 | 1188 | ||
1189 | void ftrace_free_filter(struct ftrace_ops *ops) | ||
1190 | { | ||
1191 | free_ftrace_hash(ops->filter_hash); | ||
1192 | free_ftrace_hash(ops->notrace_hash); | ||
1193 | } | ||
1194 | |||
1122 | static struct ftrace_hash *alloc_ftrace_hash(int size_bits) | 1195 | static struct ftrace_hash *alloc_ftrace_hash(int size_bits) |
1123 | { | 1196 | { |
1124 | struct ftrace_hash *hash; | 1197 | struct ftrace_hash *hash; |
@@ -1129,7 +1202,7 @@ static struct ftrace_hash *alloc_ftrace_hash(int size_bits) | |||
1129 | return NULL; | 1202 | return NULL; |
1130 | 1203 | ||
1131 | size = 1 << size_bits; | 1204 | size = 1 << size_bits; |
1132 | hash->buckets = kzalloc(sizeof(*hash->buckets) * size, GFP_KERNEL); | 1205 | hash->buckets = kcalloc(size, sizeof(*hash->buckets), GFP_KERNEL); |
1133 | 1206 | ||
1134 | if (!hash->buckets) { | 1207 | if (!hash->buckets) { |
1135 | kfree(hash); | 1208 | kfree(hash); |
@@ -3146,8 +3219,10 @@ ftrace_set_regex(struct ftrace_ops *ops, unsigned char *buf, int len, | |||
3146 | mutex_lock(&ftrace_regex_lock); | 3219 | mutex_lock(&ftrace_regex_lock); |
3147 | if (reset) | 3220 | if (reset) |
3148 | ftrace_filter_reset(hash); | 3221 | ftrace_filter_reset(hash); |
3149 | if (buf) | 3222 | if (buf && !ftrace_match_records(hash, buf, len)) { |
3150 | ftrace_match_records(hash, buf, len); | 3223 | ret = -EINVAL; |
3224 | goto out_regex_unlock; | ||
3225 | } | ||
3151 | 3226 | ||
3152 | mutex_lock(&ftrace_lock); | 3227 | mutex_lock(&ftrace_lock); |
3153 | ret = ftrace_hash_move(ops, enable, orig_hash, hash); | 3228 | ret = ftrace_hash_move(ops, enable, orig_hash, hash); |
@@ -3157,6 +3232,7 @@ ftrace_set_regex(struct ftrace_ops *ops, unsigned char *buf, int len, | |||
3157 | 3232 | ||
3158 | mutex_unlock(&ftrace_lock); | 3233 | mutex_unlock(&ftrace_lock); |
3159 | 3234 | ||
3235 | out_regex_unlock: | ||
3160 | mutex_unlock(&ftrace_regex_lock); | 3236 | mutex_unlock(&ftrace_regex_lock); |
3161 | 3237 | ||
3162 | free_ftrace_hash(hash); | 3238 | free_ftrace_hash(hash); |
@@ -3173,10 +3249,10 @@ ftrace_set_regex(struct ftrace_ops *ops, unsigned char *buf, int len, | |||
3173 | * Filters denote which functions should be enabled when tracing is enabled. | 3249 | * Filters denote which functions should be enabled when tracing is enabled. |
3174 | * If @buf is NULL and reset is set, all functions will be enabled for tracing. | 3250 | * If @buf is NULL and reset is set, all functions will be enabled for tracing. |
3175 | */ | 3251 | */ |
3176 | void ftrace_set_filter(struct ftrace_ops *ops, unsigned char *buf, | 3252 | int ftrace_set_filter(struct ftrace_ops *ops, unsigned char *buf, |
3177 | int len, int reset) | 3253 | int len, int reset) |
3178 | { | 3254 | { |
3179 | ftrace_set_regex(ops, buf, len, reset, 1); | 3255 | return ftrace_set_regex(ops, buf, len, reset, 1); |
3180 | } | 3256 | } |
3181 | EXPORT_SYMBOL_GPL(ftrace_set_filter); | 3257 | EXPORT_SYMBOL_GPL(ftrace_set_filter); |
3182 | 3258 | ||
@@ -3191,10 +3267,10 @@ EXPORT_SYMBOL_GPL(ftrace_set_filter); | |||
3191 | * is enabled. If @buf is NULL and reset is set, all functions will be enabled | 3267 | * is enabled. If @buf is NULL and reset is set, all functions will be enabled |
3192 | * for tracing. | 3268 | * for tracing. |
3193 | */ | 3269 | */ |
3194 | void ftrace_set_notrace(struct ftrace_ops *ops, unsigned char *buf, | 3270 | int ftrace_set_notrace(struct ftrace_ops *ops, unsigned char *buf, |
3195 | int len, int reset) | 3271 | int len, int reset) |
3196 | { | 3272 | { |
3197 | ftrace_set_regex(ops, buf, len, reset, 0); | 3273 | return ftrace_set_regex(ops, buf, len, reset, 0); |
3198 | } | 3274 | } |
3199 | EXPORT_SYMBOL_GPL(ftrace_set_notrace); | 3275 | EXPORT_SYMBOL_GPL(ftrace_set_notrace); |
3200 | /** | 3276 | /** |
@@ -3871,6 +3947,36 @@ ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip) | |||
3871 | #endif /* CONFIG_DYNAMIC_FTRACE */ | 3947 | #endif /* CONFIG_DYNAMIC_FTRACE */ |
3872 | 3948 | ||
3873 | static void | 3949 | static void |
3950 | ftrace_ops_control_func(unsigned long ip, unsigned long parent_ip) | ||
3951 | { | ||
3952 | struct ftrace_ops *op; | ||
3953 | |||
3954 | if (unlikely(trace_recursion_test(TRACE_CONTROL_BIT))) | ||
3955 | return; | ||
3956 | |||
3957 | /* | ||
3958 | * Some of the ops may be dynamically allocated, | ||
3959 | * they must be freed after a synchronize_sched(). | ||
3960 | */ | ||
3961 | preempt_disable_notrace(); | ||
3962 | trace_recursion_set(TRACE_CONTROL_BIT); | ||
3963 | op = rcu_dereference_raw(ftrace_control_list); | ||
3964 | while (op != &ftrace_list_end) { | ||
3965 | if (!ftrace_function_local_disabled(op) && | ||
3966 | ftrace_ops_test(op, ip)) | ||
3967 | op->func(ip, parent_ip); | ||
3968 | |||
3969 | op = rcu_dereference_raw(op->next); | ||
3970 | }; | ||
3971 | trace_recursion_clear(TRACE_CONTROL_BIT); | ||
3972 | preempt_enable_notrace(); | ||
3973 | } | ||
3974 | |||
3975 | static struct ftrace_ops control_ops = { | ||
3976 | .func = ftrace_ops_control_func, | ||
3977 | }; | ||
3978 | |||
3979 | static void | ||
3874 | ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip) | 3980 | ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip) |
3875 | { | 3981 | { |
3876 | struct ftrace_ops *op; | 3982 | struct ftrace_ops *op; |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index a3f1bc5d2a00..10d5503f0d04 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
@@ -2764,12 +2764,12 @@ static const char readme_msg[] = | |||
2764 | "tracing mini-HOWTO:\n\n" | 2764 | "tracing mini-HOWTO:\n\n" |
2765 | "# mount -t debugfs nodev /sys/kernel/debug\n\n" | 2765 | "# mount -t debugfs nodev /sys/kernel/debug\n\n" |
2766 | "# cat /sys/kernel/debug/tracing/available_tracers\n" | 2766 | "# cat /sys/kernel/debug/tracing/available_tracers\n" |
2767 | "wakeup preemptirqsoff preemptoff irqsoff function sched_switch nop\n\n" | 2767 | "wakeup wakeup_rt preemptirqsoff preemptoff irqsoff function nop\n\n" |
2768 | "# cat /sys/kernel/debug/tracing/current_tracer\n" | 2768 | "# cat /sys/kernel/debug/tracing/current_tracer\n" |
2769 | "nop\n" | 2769 | "nop\n" |
2770 | "# echo sched_switch > /sys/kernel/debug/tracing/current_tracer\n" | 2770 | "# echo wakeup > /sys/kernel/debug/tracing/current_tracer\n" |
2771 | "# cat /sys/kernel/debug/tracing/current_tracer\n" | 2771 | "# cat /sys/kernel/debug/tracing/current_tracer\n" |
2772 | "sched_switch\n" | 2772 | "wakeup\n" |
2773 | "# cat /sys/kernel/debug/tracing/trace_options\n" | 2773 | "# cat /sys/kernel/debug/tracing/trace_options\n" |
2774 | "noprint-parent nosym-offset nosym-addr noverbose\n" | 2774 | "noprint-parent nosym-offset nosym-addr noverbose\n" |
2775 | "# echo print-parent > /sys/kernel/debug/tracing/trace_options\n" | 2775 | "# echo print-parent > /sys/kernel/debug/tracing/trace_options\n" |
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h index b93ecbadad6d..54faec790bc1 100644 --- a/kernel/trace/trace.h +++ b/kernel/trace/trace.h | |||
@@ -56,17 +56,23 @@ enum trace_type { | |||
56 | #define F_STRUCT(args...) args | 56 | #define F_STRUCT(args...) args |
57 | 57 | ||
58 | #undef FTRACE_ENTRY | 58 | #undef FTRACE_ENTRY |
59 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ | 59 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print, filter) \ |
60 | struct struct_name { \ | 60 | struct struct_name { \ |
61 | struct trace_entry ent; \ | 61 | struct trace_entry ent; \ |
62 | tstruct \ | 62 | tstruct \ |
63 | } | 63 | } |
64 | 64 | ||
65 | #undef TP_ARGS | 65 | #undef TP_ARGS |
66 | #define TP_ARGS(args...) args | 66 | #define TP_ARGS(args...) args |
67 | 67 | ||
68 | #undef FTRACE_ENTRY_DUP | 68 | #undef FTRACE_ENTRY_DUP |
69 | #define FTRACE_ENTRY_DUP(name, name_struct, id, tstruct, printk) | 69 | #define FTRACE_ENTRY_DUP(name, name_struct, id, tstruct, printk, filter) |
70 | |||
71 | #undef FTRACE_ENTRY_REG | ||
72 | #define FTRACE_ENTRY_REG(name, struct_name, id, tstruct, print, \ | ||
73 | filter, regfn) \ | ||
74 | FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print), \ | ||
75 | filter) | ||
70 | 76 | ||
71 | #include "trace_entries.h" | 77 | #include "trace_entries.h" |
72 | 78 | ||
@@ -288,6 +294,8 @@ struct tracer { | |||
288 | /* for function tracing recursion */ | 294 | /* for function tracing recursion */ |
289 | #define TRACE_INTERNAL_BIT (1<<11) | 295 | #define TRACE_INTERNAL_BIT (1<<11) |
290 | #define TRACE_GLOBAL_BIT (1<<12) | 296 | #define TRACE_GLOBAL_BIT (1<<12) |
297 | #define TRACE_CONTROL_BIT (1<<13) | ||
298 | |||
291 | /* | 299 | /* |
292 | * Abuse of the trace_recursion. | 300 | * Abuse of the trace_recursion. |
293 | * As we need a way to maintain state if we are tracing the function | 301 | * As we need a way to maintain state if we are tracing the function |
@@ -589,6 +597,8 @@ static inline int ftrace_trace_task(struct task_struct *task) | |||
589 | static inline int ftrace_is_dead(void) { return 0; } | 597 | static inline int ftrace_is_dead(void) { return 0; } |
590 | #endif | 598 | #endif |
591 | 599 | ||
600 | int ftrace_event_is_function(struct ftrace_event_call *call); | ||
601 | |||
592 | /* | 602 | /* |
593 | * struct trace_parser - servers for reading the user input separated by spaces | 603 | * struct trace_parser - servers for reading the user input separated by spaces |
594 | * @cont: set if the input is not complete - no final space char was found | 604 | * @cont: set if the input is not complete - no final space char was found |
@@ -766,9 +776,7 @@ struct filter_pred { | |||
766 | u64 val; | 776 | u64 val; |
767 | struct regex regex; | 777 | struct regex regex; |
768 | unsigned short *ops; | 778 | unsigned short *ops; |
769 | #ifdef CONFIG_FTRACE_STARTUP_TEST | ||
770 | struct ftrace_event_field *field; | 779 | struct ftrace_event_field *field; |
771 | #endif | ||
772 | int offset; | 780 | int offset; |
773 | int not; | 781 | int not; |
774 | int op; | 782 | int op; |
@@ -818,12 +826,22 @@ extern const char *__start___trace_bprintk_fmt[]; | |||
818 | extern const char *__stop___trace_bprintk_fmt[]; | 826 | extern const char *__stop___trace_bprintk_fmt[]; |
819 | 827 | ||
820 | #undef FTRACE_ENTRY | 828 | #undef FTRACE_ENTRY |
821 | #define FTRACE_ENTRY(call, struct_name, id, tstruct, print) \ | 829 | #define FTRACE_ENTRY(call, struct_name, id, tstruct, print, filter) \ |
822 | extern struct ftrace_event_call \ | 830 | extern struct ftrace_event_call \ |
823 | __attribute__((__aligned__(4))) event_##call; | 831 | __attribute__((__aligned__(4))) event_##call; |
824 | #undef FTRACE_ENTRY_DUP | 832 | #undef FTRACE_ENTRY_DUP |
825 | #define FTRACE_ENTRY_DUP(call, struct_name, id, tstruct, print) \ | 833 | #define FTRACE_ENTRY_DUP(call, struct_name, id, tstruct, print, filter) \ |
826 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print)) | 834 | FTRACE_ENTRY(call, struct_name, id, PARAMS(tstruct), PARAMS(print), \ |
835 | filter) | ||
827 | #include "trace_entries.h" | 836 | #include "trace_entries.h" |
828 | 837 | ||
838 | #ifdef CONFIG_PERF_EVENTS | ||
839 | #ifdef CONFIG_FUNCTION_TRACER | ||
840 | int perf_ftrace_event_register(struct ftrace_event_call *call, | ||
841 | enum trace_reg type, void *data); | ||
842 | #else | ||
843 | #define perf_ftrace_event_register NULL | ||
844 | #endif /* CONFIG_FUNCTION_TRACER */ | ||
845 | #endif /* CONFIG_PERF_EVENTS */ | ||
846 | |||
829 | #endif /* _LINUX_KERNEL_TRACE_H */ | 847 | #endif /* _LINUX_KERNEL_TRACE_H */ |
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h index 93365907f219..d91eb0541b3a 100644 --- a/kernel/trace/trace_entries.h +++ b/kernel/trace/trace_entries.h | |||
@@ -55,7 +55,7 @@ | |||
55 | /* | 55 | /* |
56 | * Function trace entry - function address and parent function address: | 56 | * Function trace entry - function address and parent function address: |
57 | */ | 57 | */ |
58 | FTRACE_ENTRY(function, ftrace_entry, | 58 | FTRACE_ENTRY_REG(function, ftrace_entry, |
59 | 59 | ||
60 | TRACE_FN, | 60 | TRACE_FN, |
61 | 61 | ||
@@ -64,7 +64,11 @@ FTRACE_ENTRY(function, ftrace_entry, | |||
64 | __field( unsigned long, parent_ip ) | 64 | __field( unsigned long, parent_ip ) |
65 | ), | 65 | ), |
66 | 66 | ||
67 | F_printk(" %lx <-- %lx", __entry->ip, __entry->parent_ip) | 67 | F_printk(" %lx <-- %lx", __entry->ip, __entry->parent_ip), |
68 | |||
69 | FILTER_TRACE_FN, | ||
70 | |||
71 | perf_ftrace_event_register | ||
68 | ); | 72 | ); |
69 | 73 | ||
70 | /* Function call entry */ | 74 | /* Function call entry */ |
@@ -78,7 +82,9 @@ FTRACE_ENTRY(funcgraph_entry, ftrace_graph_ent_entry, | |||
78 | __field_desc( int, graph_ent, depth ) | 82 | __field_desc( int, graph_ent, depth ) |
79 | ), | 83 | ), |
80 | 84 | ||
81 | F_printk("--> %lx (%d)", __entry->func, __entry->depth) | 85 | F_printk("--> %lx (%d)", __entry->func, __entry->depth), |
86 | |||
87 | FILTER_OTHER | ||
82 | ); | 88 | ); |
83 | 89 | ||
84 | /* Function return entry */ | 90 | /* Function return entry */ |
@@ -98,7 +104,9 @@ FTRACE_ENTRY(funcgraph_exit, ftrace_graph_ret_entry, | |||
98 | F_printk("<-- %lx (%d) (start: %llx end: %llx) over: %d", | 104 | F_printk("<-- %lx (%d) (start: %llx end: %llx) over: %d", |
99 | __entry->func, __entry->depth, | 105 | __entry->func, __entry->depth, |
100 | __entry->calltime, __entry->rettime, | 106 | __entry->calltime, __entry->rettime, |
101 | __entry->depth) | 107 | __entry->depth), |
108 | |||
109 | FILTER_OTHER | ||
102 | ); | 110 | ); |
103 | 111 | ||
104 | /* | 112 | /* |
@@ -127,8 +135,9 @@ FTRACE_ENTRY(context_switch, ctx_switch_entry, | |||
127 | F_printk("%u:%u:%u ==> %u:%u:%u [%03u]", | 135 | F_printk("%u:%u:%u ==> %u:%u:%u [%03u]", |
128 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, | 136 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, |
129 | __entry->next_pid, __entry->next_prio, __entry->next_state, | 137 | __entry->next_pid, __entry->next_prio, __entry->next_state, |
130 | __entry->next_cpu | 138 | __entry->next_cpu), |
131 | ) | 139 | |
140 | FILTER_OTHER | ||
132 | ); | 141 | ); |
133 | 142 | ||
134 | /* | 143 | /* |
@@ -146,8 +155,9 @@ FTRACE_ENTRY_DUP(wakeup, ctx_switch_entry, | |||
146 | F_printk("%u:%u:%u ==+ %u:%u:%u [%03u]", | 155 | F_printk("%u:%u:%u ==+ %u:%u:%u [%03u]", |
147 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, | 156 | __entry->prev_pid, __entry->prev_prio, __entry->prev_state, |
148 | __entry->next_pid, __entry->next_prio, __entry->next_state, | 157 | __entry->next_pid, __entry->next_prio, __entry->next_state, |
149 | __entry->next_cpu | 158 | __entry->next_cpu), |
150 | ) | 159 | |
160 | FILTER_OTHER | ||
151 | ); | 161 | ); |
152 | 162 | ||
153 | /* | 163 | /* |
@@ -169,7 +179,9 @@ FTRACE_ENTRY(kernel_stack, stack_entry, | |||
169 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", | 179 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", |
170 | __entry->caller[0], __entry->caller[1], __entry->caller[2], | 180 | __entry->caller[0], __entry->caller[1], __entry->caller[2], |
171 | __entry->caller[3], __entry->caller[4], __entry->caller[5], | 181 | __entry->caller[3], __entry->caller[4], __entry->caller[5], |
172 | __entry->caller[6], __entry->caller[7]) | 182 | __entry->caller[6], __entry->caller[7]), |
183 | |||
184 | FILTER_OTHER | ||
173 | ); | 185 | ); |
174 | 186 | ||
175 | FTRACE_ENTRY(user_stack, userstack_entry, | 187 | FTRACE_ENTRY(user_stack, userstack_entry, |
@@ -185,7 +197,9 @@ FTRACE_ENTRY(user_stack, userstack_entry, | |||
185 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", | 197 | "\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n\t=> (%08lx)\n", |
186 | __entry->caller[0], __entry->caller[1], __entry->caller[2], | 198 | __entry->caller[0], __entry->caller[1], __entry->caller[2], |
187 | __entry->caller[3], __entry->caller[4], __entry->caller[5], | 199 | __entry->caller[3], __entry->caller[4], __entry->caller[5], |
188 | __entry->caller[6], __entry->caller[7]) | 200 | __entry->caller[6], __entry->caller[7]), |
201 | |||
202 | FILTER_OTHER | ||
189 | ); | 203 | ); |
190 | 204 | ||
191 | /* | 205 | /* |
@@ -202,7 +216,9 @@ FTRACE_ENTRY(bprint, bprint_entry, | |||
202 | ), | 216 | ), |
203 | 217 | ||
204 | F_printk("%08lx fmt:%p", | 218 | F_printk("%08lx fmt:%p", |
205 | __entry->ip, __entry->fmt) | 219 | __entry->ip, __entry->fmt), |
220 | |||
221 | FILTER_OTHER | ||
206 | ); | 222 | ); |
207 | 223 | ||
208 | FTRACE_ENTRY(print, print_entry, | 224 | FTRACE_ENTRY(print, print_entry, |
@@ -215,7 +231,9 @@ FTRACE_ENTRY(print, print_entry, | |||
215 | ), | 231 | ), |
216 | 232 | ||
217 | F_printk("%08lx %s", | 233 | F_printk("%08lx %s", |
218 | __entry->ip, __entry->buf) | 234 | __entry->ip, __entry->buf), |
235 | |||
236 | FILTER_OTHER | ||
219 | ); | 237 | ); |
220 | 238 | ||
221 | FTRACE_ENTRY(mmiotrace_rw, trace_mmiotrace_rw, | 239 | FTRACE_ENTRY(mmiotrace_rw, trace_mmiotrace_rw, |
@@ -234,7 +252,9 @@ FTRACE_ENTRY(mmiotrace_rw, trace_mmiotrace_rw, | |||
234 | 252 | ||
235 | F_printk("%lx %lx %lx %d %x %x", | 253 | F_printk("%lx %lx %lx %d %x %x", |
236 | (unsigned long)__entry->phys, __entry->value, __entry->pc, | 254 | (unsigned long)__entry->phys, __entry->value, __entry->pc, |
237 | __entry->map_id, __entry->opcode, __entry->width) | 255 | __entry->map_id, __entry->opcode, __entry->width), |
256 | |||
257 | FILTER_OTHER | ||
238 | ); | 258 | ); |
239 | 259 | ||
240 | FTRACE_ENTRY(mmiotrace_map, trace_mmiotrace_map, | 260 | FTRACE_ENTRY(mmiotrace_map, trace_mmiotrace_map, |
@@ -252,7 +272,9 @@ FTRACE_ENTRY(mmiotrace_map, trace_mmiotrace_map, | |||
252 | 272 | ||
253 | F_printk("%lx %lx %lx %d %x", | 273 | F_printk("%lx %lx %lx %d %x", |
254 | (unsigned long)__entry->phys, __entry->virt, __entry->len, | 274 | (unsigned long)__entry->phys, __entry->virt, __entry->len, |
255 | __entry->map_id, __entry->opcode) | 275 | __entry->map_id, __entry->opcode), |
276 | |||
277 | FILTER_OTHER | ||
256 | ); | 278 | ); |
257 | 279 | ||
258 | 280 | ||
@@ -272,6 +294,8 @@ FTRACE_ENTRY(branch, trace_branch, | |||
272 | 294 | ||
273 | F_printk("%u:%s:%s (%u)", | 295 | F_printk("%u:%s:%s (%u)", |
274 | __entry->line, | 296 | __entry->line, |
275 | __entry->func, __entry->file, __entry->correct) | 297 | __entry->func, __entry->file, __entry->correct), |
298 | |||
299 | FILTER_OTHER | ||
276 | ); | 300 | ); |
277 | 301 | ||
diff --git a/kernel/trace/trace_event_perf.c b/kernel/trace/trace_event_perf.c index 19a359d5e6d5..fee3752ae8f6 100644 --- a/kernel/trace/trace_event_perf.c +++ b/kernel/trace/trace_event_perf.c | |||
@@ -24,6 +24,11 @@ static int total_ref_count; | |||
24 | static int perf_trace_event_perm(struct ftrace_event_call *tp_event, | 24 | static int perf_trace_event_perm(struct ftrace_event_call *tp_event, |
25 | struct perf_event *p_event) | 25 | struct perf_event *p_event) |
26 | { | 26 | { |
27 | /* The ftrace function trace is allowed only for root. */ | ||
28 | if (ftrace_event_is_function(tp_event) && | ||
29 | perf_paranoid_kernel() && !capable(CAP_SYS_ADMIN)) | ||
30 | return -EPERM; | ||
31 | |||
27 | /* No tracing, just counting, so no obvious leak */ | 32 | /* No tracing, just counting, so no obvious leak */ |
28 | if (!(p_event->attr.sample_type & PERF_SAMPLE_RAW)) | 33 | if (!(p_event->attr.sample_type & PERF_SAMPLE_RAW)) |
29 | return 0; | 34 | return 0; |
@@ -44,23 +49,17 @@ static int perf_trace_event_perm(struct ftrace_event_call *tp_event, | |||
44 | return 0; | 49 | return 0; |
45 | } | 50 | } |
46 | 51 | ||
47 | static int perf_trace_event_init(struct ftrace_event_call *tp_event, | 52 | static int perf_trace_event_reg(struct ftrace_event_call *tp_event, |
48 | struct perf_event *p_event) | 53 | struct perf_event *p_event) |
49 | { | 54 | { |
50 | struct hlist_head __percpu *list; | 55 | struct hlist_head __percpu *list; |
51 | int ret; | 56 | int ret = -ENOMEM; |
52 | int cpu; | 57 | int cpu; |
53 | 58 | ||
54 | ret = perf_trace_event_perm(tp_event, p_event); | ||
55 | if (ret) | ||
56 | return ret; | ||
57 | |||
58 | p_event->tp_event = tp_event; | 59 | p_event->tp_event = tp_event; |
59 | if (tp_event->perf_refcount++ > 0) | 60 | if (tp_event->perf_refcount++ > 0) |
60 | return 0; | 61 | return 0; |
61 | 62 | ||
62 | ret = -ENOMEM; | ||
63 | |||
64 | list = alloc_percpu(struct hlist_head); | 63 | list = alloc_percpu(struct hlist_head); |
65 | if (!list) | 64 | if (!list) |
66 | goto fail; | 65 | goto fail; |
@@ -83,7 +82,7 @@ static int perf_trace_event_init(struct ftrace_event_call *tp_event, | |||
83 | } | 82 | } |
84 | } | 83 | } |
85 | 84 | ||
86 | ret = tp_event->class->reg(tp_event, TRACE_REG_PERF_REGISTER); | 85 | ret = tp_event->class->reg(tp_event, TRACE_REG_PERF_REGISTER, NULL); |
87 | if (ret) | 86 | if (ret) |
88 | goto fail; | 87 | goto fail; |
89 | 88 | ||
@@ -108,6 +107,69 @@ fail: | |||
108 | return ret; | 107 | return ret; |
109 | } | 108 | } |
110 | 109 | ||
110 | static void perf_trace_event_unreg(struct perf_event *p_event) | ||
111 | { | ||
112 | struct ftrace_event_call *tp_event = p_event->tp_event; | ||
113 | int i; | ||
114 | |||
115 | if (--tp_event->perf_refcount > 0) | ||
116 | goto out; | ||
117 | |||
118 | tp_event->class->reg(tp_event, TRACE_REG_PERF_UNREGISTER, NULL); | ||
119 | |||
120 | /* | ||
121 | * Ensure our callback won't be called anymore. The buffers | ||
122 | * will be freed after that. | ||
123 | */ | ||
124 | tracepoint_synchronize_unregister(); | ||
125 | |||
126 | free_percpu(tp_event->perf_events); | ||
127 | tp_event->perf_events = NULL; | ||
128 | |||
129 | if (!--total_ref_count) { | ||
130 | for (i = 0; i < PERF_NR_CONTEXTS; i++) { | ||
131 | free_percpu(perf_trace_buf[i]); | ||
132 | perf_trace_buf[i] = NULL; | ||
133 | } | ||
134 | } | ||
135 | out: | ||
136 | module_put(tp_event->mod); | ||
137 | } | ||
138 | |||
139 | static int perf_trace_event_open(struct perf_event *p_event) | ||
140 | { | ||
141 | struct ftrace_event_call *tp_event = p_event->tp_event; | ||
142 | return tp_event->class->reg(tp_event, TRACE_REG_PERF_OPEN, p_event); | ||
143 | } | ||
144 | |||
145 | static void perf_trace_event_close(struct perf_event *p_event) | ||
146 | { | ||
147 | struct ftrace_event_call *tp_event = p_event->tp_event; | ||
148 | tp_event->class->reg(tp_event, TRACE_REG_PERF_CLOSE, p_event); | ||
149 | } | ||
150 | |||
151 | static int perf_trace_event_init(struct ftrace_event_call *tp_event, | ||
152 | struct perf_event *p_event) | ||
153 | { | ||
154 | int ret; | ||
155 | |||
156 | ret = perf_trace_event_perm(tp_event, p_event); | ||
157 | if (ret) | ||
158 | return ret; | ||
159 | |||
160 | ret = perf_trace_event_reg(tp_event, p_event); | ||
161 | if (ret) | ||
162 | return ret; | ||
163 | |||
164 | ret = perf_trace_event_open(p_event); | ||
165 | if (ret) { | ||
166 | perf_trace_event_unreg(p_event); | ||
167 | return ret; | ||
168 | } | ||
169 | |||
170 | return 0; | ||
171 | } | ||
172 | |||
111 | int perf_trace_init(struct perf_event *p_event) | 173 | int perf_trace_init(struct perf_event *p_event) |
112 | { | 174 | { |
113 | struct ftrace_event_call *tp_event; | 175 | struct ftrace_event_call *tp_event; |
@@ -130,6 +192,14 @@ int perf_trace_init(struct perf_event *p_event) | |||
130 | return ret; | 192 | return ret; |
131 | } | 193 | } |
132 | 194 | ||
195 | void perf_trace_destroy(struct perf_event *p_event) | ||
196 | { | ||
197 | mutex_lock(&event_mutex); | ||
198 | perf_trace_event_close(p_event); | ||
199 | perf_trace_event_unreg(p_event); | ||
200 | mutex_unlock(&event_mutex); | ||
201 | } | ||
202 | |||
133 | int perf_trace_add(struct perf_event *p_event, int flags) | 203 | int perf_trace_add(struct perf_event *p_event, int flags) |
134 | { | 204 | { |
135 | struct ftrace_event_call *tp_event = p_event->tp_event; | 205 | struct ftrace_event_call *tp_event = p_event->tp_event; |
@@ -146,43 +216,14 @@ int perf_trace_add(struct perf_event *p_event, int flags) | |||
146 | list = this_cpu_ptr(pcpu_list); | 216 | list = this_cpu_ptr(pcpu_list); |
147 | hlist_add_head_rcu(&p_event->hlist_entry, list); | 217 | hlist_add_head_rcu(&p_event->hlist_entry, list); |
148 | 218 | ||
149 | return 0; | 219 | return tp_event->class->reg(tp_event, TRACE_REG_PERF_ADD, p_event); |
150 | } | 220 | } |
151 | 221 | ||
152 | void perf_trace_del(struct perf_event *p_event, int flags) | 222 | void perf_trace_del(struct perf_event *p_event, int flags) |
153 | { | 223 | { |
154 | hlist_del_rcu(&p_event->hlist_entry); | ||
155 | } | ||
156 | |||
157 | void perf_trace_destroy(struct perf_event *p_event) | ||
158 | { | ||
159 | struct ftrace_event_call *tp_event = p_event->tp_event; | 224 | struct ftrace_event_call *tp_event = p_event->tp_event; |
160 | int i; | 225 | hlist_del_rcu(&p_event->hlist_entry); |
161 | 226 | tp_event->class->reg(tp_event, TRACE_REG_PERF_DEL, p_event); | |
162 | mutex_lock(&event_mutex); | ||
163 | if (--tp_event->perf_refcount > 0) | ||
164 | goto out; | ||
165 | |||
166 | tp_event->class->reg(tp_event, TRACE_REG_PERF_UNREGISTER); | ||
167 | |||
168 | /* | ||
169 | * Ensure our callback won't be called anymore. The buffers | ||
170 | * will be freed after that. | ||
171 | */ | ||
172 | tracepoint_synchronize_unregister(); | ||
173 | |||
174 | free_percpu(tp_event->perf_events); | ||
175 | tp_event->perf_events = NULL; | ||
176 | |||
177 | if (!--total_ref_count) { | ||
178 | for (i = 0; i < PERF_NR_CONTEXTS; i++) { | ||
179 | free_percpu(perf_trace_buf[i]); | ||
180 | perf_trace_buf[i] = NULL; | ||
181 | } | ||
182 | } | ||
183 | out: | ||
184 | module_put(tp_event->mod); | ||
185 | mutex_unlock(&event_mutex); | ||
186 | } | 227 | } |
187 | 228 | ||
188 | __kprobes void *perf_trace_buf_prepare(int size, unsigned short type, | 229 | __kprobes void *perf_trace_buf_prepare(int size, unsigned short type, |
@@ -214,3 +255,86 @@ __kprobes void *perf_trace_buf_prepare(int size, unsigned short type, | |||
214 | return raw_data; | 255 | return raw_data; |
215 | } | 256 | } |
216 | EXPORT_SYMBOL_GPL(perf_trace_buf_prepare); | 257 | EXPORT_SYMBOL_GPL(perf_trace_buf_prepare); |
258 | |||
259 | #ifdef CONFIG_FUNCTION_TRACER | ||
260 | static void | ||
261 | perf_ftrace_function_call(unsigned long ip, unsigned long parent_ip) | ||
262 | { | ||
263 | struct ftrace_entry *entry; | ||
264 | struct hlist_head *head; | ||
265 | struct pt_regs regs; | ||
266 | int rctx; | ||
267 | |||
268 | #define ENTRY_SIZE (ALIGN(sizeof(struct ftrace_entry) + sizeof(u32), \ | ||
269 | sizeof(u64)) - sizeof(u32)) | ||
270 | |||
271 | BUILD_BUG_ON(ENTRY_SIZE > PERF_MAX_TRACE_SIZE); | ||
272 | |||
273 | perf_fetch_caller_regs(®s); | ||
274 | |||
275 | entry = perf_trace_buf_prepare(ENTRY_SIZE, TRACE_FN, NULL, &rctx); | ||
276 | if (!entry) | ||
277 | return; | ||
278 | |||
279 | entry->ip = ip; | ||
280 | entry->parent_ip = parent_ip; | ||
281 | |||
282 | head = this_cpu_ptr(event_function.perf_events); | ||
283 | perf_trace_buf_submit(entry, ENTRY_SIZE, rctx, 0, | ||
284 | 1, ®s, head); | ||
285 | |||
286 | #undef ENTRY_SIZE | ||
287 | } | ||
288 | |||
289 | static int perf_ftrace_function_register(struct perf_event *event) | ||
290 | { | ||
291 | struct ftrace_ops *ops = &event->ftrace_ops; | ||
292 | |||
293 | ops->flags |= FTRACE_OPS_FL_CONTROL; | ||
294 | ops->func = perf_ftrace_function_call; | ||
295 | return register_ftrace_function(ops); | ||
296 | } | ||
297 | |||
298 | static int perf_ftrace_function_unregister(struct perf_event *event) | ||
299 | { | ||
300 | struct ftrace_ops *ops = &event->ftrace_ops; | ||
301 | int ret = unregister_ftrace_function(ops); | ||
302 | ftrace_free_filter(ops); | ||
303 | return ret; | ||
304 | } | ||
305 | |||
306 | static void perf_ftrace_function_enable(struct perf_event *event) | ||
307 | { | ||
308 | ftrace_function_local_enable(&event->ftrace_ops); | ||
309 | } | ||
310 | |||
311 | static void perf_ftrace_function_disable(struct perf_event *event) | ||
312 | { | ||
313 | ftrace_function_local_disable(&event->ftrace_ops); | ||
314 | } | ||
315 | |||
316 | int perf_ftrace_event_register(struct ftrace_event_call *call, | ||
317 | enum trace_reg type, void *data) | ||
318 | { | ||
319 | switch (type) { | ||
320 | case TRACE_REG_REGISTER: | ||
321 | case TRACE_REG_UNREGISTER: | ||
322 | break; | ||
323 | case TRACE_REG_PERF_REGISTER: | ||
324 | case TRACE_REG_PERF_UNREGISTER: | ||
325 | return 0; | ||
326 | case TRACE_REG_PERF_OPEN: | ||
327 | return perf_ftrace_function_register(data); | ||
328 | case TRACE_REG_PERF_CLOSE: | ||
329 | return perf_ftrace_function_unregister(data); | ||
330 | case TRACE_REG_PERF_ADD: | ||
331 | perf_ftrace_function_enable(data); | ||
332 | return 0; | ||
333 | case TRACE_REG_PERF_DEL: | ||
334 | perf_ftrace_function_disable(data); | ||
335 | return 0; | ||
336 | } | ||
337 | |||
338 | return -EINVAL; | ||
339 | } | ||
340 | #endif /* CONFIG_FUNCTION_TRACER */ | ||
diff --git a/kernel/trace/trace_events.c b/kernel/trace/trace_events.c index c212a7f934ec..079a93ae8a9d 100644 --- a/kernel/trace/trace_events.c +++ b/kernel/trace/trace_events.c | |||
@@ -147,7 +147,8 @@ int trace_event_raw_init(struct ftrace_event_call *call) | |||
147 | } | 147 | } |
148 | EXPORT_SYMBOL_GPL(trace_event_raw_init); | 148 | EXPORT_SYMBOL_GPL(trace_event_raw_init); |
149 | 149 | ||
150 | int ftrace_event_reg(struct ftrace_event_call *call, enum trace_reg type) | 150 | int ftrace_event_reg(struct ftrace_event_call *call, |
151 | enum trace_reg type, void *data) | ||
151 | { | 152 | { |
152 | switch (type) { | 153 | switch (type) { |
153 | case TRACE_REG_REGISTER: | 154 | case TRACE_REG_REGISTER: |
@@ -170,6 +171,11 @@ int ftrace_event_reg(struct ftrace_event_call *call, enum trace_reg type) | |||
170 | call->class->perf_probe, | 171 | call->class->perf_probe, |
171 | call); | 172 | call); |
172 | return 0; | 173 | return 0; |
174 | case TRACE_REG_PERF_OPEN: | ||
175 | case TRACE_REG_PERF_CLOSE: | ||
176 | case TRACE_REG_PERF_ADD: | ||
177 | case TRACE_REG_PERF_DEL: | ||
178 | return 0; | ||
173 | #endif | 179 | #endif |
174 | } | 180 | } |
175 | return 0; | 181 | return 0; |
@@ -209,7 +215,7 @@ static int ftrace_event_enable_disable(struct ftrace_event_call *call, | |||
209 | tracing_stop_cmdline_record(); | 215 | tracing_stop_cmdline_record(); |
210 | call->flags &= ~TRACE_EVENT_FL_RECORDED_CMD; | 216 | call->flags &= ~TRACE_EVENT_FL_RECORDED_CMD; |
211 | } | 217 | } |
212 | call->class->reg(call, TRACE_REG_UNREGISTER); | 218 | call->class->reg(call, TRACE_REG_UNREGISTER, NULL); |
213 | } | 219 | } |
214 | break; | 220 | break; |
215 | case 1: | 221 | case 1: |
@@ -218,7 +224,7 @@ static int ftrace_event_enable_disable(struct ftrace_event_call *call, | |||
218 | tracing_start_cmdline_record(); | 224 | tracing_start_cmdline_record(); |
219 | call->flags |= TRACE_EVENT_FL_RECORDED_CMD; | 225 | call->flags |= TRACE_EVENT_FL_RECORDED_CMD; |
220 | } | 226 | } |
221 | ret = call->class->reg(call, TRACE_REG_REGISTER); | 227 | ret = call->class->reg(call, TRACE_REG_REGISTER, NULL); |
222 | if (ret) { | 228 | if (ret) { |
223 | tracing_stop_cmdline_record(); | 229 | tracing_stop_cmdline_record(); |
224 | pr_info("event trace: Could not enable event " | 230 | pr_info("event trace: Could not enable event " |
diff --git a/kernel/trace/trace_events_filter.c b/kernel/trace/trace_events_filter.c index 24aee7127451..431dba8b7542 100644 --- a/kernel/trace/trace_events_filter.c +++ b/kernel/trace/trace_events_filter.c | |||
@@ -81,6 +81,7 @@ enum { | |||
81 | FILT_ERR_TOO_MANY_PREDS, | 81 | FILT_ERR_TOO_MANY_PREDS, |
82 | FILT_ERR_MISSING_FIELD, | 82 | FILT_ERR_MISSING_FIELD, |
83 | FILT_ERR_INVALID_FILTER, | 83 | FILT_ERR_INVALID_FILTER, |
84 | FILT_ERR_IP_FIELD_ONLY, | ||
84 | }; | 85 | }; |
85 | 86 | ||
86 | static char *err_text[] = { | 87 | static char *err_text[] = { |
@@ -96,6 +97,7 @@ static char *err_text[] = { | |||
96 | "Too many terms in predicate expression", | 97 | "Too many terms in predicate expression", |
97 | "Missing field name and/or value", | 98 | "Missing field name and/or value", |
98 | "Meaningless filter expression", | 99 | "Meaningless filter expression", |
100 | "Only 'ip' field is supported for function trace", | ||
99 | }; | 101 | }; |
100 | 102 | ||
101 | struct opstack_op { | 103 | struct opstack_op { |
@@ -685,7 +687,7 @@ find_event_field(struct ftrace_event_call *call, char *name) | |||
685 | 687 | ||
686 | static int __alloc_pred_stack(struct pred_stack *stack, int n_preds) | 688 | static int __alloc_pred_stack(struct pred_stack *stack, int n_preds) |
687 | { | 689 | { |
688 | stack->preds = kzalloc(sizeof(*stack->preds)*(n_preds + 1), GFP_KERNEL); | 690 | stack->preds = kcalloc(n_preds + 1, sizeof(*stack->preds), GFP_KERNEL); |
689 | if (!stack->preds) | 691 | if (!stack->preds) |
690 | return -ENOMEM; | 692 | return -ENOMEM; |
691 | stack->index = n_preds; | 693 | stack->index = n_preds; |
@@ -826,8 +828,7 @@ static int __alloc_preds(struct event_filter *filter, int n_preds) | |||
826 | if (filter->preds) | 828 | if (filter->preds) |
827 | __free_preds(filter); | 829 | __free_preds(filter); |
828 | 830 | ||
829 | filter->preds = | 831 | filter->preds = kcalloc(n_preds, sizeof(*filter->preds), GFP_KERNEL); |
830 | kzalloc(sizeof(*filter->preds) * n_preds, GFP_KERNEL); | ||
831 | 832 | ||
832 | if (!filter->preds) | 833 | if (!filter->preds) |
833 | return -ENOMEM; | 834 | return -ENOMEM; |
@@ -900,6 +901,11 @@ int filter_assign_type(const char *type) | |||
900 | return FILTER_OTHER; | 901 | return FILTER_OTHER; |
901 | } | 902 | } |
902 | 903 | ||
904 | static bool is_function_field(struct ftrace_event_field *field) | ||
905 | { | ||
906 | return field->filter_type == FILTER_TRACE_FN; | ||
907 | } | ||
908 | |||
903 | static bool is_string_field(struct ftrace_event_field *field) | 909 | static bool is_string_field(struct ftrace_event_field *field) |
904 | { | 910 | { |
905 | return field->filter_type == FILTER_DYN_STRING || | 911 | return field->filter_type == FILTER_DYN_STRING || |
@@ -987,6 +993,11 @@ static int init_pred(struct filter_parse_state *ps, | |||
987 | fn = filter_pred_strloc; | 993 | fn = filter_pred_strloc; |
988 | else | 994 | else |
989 | fn = filter_pred_pchar; | 995 | fn = filter_pred_pchar; |
996 | } else if (is_function_field(field)) { | ||
997 | if (strcmp(field->name, "ip")) { | ||
998 | parse_error(ps, FILT_ERR_IP_FIELD_ONLY, 0); | ||
999 | return -EINVAL; | ||
1000 | } | ||
990 | } else { | 1001 | } else { |
991 | if (field->is_signed) | 1002 | if (field->is_signed) |
992 | ret = strict_strtoll(pred->regex.pattern, 0, &val); | 1003 | ret = strict_strtoll(pred->regex.pattern, 0, &val); |
@@ -1334,10 +1345,7 @@ static struct filter_pred *create_pred(struct filter_parse_state *ps, | |||
1334 | 1345 | ||
1335 | strcpy(pred.regex.pattern, operand2); | 1346 | strcpy(pred.regex.pattern, operand2); |
1336 | pred.regex.len = strlen(pred.regex.pattern); | 1347 | pred.regex.len = strlen(pred.regex.pattern); |
1337 | |||
1338 | #ifdef CONFIG_FTRACE_STARTUP_TEST | ||
1339 | pred.field = field; | 1348 | pred.field = field; |
1340 | #endif | ||
1341 | return init_pred(ps, field, &pred) ? NULL : &pred; | 1349 | return init_pred(ps, field, &pred) ? NULL : &pred; |
1342 | } | 1350 | } |
1343 | 1351 | ||
@@ -1486,7 +1494,7 @@ static int fold_pred(struct filter_pred *preds, struct filter_pred *root) | |||
1486 | children = count_leafs(preds, &preds[root->left]); | 1494 | children = count_leafs(preds, &preds[root->left]); |
1487 | children += count_leafs(preds, &preds[root->right]); | 1495 | children += count_leafs(preds, &preds[root->right]); |
1488 | 1496 | ||
1489 | root->ops = kzalloc(sizeof(*root->ops) * children, GFP_KERNEL); | 1497 | root->ops = kcalloc(children, sizeof(*root->ops), GFP_KERNEL); |
1490 | if (!root->ops) | 1498 | if (!root->ops) |
1491 | return -ENOMEM; | 1499 | return -ENOMEM; |
1492 | 1500 | ||
@@ -1950,6 +1958,148 @@ void ftrace_profile_free_filter(struct perf_event *event) | |||
1950 | __free_filter(filter); | 1958 | __free_filter(filter); |
1951 | } | 1959 | } |
1952 | 1960 | ||
1961 | struct function_filter_data { | ||
1962 | struct ftrace_ops *ops; | ||
1963 | int first_filter; | ||
1964 | int first_notrace; | ||
1965 | }; | ||
1966 | |||
1967 | #ifdef CONFIG_FUNCTION_TRACER | ||
1968 | static char ** | ||
1969 | ftrace_function_filter_re(char *buf, int len, int *count) | ||
1970 | { | ||
1971 | char *str, *sep, **re; | ||
1972 | |||
1973 | str = kstrndup(buf, len, GFP_KERNEL); | ||
1974 | if (!str) | ||
1975 | return NULL; | ||
1976 | |||
1977 | /* | ||
1978 | * The argv_split function takes white space | ||
1979 | * as a separator, so convert ',' into spaces. | ||
1980 | */ | ||
1981 | while ((sep = strchr(str, ','))) | ||
1982 | *sep = ' '; | ||
1983 | |||
1984 | re = argv_split(GFP_KERNEL, str, count); | ||
1985 | kfree(str); | ||
1986 | return re; | ||
1987 | } | ||
1988 | |||
1989 | static int ftrace_function_set_regexp(struct ftrace_ops *ops, int filter, | ||
1990 | int reset, char *re, int len) | ||
1991 | { | ||
1992 | int ret; | ||
1993 | |||
1994 | if (filter) | ||
1995 | ret = ftrace_set_filter(ops, re, len, reset); | ||
1996 | else | ||
1997 | ret = ftrace_set_notrace(ops, re, len, reset); | ||
1998 | |||
1999 | return ret; | ||
2000 | } | ||
2001 | |||
2002 | static int __ftrace_function_set_filter(int filter, char *buf, int len, | ||
2003 | struct function_filter_data *data) | ||
2004 | { | ||
2005 | int i, re_cnt, ret; | ||
2006 | int *reset; | ||
2007 | char **re; | ||
2008 | |||
2009 | reset = filter ? &data->first_filter : &data->first_notrace; | ||
2010 | |||
2011 | /* | ||
2012 | * The 'ip' field could have multiple filters set, separated | ||
2013 | * either by space or comma. We first cut the filter and apply | ||
2014 | * all pieces separatelly. | ||
2015 | */ | ||
2016 | re = ftrace_function_filter_re(buf, len, &re_cnt); | ||
2017 | if (!re) | ||
2018 | return -EINVAL; | ||
2019 | |||
2020 | for (i = 0; i < re_cnt; i++) { | ||
2021 | ret = ftrace_function_set_regexp(data->ops, filter, *reset, | ||
2022 | re[i], strlen(re[i])); | ||
2023 | if (ret) | ||
2024 | break; | ||
2025 | |||
2026 | if (*reset) | ||
2027 | *reset = 0; | ||
2028 | } | ||
2029 | |||
2030 | argv_free(re); | ||
2031 | return ret; | ||
2032 | } | ||
2033 | |||
2034 | static int ftrace_function_check_pred(struct filter_pred *pred, int leaf) | ||
2035 | { | ||
2036 | struct ftrace_event_field *field = pred->field; | ||
2037 | |||
2038 | if (leaf) { | ||
2039 | /* | ||
2040 | * Check the leaf predicate for function trace, verify: | ||
2041 | * - only '==' and '!=' is used | ||
2042 | * - the 'ip' field is used | ||
2043 | */ | ||
2044 | if ((pred->op != OP_EQ) && (pred->op != OP_NE)) | ||
2045 | return -EINVAL; | ||
2046 | |||
2047 | if (strcmp(field->name, "ip")) | ||
2048 | return -EINVAL; | ||
2049 | } else { | ||
2050 | /* | ||
2051 | * Check the non leaf predicate for function trace, verify: | ||
2052 | * - only '||' is used | ||
2053 | */ | ||
2054 | if (pred->op != OP_OR) | ||
2055 | return -EINVAL; | ||
2056 | } | ||
2057 | |||
2058 | return 0; | ||
2059 | } | ||
2060 | |||
2061 | static int ftrace_function_set_filter_cb(enum move_type move, | ||
2062 | struct filter_pred *pred, | ||
2063 | int *err, void *data) | ||
2064 | { | ||
2065 | /* Checking the node is valid for function trace. */ | ||
2066 | if ((move != MOVE_DOWN) || | ||
2067 | (pred->left != FILTER_PRED_INVALID)) { | ||
2068 | *err = ftrace_function_check_pred(pred, 0); | ||
2069 | } else { | ||
2070 | *err = ftrace_function_check_pred(pred, 1); | ||
2071 | if (*err) | ||
2072 | return WALK_PRED_ABORT; | ||
2073 | |||
2074 | *err = __ftrace_function_set_filter(pred->op == OP_EQ, | ||
2075 | pred->regex.pattern, | ||
2076 | pred->regex.len, | ||
2077 | data); | ||
2078 | } | ||
2079 | |||
2080 | return (*err) ? WALK_PRED_ABORT : WALK_PRED_DEFAULT; | ||
2081 | } | ||
2082 | |||
2083 | static int ftrace_function_set_filter(struct perf_event *event, | ||
2084 | struct event_filter *filter) | ||
2085 | { | ||
2086 | struct function_filter_data data = { | ||
2087 | .first_filter = 1, | ||
2088 | .first_notrace = 1, | ||
2089 | .ops = &event->ftrace_ops, | ||
2090 | }; | ||
2091 | |||
2092 | return walk_pred_tree(filter->preds, filter->root, | ||
2093 | ftrace_function_set_filter_cb, &data); | ||
2094 | } | ||
2095 | #else | ||
2096 | static int ftrace_function_set_filter(struct perf_event *event, | ||
2097 | struct event_filter *filter) | ||
2098 | { | ||
2099 | return -ENODEV; | ||
2100 | } | ||
2101 | #endif /* CONFIG_FUNCTION_TRACER */ | ||
2102 | |||
1953 | int ftrace_profile_set_filter(struct perf_event *event, int event_id, | 2103 | int ftrace_profile_set_filter(struct perf_event *event, int event_id, |
1954 | char *filter_str) | 2104 | char *filter_str) |
1955 | { | 2105 | { |
@@ -1970,9 +2120,16 @@ int ftrace_profile_set_filter(struct perf_event *event, int event_id, | |||
1970 | goto out_unlock; | 2120 | goto out_unlock; |
1971 | 2121 | ||
1972 | err = create_filter(call, filter_str, false, &filter); | 2122 | err = create_filter(call, filter_str, false, &filter); |
1973 | if (!err) | 2123 | if (err) |
1974 | event->filter = filter; | 2124 | goto free_filter; |
2125 | |||
2126 | if (ftrace_event_is_function(call)) | ||
2127 | err = ftrace_function_set_filter(event, filter); | ||
1975 | else | 2128 | else |
2129 | event->filter = filter; | ||
2130 | |||
2131 | free_filter: | ||
2132 | if (err || ftrace_event_is_function(call)) | ||
1976 | __free_filter(filter); | 2133 | __free_filter(filter); |
1977 | 2134 | ||
1978 | out_unlock: | 2135 | out_unlock: |
diff --git a/kernel/trace/trace_export.c b/kernel/trace/trace_export.c index bbeec31e0ae3..7b46c9bd22ae 100644 --- a/kernel/trace/trace_export.c +++ b/kernel/trace/trace_export.c | |||
@@ -18,6 +18,16 @@ | |||
18 | #undef TRACE_SYSTEM | 18 | #undef TRACE_SYSTEM |
19 | #define TRACE_SYSTEM ftrace | 19 | #define TRACE_SYSTEM ftrace |
20 | 20 | ||
21 | /* | ||
22 | * The FTRACE_ENTRY_REG macro allows ftrace entry to define register | ||
23 | * function and thus become accesible via perf. | ||
24 | */ | ||
25 | #undef FTRACE_ENTRY_REG | ||
26 | #define FTRACE_ENTRY_REG(name, struct_name, id, tstruct, print, \ | ||
27 | filter, regfn) \ | ||
28 | FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print), \ | ||
29 | filter) | ||
30 | |||
21 | /* not needed for this file */ | 31 | /* not needed for this file */ |
22 | #undef __field_struct | 32 | #undef __field_struct |
23 | #define __field_struct(type, item) | 33 | #define __field_struct(type, item) |
@@ -44,21 +54,22 @@ | |||
44 | #define F_printk(fmt, args...) fmt, args | 54 | #define F_printk(fmt, args...) fmt, args |
45 | 55 | ||
46 | #undef FTRACE_ENTRY | 56 | #undef FTRACE_ENTRY |
47 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ | 57 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print, filter) \ |
48 | struct ____ftrace_##name { \ | 58 | struct ____ftrace_##name { \ |
49 | tstruct \ | 59 | tstruct \ |
50 | }; \ | 60 | }; \ |
51 | static void __always_unused ____ftrace_check_##name(void) \ | 61 | static void __always_unused ____ftrace_check_##name(void) \ |
52 | { \ | 62 | { \ |
53 | struct ____ftrace_##name *__entry = NULL; \ | 63 | struct ____ftrace_##name *__entry = NULL; \ |
54 | \ | 64 | \ |
55 | /* force compile-time check on F_printk() */ \ | 65 | /* force compile-time check on F_printk() */ \ |
56 | printk(print); \ | 66 | printk(print); \ |
57 | } | 67 | } |
58 | 68 | ||
59 | #undef FTRACE_ENTRY_DUP | 69 | #undef FTRACE_ENTRY_DUP |
60 | #define FTRACE_ENTRY_DUP(name, struct_name, id, tstruct, print) \ | 70 | #define FTRACE_ENTRY_DUP(name, struct_name, id, tstruct, print, filter) \ |
61 | FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print)) | 71 | FTRACE_ENTRY(name, struct_name, id, PARAMS(tstruct), PARAMS(print), \ |
72 | filter) | ||
62 | 73 | ||
63 | #include "trace_entries.h" | 74 | #include "trace_entries.h" |
64 | 75 | ||
@@ -67,7 +78,7 @@ static void __always_unused ____ftrace_check_##name(void) \ | |||
67 | ret = trace_define_field(event_call, #type, #item, \ | 78 | ret = trace_define_field(event_call, #type, #item, \ |
68 | offsetof(typeof(field), item), \ | 79 | offsetof(typeof(field), item), \ |
69 | sizeof(field.item), \ | 80 | sizeof(field.item), \ |
70 | is_signed_type(type), FILTER_OTHER); \ | 81 | is_signed_type(type), filter_type); \ |
71 | if (ret) \ | 82 | if (ret) \ |
72 | return ret; | 83 | return ret; |
73 | 84 | ||
@@ -77,7 +88,7 @@ static void __always_unused ____ftrace_check_##name(void) \ | |||
77 | offsetof(typeof(field), \ | 88 | offsetof(typeof(field), \ |
78 | container.item), \ | 89 | container.item), \ |
79 | sizeof(field.container.item), \ | 90 | sizeof(field.container.item), \ |
80 | is_signed_type(type), FILTER_OTHER); \ | 91 | is_signed_type(type), filter_type); \ |
81 | if (ret) \ | 92 | if (ret) \ |
82 | return ret; | 93 | return ret; |
83 | 94 | ||
@@ -91,7 +102,7 @@ static void __always_unused ____ftrace_check_##name(void) \ | |||
91 | ret = trace_define_field(event_call, event_storage, #item, \ | 102 | ret = trace_define_field(event_call, event_storage, #item, \ |
92 | offsetof(typeof(field), item), \ | 103 | offsetof(typeof(field), item), \ |
93 | sizeof(field.item), \ | 104 | sizeof(field.item), \ |
94 | is_signed_type(type), FILTER_OTHER); \ | 105 | is_signed_type(type), filter_type); \ |
95 | mutex_unlock(&event_storage_mutex); \ | 106 | mutex_unlock(&event_storage_mutex); \ |
96 | if (ret) \ | 107 | if (ret) \ |
97 | return ret; \ | 108 | return ret; \ |
@@ -104,7 +115,7 @@ static void __always_unused ____ftrace_check_##name(void) \ | |||
104 | offsetof(typeof(field), \ | 115 | offsetof(typeof(field), \ |
105 | container.item), \ | 116 | container.item), \ |
106 | sizeof(field.container.item), \ | 117 | sizeof(field.container.item), \ |
107 | is_signed_type(type), FILTER_OTHER); \ | 118 | is_signed_type(type), filter_type); \ |
108 | if (ret) \ | 119 | if (ret) \ |
109 | return ret; | 120 | return ret; |
110 | 121 | ||
@@ -112,17 +123,18 @@ static void __always_unused ____ftrace_check_##name(void) \ | |||
112 | #define __dynamic_array(type, item) \ | 123 | #define __dynamic_array(type, item) \ |
113 | ret = trace_define_field(event_call, #type, #item, \ | 124 | ret = trace_define_field(event_call, #type, #item, \ |
114 | offsetof(typeof(field), item), \ | 125 | offsetof(typeof(field), item), \ |
115 | 0, is_signed_type(type), FILTER_OTHER);\ | 126 | 0, is_signed_type(type), filter_type);\ |
116 | if (ret) \ | 127 | if (ret) \ |
117 | return ret; | 128 | return ret; |
118 | 129 | ||
119 | #undef FTRACE_ENTRY | 130 | #undef FTRACE_ENTRY |
120 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print) \ | 131 | #define FTRACE_ENTRY(name, struct_name, id, tstruct, print, filter) \ |
121 | int \ | 132 | int \ |
122 | ftrace_define_fields_##name(struct ftrace_event_call *event_call) \ | 133 | ftrace_define_fields_##name(struct ftrace_event_call *event_call) \ |
123 | { \ | 134 | { \ |
124 | struct struct_name field; \ | 135 | struct struct_name field; \ |
125 | int ret; \ | 136 | int ret; \ |
137 | int filter_type = filter; \ | ||
126 | \ | 138 | \ |
127 | tstruct; \ | 139 | tstruct; \ |
128 | \ | 140 | \ |
@@ -152,13 +164,15 @@ ftrace_define_fields_##name(struct ftrace_event_call *event_call) \ | |||
152 | #undef F_printk | 164 | #undef F_printk |
153 | #define F_printk(fmt, args...) #fmt ", " __stringify(args) | 165 | #define F_printk(fmt, args...) #fmt ", " __stringify(args) |
154 | 166 | ||
155 | #undef FTRACE_ENTRY | 167 | #undef FTRACE_ENTRY_REG |
156 | #define FTRACE_ENTRY(call, struct_name, etype, tstruct, print) \ | 168 | #define FTRACE_ENTRY_REG(call, struct_name, etype, tstruct, print, filter,\ |
169 | regfn) \ | ||
157 | \ | 170 | \ |
158 | struct ftrace_event_class event_class_ftrace_##call = { \ | 171 | struct ftrace_event_class event_class_ftrace_##call = { \ |
159 | .system = __stringify(TRACE_SYSTEM), \ | 172 | .system = __stringify(TRACE_SYSTEM), \ |
160 | .define_fields = ftrace_define_fields_##call, \ | 173 | .define_fields = ftrace_define_fields_##call, \ |
161 | .fields = LIST_HEAD_INIT(event_class_ftrace_##call.fields),\ | 174 | .fields = LIST_HEAD_INIT(event_class_ftrace_##call.fields),\ |
175 | .reg = regfn, \ | ||
162 | }; \ | 176 | }; \ |
163 | \ | 177 | \ |
164 | struct ftrace_event_call __used event_##call = { \ | 178 | struct ftrace_event_call __used event_##call = { \ |
@@ -170,4 +184,14 @@ struct ftrace_event_call __used event_##call = { \ | |||
170 | struct ftrace_event_call __used \ | 184 | struct ftrace_event_call __used \ |
171 | __attribute__((section("_ftrace_events"))) *__event_##call = &event_##call; | 185 | __attribute__((section("_ftrace_events"))) *__event_##call = &event_##call; |
172 | 186 | ||
187 | #undef FTRACE_ENTRY | ||
188 | #define FTRACE_ENTRY(call, struct_name, etype, tstruct, print, filter) \ | ||
189 | FTRACE_ENTRY_REG(call, struct_name, etype, \ | ||
190 | PARAMS(tstruct), PARAMS(print), filter, NULL) | ||
191 | |||
192 | int ftrace_event_is_function(struct ftrace_event_call *call) | ||
193 | { | ||
194 | return call == &event_function; | ||
195 | } | ||
196 | |||
173 | #include "trace_entries.h" | 197 | #include "trace_entries.h" |
diff --git a/kernel/trace/trace_kprobe.c b/kernel/trace/trace_kprobe.c index 00d527c945a4..580a05ec926b 100644 --- a/kernel/trace/trace_kprobe.c +++ b/kernel/trace/trace_kprobe.c | |||
@@ -1892,7 +1892,8 @@ static __kprobes void kretprobe_perf_func(struct kretprobe_instance *ri, | |||
1892 | #endif /* CONFIG_PERF_EVENTS */ | 1892 | #endif /* CONFIG_PERF_EVENTS */ |
1893 | 1893 | ||
1894 | static __kprobes | 1894 | static __kprobes |
1895 | int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) | 1895 | int kprobe_register(struct ftrace_event_call *event, |
1896 | enum trace_reg type, void *data) | ||
1896 | { | 1897 | { |
1897 | struct trace_probe *tp = (struct trace_probe *)event->data; | 1898 | struct trace_probe *tp = (struct trace_probe *)event->data; |
1898 | 1899 | ||
@@ -1909,6 +1910,11 @@ int kprobe_register(struct ftrace_event_call *event, enum trace_reg type) | |||
1909 | case TRACE_REG_PERF_UNREGISTER: | 1910 | case TRACE_REG_PERF_UNREGISTER: |
1910 | disable_trace_probe(tp, TP_FLAG_PROFILE); | 1911 | disable_trace_probe(tp, TP_FLAG_PROFILE); |
1911 | return 0; | 1912 | return 0; |
1913 | case TRACE_REG_PERF_OPEN: | ||
1914 | case TRACE_REG_PERF_CLOSE: | ||
1915 | case TRACE_REG_PERF_ADD: | ||
1916 | case TRACE_REG_PERF_DEL: | ||
1917 | return 0; | ||
1912 | #endif | 1918 | #endif |
1913 | } | 1919 | } |
1914 | return 0; | 1920 | return 0; |
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c index 0d6ff3555942..c5a01873567d 100644 --- a/kernel/trace/trace_output.c +++ b/kernel/trace/trace_output.c | |||
@@ -300,7 +300,7 @@ ftrace_print_flags_seq(struct trace_seq *p, const char *delim, | |||
300 | unsigned long mask; | 300 | unsigned long mask; |
301 | const char *str; | 301 | const char *str; |
302 | const char *ret = p->buffer + p->len; | 302 | const char *ret = p->buffer + p->len; |
303 | int i; | 303 | int i, first = 1; |
304 | 304 | ||
305 | for (i = 0; flag_array[i].name && flags; i++) { | 305 | for (i = 0; flag_array[i].name && flags; i++) { |
306 | 306 | ||
@@ -310,14 +310,16 @@ ftrace_print_flags_seq(struct trace_seq *p, const char *delim, | |||
310 | 310 | ||
311 | str = flag_array[i].name; | 311 | str = flag_array[i].name; |
312 | flags &= ~mask; | 312 | flags &= ~mask; |
313 | if (p->len && delim) | 313 | if (!first && delim) |
314 | trace_seq_puts(p, delim); | 314 | trace_seq_puts(p, delim); |
315 | else | ||
316 | first = 0; | ||
315 | trace_seq_puts(p, str); | 317 | trace_seq_puts(p, str); |
316 | } | 318 | } |
317 | 319 | ||
318 | /* check for left over flags */ | 320 | /* check for left over flags */ |
319 | if (flags) { | 321 | if (flags) { |
320 | if (p->len && delim) | 322 | if (!first && delim) |
321 | trace_seq_puts(p, delim); | 323 | trace_seq_puts(p, delim); |
322 | trace_seq_printf(p, "0x%lx", flags); | 324 | trace_seq_printf(p, "0x%lx", flags); |
323 | } | 325 | } |
@@ -344,7 +346,7 @@ ftrace_print_symbols_seq(struct trace_seq *p, unsigned long val, | |||
344 | break; | 346 | break; |
345 | } | 347 | } |
346 | 348 | ||
347 | if (!p->len) | 349 | if (ret == (const char *)(p->buffer + p->len)) |
348 | trace_seq_printf(p, "0x%lx", val); | 350 | trace_seq_printf(p, "0x%lx", val); |
349 | 351 | ||
350 | trace_seq_putc(p, 0); | 352 | trace_seq_putc(p, 0); |
@@ -370,7 +372,7 @@ ftrace_print_symbols_seq_u64(struct trace_seq *p, unsigned long long val, | |||
370 | break; | 372 | break; |
371 | } | 373 | } |
372 | 374 | ||
373 | if (!p->len) | 375 | if (ret == (const char *)(p->buffer + p->len)) |
374 | trace_seq_printf(p, "0x%llx", val); | 376 | trace_seq_printf(p, "0x%llx", val); |
375 | 377 | ||
376 | trace_seq_putc(p, 0); | 378 | trace_seq_putc(p, 0); |
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index cb654542c1a1..96fc73369099 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
@@ -17,9 +17,9 @@ static DECLARE_BITMAP(enabled_enter_syscalls, NR_syscalls); | |||
17 | static DECLARE_BITMAP(enabled_exit_syscalls, NR_syscalls); | 17 | static DECLARE_BITMAP(enabled_exit_syscalls, NR_syscalls); |
18 | 18 | ||
19 | static int syscall_enter_register(struct ftrace_event_call *event, | 19 | static int syscall_enter_register(struct ftrace_event_call *event, |
20 | enum trace_reg type); | 20 | enum trace_reg type, void *data); |
21 | static int syscall_exit_register(struct ftrace_event_call *event, | 21 | static int syscall_exit_register(struct ftrace_event_call *event, |
22 | enum trace_reg type); | 22 | enum trace_reg type, void *data); |
23 | 23 | ||
24 | static int syscall_enter_define_fields(struct ftrace_event_call *call); | 24 | static int syscall_enter_define_fields(struct ftrace_event_call *call); |
25 | static int syscall_exit_define_fields(struct ftrace_event_call *call); | 25 | static int syscall_exit_define_fields(struct ftrace_event_call *call); |
@@ -468,8 +468,8 @@ int __init init_ftrace_syscalls(void) | |||
468 | unsigned long addr; | 468 | unsigned long addr; |
469 | int i; | 469 | int i; |
470 | 470 | ||
471 | syscalls_metadata = kzalloc(sizeof(*syscalls_metadata) * | 471 | syscalls_metadata = kcalloc(NR_syscalls, sizeof(*syscalls_metadata), |
472 | NR_syscalls, GFP_KERNEL); | 472 | GFP_KERNEL); |
473 | if (!syscalls_metadata) { | 473 | if (!syscalls_metadata) { |
474 | WARN_ON(1); | 474 | WARN_ON(1); |
475 | return -ENOMEM; | 475 | return -ENOMEM; |
@@ -649,7 +649,7 @@ void perf_sysexit_disable(struct ftrace_event_call *call) | |||
649 | #endif /* CONFIG_PERF_EVENTS */ | 649 | #endif /* CONFIG_PERF_EVENTS */ |
650 | 650 | ||
651 | static int syscall_enter_register(struct ftrace_event_call *event, | 651 | static int syscall_enter_register(struct ftrace_event_call *event, |
652 | enum trace_reg type) | 652 | enum trace_reg type, void *data) |
653 | { | 653 | { |
654 | switch (type) { | 654 | switch (type) { |
655 | case TRACE_REG_REGISTER: | 655 | case TRACE_REG_REGISTER: |
@@ -664,13 +664,18 @@ static int syscall_enter_register(struct ftrace_event_call *event, | |||
664 | case TRACE_REG_PERF_UNREGISTER: | 664 | case TRACE_REG_PERF_UNREGISTER: |
665 | perf_sysenter_disable(event); | 665 | perf_sysenter_disable(event); |
666 | return 0; | 666 | return 0; |
667 | case TRACE_REG_PERF_OPEN: | ||
668 | case TRACE_REG_PERF_CLOSE: | ||
669 | case TRACE_REG_PERF_ADD: | ||
670 | case TRACE_REG_PERF_DEL: | ||
671 | return 0; | ||
667 | #endif | 672 | #endif |
668 | } | 673 | } |
669 | return 0; | 674 | return 0; |
670 | } | 675 | } |
671 | 676 | ||
672 | static int syscall_exit_register(struct ftrace_event_call *event, | 677 | static int syscall_exit_register(struct ftrace_event_call *event, |
673 | enum trace_reg type) | 678 | enum trace_reg type, void *data) |
674 | { | 679 | { |
675 | switch (type) { | 680 | switch (type) { |
676 | case TRACE_REG_REGISTER: | 681 | case TRACE_REG_REGISTER: |
@@ -685,6 +690,11 @@ static int syscall_exit_register(struct ftrace_event_call *event, | |||
685 | case TRACE_REG_PERF_UNREGISTER: | 690 | case TRACE_REG_PERF_UNREGISTER: |
686 | perf_sysexit_disable(event); | 691 | perf_sysexit_disable(event); |
687 | return 0; | 692 | return 0; |
693 | case TRACE_REG_PERF_OPEN: | ||
694 | case TRACE_REG_PERF_CLOSE: | ||
695 | case TRACE_REG_PERF_ADD: | ||
696 | case TRACE_REG_PERF_DEL: | ||
697 | return 0; | ||
688 | #endif | 698 | #endif |
689 | } | 699 | } |
690 | return 0; | 700 | return 0; |
diff --git a/kernel/tracepoint.c b/kernel/tracepoint.c index f1539decd99d..d96ba22dabfa 100644 --- a/kernel/tracepoint.c +++ b/kernel/tracepoint.c | |||
@@ -25,7 +25,7 @@ | |||
25 | #include <linux/err.h> | 25 | #include <linux/err.h> |
26 | #include <linux/slab.h> | 26 | #include <linux/slab.h> |
27 | #include <linux/sched.h> | 27 | #include <linux/sched.h> |
28 | #include <linux/jump_label.h> | 28 | #include <linux/static_key.h> |
29 | 29 | ||
30 | extern struct tracepoint * const __start___tracepoints_ptrs[]; | 30 | extern struct tracepoint * const __start___tracepoints_ptrs[]; |
31 | extern struct tracepoint * const __stop___tracepoints_ptrs[]; | 31 | extern struct tracepoint * const __stop___tracepoints_ptrs[]; |
@@ -256,9 +256,9 @@ static void set_tracepoint(struct tracepoint_entry **entry, | |||
256 | { | 256 | { |
257 | WARN_ON(strcmp((*entry)->name, elem->name) != 0); | 257 | WARN_ON(strcmp((*entry)->name, elem->name) != 0); |
258 | 258 | ||
259 | if (elem->regfunc && !jump_label_enabled(&elem->key) && active) | 259 | if (elem->regfunc && !static_key_enabled(&elem->key) && active) |
260 | elem->regfunc(); | 260 | elem->regfunc(); |
261 | else if (elem->unregfunc && jump_label_enabled(&elem->key) && !active) | 261 | else if (elem->unregfunc && static_key_enabled(&elem->key) && !active) |
262 | elem->unregfunc(); | 262 | elem->unregfunc(); |
263 | 263 | ||
264 | /* | 264 | /* |
@@ -269,10 +269,10 @@ static void set_tracepoint(struct tracepoint_entry **entry, | |||
269 | * is used. | 269 | * is used. |
270 | */ | 270 | */ |
271 | rcu_assign_pointer(elem->funcs, (*entry)->funcs); | 271 | rcu_assign_pointer(elem->funcs, (*entry)->funcs); |
272 | if (active && !jump_label_enabled(&elem->key)) | 272 | if (active && !static_key_enabled(&elem->key)) |
273 | jump_label_inc(&elem->key); | 273 | static_key_slow_inc(&elem->key); |
274 | else if (!active && jump_label_enabled(&elem->key)) | 274 | else if (!active && static_key_enabled(&elem->key)) |
275 | jump_label_dec(&elem->key); | 275 | static_key_slow_dec(&elem->key); |
276 | } | 276 | } |
277 | 277 | ||
278 | /* | 278 | /* |
@@ -283,11 +283,11 @@ static void set_tracepoint(struct tracepoint_entry **entry, | |||
283 | */ | 283 | */ |
284 | static void disable_tracepoint(struct tracepoint *elem) | 284 | static void disable_tracepoint(struct tracepoint *elem) |
285 | { | 285 | { |
286 | if (elem->unregfunc && jump_label_enabled(&elem->key)) | 286 | if (elem->unregfunc && static_key_enabled(&elem->key)) |
287 | elem->unregfunc(); | 287 | elem->unregfunc(); |
288 | 288 | ||
289 | if (jump_label_enabled(&elem->key)) | 289 | if (static_key_enabled(&elem->key)) |
290 | jump_label_dec(&elem->key); | 290 | static_key_slow_dec(&elem->key); |
291 | rcu_assign_pointer(elem->funcs, NULL); | 291 | rcu_assign_pointer(elem->funcs, NULL); |
292 | } | 292 | } |
293 | 293 | ||
diff --git a/kernel/watchdog.c b/kernel/watchdog.c index d117262deba3..14bc092fb12c 100644 --- a/kernel/watchdog.c +++ b/kernel/watchdog.c | |||
@@ -3,12 +3,9 @@ | |||
3 | * | 3 | * |
4 | * started by Don Zickus, Copyright (C) 2010 Red Hat, Inc. | 4 | * started by Don Zickus, Copyright (C) 2010 Red Hat, Inc. |
5 | * | 5 | * |
6 | * this code detects hard lockups: incidents in where on a CPU | 6 | * Note: Most of this code is borrowed heavily from the original softlockup |
7 | * the kernel does not respond to anything except NMI. | 7 | * detector, so thanks to Ingo for the initial implementation. |
8 | * | 8 | * Some chunks also taken from the old x86-specific nmi watchdog code, thanks |
9 | * Note: Most of this code is borrowed heavily from softlockup.c, | ||
10 | * so thanks to Ingo for the initial implementation. | ||
11 | * Some chunks also taken from arch/x86/kernel/apic/nmi.c, thanks | ||
12 | * to those contributors as well. | 9 | * to those contributors as well. |
13 | */ | 10 | */ |
14 | 11 | ||
@@ -117,9 +114,10 @@ static unsigned long get_sample_period(void) | |||
117 | { | 114 | { |
118 | /* | 115 | /* |
119 | * convert watchdog_thresh from seconds to ns | 116 | * convert watchdog_thresh from seconds to ns |
120 | * the divide by 5 is to give hrtimer 5 chances to | 117 | * the divide by 5 is to give hrtimer several chances (two |
121 | * increment before the hardlockup detector generates | 118 | * or three with the current relation between the soft |
122 | * a warning | 119 | * and hard thresholds) to increment before the |
120 | * hardlockup detector generates a warning | ||
123 | */ | 121 | */ |
124 | return get_softlockup_thresh() * (NSEC_PER_SEC / 5); | 122 | return get_softlockup_thresh() * (NSEC_PER_SEC / 5); |
125 | } | 123 | } |
@@ -336,9 +334,11 @@ static int watchdog(void *unused) | |||
336 | 334 | ||
337 | set_current_state(TASK_INTERRUPTIBLE); | 335 | set_current_state(TASK_INTERRUPTIBLE); |
338 | /* | 336 | /* |
339 | * Run briefly once per second to reset the softlockup timestamp. | 337 | * Run briefly (kicked by the hrtimer callback function) once every |
340 | * If this gets delayed for more than 60 seconds then the | 338 | * get_sample_period() seconds (4 seconds by default) to reset the |
341 | * debug-printout triggers in watchdog_timer_fn(). | 339 | * softlockup timestamp. If this gets delayed for more than |
340 | * 2*watchdog_thresh seconds then the debug-printout triggers in | ||
341 | * watchdog_timer_fn(). | ||
342 | */ | 342 | */ |
343 | while (!kthread_should_stop()) { | 343 | while (!kthread_should_stop()) { |
344 | __touch_watchdog(); | 344 | __touch_watchdog(); |