diff options
| author | Russell King <rmk+kernel@arm.linux.org.uk> | 2010-12-18 09:27:55 -0500 |
|---|---|---|
| committer | Russell King <rmk+kernel@arm.linux.org.uk> | 2010-12-18 09:27:55 -0500 |
| commit | 2f841ed13b9f10037e25ddf417d01700ecd886d0 (patch) | |
| tree | 123448d98b3be03ac90fbb6e32f224235063c8bf /kernel | |
| parent | 961ec6daa7b14f376c30d447a830fa4783a2112c (diff) | |
| parent | 8fbf397c3389c1dedfa9ee412715046ab28fd82d (diff) | |
Merge branch 'hw-breakpoint' of git://repo.or.cz/linux-2.6/linux-wd into devel-stable
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/hw_breakpoint.c | 3 | ||||
| -rw-r--r-- | kernel/irq_work.c | 4 | ||||
| -rw-r--r-- | kernel/module.c | 12 | ||||
| -rw-r--r-- | kernel/perf_event.c | 93 | ||||
| -rw-r--r-- | kernel/posix-cpu-timers.c | 12 | ||||
| -rw-r--r-- | kernel/sched_fair.c | 8 | ||||
| -rw-r--r-- | kernel/trace/trace.c | 19 |
7 files changed, 121 insertions, 30 deletions
diff --git a/kernel/hw_breakpoint.c b/kernel/hw_breakpoint.c index 2c9120f0afca..e5325825aeb6 100644 --- a/kernel/hw_breakpoint.c +++ b/kernel/hw_breakpoint.c | |||
| @@ -620,7 +620,7 @@ static struct pmu perf_breakpoint = { | |||
| 620 | .read = hw_breakpoint_pmu_read, | 620 | .read = hw_breakpoint_pmu_read, |
| 621 | }; | 621 | }; |
| 622 | 622 | ||
| 623 | static int __init init_hw_breakpoint(void) | 623 | int __init init_hw_breakpoint(void) |
| 624 | { | 624 | { |
| 625 | unsigned int **task_bp_pinned; | 625 | unsigned int **task_bp_pinned; |
| 626 | int cpu, err_cpu; | 626 | int cpu, err_cpu; |
| @@ -655,6 +655,5 @@ static int __init init_hw_breakpoint(void) | |||
| 655 | 655 | ||
| 656 | return -ENOMEM; | 656 | return -ENOMEM; |
| 657 | } | 657 | } |
| 658 | core_initcall(init_hw_breakpoint); | ||
| 659 | 658 | ||
| 660 | 659 | ||
diff --git a/kernel/irq_work.c b/kernel/irq_work.c index f16763ff8481..90f881904bb1 100644 --- a/kernel/irq_work.c +++ b/kernel/irq_work.c | |||
| @@ -145,7 +145,9 @@ void irq_work_run(void) | |||
| 145 | * Clear the BUSY bit and return to the free state if | 145 | * Clear the BUSY bit and return to the free state if |
| 146 | * no-one else claimed it meanwhile. | 146 | * no-one else claimed it meanwhile. |
| 147 | */ | 147 | */ |
| 148 | cmpxchg(&entry->next, next_flags(NULL, IRQ_WORK_BUSY), NULL); | 148 | (void)cmpxchg(&entry->next, |
| 149 | next_flags(NULL, IRQ_WORK_BUSY), | ||
| 150 | NULL); | ||
| 149 | } | 151 | } |
| 150 | } | 152 | } |
| 151 | EXPORT_SYMBOL_GPL(irq_work_run); | 153 | EXPORT_SYMBOL_GPL(irq_work_run); |
diff --git a/kernel/module.c b/kernel/module.c index 437a74a7524a..d190664f25ff 100644 --- a/kernel/module.c +++ b/kernel/module.c | |||
| @@ -2326,6 +2326,18 @@ static void find_module_sections(struct module *mod, struct load_info *info) | |||
| 2326 | kmemleak_scan_area(mod->trace_events, sizeof(*mod->trace_events) * | 2326 | kmemleak_scan_area(mod->trace_events, sizeof(*mod->trace_events) * |
| 2327 | mod->num_trace_events, GFP_KERNEL); | 2327 | mod->num_trace_events, GFP_KERNEL); |
| 2328 | #endif | 2328 | #endif |
| 2329 | #ifdef CONFIG_TRACING | ||
| 2330 | mod->trace_bprintk_fmt_start = section_objs(info, "__trace_printk_fmt", | ||
| 2331 | sizeof(*mod->trace_bprintk_fmt_start), | ||
| 2332 | &mod->num_trace_bprintk_fmt); | ||
| 2333 | /* | ||
| 2334 | * This section contains pointers to allocated objects in the trace | ||
| 2335 | * code and not scanning it leads to false positives. | ||
| 2336 | */ | ||
| 2337 | kmemleak_scan_area(mod->trace_bprintk_fmt_start, | ||
| 2338 | sizeof(*mod->trace_bprintk_fmt_start) * | ||
| 2339 | mod->num_trace_bprintk_fmt, GFP_KERNEL); | ||
| 2340 | #endif | ||
| 2329 | #ifdef CONFIG_FTRACE_MCOUNT_RECORD | 2341 | #ifdef CONFIG_FTRACE_MCOUNT_RECORD |
| 2330 | /* sechdrs[0].sh_size is always zero */ | 2342 | /* sechdrs[0].sh_size is always zero */ |
| 2331 | mod->ftrace_callsites = section_objs(info, "__mcount_loc", | 2343 | mod->ftrace_callsites = section_objs(info, "__mcount_loc", |
diff --git a/kernel/perf_event.c b/kernel/perf_event.c index cb6c0d2af68f..eac7e3364335 100644 --- a/kernel/perf_event.c +++ b/kernel/perf_event.c | |||
| @@ -31,6 +31,7 @@ | |||
| 31 | #include <linux/kernel_stat.h> | 31 | #include <linux/kernel_stat.h> |
| 32 | #include <linux/perf_event.h> | 32 | #include <linux/perf_event.h> |
| 33 | #include <linux/ftrace_event.h> | 33 | #include <linux/ftrace_event.h> |
| 34 | #include <linux/hw_breakpoint.h> | ||
| 34 | 35 | ||
| 35 | #include <asm/irq_regs.h> | 36 | #include <asm/irq_regs.h> |
| 36 | 37 | ||
| @@ -1286,8 +1287,6 @@ void __perf_event_task_sched_out(struct task_struct *task, | |||
| 1286 | { | 1287 | { |
| 1287 | int ctxn; | 1288 | int ctxn; |
| 1288 | 1289 | ||
| 1289 | perf_sw_event(PERF_COUNT_SW_CONTEXT_SWITCHES, 1, 1, NULL, 0); | ||
| 1290 | |||
| 1291 | for_each_task_context_nr(ctxn) | 1290 | for_each_task_context_nr(ctxn) |
| 1292 | perf_event_context_sched_out(task, ctxn, next); | 1291 | perf_event_context_sched_out(task, ctxn, next); |
| 1293 | } | 1292 | } |
| @@ -1621,8 +1620,12 @@ static void rotate_ctx(struct perf_event_context *ctx) | |||
| 1621 | { | 1620 | { |
| 1622 | raw_spin_lock(&ctx->lock); | 1621 | raw_spin_lock(&ctx->lock); |
| 1623 | 1622 | ||
| 1624 | /* Rotate the first entry last of non-pinned groups */ | 1623 | /* |
| 1625 | list_rotate_left(&ctx->flexible_groups); | 1624 | * Rotate the first entry last of non-pinned groups. Rotation might be |
| 1625 | * disabled by the inheritance code. | ||
| 1626 | */ | ||
| 1627 | if (!ctx->rotate_disable) | ||
| 1628 | list_rotate_left(&ctx->flexible_groups); | ||
| 1626 | 1629 | ||
| 1627 | raw_spin_unlock(&ctx->lock); | 1630 | raw_spin_unlock(&ctx->lock); |
| 1628 | } | 1631 | } |
| @@ -2234,11 +2237,6 @@ int perf_event_release_kernel(struct perf_event *event) | |||
| 2234 | raw_spin_unlock_irq(&ctx->lock); | 2237 | raw_spin_unlock_irq(&ctx->lock); |
| 2235 | mutex_unlock(&ctx->mutex); | 2238 | mutex_unlock(&ctx->mutex); |
| 2236 | 2239 | ||
| 2237 | mutex_lock(&event->owner->perf_event_mutex); | ||
| 2238 | list_del_init(&event->owner_entry); | ||
| 2239 | mutex_unlock(&event->owner->perf_event_mutex); | ||
| 2240 | put_task_struct(event->owner); | ||
| 2241 | |||
| 2242 | free_event(event); | 2240 | free_event(event); |
| 2243 | 2241 | ||
| 2244 | return 0; | 2242 | return 0; |
| @@ -2251,9 +2249,43 @@ EXPORT_SYMBOL_GPL(perf_event_release_kernel); | |||
| 2251 | static int perf_release(struct inode *inode, struct file *file) | 2249 | static int perf_release(struct inode *inode, struct file *file) |
| 2252 | { | 2250 | { |
| 2253 | struct perf_event *event = file->private_data; | 2251 | struct perf_event *event = file->private_data; |
| 2252 | struct task_struct *owner; | ||
| 2254 | 2253 | ||
| 2255 | file->private_data = NULL; | 2254 | file->private_data = NULL; |
| 2256 | 2255 | ||
| 2256 | rcu_read_lock(); | ||
| 2257 | owner = ACCESS_ONCE(event->owner); | ||
| 2258 | /* | ||
| 2259 | * Matches the smp_wmb() in perf_event_exit_task(). If we observe | ||
| 2260 | * !owner it means the list deletion is complete and we can indeed | ||
| 2261 | * free this event, otherwise we need to serialize on | ||
| 2262 | * owner->perf_event_mutex. | ||
| 2263 | */ | ||
| 2264 | smp_read_barrier_depends(); | ||
| 2265 | if (owner) { | ||
| 2266 | /* | ||
| 2267 | * Since delayed_put_task_struct() also drops the last | ||
| 2268 | * task reference we can safely take a new reference | ||
| 2269 | * while holding the rcu_read_lock(). | ||
| 2270 | */ | ||
| 2271 | get_task_struct(owner); | ||
| 2272 | } | ||
| 2273 | rcu_read_unlock(); | ||
| 2274 | |||
| 2275 | if (owner) { | ||
| 2276 | mutex_lock(&owner->perf_event_mutex); | ||
| 2277 | /* | ||
| 2278 | * We have to re-check the event->owner field, if it is cleared | ||
| 2279 | * we raced with perf_event_exit_task(), acquiring the mutex | ||
| 2280 | * ensured they're done, and we can proceed with freeing the | ||
| 2281 | * event. | ||
| 2282 | */ | ||
| 2283 | if (event->owner) | ||
| 2284 | list_del_init(&event->owner_entry); | ||
| 2285 | mutex_unlock(&owner->perf_event_mutex); | ||
| 2286 | put_task_struct(owner); | ||
| 2287 | } | ||
| 2288 | |||
| 2257 | return perf_event_release_kernel(event); | 2289 | return perf_event_release_kernel(event); |
| 2258 | } | 2290 | } |
| 2259 | 2291 | ||
| @@ -5677,7 +5709,7 @@ SYSCALL_DEFINE5(perf_event_open, | |||
| 5677 | mutex_unlock(&ctx->mutex); | 5709 | mutex_unlock(&ctx->mutex); |
| 5678 | 5710 | ||
| 5679 | event->owner = current; | 5711 | event->owner = current; |
| 5680 | get_task_struct(current); | 5712 | |
| 5681 | mutex_lock(¤t->perf_event_mutex); | 5713 | mutex_lock(¤t->perf_event_mutex); |
| 5682 | list_add_tail(&event->owner_entry, ¤t->perf_event_list); | 5714 | list_add_tail(&event->owner_entry, ¤t->perf_event_list); |
| 5683 | mutex_unlock(¤t->perf_event_mutex); | 5715 | mutex_unlock(¤t->perf_event_mutex); |
| @@ -5745,12 +5777,6 @@ perf_event_create_kernel_counter(struct perf_event_attr *attr, int cpu, | |||
| 5745 | ++ctx->generation; | 5777 | ++ctx->generation; |
| 5746 | mutex_unlock(&ctx->mutex); | 5778 | mutex_unlock(&ctx->mutex); |
| 5747 | 5779 | ||
| 5748 | event->owner = current; | ||
| 5749 | get_task_struct(current); | ||
| 5750 | mutex_lock(¤t->perf_event_mutex); | ||
| 5751 | list_add_tail(&event->owner_entry, ¤t->perf_event_list); | ||
| 5752 | mutex_unlock(¤t->perf_event_mutex); | ||
| 5753 | |||
| 5754 | return event; | 5780 | return event; |
| 5755 | 5781 | ||
| 5756 | err_free: | 5782 | err_free: |
| @@ -5901,8 +5927,24 @@ again: | |||
| 5901 | */ | 5927 | */ |
| 5902 | void perf_event_exit_task(struct task_struct *child) | 5928 | void perf_event_exit_task(struct task_struct *child) |
| 5903 | { | 5929 | { |
| 5930 | struct perf_event *event, *tmp; | ||
| 5904 | int ctxn; | 5931 | int ctxn; |
| 5905 | 5932 | ||
| 5933 | mutex_lock(&child->perf_event_mutex); | ||
| 5934 | list_for_each_entry_safe(event, tmp, &child->perf_event_list, | ||
| 5935 | owner_entry) { | ||
| 5936 | list_del_init(&event->owner_entry); | ||
| 5937 | |||
| 5938 | /* | ||
| 5939 | * Ensure the list deletion is visible before we clear | ||
| 5940 | * the owner, closes a race against perf_release() where | ||
| 5941 | * we need to serialize on the owner->perf_event_mutex. | ||
| 5942 | */ | ||
| 5943 | smp_wmb(); | ||
| 5944 | event->owner = NULL; | ||
| 5945 | } | ||
| 5946 | mutex_unlock(&child->perf_event_mutex); | ||
| 5947 | |||
| 5906 | for_each_task_context_nr(ctxn) | 5948 | for_each_task_context_nr(ctxn) |
| 5907 | perf_event_exit_task_context(child, ctxn); | 5949 | perf_event_exit_task_context(child, ctxn); |
| 5908 | } | 5950 | } |
| @@ -6122,6 +6164,7 @@ int perf_event_init_context(struct task_struct *child, int ctxn) | |||
| 6122 | struct perf_event *event; | 6164 | struct perf_event *event; |
| 6123 | struct task_struct *parent = current; | 6165 | struct task_struct *parent = current; |
| 6124 | int inherited_all = 1; | 6166 | int inherited_all = 1; |
| 6167 | unsigned long flags; | ||
| 6125 | int ret = 0; | 6168 | int ret = 0; |
| 6126 | 6169 | ||
| 6127 | child->perf_event_ctxp[ctxn] = NULL; | 6170 | child->perf_event_ctxp[ctxn] = NULL; |
| @@ -6162,6 +6205,15 @@ int perf_event_init_context(struct task_struct *child, int ctxn) | |||
| 6162 | break; | 6205 | break; |
| 6163 | } | 6206 | } |
| 6164 | 6207 | ||
| 6208 | /* | ||
| 6209 | * We can't hold ctx->lock when iterating the ->flexible_group list due | ||
| 6210 | * to allocations, but we need to prevent rotation because | ||
| 6211 | * rotate_ctx() will change the list from interrupt context. | ||
| 6212 | */ | ||
| 6213 | raw_spin_lock_irqsave(&parent_ctx->lock, flags); | ||
| 6214 | parent_ctx->rotate_disable = 1; | ||
| 6215 | raw_spin_unlock_irqrestore(&parent_ctx->lock, flags); | ||
| 6216 | |||
| 6165 | list_for_each_entry(event, &parent_ctx->flexible_groups, group_entry) { | 6217 | list_for_each_entry(event, &parent_ctx->flexible_groups, group_entry) { |
| 6166 | ret = inherit_task_group(event, parent, parent_ctx, | 6218 | ret = inherit_task_group(event, parent, parent_ctx, |
| 6167 | child, ctxn, &inherited_all); | 6219 | child, ctxn, &inherited_all); |
| @@ -6169,6 +6221,10 @@ int perf_event_init_context(struct task_struct *child, int ctxn) | |||
| 6169 | break; | 6221 | break; |
| 6170 | } | 6222 | } |
| 6171 | 6223 | ||
| 6224 | raw_spin_lock_irqsave(&parent_ctx->lock, flags); | ||
| 6225 | parent_ctx->rotate_disable = 0; | ||
| 6226 | raw_spin_unlock_irqrestore(&parent_ctx->lock, flags); | ||
| 6227 | |||
| 6172 | child_ctx = child->perf_event_ctxp[ctxn]; | 6228 | child_ctx = child->perf_event_ctxp[ctxn]; |
| 6173 | 6229 | ||
| 6174 | if (child_ctx && inherited_all) { | 6230 | if (child_ctx && inherited_all) { |
| @@ -6321,6 +6377,8 @@ perf_cpu_notify(struct notifier_block *self, unsigned long action, void *hcpu) | |||
| 6321 | 6377 | ||
| 6322 | void __init perf_event_init(void) | 6378 | void __init perf_event_init(void) |
| 6323 | { | 6379 | { |
| 6380 | int ret; | ||
| 6381 | |||
| 6324 | perf_event_init_all_cpus(); | 6382 | perf_event_init_all_cpus(); |
| 6325 | init_srcu_struct(&pmus_srcu); | 6383 | init_srcu_struct(&pmus_srcu); |
| 6326 | perf_pmu_register(&perf_swevent); | 6384 | perf_pmu_register(&perf_swevent); |
| @@ -6328,4 +6386,7 @@ void __init perf_event_init(void) | |||
| 6328 | perf_pmu_register(&perf_task_clock); | 6386 | perf_pmu_register(&perf_task_clock); |
| 6329 | perf_tp_register(); | 6387 | perf_tp_register(); |
| 6330 | perf_cpu_notifier(perf_cpu_notify); | 6388 | perf_cpu_notifier(perf_cpu_notify); |
| 6389 | |||
| 6390 | ret = init_hw_breakpoint(); | ||
| 6391 | WARN(ret, "hw_breakpoint initialization failed with: %d", ret); | ||
| 6331 | } | 6392 | } |
diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c index 6842eeba5879..05bb7173850e 100644 --- a/kernel/posix-cpu-timers.c +++ b/kernel/posix-cpu-timers.c | |||
| @@ -37,13 +37,13 @@ static int check_clock(const clockid_t which_clock) | |||
| 37 | if (pid == 0) | 37 | if (pid == 0) |
| 38 | return 0; | 38 | return 0; |
| 39 | 39 | ||
| 40 | read_lock(&tasklist_lock); | 40 | rcu_read_lock(); |
| 41 | p = find_task_by_vpid(pid); | 41 | p = find_task_by_vpid(pid); |
| 42 | if (!p || !(CPUCLOCK_PERTHREAD(which_clock) ? | 42 | if (!p || !(CPUCLOCK_PERTHREAD(which_clock) ? |
| 43 | same_thread_group(p, current) : thread_group_leader(p))) { | 43 | same_thread_group(p, current) : has_group_leader_pid(p))) { |
| 44 | error = -EINVAL; | 44 | error = -EINVAL; |
| 45 | } | 45 | } |
| 46 | read_unlock(&tasklist_lock); | 46 | rcu_read_unlock(); |
| 47 | 47 | ||
| 48 | return error; | 48 | return error; |
| 49 | } | 49 | } |
| @@ -390,7 +390,7 @@ int posix_cpu_timer_create(struct k_itimer *new_timer) | |||
| 390 | 390 | ||
| 391 | INIT_LIST_HEAD(&new_timer->it.cpu.entry); | 391 | INIT_LIST_HEAD(&new_timer->it.cpu.entry); |
| 392 | 392 | ||
| 393 | read_lock(&tasklist_lock); | 393 | rcu_read_lock(); |
| 394 | if (CPUCLOCK_PERTHREAD(new_timer->it_clock)) { | 394 | if (CPUCLOCK_PERTHREAD(new_timer->it_clock)) { |
| 395 | if (pid == 0) { | 395 | if (pid == 0) { |
| 396 | p = current; | 396 | p = current; |
| @@ -404,7 +404,7 @@ int posix_cpu_timer_create(struct k_itimer *new_timer) | |||
| 404 | p = current->group_leader; | 404 | p = current->group_leader; |
| 405 | } else { | 405 | } else { |
| 406 | p = find_task_by_vpid(pid); | 406 | p = find_task_by_vpid(pid); |
| 407 | if (p && !thread_group_leader(p)) | 407 | if (p && !has_group_leader_pid(p)) |
| 408 | p = NULL; | 408 | p = NULL; |
| 409 | } | 409 | } |
| 410 | } | 410 | } |
| @@ -414,7 +414,7 @@ int posix_cpu_timer_create(struct k_itimer *new_timer) | |||
| 414 | } else { | 414 | } else { |
| 415 | ret = -EINVAL; | 415 | ret = -EINVAL; |
| 416 | } | 416 | } |
| 417 | read_unlock(&tasklist_lock); | 417 | rcu_read_unlock(); |
| 418 | 418 | ||
| 419 | return ret; | 419 | return ret; |
| 420 | } | 420 | } |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 52ab113d8bb9..00ebd7686676 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
| @@ -1758,10 +1758,6 @@ static void pull_task(struct rq *src_rq, struct task_struct *p, | |||
| 1758 | set_task_cpu(p, this_cpu); | 1758 | set_task_cpu(p, this_cpu); |
| 1759 | activate_task(this_rq, p, 0); | 1759 | activate_task(this_rq, p, 0); |
| 1760 | check_preempt_curr(this_rq, p, 0); | 1760 | check_preempt_curr(this_rq, p, 0); |
| 1761 | |||
| 1762 | /* re-arm NEWIDLE balancing when moving tasks */ | ||
| 1763 | src_rq->avg_idle = this_rq->avg_idle = 2*sysctl_sched_migration_cost; | ||
| 1764 | this_rq->idle_stamp = 0; | ||
| 1765 | } | 1761 | } |
| 1766 | 1762 | ||
| 1767 | /* | 1763 | /* |
| @@ -3219,8 +3215,10 @@ static void idle_balance(int this_cpu, struct rq *this_rq) | |||
| 3219 | interval = msecs_to_jiffies(sd->balance_interval); | 3215 | interval = msecs_to_jiffies(sd->balance_interval); |
| 3220 | if (time_after(next_balance, sd->last_balance + interval)) | 3216 | if (time_after(next_balance, sd->last_balance + interval)) |
| 3221 | next_balance = sd->last_balance + interval; | 3217 | next_balance = sd->last_balance + interval; |
| 3222 | if (pulled_task) | 3218 | if (pulled_task) { |
| 3219 | this_rq->idle_stamp = 0; | ||
| 3223 | break; | 3220 | break; |
| 3221 | } | ||
| 3224 | } | 3222 | } |
| 3225 | 3223 | ||
| 3226 | raw_spin_lock(&this_rq->lock); | 3224 | raw_spin_lock(&this_rq->lock); |
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index 042084157980..c380612273bf 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c | |||
| @@ -1283,6 +1283,8 @@ void trace_dump_stack(void) | |||
| 1283 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count()); | 1283 | __ftrace_trace_stack(global_trace.buffer, flags, 3, preempt_count()); |
| 1284 | } | 1284 | } |
| 1285 | 1285 | ||
| 1286 | static DEFINE_PER_CPU(int, user_stack_count); | ||
| 1287 | |||
| 1286 | void | 1288 | void |
| 1287 | ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) | 1289 | ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) |
| 1288 | { | 1290 | { |
| @@ -1301,6 +1303,18 @@ ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) | |||
| 1301 | if (unlikely(in_nmi())) | 1303 | if (unlikely(in_nmi())) |
| 1302 | return; | 1304 | return; |
| 1303 | 1305 | ||
| 1306 | /* | ||
| 1307 | * prevent recursion, since the user stack tracing may | ||
| 1308 | * trigger other kernel events. | ||
| 1309 | */ | ||
| 1310 | preempt_disable(); | ||
| 1311 | if (__this_cpu_read(user_stack_count)) | ||
| 1312 | goto out; | ||
| 1313 | |||
| 1314 | __this_cpu_inc(user_stack_count); | ||
| 1315 | |||
| 1316 | |||
| 1317 | |||
| 1304 | event = trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, | 1318 | event = trace_buffer_lock_reserve(buffer, TRACE_USER_STACK, |
| 1305 | sizeof(*entry), flags, pc); | 1319 | sizeof(*entry), flags, pc); |
| 1306 | if (!event) | 1320 | if (!event) |
| @@ -1318,6 +1332,11 @@ ftrace_trace_userstack(struct ring_buffer *buffer, unsigned long flags, int pc) | |||
| 1318 | save_stack_trace_user(&trace); | 1332 | save_stack_trace_user(&trace); |
| 1319 | if (!filter_check_discard(call, entry, buffer, event)) | 1333 | if (!filter_check_discard(call, entry, buffer, event)) |
| 1320 | ring_buffer_unlock_commit(buffer, event); | 1334 | ring_buffer_unlock_commit(buffer, event); |
| 1335 | |||
| 1336 | __this_cpu_dec(user_stack_count); | ||
| 1337 | |||
| 1338 | out: | ||
| 1339 | preempt_enable(); | ||
| 1321 | } | 1340 | } |
| 1322 | 1341 | ||
| 1323 | #ifdef UNUSED | 1342 | #ifdef UNUSED |
