diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2009-05-29 08:51:57 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-05-29 10:21:51 -0400 |
commit | 665c2142a94202881a3c11cbaee6506cb10ada2d (patch) | |
tree | 14a518bf3afe81e7300a822e9b6f55980ff89d46 /kernel | |
parent | efb3d17240d80e27508d238809168120fe4b93a4 (diff) |
perf_counter: Clean up task_ctx vs interrupts
Remove the local_irq_save() etc.. in routines that are smp function
calls, or have IRQs disabled by other means.
Then change the COMM, MMAP, and swcounter context iteration to
current->perf_counter_ctxp and RCU, since it really doesn't matter
which context they iterate, they're all folded.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Corey Ashford <cjashfor@linux.vnet.ibm.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: John Kacur <jkacur@redhat.com>
LKML-Reference: <new-submission>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/perf_counter.c | 82 |
1 files changed, 50 insertions, 32 deletions
diff --git a/kernel/perf_counter.c b/kernel/perf_counter.c index 58d6d198faa2..0c000d305e0e 100644 --- a/kernel/perf_counter.c +++ b/kernel/perf_counter.c | |||
@@ -232,18 +232,14 @@ static void __perf_counter_remove_from_context(void *info) | |||
232 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); | 232 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
233 | struct perf_counter *counter = info; | 233 | struct perf_counter *counter = info; |
234 | struct perf_counter_context *ctx = counter->ctx; | 234 | struct perf_counter_context *ctx = counter->ctx; |
235 | unsigned long flags; | ||
236 | 235 | ||
237 | local_irq_save(flags); | ||
238 | /* | 236 | /* |
239 | * If this is a task context, we need to check whether it is | 237 | * If this is a task context, we need to check whether it is |
240 | * the current task context of this cpu. If not it has been | 238 | * the current task context of this cpu. If not it has been |
241 | * scheduled out before the smp call arrived. | 239 | * scheduled out before the smp call arrived. |
242 | */ | 240 | */ |
243 | if (ctx->task && cpuctx->task_ctx != ctx) { | 241 | if (ctx->task && cpuctx->task_ctx != ctx) |
244 | local_irq_restore(flags); | ||
245 | return; | 242 | return; |
246 | } | ||
247 | 243 | ||
248 | spin_lock(&ctx->lock); | 244 | spin_lock(&ctx->lock); |
249 | /* | 245 | /* |
@@ -267,7 +263,7 @@ static void __perf_counter_remove_from_context(void *info) | |||
267 | } | 263 | } |
268 | 264 | ||
269 | perf_enable(); | 265 | perf_enable(); |
270 | spin_unlock_irqrestore(&ctx->lock, flags); | 266 | spin_unlock(&ctx->lock); |
271 | } | 267 | } |
272 | 268 | ||
273 | 269 | ||
@@ -383,17 +379,13 @@ static void __perf_counter_disable(void *info) | |||
383 | struct perf_counter *counter = info; | 379 | struct perf_counter *counter = info; |
384 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); | 380 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
385 | struct perf_counter_context *ctx = counter->ctx; | 381 | struct perf_counter_context *ctx = counter->ctx; |
386 | unsigned long flags; | ||
387 | 382 | ||
388 | local_irq_save(flags); | ||
389 | /* | 383 | /* |
390 | * If this is a per-task counter, need to check whether this | 384 | * If this is a per-task counter, need to check whether this |
391 | * counter's task is the current task on this cpu. | 385 | * counter's task is the current task on this cpu. |
392 | */ | 386 | */ |
393 | if (ctx->task && cpuctx->task_ctx != ctx) { | 387 | if (ctx->task && cpuctx->task_ctx != ctx) |
394 | local_irq_restore(flags); | ||
395 | return; | 388 | return; |
396 | } | ||
397 | 389 | ||
398 | spin_lock(&ctx->lock); | 390 | spin_lock(&ctx->lock); |
399 | 391 | ||
@@ -411,7 +403,7 @@ static void __perf_counter_disable(void *info) | |||
411 | counter->state = PERF_COUNTER_STATE_OFF; | 403 | counter->state = PERF_COUNTER_STATE_OFF; |
412 | } | 404 | } |
413 | 405 | ||
414 | spin_unlock_irqrestore(&ctx->lock, flags); | 406 | spin_unlock(&ctx->lock); |
415 | } | 407 | } |
416 | 408 | ||
417 | /* | 409 | /* |
@@ -618,10 +610,8 @@ static void __perf_install_in_context(void *info) | |||
618 | struct perf_counter_context *ctx = counter->ctx; | 610 | struct perf_counter_context *ctx = counter->ctx; |
619 | struct perf_counter *leader = counter->group_leader; | 611 | struct perf_counter *leader = counter->group_leader; |
620 | int cpu = smp_processor_id(); | 612 | int cpu = smp_processor_id(); |
621 | unsigned long flags; | ||
622 | int err; | 613 | int err; |
623 | 614 | ||
624 | local_irq_save(flags); | ||
625 | /* | 615 | /* |
626 | * If this is a task context, we need to check whether it is | 616 | * If this is a task context, we need to check whether it is |
627 | * the current task context of this cpu. If not it has been | 617 | * the current task context of this cpu. If not it has been |
@@ -630,10 +620,8 @@ static void __perf_install_in_context(void *info) | |||
630 | * on this cpu because it had no counters. | 620 | * on this cpu because it had no counters. |
631 | */ | 621 | */ |
632 | if (ctx->task && cpuctx->task_ctx != ctx) { | 622 | if (ctx->task && cpuctx->task_ctx != ctx) { |
633 | if (cpuctx->task_ctx || ctx->task != current) { | 623 | if (cpuctx->task_ctx || ctx->task != current) |
634 | local_irq_restore(flags); | ||
635 | return; | 624 | return; |
636 | } | ||
637 | cpuctx->task_ctx = ctx; | 625 | cpuctx->task_ctx = ctx; |
638 | } | 626 | } |
639 | 627 | ||
@@ -687,7 +675,7 @@ static void __perf_install_in_context(void *info) | |||
687 | unlock: | 675 | unlock: |
688 | perf_enable(); | 676 | perf_enable(); |
689 | 677 | ||
690 | spin_unlock_irqrestore(&ctx->lock, flags); | 678 | spin_unlock(&ctx->lock); |
691 | } | 679 | } |
692 | 680 | ||
693 | /* | 681 | /* |
@@ -751,19 +739,15 @@ static void __perf_counter_enable(void *info) | |||
751 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); | 739 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
752 | struct perf_counter_context *ctx = counter->ctx; | 740 | struct perf_counter_context *ctx = counter->ctx; |
753 | struct perf_counter *leader = counter->group_leader; | 741 | struct perf_counter *leader = counter->group_leader; |
754 | unsigned long flags; | ||
755 | int err; | 742 | int err; |
756 | 743 | ||
757 | local_irq_save(flags); | ||
758 | /* | 744 | /* |
759 | * If this is a per-task counter, need to check whether this | 745 | * If this is a per-task counter, need to check whether this |
760 | * counter's task is the current task on this cpu. | 746 | * counter's task is the current task on this cpu. |
761 | */ | 747 | */ |
762 | if (ctx->task && cpuctx->task_ctx != ctx) { | 748 | if (ctx->task && cpuctx->task_ctx != ctx) { |
763 | if (cpuctx->task_ctx || ctx->task != current) { | 749 | if (cpuctx->task_ctx || ctx->task != current) |
764 | local_irq_restore(flags); | ||
765 | return; | 750 | return; |
766 | } | ||
767 | cpuctx->task_ctx = ctx; | 751 | cpuctx->task_ctx = ctx; |
768 | } | 752 | } |
769 | 753 | ||
@@ -811,7 +795,7 @@ static void __perf_counter_enable(void *info) | |||
811 | } | 795 | } |
812 | 796 | ||
813 | unlock: | 797 | unlock: |
814 | spin_unlock_irqrestore(&ctx->lock, flags); | 798 | spin_unlock(&ctx->lock); |
815 | } | 799 | } |
816 | 800 | ||
817 | /* | 801 | /* |
@@ -981,6 +965,10 @@ void perf_counter_task_sched_out(struct task_struct *task, | |||
981 | spin_lock(&ctx->lock); | 965 | spin_lock(&ctx->lock); |
982 | spin_lock_nested(&next_ctx->lock, SINGLE_DEPTH_NESTING); | 966 | spin_lock_nested(&next_ctx->lock, SINGLE_DEPTH_NESTING); |
983 | if (context_equiv(ctx, next_ctx)) { | 967 | if (context_equiv(ctx, next_ctx)) { |
968 | /* | ||
969 | * XXX do we need a memory barrier of sorts | ||
970 | * wrt to rcu_dereference() of perf_counter_ctxp | ||
971 | */ | ||
984 | task->perf_counter_ctxp = next_ctx; | 972 | task->perf_counter_ctxp = next_ctx; |
985 | next->perf_counter_ctxp = ctx; | 973 | next->perf_counter_ctxp = ctx; |
986 | ctx->task = next; | 974 | ctx->task = next; |
@@ -998,6 +986,9 @@ void perf_counter_task_sched_out(struct task_struct *task, | |||
998 | } | 986 | } |
999 | } | 987 | } |
1000 | 988 | ||
989 | /* | ||
990 | * Called with IRQs disabled | ||
991 | */ | ||
1001 | static void __perf_counter_task_sched_out(struct perf_counter_context *ctx) | 992 | static void __perf_counter_task_sched_out(struct perf_counter_context *ctx) |
1002 | { | 993 | { |
1003 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); | 994 | struct perf_cpu_context *cpuctx = &__get_cpu_var(perf_cpu_context); |
@@ -1012,6 +1003,9 @@ static void __perf_counter_task_sched_out(struct perf_counter_context *ctx) | |||
1012 | cpuctx->task_ctx = NULL; | 1003 | cpuctx->task_ctx = NULL; |
1013 | } | 1004 | } |
1014 | 1005 | ||
1006 | /* | ||
1007 | * Called with IRQs disabled | ||
1008 | */ | ||
1015 | static void perf_counter_cpu_sched_out(struct perf_cpu_context *cpuctx) | 1009 | static void perf_counter_cpu_sched_out(struct perf_cpu_context *cpuctx) |
1016 | { | 1010 | { |
1017 | __perf_counter_sched_out(&cpuctx->ctx, cpuctx); | 1011 | __perf_counter_sched_out(&cpuctx->ctx, cpuctx); |
@@ -2431,6 +2425,7 @@ static void perf_counter_comm_ctx(struct perf_counter_context *ctx, | |||
2431 | static void perf_counter_comm_event(struct perf_comm_event *comm_event) | 2425 | static void perf_counter_comm_event(struct perf_comm_event *comm_event) |
2432 | { | 2426 | { |
2433 | struct perf_cpu_context *cpuctx; | 2427 | struct perf_cpu_context *cpuctx; |
2428 | struct perf_counter_context *ctx; | ||
2434 | unsigned int size; | 2429 | unsigned int size; |
2435 | char *comm = comm_event->task->comm; | 2430 | char *comm = comm_event->task->comm; |
2436 | 2431 | ||
@@ -2443,9 +2438,17 @@ static void perf_counter_comm_event(struct perf_comm_event *comm_event) | |||
2443 | 2438 | ||
2444 | cpuctx = &get_cpu_var(perf_cpu_context); | 2439 | cpuctx = &get_cpu_var(perf_cpu_context); |
2445 | perf_counter_comm_ctx(&cpuctx->ctx, comm_event); | 2440 | perf_counter_comm_ctx(&cpuctx->ctx, comm_event); |
2446 | if (cpuctx->task_ctx) | ||
2447 | perf_counter_comm_ctx(cpuctx->task_ctx, comm_event); | ||
2448 | put_cpu_var(perf_cpu_context); | 2441 | put_cpu_var(perf_cpu_context); |
2442 | |||
2443 | rcu_read_lock(); | ||
2444 | /* | ||
2445 | * doesn't really matter which of the child contexts the | ||
2446 | * events ends up in. | ||
2447 | */ | ||
2448 | ctx = rcu_dereference(current->perf_counter_ctxp); | ||
2449 | if (ctx) | ||
2450 | perf_counter_comm_ctx(ctx, comm_event); | ||
2451 | rcu_read_unlock(); | ||
2449 | } | 2452 | } |
2450 | 2453 | ||
2451 | void perf_counter_comm(struct task_struct *task) | 2454 | void perf_counter_comm(struct task_struct *task) |
@@ -2536,6 +2539,7 @@ static void perf_counter_mmap_ctx(struct perf_counter_context *ctx, | |||
2536 | static void perf_counter_mmap_event(struct perf_mmap_event *mmap_event) | 2539 | static void perf_counter_mmap_event(struct perf_mmap_event *mmap_event) |
2537 | { | 2540 | { |
2538 | struct perf_cpu_context *cpuctx; | 2541 | struct perf_cpu_context *cpuctx; |
2542 | struct perf_counter_context *ctx; | ||
2539 | struct file *file = mmap_event->file; | 2543 | struct file *file = mmap_event->file; |
2540 | unsigned int size; | 2544 | unsigned int size; |
2541 | char tmp[16]; | 2545 | char tmp[16]; |
@@ -2568,10 +2572,18 @@ got_name: | |||
2568 | 2572 | ||
2569 | cpuctx = &get_cpu_var(perf_cpu_context); | 2573 | cpuctx = &get_cpu_var(perf_cpu_context); |
2570 | perf_counter_mmap_ctx(&cpuctx->ctx, mmap_event); | 2574 | perf_counter_mmap_ctx(&cpuctx->ctx, mmap_event); |
2571 | if (cpuctx->task_ctx) | ||
2572 | perf_counter_mmap_ctx(cpuctx->task_ctx, mmap_event); | ||
2573 | put_cpu_var(perf_cpu_context); | 2575 | put_cpu_var(perf_cpu_context); |
2574 | 2576 | ||
2577 | rcu_read_lock(); | ||
2578 | /* | ||
2579 | * doesn't really matter which of the child contexts the | ||
2580 | * events ends up in. | ||
2581 | */ | ||
2582 | ctx = rcu_dereference(current->perf_counter_ctxp); | ||
2583 | if (ctx) | ||
2584 | perf_counter_mmap_ctx(ctx, mmap_event); | ||
2585 | rcu_read_unlock(); | ||
2586 | |||
2575 | kfree(buf); | 2587 | kfree(buf); |
2576 | } | 2588 | } |
2577 | 2589 | ||
@@ -2882,6 +2894,7 @@ static void __perf_swcounter_event(enum perf_event_types type, u32 event, | |||
2882 | { | 2894 | { |
2883 | struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context); | 2895 | struct perf_cpu_context *cpuctx = &get_cpu_var(perf_cpu_context); |
2884 | int *recursion = perf_swcounter_recursion_context(cpuctx); | 2896 | int *recursion = perf_swcounter_recursion_context(cpuctx); |
2897 | struct perf_counter_context *ctx; | ||
2885 | 2898 | ||
2886 | if (*recursion) | 2899 | if (*recursion) |
2887 | goto out; | 2900 | goto out; |
@@ -2891,10 +2904,15 @@ static void __perf_swcounter_event(enum perf_event_types type, u32 event, | |||
2891 | 2904 | ||
2892 | perf_swcounter_ctx_event(&cpuctx->ctx, type, event, | 2905 | perf_swcounter_ctx_event(&cpuctx->ctx, type, event, |
2893 | nr, nmi, regs, addr); | 2906 | nr, nmi, regs, addr); |
2894 | if (cpuctx->task_ctx) { | 2907 | rcu_read_lock(); |
2895 | perf_swcounter_ctx_event(cpuctx->task_ctx, type, event, | 2908 | /* |
2896 | nr, nmi, regs, addr); | 2909 | * doesn't really matter which of the child contexts the |
2897 | } | 2910 | * events ends up in. |
2911 | */ | ||
2912 | ctx = rcu_dereference(current->perf_counter_ctxp); | ||
2913 | if (ctx) | ||
2914 | perf_swcounter_ctx_event(ctx, type, event, nr, nmi, regs, addr); | ||
2915 | rcu_read_unlock(); | ||
2898 | 2916 | ||
2899 | barrier(); | 2917 | barrier(); |
2900 | (*recursion)--; | 2918 | (*recursion)--; |