aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2009-03-19 15:26:19 -0400
committerIngo Molnar <mingo@elte.hu>2009-04-06 03:30:17 -0400
commit0322cd6ec504b0bf08ca7b2c3d7f43bda37d79c9 (patch)
tree8e5487e1a6700eb865ec1f268d51e9d6d3d38a71 /arch
parentb8e83514b64577b48bfb794fe85fcde40a9343ca (diff)
perf_counter: unify irq output code
Impact: cleanup Having 3 slightly different copies of the same code around does nobody any good. First step in revamping the output format. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Paul Mackerras <paulus@samba.org> Cc: Steven Rostedt <rostedt@goodmis.org> Orig-LKML-Reference: <20090319194233.929962222@chello.nl> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch')
-rw-r--r--arch/powerpc/kernel/perf_counter.c51
-rw-r--r--arch/x86/kernel/cpu/perf_counter.c53
2 files changed, 3 insertions, 101 deletions
diff --git a/arch/powerpc/kernel/perf_counter.c b/arch/powerpc/kernel/perf_counter.c
index 26f69dc7130e..88b72eb4af12 100644
--- a/arch/powerpc/kernel/perf_counter.c
+++ b/arch/powerpc/kernel/perf_counter.c
@@ -663,41 +663,6 @@ void perf_counter_do_pending(void)
663} 663}
664 664
665/* 665/*
666 * Record data for an irq counter.
667 * This function was lifted from the x86 code; maybe it should
668 * go in the core?
669 */
670static void perf_store_irq_data(struct perf_counter *counter, u64 data)
671{
672 struct perf_data *irqdata = counter->irqdata;
673
674 if (irqdata->len > PERF_DATA_BUFLEN - sizeof(u64)) {
675 irqdata->overrun++;
676 } else {
677 u64 *p = (u64 *) &irqdata->data[irqdata->len];
678
679 *p = data;
680 irqdata->len += sizeof(u64);
681 }
682}
683
684/*
685 * Record all the values of the counters in a group
686 */
687static void perf_handle_group(struct perf_counter *counter)
688{
689 struct perf_counter *leader, *sub;
690
691 leader = counter->group_leader;
692 list_for_each_entry(sub, &leader->sibling_list, list_entry) {
693 if (sub != counter)
694 sub->hw_ops->read(sub);
695 perf_store_irq_data(counter, sub->hw_event.event_config);
696 perf_store_irq_data(counter, atomic64_read(&sub->count));
697 }
698}
699
700/*
701 * A counter has overflowed; update its count and record 666 * A counter has overflowed; update its count and record
702 * things if requested. Note that interrupts are hard-disabled 667 * things if requested. Note that interrupts are hard-disabled
703 * here so there is no possibility of being interrupted. 668 * here so there is no possibility of being interrupted.
@@ -736,20 +701,8 @@ static void record_and_restart(struct perf_counter *counter, long val,
736 /* 701 /*
737 * Finally record data if requested. 702 * Finally record data if requested.
738 */ 703 */
739 if (record) { 704 if (record)
740 switch (counter->hw_event.record_type) { 705 perf_counter_output(counter, 1, regs);
741 case PERF_RECORD_SIMPLE:
742 break;
743 case PERF_RECORD_IRQ:
744 perf_store_irq_data(counter, instruction_pointer(regs));
745 counter->wakeup_pending = 1;
746 break;
747 case PERF_RECORD_GROUP:
748 perf_handle_group(counter);
749 counter->wakeup_pending = 1;
750 break;
751 }
752 }
753} 706}
754 707
755/* 708/*
diff --git a/arch/x86/kernel/cpu/perf_counter.c b/arch/x86/kernel/cpu/perf_counter.c
index d844ae41d5a3..902282d68b0c 100644
--- a/arch/x86/kernel/cpu/perf_counter.c
+++ b/arch/x86/kernel/cpu/perf_counter.c
@@ -674,20 +674,6 @@ static void pmc_generic_disable(struct perf_counter *counter)
674 x86_perf_counter_update(counter, hwc, idx); 674 x86_perf_counter_update(counter, hwc, idx);
675} 675}
676 676
677static void perf_store_irq_data(struct perf_counter *counter, u64 data)
678{
679 struct perf_data *irqdata = counter->irqdata;
680
681 if (irqdata->len > PERF_DATA_BUFLEN - sizeof(u64)) {
682 irqdata->overrun++;
683 } else {
684 u64 *p = (u64 *) &irqdata->data[irqdata->len];
685
686 *p = data;
687 irqdata->len += sizeof(u64);
688 }
689}
690
691/* 677/*
692 * Save and restart an expired counter. Called by NMI contexts, 678 * Save and restart an expired counter. Called by NMI contexts,
693 * so it has to be careful about preempting normal counter ops: 679 * so it has to be careful about preempting normal counter ops:
@@ -704,22 +690,6 @@ static void perf_save_and_restart(struct perf_counter *counter)
704 __pmc_generic_enable(counter, hwc, idx); 690 __pmc_generic_enable(counter, hwc, idx);
705} 691}
706 692
707static void
708perf_handle_group(struct perf_counter *sibling, u64 *status, u64 *overflown)
709{
710 struct perf_counter *counter, *group_leader = sibling->group_leader;
711
712 /*
713 * Store sibling timestamps (if any):
714 */
715 list_for_each_entry(counter, &group_leader->sibling_list, list_entry) {
716
717 x86_perf_counter_update(counter, &counter->hw, counter->hw.idx);
718 perf_store_irq_data(sibling, counter->hw_event.event_config);
719 perf_store_irq_data(sibling, atomic64_read(&counter->count));
720 }
721}
722
723/* 693/*
724 * Maximum interrupt frequency of 100KHz per CPU 694 * Maximum interrupt frequency of 100KHz per CPU
725 */ 695 */
@@ -754,28 +724,7 @@ again:
754 continue; 724 continue;
755 725
756 perf_save_and_restart(counter); 726 perf_save_and_restart(counter);
757 727 perf_counter_output(counter, nmi, regs);
758 switch (counter->hw_event.record_type) {
759 case PERF_RECORD_SIMPLE:
760 continue;
761 case PERF_RECORD_IRQ:
762 perf_store_irq_data(counter, instruction_pointer(regs));
763 break;
764 case PERF_RECORD_GROUP:
765 perf_handle_group(counter, &status, &ack);
766 break;
767 }
768 /*
769 * From NMI context we cannot call into the scheduler to
770 * do a task wakeup - but we mark these generic as
771 * wakeup_pending and initate a wakeup callback:
772 */
773 if (nmi) {
774 counter->wakeup_pending = 1;
775 set_tsk_thread_flag(current, TIF_PERF_COUNTERS);
776 } else {
777 wake_up(&counter->waitq);
778 }
779 } 728 }
780 729
781 hw_perf_ack_status(ack); 730 hw_perf_ack_status(ack);