aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/oprofile
diff options
context:
space:
mode:
Diffstat (limited to 'arch/powerpc/oprofile')
-rw-r--r--arch/powerpc/oprofile/cell/pr_util.h7
-rw-r--r--arch/powerpc/oprofile/cell/spu_profiler.c34
-rw-r--r--arch/powerpc/oprofile/common.c22
-rw-r--r--arch/powerpc/oprofile/op_model_cell.c490
4 files changed, 537 insertions, 16 deletions
diff --git a/arch/powerpc/oprofile/cell/pr_util.h b/arch/powerpc/oprofile/cell/pr_util.h
index bca7207bd92a..a048b0b72be3 100644
--- a/arch/powerpc/oprofile/cell/pr_util.h
+++ b/arch/powerpc/oprofile/cell/pr_util.h
@@ -30,6 +30,10 @@
30extern struct delayed_work spu_work; 30extern struct delayed_work spu_work;
31extern int spu_prof_running; 31extern int spu_prof_running;
32 32
33#define TRACE_ARRAY_SIZE 1024
34
35extern spinlock_t oprof_spu_smpl_arry_lck;
36
33struct spu_overlay_info { /* map of sections within an SPU overlay */ 37struct spu_overlay_info { /* map of sections within an SPU overlay */
34 unsigned int vma; /* SPU virtual memory address from elf */ 38 unsigned int vma; /* SPU virtual memory address from elf */
35 unsigned int size; /* size of section from elf */ 39 unsigned int size; /* size of section from elf */
@@ -90,9 +94,10 @@ void vma_map_free(struct vma_to_fileoffset_map *map);
90 * cycles_reset is the SPU_CYCLES count value specified by the user. 94 * cycles_reset is the SPU_CYCLES count value specified by the user.
91 */ 95 */
92int start_spu_profiling_cycles(unsigned int cycles_reset); 96int start_spu_profiling_cycles(unsigned int cycles_reset);
97void start_spu_profiling_events(void);
93 98
94void stop_spu_profiling_cycles(void); 99void stop_spu_profiling_cycles(void);
95 100void stop_spu_profiling_events(void);
96 101
97/* add the necessary profiling hooks */ 102/* add the necessary profiling hooks */
98int spu_sync_start(void); 103int spu_sync_start(void);
diff --git a/arch/powerpc/oprofile/cell/spu_profiler.c b/arch/powerpc/oprofile/cell/spu_profiler.c
index 8b1b9ccaff9f..de170b7ae71b 100644
--- a/arch/powerpc/oprofile/cell/spu_profiler.c
+++ b/arch/powerpc/oprofile/cell/spu_profiler.c
@@ -18,11 +18,21 @@
18#include <asm/cell-pmu.h> 18#include <asm/cell-pmu.h>
19#include "pr_util.h" 19#include "pr_util.h"
20 20
21#define TRACE_ARRAY_SIZE 1024
22#define SCALE_SHIFT 14 21#define SCALE_SHIFT 14
23 22
24static u32 *samples; 23static u32 *samples;
25 24
25/* spu_prof_running is a flag used to indicate if spu profiling is enabled
26 * or not. It is set by the routines start_spu_profiling_cycles() and
27 * start_spu_profiling_events(). The flag is cleared by the routines
28 * stop_spu_profiling_cycles() and stop_spu_profiling_events(). These
29 * routines are called via global_start() and global_stop() which are called in
30 * op_powerpc_start() and op_powerpc_stop(). These routines are called once
31 * per system as a result of the user starting/stopping oprofile. Hence, only
32 * one CPU per user at a time will be changing the value of spu_prof_running.
33 * In general, OProfile does not protect against multiple users trying to run
34 * OProfile at a time.
35 */
26int spu_prof_running; 36int spu_prof_running;
27static unsigned int profiling_interval; 37static unsigned int profiling_interval;
28 38
@@ -31,7 +41,7 @@ static unsigned int profiling_interval;
31 41
32#define SPU_PC_MASK 0xFFFF 42#define SPU_PC_MASK 0xFFFF
33 43
34static DEFINE_SPINLOCK(oprof_spu_smpl_arry_lck); 44DEFINE_SPINLOCK(oprof_spu_smpl_arry_lck);
35unsigned long oprof_spu_smpl_arry_lck_flags; 45unsigned long oprof_spu_smpl_arry_lck_flags;
36 46
37void set_spu_profiling_frequency(unsigned int freq_khz, unsigned int cycles_reset) 47void set_spu_profiling_frequency(unsigned int freq_khz, unsigned int cycles_reset)
@@ -212,6 +222,21 @@ int start_spu_profiling_cycles(unsigned int cycles_reset)
212 return 0; 222 return 0;
213} 223}
214 224
225/*
226 * Entry point for SPU event profiling.
227 * NOTE: SPU profiling is done system-wide, not per-CPU.
228 *
229 * cycles_reset is the count value specified by the user when
230 * setting up OProfile to count SPU_CYCLES.
231 */
232void start_spu_profiling_events(void)
233{
234 spu_prof_running = 1;
235 schedule_delayed_work(&spu_work, DEFAULT_TIMER_EXPIRE);
236
237 return;
238}
239
215void stop_spu_profiling_cycles(void) 240void stop_spu_profiling_cycles(void)
216{ 241{
217 spu_prof_running = 0; 242 spu_prof_running = 0;
@@ -219,3 +244,8 @@ void stop_spu_profiling_cycles(void)
219 kfree(samples); 244 kfree(samples);
220 pr_debug("SPU_PROF: stop_spu_profiling_cycles issued\n"); 245 pr_debug("SPU_PROF: stop_spu_profiling_cycles issued\n");
221} 246}
247
248void stop_spu_profiling_events(void)
249{
250 spu_prof_running = 0;
251}
diff --git a/arch/powerpc/oprofile/common.c b/arch/powerpc/oprofile/common.c
index 17807acb05d9..21f16edf6c8d 100644
--- a/arch/powerpc/oprofile/common.c
+++ b/arch/powerpc/oprofile/common.c
@@ -132,6 +132,28 @@ static int op_powerpc_create_files(struct super_block *sb, struct dentry *root)
132 oprofilefs_create_ulong(sb, root, "mmcr0", &sys.mmcr0); 132 oprofilefs_create_ulong(sb, root, "mmcr0", &sys.mmcr0);
133 oprofilefs_create_ulong(sb, root, "mmcr1", &sys.mmcr1); 133 oprofilefs_create_ulong(sb, root, "mmcr1", &sys.mmcr1);
134 oprofilefs_create_ulong(sb, root, "mmcra", &sys.mmcra); 134 oprofilefs_create_ulong(sb, root, "mmcra", &sys.mmcra);
135#ifdef CONFIG_OPROFILE_CELL
136 /* create a file the user tool can check to see what level of profiling
137 * support exits with this kernel. Initialize bit mask to indicate
138 * what support the kernel has:
139 * bit 0 - Supports SPU event profiling in addition to PPU
140 * event and cycles; and SPU cycle profiling
141 * bits 1-31 - Currently unused.
142 *
143 * If the file does not exist, then the kernel only supports SPU
144 * cycle profiling, PPU event and cycle profiling.
145 */
146 oprofilefs_create_ulong(sb, root, "cell_support", &sys.cell_support);
147 sys.cell_support = 0x1; /* Note, the user OProfile tool must check
148 * that this bit is set before attempting to
149 * user SPU event profiling. Older kernels
150 * will not have this file, hence the user
151 * tool is not allowed to do SPU event
152 * profiling on older kernels. Older kernels
153 * will accept SPU events but collected data
154 * is garbage.
155 */
156#endif
135#endif 157#endif
136 158
137 for (i = 0; i < model->num_counters; ++i) { 159 for (i = 0; i < model->num_counters; ++i) {
diff --git a/arch/powerpc/oprofile/op_model_cell.c b/arch/powerpc/oprofile/op_model_cell.c
index ad7f32c848f8..ff96cbfb89bb 100644
--- a/arch/powerpc/oprofile/op_model_cell.c
+++ b/arch/powerpc/oprofile/op_model_cell.c
@@ -44,6 +44,12 @@
44#define SPU_PROFILING_CYCLES 1 44#define SPU_PROFILING_CYCLES 1
45#define SPU_PROFILING_EVENTS 2 45#define SPU_PROFILING_EVENTS 2
46 46
47#define SPU_EVENT_NUM_START 4100
48#define SPU_EVENT_NUM_STOP 4399
49#define SPU_PROFILE_EVENT_ADDR 4363 /* spu, address trace, decimal */
50#define SPU_PROFILE_EVENT_ADDR_MASK_A 0x146 /* sub unit set to zero */
51#define SPU_PROFILE_EVENT_ADDR_MASK_B 0x186 /* sub unit set to zero */
52
47#define NUM_SPUS_PER_NODE 8 53#define NUM_SPUS_PER_NODE 8
48#define SPU_CYCLES_EVENT_NUM 2 /* event number for SPU_CYCLES */ 54#define SPU_CYCLES_EVENT_NUM 2 /* event number for SPU_CYCLES */
49 55
@@ -61,6 +67,12 @@
61 67
62#define MAX_SPU_COUNT 0xFFFFFF /* maximum 24 bit LFSR value */ 68#define MAX_SPU_COUNT 0xFFFFFF /* maximum 24 bit LFSR value */
63 69
70/* Minumum HW interval timer setting to send value to trace buffer is 10 cycle.
71 * To configure counter to send value every N cycles set counter to
72 * 2^32 - 1 - N.
73 */
74#define NUM_INTERVAL_CYC 0xFFFFFFFF - 10
75
64/* 76/*
65 * spu_cycle_reset is the number of cycles between samples. 77 * spu_cycle_reset is the number of cycles between samples.
66 * This variable is used for SPU profiling and should ONLY be set 78 * This variable is used for SPU profiling and should ONLY be set
@@ -68,6 +80,7 @@
68 */ 80 */
69static unsigned int spu_cycle_reset; 81static unsigned int spu_cycle_reset;
70static unsigned int profiling_mode; 82static unsigned int profiling_mode;
83static int spu_evnt_phys_spu_indx;
71 84
72struct pmc_cntrl_data { 85struct pmc_cntrl_data {
73 unsigned long vcntr; 86 unsigned long vcntr;
@@ -108,6 +121,8 @@ struct pm_cntrl {
108 u16 trace_mode; 121 u16 trace_mode;
109 u16 freeze; 122 u16 freeze;
110 u16 count_mode; 123 u16 count_mode;
124 u16 spu_addr_trace;
125 u8 trace_buf_ovflw;
111}; 126};
112 127
113static struct { 128static struct {
@@ -125,6 +140,7 @@ static struct {
125#define GET_INPUT_CONTROL(x) ((x & 0x00000004) >> 2) 140#define GET_INPUT_CONTROL(x) ((x & 0x00000004) >> 2)
126 141
127static DEFINE_PER_CPU(unsigned long[NR_PHYS_CTRS], pmc_values); 142static DEFINE_PER_CPU(unsigned long[NR_PHYS_CTRS], pmc_values);
143static unsigned long spu_pm_cnt[MAX_NUMNODES * NUM_SPUS_PER_NODE];
128static struct pmc_cntrl_data pmc_cntrl[NUM_THREADS][NR_PHYS_CTRS]; 144static struct pmc_cntrl_data pmc_cntrl[NUM_THREADS][NR_PHYS_CTRS];
129 145
130/* 146/*
@@ -154,6 +170,7 @@ static u32 hdw_thread;
154 170
155static u32 virt_cntr_inter_mask; 171static u32 virt_cntr_inter_mask;
156static struct timer_list timer_virt_cntr; 172static struct timer_list timer_virt_cntr;
173static struct timer_list timer_spu_event_swap;
157 174
158/* 175/*
159 * pm_signal needs to be global since it is initialized in 176 * pm_signal needs to be global since it is initialized in
@@ -372,9 +389,13 @@ static void write_pm_cntrl(int cpu)
372 if (pm_regs.pm_cntrl.trace_mode != 0) 389 if (pm_regs.pm_cntrl.trace_mode != 0)
373 val |= CBE_PM_TRACE_MODE_SET(pm_regs.pm_cntrl.trace_mode); 390 val |= CBE_PM_TRACE_MODE_SET(pm_regs.pm_cntrl.trace_mode);
374 391
392 if (pm_regs.pm_cntrl.trace_buf_ovflw == 1)
393 val |= CBE_PM_TRACE_BUF_OVFLW(pm_regs.pm_cntrl.trace_buf_ovflw);
375 if (pm_regs.pm_cntrl.freeze == 1) 394 if (pm_regs.pm_cntrl.freeze == 1)
376 val |= CBE_PM_FREEZE_ALL_CTRS; 395 val |= CBE_PM_FREEZE_ALL_CTRS;
377 396
397 val |= CBE_PM_SPU_ADDR_TRACE_SET(pm_regs.pm_cntrl.spu_addr_trace);
398
378 /* 399 /*
379 * Routine set_count_mode must be called previously to set 400 * Routine set_count_mode must be called previously to set
380 * the count mode based on the user selection of user and kernel. 401 * the count mode based on the user selection of user and kernel.
@@ -563,9 +584,184 @@ static int cell_reg_setup_spu_cycles(struct op_counter_config *ctr,
563 return 0; 584 return 0;
564} 585}
565 586
587/* Unfortunately, the hardware will only support event profiling
588 * on one SPU per node at a time. Therefore, we must time slice
589 * the profiling across all SPUs in the node. Note, we do this
590 * in parallel for each node. The following routine is called
591 * periodically based on kernel timer to switch which SPU is
592 * being monitored in a round robbin fashion.
593 */
594static void spu_evnt_swap(unsigned long data)
595{
596 int node;
597 int cur_phys_spu, nxt_phys_spu, cur_spu_evnt_phys_spu_indx;
598 unsigned long flags;
599 int cpu;
600 int ret;
601 u32 interrupt_mask;
602
603
604 /* enable interrupts on cntr 0 */
605 interrupt_mask = CBE_PM_CTR_OVERFLOW_INTR(0);
606
607 hdw_thread = 0;
608
609 /* Make sure spu event interrupt handler and spu event swap
610 * don't access the counters simultaneously.
611 */
612 spin_lock_irqsave(&cntr_lock, flags);
613
614 cur_spu_evnt_phys_spu_indx = spu_evnt_phys_spu_indx;
615
616 if (++(spu_evnt_phys_spu_indx) == NUM_SPUS_PER_NODE)
617 spu_evnt_phys_spu_indx = 0;
618
619 pm_signal[0].sub_unit = spu_evnt_phys_spu_indx;
620 pm_signal[1].sub_unit = spu_evnt_phys_spu_indx;
621 pm_signal[2].sub_unit = spu_evnt_phys_spu_indx;
622
623 /* switch the SPU being profiled on each node */
624 for_each_online_cpu(cpu) {
625 if (cbe_get_hw_thread_id(cpu))
626 continue;
627
628 node = cbe_cpu_to_node(cpu);
629 cur_phys_spu = (node * NUM_SPUS_PER_NODE)
630 + cur_spu_evnt_phys_spu_indx;
631 nxt_phys_spu = (node * NUM_SPUS_PER_NODE)
632 + spu_evnt_phys_spu_indx;
633
634 /*
635 * stop counters, save counter values, restore counts
636 * for previous physical SPU
637 */
638 cbe_disable_pm(cpu);
639 cbe_disable_pm_interrupts(cpu);
640
641 spu_pm_cnt[cur_phys_spu]
642 = cbe_read_ctr(cpu, 0);
643
644 /* restore previous count for the next spu to sample */
645 /* NOTE, hardware issue, counter will not start if the
646 * counter value is at max (0xFFFFFFFF).
647 */
648 if (spu_pm_cnt[nxt_phys_spu] >= 0xFFFFFFFF)
649 cbe_write_ctr(cpu, 0, 0xFFFFFFF0);
650 else
651 cbe_write_ctr(cpu, 0, spu_pm_cnt[nxt_phys_spu]);
652
653 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
654
655 /* setup the debug bus measure the one event and
656 * the two events to route the next SPU's PC on
657 * the debug bus
658 */
659 ret = pm_rtas_activate_signals(cbe_cpu_to_node(cpu), 3);
660 if (ret)
661 printk(KERN_ERR
662 "%s: pm_rtas_activate_signals failed, SPU event swap\n",
663 __func__);
664
665 /* clear the trace buffer, don't want to take PC for
666 * previous SPU*/
667 cbe_write_pm(cpu, trace_address, 0);
668
669 enable_ctr(cpu, 0, pm_regs.pm07_cntrl);
670
671 /* Enable interrupts on the CPU thread that is starting */
672 cbe_enable_pm_interrupts(cpu, hdw_thread,
673 interrupt_mask);
674 cbe_enable_pm(cpu);
675 }
676
677 spin_unlock_irqrestore(&cntr_lock, flags);
678
679 /* swap approximately every 0.1 seconds */
680 mod_timer(&timer_spu_event_swap, jiffies + HZ / 25);
681}
682
683static void start_spu_event_swap(void)
684{
685 init_timer(&timer_spu_event_swap);
686 timer_spu_event_swap.function = spu_evnt_swap;
687 timer_spu_event_swap.data = 0UL;
688 timer_spu_event_swap.expires = jiffies + HZ / 25;
689 add_timer(&timer_spu_event_swap);
690}
691
692static int cell_reg_setup_spu_events(struct op_counter_config *ctr,
693 struct op_system_config *sys, int num_ctrs)
694{
695 int i;
696
697 /* routine is called once for all nodes */
698
699 spu_evnt_phys_spu_indx = 0;
700 /*
701 * For all events except PPU CYCLEs, each node will need to make
702 * the rtas cbe-perftools call to setup and reset the debug bus.
703 * Make the token lookup call once and store it in the global
704 * variable pm_rtas_token.
705 */
706 pm_rtas_token = rtas_token("ibm,cbe-perftools");
707
708 if (unlikely(pm_rtas_token == RTAS_UNKNOWN_SERVICE)) {
709 printk(KERN_ERR
710 "%s: rtas token ibm,cbe-perftools unknown\n",
711 __func__);
712 return -EIO;
713 }
714
715 /* setup the pm_control register settings,
716 * settings will be written per node by the
717 * cell_cpu_setup() function.
718 */
719 pm_regs.pm_cntrl.trace_buf_ovflw = 1;
720
721 /* Use the occurrence trace mode to have SPU PC saved
722 * to the trace buffer. Occurrence data in trace buffer
723 * is not used. Bit 2 must be set to store SPU addresses.
724 */
725 pm_regs.pm_cntrl.trace_mode = 2;
726
727 pm_regs.pm_cntrl.spu_addr_trace = 0x1; /* using debug bus
728 event 2 & 3 */
729
730 /* setup the debug bus event array with the SPU PC routing events.
731 * Note, pm_signal[0] will be filled in by set_pm_event() call below.
732 */
733 pm_signal[1].signal_group = SPU_PROFILE_EVENT_ADDR / 100;
734 pm_signal[1].bus_word = GET_BUS_WORD(SPU_PROFILE_EVENT_ADDR_MASK_A);
735 pm_signal[1].bit = SPU_PROFILE_EVENT_ADDR % 100;
736 pm_signal[1].sub_unit = spu_evnt_phys_spu_indx;
737
738 pm_signal[2].signal_group = SPU_PROFILE_EVENT_ADDR / 100;
739 pm_signal[2].bus_word = GET_BUS_WORD(SPU_PROFILE_EVENT_ADDR_MASK_B);
740 pm_signal[2].bit = SPU_PROFILE_EVENT_ADDR % 100;
741 pm_signal[2].sub_unit = spu_evnt_phys_spu_indx;
742
743 /* Set the user selected spu event to profile on,
744 * note, only one SPU profiling event is supported
745 */
746 num_counters = 1; /* Only support one SPU event at a time */
747 set_pm_event(0, ctr[0].event, ctr[0].unit_mask);
748
749 reset_value[0] = 0xFFFFFFFF - ctr[0].count;
750
751 /* global, used by cell_cpu_setup */
752 ctr_enabled |= 1;
753
754 /* Initialize the count for each SPU to the reset value */
755 for (i=0; i < MAX_NUMNODES * NUM_SPUS_PER_NODE; i++)
756 spu_pm_cnt[i] = reset_value[0];
757
758 return 0;
759}
760
566static int cell_reg_setup_ppu(struct op_counter_config *ctr, 761static int cell_reg_setup_ppu(struct op_counter_config *ctr,
567 struct op_system_config *sys, int num_ctrs) 762 struct op_system_config *sys, int num_ctrs)
568{ 763{
764 /* routine is called once for all nodes */
569 int i, j, cpu; 765 int i, j, cpu;
570 766
571 num_counters = num_ctrs; 767 num_counters = num_ctrs;
@@ -577,14 +773,6 @@ static int cell_reg_setup_ppu(struct op_counter_config *ctr,
577 __func__); 773 __func__);
578 return -EIO; 774 return -EIO;
579 } 775 }
580 pm_regs.group_control = 0;
581 pm_regs.debug_bus_control = 0;
582
583 /* setup the pm_control register */
584 memset(&pm_regs.pm_cntrl, 0, sizeof(struct pm_cntrl));
585 pm_regs.pm_cntrl.stop_at_max = 1;
586 pm_regs.pm_cntrl.trace_mode = 0;
587 pm_regs.pm_cntrl.freeze = 1;
588 776
589 set_count_mode(sys->enable_kernel, sys->enable_user); 777 set_count_mode(sys->enable_kernel, sys->enable_user);
590 778
@@ -657,10 +845,20 @@ static int cell_reg_setup_ppu(struct op_counter_config *ctr,
657static int cell_reg_setup(struct op_counter_config *ctr, 845static int cell_reg_setup(struct op_counter_config *ctr,
658 struct op_system_config *sys, int num_ctrs) 846 struct op_system_config *sys, int num_ctrs)
659{ 847{
660 int ret; 848 int ret=0;
661
662 spu_cycle_reset = 0; 849 spu_cycle_reset = 0;
663 850
851 /* initialize the spu_arr_trace value, will be reset if
852 * doing spu event profiling.
853 */
854 pm_regs.group_control = 0;
855 pm_regs.debug_bus_control = 0;
856 pm_regs.pm_cntrl.stop_at_max = 1;
857 pm_regs.pm_cntrl.trace_mode = 0;
858 pm_regs.pm_cntrl.freeze = 1;
859 pm_regs.pm_cntrl.trace_buf_ovflw = 0;
860 pm_regs.pm_cntrl.spu_addr_trace = 0;
861
664 /* 862 /*
665 * For all events except PPU CYCLEs, each node will need to make 863 * For all events except PPU CYCLEs, each node will need to make
666 * the rtas cbe-perftools call to setup and reset the debug bus. 864 * the rtas cbe-perftools call to setup and reset the debug bus.
@@ -679,6 +877,18 @@ static int cell_reg_setup(struct op_counter_config *ctr,
679 if (ctr[0].event == SPU_CYCLES_EVENT_NUM) { 877 if (ctr[0].event == SPU_CYCLES_EVENT_NUM) {
680 profiling_mode = SPU_PROFILING_CYCLES; 878 profiling_mode = SPU_PROFILING_CYCLES;
681 ret = cell_reg_setup_spu_cycles(ctr, sys, num_ctrs); 879 ret = cell_reg_setup_spu_cycles(ctr, sys, num_ctrs);
880 } else if ((ctr[0].event >= SPU_EVENT_NUM_START) &&
881 (ctr[0].event <= SPU_EVENT_NUM_STOP)) {
882 profiling_mode = SPU_PROFILING_EVENTS;
883 spu_cycle_reset = ctr[0].count;
884
885 /* for SPU event profiling, need to setup the
886 * pm_signal array with the events to route the
887 * SPU PC before making the FW call. Note, only
888 * one SPU event for profiling can be specified
889 * at a time.
890 */
891 cell_reg_setup_spu_events(ctr, sys, num_ctrs);
682 } else { 892 } else {
683 profiling_mode = PPU_PROFILING; 893 profiling_mode = PPU_PROFILING;
684 ret = cell_reg_setup_ppu(ctr, sys, num_ctrs); 894 ret = cell_reg_setup_ppu(ctr, sys, num_ctrs);
@@ -695,6 +905,7 @@ static int cell_cpu_setup(struct op_counter_config *cntr)
695 u32 cpu = smp_processor_id(); 905 u32 cpu = smp_processor_id();
696 u32 num_enabled = 0; 906 u32 num_enabled = 0;
697 int i; 907 int i;
908 int ret;
698 909
699 /* Cycle based SPU profiling does not use the performance 910 /* Cycle based SPU profiling does not use the performance
700 * counters. The trace array is configured to collect 911 * counters. The trace array is configured to collect
@@ -729,7 +940,20 @@ static int cell_cpu_setup(struct op_counter_config *cntr)
729 * The pm_rtas_activate_signals will return -EIO if the FW 940 * The pm_rtas_activate_signals will return -EIO if the FW
730 * call failed. 941 * call failed.
731 */ 942 */
732 return pm_rtas_activate_signals(cbe_cpu_to_node(cpu), num_enabled); 943 if (profiling_mode == SPU_PROFILING_EVENTS) {
944 /* For SPU event profiling also need to setup the
945 * pm interval timer
946 */
947 ret = pm_rtas_activate_signals(cbe_cpu_to_node(cpu),
948 num_enabled+2);
949 /* store PC from debug bus to Trace buffer as often
950 * as possible (every 10 cycles)
951 */
952 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
953 return ret;
954 } else
955 return pm_rtas_activate_signals(cbe_cpu_to_node(cpu),
956 num_enabled);
733} 957}
734 958
735#define ENTRIES 303 959#define ENTRIES 303
@@ -926,6 +1150,7 @@ static void cell_global_stop_spu_cycles(void)
926 int cpu; 1150 int cpu;
927 1151
928 oprofile_running = 0; 1152 oprofile_running = 0;
1153 smp_wmb();
929 1154
930#ifdef CONFIG_CPU_FREQ 1155#ifdef CONFIG_CPU_FREQ
931 cpufreq_unregister_notifier(&cpu_freq_notifier_block, 1156 cpufreq_unregister_notifier(&cpu_freq_notifier_block,
@@ -957,8 +1182,33 @@ static void cell_global_stop_spu_cycles(void)
957 pm_rtas_reset_signals(cbe_cpu_to_node(cpu)); 1182 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
958 } 1183 }
959 1184
960 if (profiling_mode == SPU_PROFILING_CYCLES) 1185 stop_spu_profiling_cycles();
961 stop_spu_profiling_cycles(); 1186}
1187
1188static void cell_global_stop_spu_events(void)
1189{
1190 int cpu;
1191 oprofile_running = 0;
1192
1193 stop_spu_profiling_events();
1194 smp_wmb();
1195
1196 for_each_online_cpu(cpu) {
1197 if (cbe_get_hw_thread_id(cpu))
1198 continue;
1199
1200 cbe_sync_irq(cbe_cpu_to_node(cpu));
1201 /* Stop the counters */
1202 cbe_disable_pm(cpu);
1203 cbe_write_pm07_control(cpu, 0, 0);
1204
1205 /* Deactivate the signals */
1206 pm_rtas_reset_signals(cbe_cpu_to_node(cpu));
1207
1208 /* Deactivate interrupts */
1209 cbe_disable_pm_interrupts(cpu);
1210 }
1211 del_timer_sync(&timer_spu_event_swap);
962} 1212}
963 1213
964static void cell_global_stop_ppu(void) 1214static void cell_global_stop_ppu(void)
@@ -994,6 +1244,8 @@ static void cell_global_stop(void)
994{ 1244{
995 if (profiling_mode == PPU_PROFILING) 1245 if (profiling_mode == PPU_PROFILING)
996 cell_global_stop_ppu(); 1246 cell_global_stop_ppu();
1247 else if (profiling_mode == SPU_PROFILING_EVENTS)
1248 cell_global_stop_spu_events();
997 else 1249 else
998 cell_global_stop_spu_cycles(); 1250 cell_global_stop_spu_cycles();
999} 1251}
@@ -1088,6 +1340,69 @@ out:
1088 return rtas_error; 1340 return rtas_error;
1089} 1341}
1090 1342
1343static int cell_global_start_spu_events(struct op_counter_config *ctr)
1344{
1345 int cpu;
1346 u32 interrupt_mask = 0;
1347 int rtn = 0;
1348
1349 hdw_thread = 0;
1350
1351 /* spu event profiling, uses the performance counters to generate
1352 * an interrupt. The hardware is setup to store the SPU program
1353 * counter into the trace array. The occurrence mode is used to
1354 * enable storing data to the trace buffer. The bits are set
1355 * to send/store the SPU address in the trace buffer. The debug
1356 * bus must be setup to route the SPU program counter onto the
1357 * debug bus. The occurrence data in the trace buffer is not used.
1358 */
1359
1360 /* This routine gets called once for the system.
1361 * There is one performance monitor per node, so we
1362 * only need to perform this function once per node.
1363 */
1364
1365 for_each_online_cpu(cpu) {
1366 if (cbe_get_hw_thread_id(cpu))
1367 continue;
1368
1369 /*
1370 * Setup SPU event-based profiling.
1371 * Set perf_mon_control bit 0 to a zero before
1372 * enabling spu collection hardware.
1373 *
1374 * Only support one SPU event on one SPU per node.
1375 */
1376 if (ctr_enabled & 1) {
1377 cbe_write_ctr(cpu, 0, reset_value[0]);
1378 enable_ctr(cpu, 0, pm_regs.pm07_cntrl);
1379 interrupt_mask |=
1380 CBE_PM_CTR_OVERFLOW_INTR(0);
1381 } else {
1382 /* Disable counter */
1383 cbe_write_pm07_control(cpu, 0, 0);
1384 }
1385
1386 cbe_get_and_clear_pm_interrupts(cpu);
1387 cbe_enable_pm_interrupts(cpu, hdw_thread, interrupt_mask);
1388 cbe_enable_pm(cpu);
1389
1390 /* clear the trace buffer */
1391 cbe_write_pm(cpu, trace_address, 0);
1392 }
1393
1394 /* Start the timer to time slice collecting the event profile
1395 * on each of the SPUs. Note, can collect profile on one SPU
1396 * per node at a time.
1397 */
1398 start_spu_event_swap();
1399 start_spu_profiling_events();
1400 oprofile_running = 1;
1401 smp_wmb();
1402
1403 return rtn;
1404}
1405
1091static int cell_global_start_ppu(struct op_counter_config *ctr) 1406static int cell_global_start_ppu(struct op_counter_config *ctr)
1092{ 1407{
1093 u32 cpu, i; 1408 u32 cpu, i;
@@ -1139,11 +1454,158 @@ static int cell_global_start(struct op_counter_config *ctr)
1139{ 1454{
1140 if (profiling_mode == SPU_PROFILING_CYCLES) 1455 if (profiling_mode == SPU_PROFILING_CYCLES)
1141 return cell_global_start_spu_cycles(ctr); 1456 return cell_global_start_spu_cycles(ctr);
1457 else if (profiling_mode == SPU_PROFILING_EVENTS)
1458 return cell_global_start_spu_events(ctr);
1142 else 1459 else
1143 return cell_global_start_ppu(ctr); 1460 return cell_global_start_ppu(ctr);
1144} 1461}
1145 1462
1146 1463
1464/* The SPU interrupt handler
1465 *
1466 * SPU event profiling works as follows:
1467 * The pm_signal[0] holds the one SPU event to be measured. It is routed on
1468 * the debug bus using word 0 or 1. The value of pm_signal[1] and
1469 * pm_signal[2] contain the necessary events to route the SPU program
1470 * counter for the selected SPU onto the debug bus using words 2 and 3.
1471 * The pm_interval register is setup to write the SPU PC value into the
1472 * trace buffer at the maximum rate possible. The trace buffer is configured
1473 * to store the PCs, wrapping when it is full. The performance counter is
1474 * intialized to the max hardware count minus the number of events, N, between
1475 * samples. Once the N events have occured, a HW counter overflow occurs
1476 * causing the generation of a HW counter interrupt which also stops the
1477 * writing of the SPU PC values to the trace buffer. Hence the last PC
1478 * written to the trace buffer is the SPU PC that we want. Unfortunately,
1479 * we have to read from the beginning of the trace buffer to get to the
1480 * last value written. We just hope the PPU has nothing better to do then
1481 * service this interrupt. The PC for the specific SPU being profiled is
1482 * extracted from the trace buffer processed and stored. The trace buffer
1483 * is cleared, interrupts are cleared, the counter is reset to max - N.
1484 * A kernel timer is used to periodically call the routine spu_evnt_swap()
1485 * to switch to the next physical SPU in the node to profile in round robbin
1486 * order. This way data is collected for all SPUs on the node. It does mean
1487 * that we need to use a relatively small value of N to ensure enough samples
1488 * on each SPU are collected each SPU is being profiled 1/8 of the time.
1489 * It may also be necessary to use a longer sample collection period.
1490 */
1491static void cell_handle_interrupt_spu(struct pt_regs *regs,
1492 struct op_counter_config *ctr)
1493{
1494 u32 cpu, cpu_tmp;
1495 u64 trace_entry;
1496 u32 interrupt_mask;
1497 u64 trace_buffer[2];
1498 u64 last_trace_buffer;
1499 u32 sample;
1500 u32 trace_addr;
1501 unsigned long sample_array_lock_flags;
1502 int spu_num;
1503 unsigned long flags;
1504
1505 /* Make sure spu event interrupt handler and spu event swap
1506 * don't access the counters simultaneously.
1507 */
1508 cpu = smp_processor_id();
1509 spin_lock_irqsave(&cntr_lock, flags);
1510
1511 cpu_tmp = cpu;
1512 cbe_disable_pm(cpu);
1513
1514 interrupt_mask = cbe_get_and_clear_pm_interrupts(cpu);
1515
1516 sample = 0xABCDEF;
1517 trace_entry = 0xfedcba;
1518 last_trace_buffer = 0xdeadbeaf;
1519
1520 if ((oprofile_running == 1) && (interrupt_mask != 0)) {
1521 /* disable writes to trace buff */
1522 cbe_write_pm(cpu, pm_interval, 0);
1523
1524 /* only have one perf cntr being used, cntr 0 */
1525 if ((interrupt_mask & CBE_PM_CTR_OVERFLOW_INTR(0))
1526 && ctr[0].enabled)
1527 /* The SPU PC values will be read
1528 * from the trace buffer, reset counter
1529 */
1530
1531 cbe_write_ctr(cpu, 0, reset_value[0]);
1532
1533 trace_addr = cbe_read_pm(cpu, trace_address);
1534
1535 while (!(trace_addr & CBE_PM_TRACE_BUF_EMPTY)) {
1536 /* There is data in the trace buffer to process
1537 * Read the buffer until you get to the last
1538 * entry. This is the value we want.
1539 */
1540
1541 cbe_read_trace_buffer(cpu, trace_buffer);
1542 trace_addr = cbe_read_pm(cpu, trace_address);
1543 }
1544
1545 /* SPU Address 16 bit count format for 128 bit
1546 * HW trace buffer is used for the SPU PC storage
1547 * HDR bits 0:15
1548 * SPU Addr 0 bits 16:31
1549 * SPU Addr 1 bits 32:47
1550 * unused bits 48:127
1551 *
1552 * HDR: bit4 = 1 SPU Address 0 valid
1553 * HDR: bit5 = 1 SPU Address 1 valid
1554 * - unfortunately, the valid bits don't seem to work
1555 *
1556 * Note trace_buffer[0] holds bits 0:63 of the HW
1557 * trace buffer, trace_buffer[1] holds bits 64:127
1558 */
1559
1560 trace_entry = trace_buffer[0]
1561 & 0x00000000FFFF0000;
1562
1563 /* only top 16 of the 18 bit SPU PC address
1564 * is stored in trace buffer, hence shift right
1565 * by 16 -2 bits */
1566 sample = trace_entry >> 14;
1567 last_trace_buffer = trace_buffer[0];
1568
1569 spu_num = spu_evnt_phys_spu_indx
1570 + (cbe_cpu_to_node(cpu) * NUM_SPUS_PER_NODE);
1571
1572 /* make sure only one process at a time is calling
1573 * spu_sync_buffer()
1574 */
1575 spin_lock_irqsave(&oprof_spu_smpl_arry_lck,
1576 sample_array_lock_flags);
1577 spu_sync_buffer(spu_num, &sample, 1);
1578 spin_unlock_irqrestore(&oprof_spu_smpl_arry_lck,
1579 sample_array_lock_flags);
1580
1581 smp_wmb(); /* insure spu event buffer updates are written
1582 * don't want events intermingled... */
1583
1584 /* The counters were frozen by the interrupt.
1585 * Reenable the interrupt and restart the counters.
1586 */
1587 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
1588 cbe_enable_pm_interrupts(cpu, hdw_thread,
1589 virt_cntr_inter_mask);
1590
1591 /* clear the trace buffer, re-enable writes to trace buff */
1592 cbe_write_pm(cpu, trace_address, 0);
1593 cbe_write_pm(cpu, pm_interval, NUM_INTERVAL_CYC);
1594
1595 /* The writes to the various performance counters only writes
1596 * to a latch. The new values (interrupt setting bits, reset
1597 * counter value etc.) are not copied to the actual registers
1598 * until the performance monitor is enabled. In order to get
1599 * this to work as desired, the permormance monitor needs to
1600 * be disabled while writing to the latches. This is a
1601 * HW design issue.
1602 */
1603 write_pm_cntrl(cpu);
1604 cbe_enable_pm(cpu);
1605 }
1606 spin_unlock_irqrestore(&cntr_lock, flags);
1607}
1608
1147static void cell_handle_interrupt_ppu(struct pt_regs *regs, 1609static void cell_handle_interrupt_ppu(struct pt_regs *regs,
1148 struct op_counter_config *ctr) 1610 struct op_counter_config *ctr)
1149{ 1611{
@@ -1222,6 +1684,8 @@ static void cell_handle_interrupt(struct pt_regs *regs,
1222{ 1684{
1223 if (profiling_mode == PPU_PROFILING) 1685 if (profiling_mode == PPU_PROFILING)
1224 cell_handle_interrupt_ppu(regs, ctr); 1686 cell_handle_interrupt_ppu(regs, ctr);
1687 else
1688 cell_handle_interrupt_spu(regs, ctr);
1225} 1689}
1226 1690
1227/* 1691/*