diff options
author | Venkatesh Pallipadi <venkatesh.pallipadi@intel.com> | 2006-01-11 16:44:18 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2006-01-11 22:04:54 -0500 |
commit | 5a07a30c3cc4dc438494d6416ffa74008a2194b3 (patch) | |
tree | eafda8ee2f58f392f5d78af7748ce58ddb6bf9a2 /arch/x86_64 | |
parent | 01b8faaef5d239aeabb1e712c5d9619f29e808dd (diff) |
[PATCH] i386/x86-64: Remove sub jiffy profile timer support
Remove the finer control of local APIC timer. We cannot provide a sub-jiffy
control like this when we use broadcast from external timer in place of
local APIC. Instead of removing this only on systems that may end up using
broadcast from external timer (due to C3), I am going the
"I'm feeling lucky" way to remove this fully. Basically, I am not sure about
usefulness of this code today. Few other architectures also don't seem to
support this today.
If you are using profiling and fine grained control and don't like this going
away in normal case, yell at me right now.
Signed-off-by: Venkatesh Pallipadi <venkatesh.pallipadi@intel.com>
Signed-off-by: Andi Kleen <ak@suse.de>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'arch/x86_64')
-rw-r--r-- | arch/x86_64/kernel/apic.c | 53 |
1 files changed, 2 insertions, 51 deletions
diff --git a/arch/x86_64/kernel/apic.c b/arch/x86_64/kernel/apic.c index 18691ce4c759..fa47ae731576 100644 --- a/arch/x86_64/kernel/apic.c +++ b/arch/x86_64/kernel/apic.c | |||
@@ -41,10 +41,6 @@ int disable_apic_timer __initdata; | |||
41 | /* Using APIC to generate smp_local_timer_interrupt? */ | 41 | /* Using APIC to generate smp_local_timer_interrupt? */ |
42 | int using_apic_timer = 0; | 42 | int using_apic_timer = 0; |
43 | 43 | ||
44 | static DEFINE_PER_CPU(int, prof_multiplier) = 1; | ||
45 | static DEFINE_PER_CPU(int, prof_old_multiplier) = 1; | ||
46 | static DEFINE_PER_CPU(int, prof_counter) = 1; | ||
47 | |||
48 | static void apic_pm_activate(void); | 44 | static void apic_pm_activate(void); |
49 | 45 | ||
50 | void enable_NMI_through_LVT0 (void * dummy) | 46 | void enable_NMI_through_LVT0 (void * dummy) |
@@ -805,32 +801,9 @@ void enable_APIC_timer(void) | |||
805 | } | 801 | } |
806 | } | 802 | } |
807 | 803 | ||
808 | /* | ||
809 | * the frequency of the profiling timer can be changed | ||
810 | * by writing a multiplier value into /proc/profile. | ||
811 | */ | ||
812 | int setup_profiling_timer(unsigned int multiplier) | 804 | int setup_profiling_timer(unsigned int multiplier) |
813 | { | 805 | { |
814 | int i; | 806 | return -EINVAL; |
815 | |||
816 | /* | ||
817 | * Sanity check. [at least 500 APIC cycles should be | ||
818 | * between APIC interrupts as a rule of thumb, to avoid | ||
819 | * irqs flooding us] | ||
820 | */ | ||
821 | if ( (!multiplier) || (calibration_result/multiplier < 500)) | ||
822 | return -EINVAL; | ||
823 | |||
824 | /* | ||
825 | * Set the new multiplier for each CPU. CPUs don't start using the | ||
826 | * new values until the next timer interrupt in which they do process | ||
827 | * accounting. At that time they also adjust their APIC timers | ||
828 | * accordingly. | ||
829 | */ | ||
830 | for (i = 0; i < NR_CPUS; ++i) | ||
831 | per_cpu(prof_multiplier, i) = multiplier; | ||
832 | |||
833 | return 0; | ||
834 | } | 807 | } |
835 | 808 | ||
836 | #ifdef CONFIG_X86_MCE_AMD | 809 | #ifdef CONFIG_X86_MCE_AMD |
@@ -857,32 +830,10 @@ void setup_threshold_lvt(unsigned long lvt_off) | |||
857 | 830 | ||
858 | void smp_local_timer_interrupt(struct pt_regs *regs) | 831 | void smp_local_timer_interrupt(struct pt_regs *regs) |
859 | { | 832 | { |
860 | int cpu = smp_processor_id(); | ||
861 | |||
862 | profile_tick(CPU_PROFILING, regs); | 833 | profile_tick(CPU_PROFILING, regs); |
863 | if (--per_cpu(prof_counter, cpu) <= 0) { | ||
864 | /* | ||
865 | * The multiplier may have changed since the last time we got | ||
866 | * to this point as a result of the user writing to | ||
867 | * /proc/profile. In this case we need to adjust the APIC | ||
868 | * timer accordingly. | ||
869 | * | ||
870 | * Interrupts are already masked off at this point. | ||
871 | */ | ||
872 | per_cpu(prof_counter, cpu) = per_cpu(prof_multiplier, cpu); | ||
873 | if (per_cpu(prof_counter, cpu) != | ||
874 | per_cpu(prof_old_multiplier, cpu)) { | ||
875 | __setup_APIC_LVTT(calibration_result/ | ||
876 | per_cpu(prof_counter, cpu)); | ||
877 | per_cpu(prof_old_multiplier, cpu) = | ||
878 | per_cpu(prof_counter, cpu); | ||
879 | } | ||
880 | |||
881 | #ifdef CONFIG_SMP | 834 | #ifdef CONFIG_SMP |
882 | update_process_times(user_mode(regs)); | 835 | update_process_times(user_mode(regs)); |
883 | #endif | 836 | #endif |
884 | } | ||
885 | |||
886 | /* | 837 | /* |
887 | * We take the 'long' return path, and there every subsystem | 838 | * We take the 'long' return path, and there every subsystem |
888 | * grabs the appropriate locks (kernel lock/ irq lock). | 839 | * grabs the appropriate locks (kernel lock/ irq lock). |