diff options
author | Andrea Bastoni <bastoni@cs.unc.edu> | 2010-05-29 23:35:01 -0400 |
---|---|---|
committer | Andrea Bastoni <bastoni@cs.unc.edu> | 2010-05-29 23:35:01 -0400 |
commit | 6ffc1fee98c4b995eb3a0285f4f8fb467cb0306e (patch) | |
tree | 69a05892a41e7f7400fa598ee0bdf8027c8f0fd6 /arch/x86/kernel | |
parent | e40152ee1e1c7a63f4777791863215e3faa37a86 (diff) | |
parent | 7c1ff4c544dd650cceff3cd69a04bcba60856678 (diff) |
Merge branch 'master' into wip-merge-2.6.34
Simple merge between master and 2.6.34 with conflicts resolved.
This commit does not compile, the following main problems are still
unresolved:
- spinlock -> raw_spinlock API changes
- kfifo API changes
- sched_class API changes
Conflicts:
Makefile
arch/x86/include/asm/hw_irq.h
arch/x86/include/asm/unistd_32.h
arch/x86/kernel/syscall_table_32.S
include/linux/hrtimer.h
kernel/sched.c
kernel/sched_fair.c
Diffstat (limited to 'arch/x86/kernel')
-rw-r--r-- | arch/x86/kernel/Makefile | 2 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/intel_cacheinfo.c | 17 | ||||
-rw-r--r-- | arch/x86/kernel/entry_64.S | 2 | ||||
-rw-r--r-- | arch/x86/kernel/ft_event.c | 112 | ||||
-rw-r--r-- | arch/x86/kernel/irqinit.c | 3 | ||||
-rw-r--r-- | arch/x86/kernel/smp.c | 28 | ||||
-rw-r--r-- | arch/x86/kernel/syscall_table_32.S | 14 |
7 files changed, 178 insertions, 0 deletions
diff --git a/arch/x86/kernel/Makefile b/arch/x86/kernel/Makefile index 4c58352209e0..d09934e22ca5 100644 --- a/arch/x86/kernel/Makefile +++ b/arch/x86/kernel/Makefile | |||
@@ -117,6 +117,8 @@ obj-$(CONFIG_X86_CHECK_BIOS_CORRUPTION) += check.o | |||
117 | 117 | ||
118 | obj-$(CONFIG_SWIOTLB) += pci-swiotlb.o | 118 | obj-$(CONFIG_SWIOTLB) += pci-swiotlb.o |
119 | 119 | ||
120 | obj-$(CONFIG_FEATHER_TRACE) += ft_event.o | ||
121 | |||
120 | ### | 122 | ### |
121 | # 64 bit specific files | 123 | # 64 bit specific files |
122 | ifeq ($(CONFIG_X86_64),y) | 124 | ifeq ($(CONFIG_X86_64),y) |
diff --git a/arch/x86/kernel/cpu/intel_cacheinfo.c b/arch/x86/kernel/cpu/intel_cacheinfo.c index 95962a93f99a..94d8e475744c 100644 --- a/arch/x86/kernel/cpu/intel_cacheinfo.c +++ b/arch/x86/kernel/cpu/intel_cacheinfo.c | |||
@@ -632,6 +632,23 @@ unsigned int __cpuinit init_intel_cacheinfo(struct cpuinfo_x86 *c) | |||
632 | static DEFINE_PER_CPU(struct _cpuid4_info *, ici_cpuid4_info); | 632 | static DEFINE_PER_CPU(struct _cpuid4_info *, ici_cpuid4_info); |
633 | #define CPUID4_INFO_IDX(x, y) (&((per_cpu(ici_cpuid4_info, x))[y])) | 633 | #define CPUID4_INFO_IDX(x, y) (&((per_cpu(ici_cpuid4_info, x))[y])) |
634 | 634 | ||
635 | /* returns CPUs that share the index cache with cpu */ | ||
636 | int get_shared_cpu_map(cpumask_var_t mask, unsigned int cpu, int index) | ||
637 | { | ||
638 | int ret = 0; | ||
639 | struct _cpuid4_info *this_leaf; | ||
640 | |||
641 | if (index >= num_cache_leaves) { | ||
642 | index = num_cache_leaves - 1; | ||
643 | ret = index; | ||
644 | } | ||
645 | |||
646 | this_leaf = CPUID4_INFO_IDX(cpu,index); | ||
647 | cpumask_copy(mask, to_cpumask(this_leaf->shared_cpu_map)); | ||
648 | |||
649 | return ret; | ||
650 | } | ||
651 | |||
635 | #ifdef CONFIG_SMP | 652 | #ifdef CONFIG_SMP |
636 | static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) | 653 | static void __cpuinit cache_shared_cpu_map_setup(unsigned int cpu, int index) |
637 | { | 654 | { |
diff --git a/arch/x86/kernel/entry_64.S b/arch/x86/kernel/entry_64.S index 0697ff139837..b9ec6cd7796f 100644 --- a/arch/x86/kernel/entry_64.S +++ b/arch/x86/kernel/entry_64.S | |||
@@ -1016,6 +1016,8 @@ apicinterrupt CALL_FUNCTION_VECTOR \ | |||
1016 | call_function_interrupt smp_call_function_interrupt | 1016 | call_function_interrupt smp_call_function_interrupt |
1017 | apicinterrupt RESCHEDULE_VECTOR \ | 1017 | apicinterrupt RESCHEDULE_VECTOR \ |
1018 | reschedule_interrupt smp_reschedule_interrupt | 1018 | reschedule_interrupt smp_reschedule_interrupt |
1019 | apicinterrupt PULL_TIMERS_VECTOR \ | ||
1020 | pull_timers_interrupt smp_pull_timers_interrupt | ||
1019 | #endif | 1021 | #endif |
1020 | 1022 | ||
1021 | apicinterrupt ERROR_APIC_VECTOR \ | 1023 | apicinterrupt ERROR_APIC_VECTOR \ |
diff --git a/arch/x86/kernel/ft_event.c b/arch/x86/kernel/ft_event.c new file mode 100644 index 000000000000..e07ee30dfff9 --- /dev/null +++ b/arch/x86/kernel/ft_event.c | |||
@@ -0,0 +1,112 @@ | |||
1 | #include <linux/types.h> | ||
2 | |||
3 | #include <litmus/feather_trace.h> | ||
4 | |||
5 | #ifdef __ARCH_HAS_FEATHER_TRACE | ||
6 | /* the feather trace management functions assume | ||
7 | * exclusive access to the event table | ||
8 | */ | ||
9 | |||
10 | |||
11 | #define BYTE_JUMP 0xeb | ||
12 | #define BYTE_JUMP_LEN 0x02 | ||
13 | |||
14 | /* for each event, there is an entry in the event table */ | ||
15 | struct trace_event { | ||
16 | long id; | ||
17 | long count; | ||
18 | long start_addr; | ||
19 | long end_addr; | ||
20 | }; | ||
21 | |||
22 | extern struct trace_event __start___event_table[]; | ||
23 | extern struct trace_event __stop___event_table[]; | ||
24 | |||
25 | int ft_enable_event(unsigned long id) | ||
26 | { | ||
27 | struct trace_event* te = __start___event_table; | ||
28 | int count = 0; | ||
29 | char* delta; | ||
30 | unsigned char* instr; | ||
31 | |||
32 | while (te < __stop___event_table) { | ||
33 | if (te->id == id && ++te->count == 1) { | ||
34 | instr = (unsigned char*) te->start_addr; | ||
35 | /* make sure we don't clobber something wrong */ | ||
36 | if (*instr == BYTE_JUMP) { | ||
37 | delta = (((unsigned char*) te->start_addr) + 1); | ||
38 | *delta = 0; | ||
39 | } | ||
40 | } | ||
41 | if (te->id == id) | ||
42 | count++; | ||
43 | te++; | ||
44 | } | ||
45 | |||
46 | printk(KERN_DEBUG "ft_enable_event: enabled %d events\n", count); | ||
47 | return count; | ||
48 | } | ||
49 | |||
50 | int ft_disable_event(unsigned long id) | ||
51 | { | ||
52 | struct trace_event* te = __start___event_table; | ||
53 | int count = 0; | ||
54 | char* delta; | ||
55 | unsigned char* instr; | ||
56 | |||
57 | while (te < __stop___event_table) { | ||
58 | if (te->id == id && --te->count == 0) { | ||
59 | instr = (unsigned char*) te->start_addr; | ||
60 | if (*instr == BYTE_JUMP) { | ||
61 | delta = (((unsigned char*) te->start_addr) + 1); | ||
62 | *delta = te->end_addr - te->start_addr - | ||
63 | BYTE_JUMP_LEN; | ||
64 | } | ||
65 | } | ||
66 | if (te->id == id) | ||
67 | count++; | ||
68 | te++; | ||
69 | } | ||
70 | |||
71 | printk(KERN_DEBUG "ft_disable_event: disabled %d events\n", count); | ||
72 | return count; | ||
73 | } | ||
74 | |||
75 | int ft_disable_all_events(void) | ||
76 | { | ||
77 | struct trace_event* te = __start___event_table; | ||
78 | int count = 0; | ||
79 | char* delta; | ||
80 | unsigned char* instr; | ||
81 | |||
82 | while (te < __stop___event_table) { | ||
83 | if (te->count) { | ||
84 | instr = (unsigned char*) te->start_addr; | ||
85 | if (*instr == BYTE_JUMP) { | ||
86 | delta = (((unsigned char*) te->start_addr) | ||
87 | + 1); | ||
88 | *delta = te->end_addr - te->start_addr - | ||
89 | BYTE_JUMP_LEN; | ||
90 | te->count = 0; | ||
91 | count++; | ||
92 | } | ||
93 | } | ||
94 | te++; | ||
95 | } | ||
96 | return count; | ||
97 | } | ||
98 | |||
99 | int ft_is_event_enabled(unsigned long id) | ||
100 | { | ||
101 | struct trace_event* te = __start___event_table; | ||
102 | |||
103 | while (te < __stop___event_table) { | ||
104 | if (te->id == id) | ||
105 | return te->count; | ||
106 | te++; | ||
107 | } | ||
108 | return 0; | ||
109 | } | ||
110 | |||
111 | #endif | ||
112 | |||
diff --git a/arch/x86/kernel/irqinit.c b/arch/x86/kernel/irqinit.c index 0ed2d300cd46..a760ce1a2c0d 100644 --- a/arch/x86/kernel/irqinit.c +++ b/arch/x86/kernel/irqinit.c | |||
@@ -189,6 +189,9 @@ static void __init smp_intr_init(void) | |||
189 | alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR, | 189 | alloc_intr_gate(CALL_FUNCTION_SINGLE_VECTOR, |
190 | call_function_single_interrupt); | 190 | call_function_single_interrupt); |
191 | 191 | ||
192 | /* IPI for hrtimer pulling on remote cpus */ | ||
193 | alloc_intr_gate(PULL_TIMERS_VECTOR, pull_timers_interrupt); | ||
194 | |||
192 | /* Low priority IPI to cleanup after moving an irq */ | 195 | /* Low priority IPI to cleanup after moving an irq */ |
193 | set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR, irq_move_cleanup_interrupt); | 196 | set_intr_gate(IRQ_MOVE_CLEANUP_VECTOR, irq_move_cleanup_interrupt); |
194 | set_bit(IRQ_MOVE_CLEANUP_VECTOR, used_vectors); | 197 | set_bit(IRQ_MOVE_CLEANUP_VECTOR, used_vectors); |
diff --git a/arch/x86/kernel/smp.c b/arch/x86/kernel/smp.c index d801210945d6..97af589a5c0c 100644 --- a/arch/x86/kernel/smp.c +++ b/arch/x86/kernel/smp.c | |||
@@ -23,6 +23,9 @@ | |||
23 | #include <linux/cpu.h> | 23 | #include <linux/cpu.h> |
24 | #include <linux/gfp.h> | 24 | #include <linux/gfp.h> |
25 | 25 | ||
26 | #include <litmus/litmus.h> | ||
27 | #include <litmus/trace.h> | ||
28 | |||
26 | #include <asm/mtrr.h> | 29 | #include <asm/mtrr.h> |
27 | #include <asm/tlbflush.h> | 30 | #include <asm/tlbflush.h> |
28 | #include <asm/mmu_context.h> | 31 | #include <asm/mmu_context.h> |
@@ -118,6 +121,7 @@ static void native_smp_send_reschedule(int cpu) | |||
118 | WARN_ON(1); | 121 | WARN_ON(1); |
119 | return; | 122 | return; |
120 | } | 123 | } |
124 | TS_SEND_RESCHED_START(cpu); | ||
121 | apic->send_IPI_mask(cpumask_of(cpu), RESCHEDULE_VECTOR); | 125 | apic->send_IPI_mask(cpumask_of(cpu), RESCHEDULE_VECTOR); |
122 | } | 126 | } |
123 | 127 | ||
@@ -147,6 +151,16 @@ void native_send_call_func_ipi(const struct cpumask *mask) | |||
147 | free_cpumask_var(allbutself); | 151 | free_cpumask_var(allbutself); |
148 | } | 152 | } |
149 | 153 | ||
154 | /* trigger timers on remote cpu */ | ||
155 | void smp_send_pull_timers(int cpu) | ||
156 | { | ||
157 | if (unlikely(cpu_is_offline(cpu))) { | ||
158 | WARN_ON(1); | ||
159 | return; | ||
160 | } | ||
161 | apic->send_IPI_mask(cpumask_of(cpu), PULL_TIMERS_VECTOR); | ||
162 | } | ||
163 | |||
150 | /* | 164 | /* |
151 | * this function calls the 'stop' function on all other CPUs in the system. | 165 | * this function calls the 'stop' function on all other CPUs in the system. |
152 | */ | 166 | */ |
@@ -198,7 +212,12 @@ static void native_smp_send_stop(void) | |||
198 | void smp_reschedule_interrupt(struct pt_regs *regs) | 212 | void smp_reschedule_interrupt(struct pt_regs *regs) |
199 | { | 213 | { |
200 | ack_APIC_irq(); | 214 | ack_APIC_irq(); |
215 | /* LITMUS^RT needs this interrupt to proper reschedule | ||
216 | * on this cpu | ||
217 | */ | ||
218 | set_tsk_need_resched(current); | ||
201 | inc_irq_stat(irq_resched_count); | 219 | inc_irq_stat(irq_resched_count); |
220 | TS_SEND_RESCHED_END; | ||
202 | /* | 221 | /* |
203 | * KVM uses this interrupt to force a cpu out of guest mode | 222 | * KVM uses this interrupt to force a cpu out of guest mode |
204 | */ | 223 | */ |
@@ -222,6 +241,15 @@ void smp_call_function_single_interrupt(struct pt_regs *regs) | |||
222 | irq_exit(); | 241 | irq_exit(); |
223 | } | 242 | } |
224 | 243 | ||
244 | extern void hrtimer_pull(void); | ||
245 | |||
246 | void smp_pull_timers_interrupt(struct pt_regs *regs) | ||
247 | { | ||
248 | ack_APIC_irq(); | ||
249 | TRACE("pull timer interrupt\n"); | ||
250 | hrtimer_pull(); | ||
251 | } | ||
252 | |||
225 | struct smp_ops smp_ops = { | 253 | struct smp_ops smp_ops = { |
226 | .smp_prepare_boot_cpu = native_smp_prepare_boot_cpu, | 254 | .smp_prepare_boot_cpu = native_smp_prepare_boot_cpu, |
227 | .smp_prepare_cpus = native_smp_prepare_cpus, | 255 | .smp_prepare_cpus = native_smp_prepare_cpus, |
diff --git a/arch/x86/kernel/syscall_table_32.S b/arch/x86/kernel/syscall_table_32.S index 8b3729341216..5da9a68546b7 100644 --- a/arch/x86/kernel/syscall_table_32.S +++ b/arch/x86/kernel/syscall_table_32.S | |||
@@ -337,3 +337,17 @@ ENTRY(sys_call_table) | |||
337 | .long sys_rt_tgsigqueueinfo /* 335 */ | 337 | .long sys_rt_tgsigqueueinfo /* 335 */ |
338 | .long sys_perf_event_open | 338 | .long sys_perf_event_open |
339 | .long sys_recvmmsg | 339 | .long sys_recvmmsg |
340 | .long sys_set_rt_task_param /* LITMUS^RT 338 */ | ||
341 | .long sys_get_rt_task_param | ||
342 | .long sys_complete_job | ||
343 | .long sys_od_open | ||
344 | .long sys_od_close | ||
345 | .long sys_fmlp_down | ||
346 | .long sys_fmlp_up | ||
347 | .long sys_srp_down | ||
348 | .long sys_srp_up | ||
349 | .long sys_query_job_no | ||
350 | .long sys_wait_for_job_release | ||
351 | .long sys_wait_for_ts_release | ||
352 | .long sys_release_ts | ||
353 | .long sys_null_call | ||