diff options
author | Jeff Garzik <jeff@garzik.org> | 2007-01-24 02:14:25 -0500 |
---|---|---|
committer | Jeff Garzik <jeff@garzik.org> | 2007-01-24 02:14:25 -0500 |
commit | e47b207a5bcb73bb097f94d9fe364f50737b0eb2 (patch) | |
tree | 57389653c37d001de3a90655cb4172ba29199ce4 /arch | |
parent | d344bff9c36db17dc4765215495aaa7212c1eb6c (diff) | |
parent | 419dd8378dfa32985672ab7927b4bc827f33b332 (diff) |
Merge branch 'master' into upstream-fixes
Diffstat (limited to 'arch')
-rw-r--r-- | arch/i386/kernel/cpu/common.c | 13 | ||||
-rw-r--r-- | arch/i386/kernel/nmi.c | 8 | ||||
-rw-r--r-- | arch/i386/kernel/paravirt.c | 9 | ||||
-rw-r--r-- | arch/i386/kernel/smpboot.c | 9 | ||||
-rw-r--r-- | arch/i386/mach-voyager/voyager_smp.c | 6 | ||||
-rw-r--r-- | arch/mips/Kconfig | 14 | ||||
-rw-r--r-- | arch/mips/kernel/smtc.c | 54 | ||||
-rw-r--r-- | arch/mips/vr41xx/common/irq.c | 12 | ||||
-rw-r--r-- | arch/ppc/platforms/ev64360.c | 3 | ||||
-rw-r--r-- | arch/x86_64/kernel/nmi.c | 2 |
10 files changed, 87 insertions, 43 deletions
diff --git a/arch/i386/kernel/cpu/common.c b/arch/i386/kernel/cpu/common.c index 8689d62abd..8a8bbdaaf3 100644 --- a/arch/i386/kernel/cpu/common.c +++ b/arch/i386/kernel/cpu/common.c | |||
@@ -710,11 +710,8 @@ __cpuinit int init_gdt(int cpu, struct task_struct *idle) | |||
710 | return 1; | 710 | return 1; |
711 | } | 711 | } |
712 | 712 | ||
713 | /* Common CPU init for both boot and secondary CPUs */ | 713 | void __cpuinit cpu_set_gdt(int cpu) |
714 | static void __cpuinit _cpu_init(int cpu, struct task_struct *curr) | ||
715 | { | 714 | { |
716 | struct tss_struct * t = &per_cpu(init_tss, cpu); | ||
717 | struct thread_struct *thread = &curr->thread; | ||
718 | struct Xgt_desc_struct *cpu_gdt_descr = &per_cpu(cpu_gdt_descr, cpu); | 715 | struct Xgt_desc_struct *cpu_gdt_descr = &per_cpu(cpu_gdt_descr, cpu); |
719 | 716 | ||
720 | /* Reinit these anyway, even if they've already been done (on | 717 | /* Reinit these anyway, even if they've already been done (on |
@@ -722,6 +719,13 @@ static void __cpuinit _cpu_init(int cpu, struct task_struct *curr) | |||
722 | the real ones). */ | 719 | the real ones). */ |
723 | load_gdt(cpu_gdt_descr); | 720 | load_gdt(cpu_gdt_descr); |
724 | set_kernel_gs(); | 721 | set_kernel_gs(); |
722 | } | ||
723 | |||
724 | /* Common CPU init for both boot and secondary CPUs */ | ||
725 | static void __cpuinit _cpu_init(int cpu, struct task_struct *curr) | ||
726 | { | ||
727 | struct tss_struct * t = &per_cpu(init_tss, cpu); | ||
728 | struct thread_struct *thread = &curr->thread; | ||
725 | 729 | ||
726 | if (cpu_test_and_set(cpu, cpu_initialized)) { | 730 | if (cpu_test_and_set(cpu, cpu_initialized)) { |
727 | printk(KERN_WARNING "CPU#%d already initialized!\n", cpu); | 731 | printk(KERN_WARNING "CPU#%d already initialized!\n", cpu); |
@@ -807,6 +811,7 @@ void __cpuinit cpu_init(void) | |||
807 | local_irq_enable(); | 811 | local_irq_enable(); |
808 | } | 812 | } |
809 | 813 | ||
814 | cpu_set_gdt(cpu); | ||
810 | _cpu_init(cpu, curr); | 815 | _cpu_init(cpu, curr); |
811 | } | 816 | } |
812 | 817 | ||
diff --git a/arch/i386/kernel/nmi.c b/arch/i386/kernel/nmi.c index a5e34d6559..1a6f8bb888 100644 --- a/arch/i386/kernel/nmi.c +++ b/arch/i386/kernel/nmi.c | |||
@@ -310,13 +310,7 @@ static int __init setup_nmi_watchdog(char *str) | |||
310 | 310 | ||
311 | if ((nmi >= NMI_INVALID) || (nmi < NMI_NONE)) | 311 | if ((nmi >= NMI_INVALID) || (nmi < NMI_NONE)) |
312 | return 0; | 312 | return 0; |
313 | /* | 313 | |
314 | * If any other x86 CPU has a local APIC, then | ||
315 | * please test the NMI stuff there and send me the | ||
316 | * missing bits. Right now Intel P6/P4 and AMD K7 only. | ||
317 | */ | ||
318 | if ((nmi == NMI_LOCAL_APIC) && (nmi_known_cpu() == 0)) | ||
319 | return 0; /* no lapic support */ | ||
320 | nmi_watchdog = nmi; | 314 | nmi_watchdog = nmi; |
321 | return 1; | 315 | return 1; |
322 | } | 316 | } |
diff --git a/arch/i386/kernel/paravirt.c b/arch/i386/kernel/paravirt.c index 3dceab5828..e55fd05da0 100644 --- a/arch/i386/kernel/paravirt.c +++ b/arch/i386/kernel/paravirt.c | |||
@@ -566,4 +566,11 @@ struct paravirt_ops paravirt_ops = { | |||
566 | .irq_enable_sysexit = native_irq_enable_sysexit, | 566 | .irq_enable_sysexit = native_irq_enable_sysexit, |
567 | .iret = native_iret, | 567 | .iret = native_iret, |
568 | }; | 568 | }; |
569 | EXPORT_SYMBOL(paravirt_ops); | 569 | |
570 | /* | ||
571 | * NOTE: CONFIG_PARAVIRT is experimental and the paravirt_ops | ||
572 | * semantics are subject to change. Hence we only do this | ||
573 | * internal-only export of this, until it gets sorted out and | ||
574 | * all lowlevel CPU ops used by modules are separately exported. | ||
575 | */ | ||
576 | EXPORT_SYMBOL_GPL(paravirt_ops); | ||
diff --git a/arch/i386/kernel/smpboot.c b/arch/i386/kernel/smpboot.c index dea7ef9d3e..8c6c8c52b9 100644 --- a/arch/i386/kernel/smpboot.c +++ b/arch/i386/kernel/smpboot.c | |||
@@ -596,6 +596,12 @@ static void __cpuinit start_secondary(void *unused) | |||
596 | void __devinit initialize_secondary(void) | 596 | void __devinit initialize_secondary(void) |
597 | { | 597 | { |
598 | /* | 598 | /* |
599 | * switch to the per CPU GDT we already set up | ||
600 | * in do_boot_cpu() | ||
601 | */ | ||
602 | cpu_set_gdt(current_thread_info()->cpu); | ||
603 | |||
604 | /* | ||
599 | * We don't actually need to load the full TSS, | 605 | * We don't actually need to load the full TSS, |
600 | * basically just the stack pointer and the eip. | 606 | * basically just the stack pointer and the eip. |
601 | */ | 607 | */ |
@@ -972,9 +978,6 @@ static int __cpuinit do_boot_cpu(int apicid, int cpu) | |||
972 | /* Stack for startup_32 can be just as for start_secondary onwards */ | 978 | /* Stack for startup_32 can be just as for start_secondary onwards */ |
973 | stack_start.esp = (void *) idle->thread.esp; | 979 | stack_start.esp = (void *) idle->thread.esp; |
974 | 980 | ||
975 | start_pda = cpu_pda(cpu); | ||
976 | cpu_gdt_descr = per_cpu(cpu_gdt_descr, cpu); | ||
977 | |||
978 | irq_ctx_init(cpu); | 981 | irq_ctx_init(cpu); |
979 | 982 | ||
980 | x86_cpu_to_apicid[cpu] = apicid; | 983 | x86_cpu_to_apicid[cpu] = apicid; |
diff --git a/arch/i386/mach-voyager/voyager_smp.c b/arch/i386/mach-voyager/voyager_smp.c index 55428e656a..74aeedf277 100644 --- a/arch/i386/mach-voyager/voyager_smp.c +++ b/arch/i386/mach-voyager/voyager_smp.c | |||
@@ -773,6 +773,12 @@ initialize_secondary(void) | |||
773 | #endif | 773 | #endif |
774 | 774 | ||
775 | /* | 775 | /* |
776 | * switch to the per CPU GDT we already set up | ||
777 | * in do_boot_cpu() | ||
778 | */ | ||
779 | cpu_set_gdt(current_thread_info()->cpu); | ||
780 | |||
781 | /* | ||
776 | * We don't actually need to load the full TSS, | 782 | * We don't actually need to load the full TSS, |
777 | * basically just the stack pointer and the eip. | 783 | * basically just the stack pointer and the eip. |
778 | */ | 784 | */ |
diff --git a/arch/mips/Kconfig b/arch/mips/Kconfig index fd2ff0698a..bbd386f572 100644 --- a/arch/mips/Kconfig +++ b/arch/mips/Kconfig | |||
@@ -1568,6 +1568,20 @@ config MIPS_MT_FPAFF | |||
1568 | depends on MIPS_MT | 1568 | depends on MIPS_MT |
1569 | default y | 1569 | default y |
1570 | 1570 | ||
1571 | config MIPS_MT_SMTC_INSTANT_REPLAY | ||
1572 | bool "Low-latency Dispatch of Deferred SMTC IPIs" | ||
1573 | depends on MIPS_MT_SMTC | ||
1574 | default y | ||
1575 | help | ||
1576 | SMTC pseudo-interrupts between TCs are deferred and queued | ||
1577 | if the target TC is interrupt-inhibited (IXMT). In the first | ||
1578 | SMTC prototypes, these queued IPIs were serviced on return | ||
1579 | to user mode, or on entry into the kernel idle loop. The | ||
1580 | INSTANT_REPLAY option dispatches them as part of local_irq_restore() | ||
1581 | processing, which adds runtime overhead (hence the option to turn | ||
1582 | it off), but ensures that IPIs are handled promptly even under | ||
1583 | heavy I/O interrupt load. | ||
1584 | |||
1571 | config MIPS_VPE_LOADER_TOM | 1585 | config MIPS_VPE_LOADER_TOM |
1572 | bool "Load VPE program into memory hidden from linux" | 1586 | bool "Load VPE program into memory hidden from linux" |
1573 | depends on MIPS_VPE_LOADER | 1587 | depends on MIPS_VPE_LOADER |
diff --git a/arch/mips/kernel/smtc.c b/arch/mips/kernel/smtc.c index a8b387197d..44238ab2fc 100644 --- a/arch/mips/kernel/smtc.c +++ b/arch/mips/kernel/smtc.c | |||
@@ -1017,6 +1017,33 @@ void setup_cross_vpe_interrupts(void) | |||
1017 | * SMTC-specific hacks invoked from elsewhere in the kernel. | 1017 | * SMTC-specific hacks invoked from elsewhere in the kernel. |
1018 | */ | 1018 | */ |
1019 | 1019 | ||
1020 | void smtc_ipi_replay(void) | ||
1021 | { | ||
1022 | /* | ||
1023 | * To the extent that we've ever turned interrupts off, | ||
1024 | * we may have accumulated deferred IPIs. This is subtle. | ||
1025 | * If we use the smtc_ipi_qdepth() macro, we'll get an | ||
1026 | * exact number - but we'll also disable interrupts | ||
1027 | * and create a window of failure where a new IPI gets | ||
1028 | * queued after we test the depth but before we re-enable | ||
1029 | * interrupts. So long as IXMT never gets set, however, | ||
1030 | * we should be OK: If we pick up something and dispatch | ||
1031 | * it here, that's great. If we see nothing, but concurrent | ||
1032 | * with this operation, another TC sends us an IPI, IXMT | ||
1033 | * is clear, and we'll handle it as a real pseudo-interrupt | ||
1034 | * and not a pseudo-pseudo interrupt. | ||
1035 | */ | ||
1036 | if (IPIQ[smp_processor_id()].depth > 0) { | ||
1037 | struct smtc_ipi *pipi; | ||
1038 | extern void self_ipi(struct smtc_ipi *); | ||
1039 | |||
1040 | while ((pipi = smtc_ipi_dq(&IPIQ[smp_processor_id()]))) { | ||
1041 | self_ipi(pipi); | ||
1042 | smtc_cpu_stats[smp_processor_id()].selfipis++; | ||
1043 | } | ||
1044 | } | ||
1045 | } | ||
1046 | |||
1020 | void smtc_idle_loop_hook(void) | 1047 | void smtc_idle_loop_hook(void) |
1021 | { | 1048 | { |
1022 | #ifdef SMTC_IDLE_HOOK_DEBUG | 1049 | #ifdef SMTC_IDLE_HOOK_DEBUG |
@@ -1113,29 +1140,14 @@ void smtc_idle_loop_hook(void) | |||
1113 | if (pdb_msg != &id_ho_db_msg[0]) | 1140 | if (pdb_msg != &id_ho_db_msg[0]) |
1114 | printk("CPU%d: %s", smp_processor_id(), id_ho_db_msg); | 1141 | printk("CPU%d: %s", smp_processor_id(), id_ho_db_msg); |
1115 | #endif /* SMTC_IDLE_HOOK_DEBUG */ | 1142 | #endif /* SMTC_IDLE_HOOK_DEBUG */ |
1143 | |||
1116 | /* | 1144 | /* |
1117 | * To the extent that we've ever turned interrupts off, | 1145 | * Replay any accumulated deferred IPIs. If "Instant Replay" |
1118 | * we may have accumulated deferred IPIs. This is subtle. | 1146 | * is in use, there should never be any. |
1119 | * If we use the smtc_ipi_qdepth() macro, we'll get an | ||
1120 | * exact number - but we'll also disable interrupts | ||
1121 | * and create a window of failure where a new IPI gets | ||
1122 | * queued after we test the depth but before we re-enable | ||
1123 | * interrupts. So long as IXMT never gets set, however, | ||
1124 | * we should be OK: If we pick up something and dispatch | ||
1125 | * it here, that's great. If we see nothing, but concurrent | ||
1126 | * with this operation, another TC sends us an IPI, IXMT | ||
1127 | * is clear, and we'll handle it as a real pseudo-interrupt | ||
1128 | * and not a pseudo-pseudo interrupt. | ||
1129 | */ | 1147 | */ |
1130 | if (IPIQ[smp_processor_id()].depth > 0) { | 1148 | #ifndef CONFIG_MIPS_MT_SMTC_INSTANT_REPLAY |
1131 | struct smtc_ipi *pipi; | 1149 | smtc_ipi_replay(); |
1132 | extern void self_ipi(struct smtc_ipi *); | 1150 | #endif /* CONFIG_MIPS_MT_SMTC_INSTANT_REPLAY */ |
1133 | |||
1134 | if ((pipi = smtc_ipi_dq(&IPIQ[smp_processor_id()])) != NULL) { | ||
1135 | self_ipi(pipi); | ||
1136 | smtc_cpu_stats[smp_processor_id()].selfipis++; | ||
1137 | } | ||
1138 | } | ||
1139 | } | 1151 | } |
1140 | 1152 | ||
1141 | void smtc_soft_dump(void) | 1153 | void smtc_soft_dump(void) |
diff --git a/arch/mips/vr41xx/common/irq.c b/arch/mips/vr41xx/common/irq.c index 397ba94cd7..16decf4ac2 100644 --- a/arch/mips/vr41xx/common/irq.c +++ b/arch/mips/vr41xx/common/irq.c | |||
@@ -1,7 +1,7 @@ | |||
1 | /* | 1 | /* |
2 | * Interrupt handing routines for NEC VR4100 series. | 2 | * Interrupt handing routines for NEC VR4100 series. |
3 | * | 3 | * |
4 | * Copyright (C) 2005 Yoichi Yuasa <yoichi_yuasa@tripeaks.co.jp> | 4 | * Copyright (C) 2005-2007 Yoichi Yuasa <yoichi_yuasa@tripeaks.co.jp> |
5 | * | 5 | * |
6 | * This program is free software; you can redistribute it and/or modify | 6 | * This program is free software; you can redistribute it and/or modify |
7 | * it under the terms of the GNU General Public License as published by | 7 | * it under the terms of the GNU General Public License as published by |
@@ -73,13 +73,19 @@ static void irq_dispatch(unsigned int irq) | |||
73 | if (cascade->get_irq != NULL) { | 73 | if (cascade->get_irq != NULL) { |
74 | unsigned int source_irq = irq; | 74 | unsigned int source_irq = irq; |
75 | desc = irq_desc + source_irq; | 75 | desc = irq_desc + source_irq; |
76 | desc->chip->ack(source_irq); | 76 | if (desc->chip->mask_ack) |
77 | desc->chip->mask_ack(source_irq); | ||
78 | else { | ||
79 | desc->chip->mask(source_irq); | ||
80 | desc->chip->ack(source_irq); | ||
81 | } | ||
77 | irq = cascade->get_irq(irq); | 82 | irq = cascade->get_irq(irq); |
78 | if (irq < 0) | 83 | if (irq < 0) |
79 | atomic_inc(&irq_err_count); | 84 | atomic_inc(&irq_err_count); |
80 | else | 85 | else |
81 | irq_dispatch(irq); | 86 | irq_dispatch(irq); |
82 | desc->chip->end(source_irq); | 87 | if (!(desc->status & IRQ_DISABLED) && desc->chip->unmask) |
88 | desc->chip->unmask(source_irq); | ||
83 | } else | 89 | } else |
84 | do_IRQ(irq); | 90 | do_IRQ(irq); |
85 | } | 91 | } |
diff --git a/arch/ppc/platforms/ev64360.c b/arch/ppc/platforms/ev64360.c index 90ed375c9b..f87e06f6ba 100644 --- a/arch/ppc/platforms/ev64360.c +++ b/arch/ppc/platforms/ev64360.c | |||
@@ -358,13 +358,12 @@ ev64360_setup_mtd(void) | |||
358 | 358 | ||
359 | ptbl_entries = 3; | 359 | ptbl_entries = 3; |
360 | 360 | ||
361 | if ((ptbl = kmalloc(ptbl_entries * sizeof(struct mtd_partition), | 361 | if ((ptbl = kzalloc(ptbl_entries * sizeof(struct mtd_partition), |
362 | GFP_KERNEL)) == NULL) { | 362 | GFP_KERNEL)) == NULL) { |
363 | 363 | ||
364 | printk(KERN_WARNING "Can't alloc MTD partition table\n"); | 364 | printk(KERN_WARNING "Can't alloc MTD partition table\n"); |
365 | return -ENOMEM; | 365 | return -ENOMEM; |
366 | } | 366 | } |
367 | memset(ptbl, 0, ptbl_entries * sizeof(struct mtd_partition)); | ||
368 | 367 | ||
369 | ptbl[0].name = "reserved"; | 368 | ptbl[0].name = "reserved"; |
370 | ptbl[0].offset = 0; | 369 | ptbl[0].offset = 0; |
diff --git a/arch/x86_64/kernel/nmi.c b/arch/x86_64/kernel/nmi.c index 186aebbae3..9cb42ecb7f 100644 --- a/arch/x86_64/kernel/nmi.c +++ b/arch/x86_64/kernel/nmi.c | |||
@@ -302,8 +302,6 @@ int __init setup_nmi_watchdog(char *str) | |||
302 | if ((nmi >= NMI_INVALID) || (nmi < NMI_NONE)) | 302 | if ((nmi >= NMI_INVALID) || (nmi < NMI_NONE)) |
303 | return 0; | 303 | return 0; |
304 | 304 | ||
305 | if ((nmi == NMI_LOCAL_APIC) && (nmi_known_cpu() == 0)) | ||
306 | return 0; /* no lapic support */ | ||
307 | nmi_watchdog = nmi; | 305 | nmi_watchdog = nmi; |
308 | return 1; | 306 | return 1; |
309 | } | 307 | } |