aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/sysdev
diff options
context:
space:
mode:
authorAnton Altaparmakov <aia21@cantab.net>2005-10-31 05:06:46 -0500
committerAnton Altaparmakov <aia21@cantab.net>2005-10-31 05:06:46 -0500
commit1f04c0a24b2f3cfe89c802a24396263623e3512d (patch)
treed7e2216b6e65b833c0c2b79b478d13ce17dbf296 /arch/powerpc/sysdev
parent07b188ab773e183871e57b33ae37bf635c9f12ba (diff)
parente2f2e58e7968f8446b1078a20a18bf8ea12b4fbc (diff)
Merge branch 'master' of /usr/src/ntfs-2.6/
Diffstat (limited to 'arch/powerpc/sysdev')
-rw-r--r--arch/powerpc/sysdev/Makefile7
-rw-r--r--arch/powerpc/sysdev/dcr.S41
-rw-r--r--arch/powerpc/sysdev/grackle.c64
-rw-r--r--arch/powerpc/sysdev/i8259.c221
-rw-r--r--arch/powerpc/sysdev/indirect_pci.c134
-rw-r--r--arch/powerpc/sysdev/mpic.c927
-rw-r--r--arch/powerpc/sysdev/u3_iommu.c327
7 files changed, 1721 insertions, 0 deletions
diff --git a/arch/powerpc/sysdev/Makefile b/arch/powerpc/sysdev/Makefile
new file mode 100644
index 000000000000..8acd21dee05d
--- /dev/null
+++ b/arch/powerpc/sysdev/Makefile
@@ -0,0 +1,7 @@
1obj-$(CONFIG_MPIC) += mpic.o
2obj-$(CONFIG_PPC_INDIRECT_PCI) += indirect_pci.o
3obj-$(CONFIG_PPC_I8259) += i8259.o
4obj-$(CONFIG_PPC_MPC106) += grackle.o
5obj-$(CONFIG_BOOKE) += dcr.o
6obj-$(CONFIG_40x) += dcr.o
7obj-$(CONFIG_U3_DART) += u3_iommu.o
diff --git a/arch/powerpc/sysdev/dcr.S b/arch/powerpc/sysdev/dcr.S
new file mode 100644
index 000000000000..895f10243a43
--- /dev/null
+++ b/arch/powerpc/sysdev/dcr.S
@@ -0,0 +1,41 @@
1/*
2 * arch/ppc/syslib/dcr.S
3 *
4 * "Indirect" DCR access
5 *
6 * Copyright (c) 2004 Eugene Surovegin <ebs@ebshome.net>
7 *
8 * This program is free software; you can redistribute it and/or modify it
9 * under the terms of the GNU General Public License as published by the
10 * Free Software Foundation; either version 2 of the License, or (at your
11 * option) any later version.
12 */
13
14#include <asm/ppc_asm.h>
15#include <asm/processor.h>
16
17#define DCR_ACCESS_PROLOG(table) \
18 rlwinm r3,r3,4,18,27; \
19 lis r5,table@h; \
20 ori r5,r5,table@l; \
21 add r3,r3,r5; \
22 mtctr r3; \
23 bctr
24
25_GLOBAL(__mfdcr)
26 DCR_ACCESS_PROLOG(__mfdcr_table)
27
28_GLOBAL(__mtdcr)
29 DCR_ACCESS_PROLOG(__mtdcr_table)
30
31__mfdcr_table:
32 mfdcr r3,0; blr
33__mtdcr_table:
34 mtdcr 0,r4; blr
35
36dcr = 1
37 .rept 1023
38 mfdcr r3,dcr; blr
39 mtdcr dcr,r4; blr
40 dcr = dcr + 1
41 .endr
diff --git a/arch/powerpc/sysdev/grackle.c b/arch/powerpc/sysdev/grackle.c
new file mode 100644
index 000000000000..b6ec793a23be
--- /dev/null
+++ b/arch/powerpc/sysdev/grackle.c
@@ -0,0 +1,64 @@
1/*
2 * Functions for setting up and using a MPC106 northbridge
3 * Extracted from arch/powerpc/platforms/powermac/pci.c.
4 *
5 * Copyright (C) 2003 Benjamin Herrenschmuidt (benh@kernel.crashing.org)
6 * Copyright (C) 1997 Paul Mackerras (paulus@samba.org)
7 *
8 * This program is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU General Public License
10 * as published by the Free Software Foundation; either version
11 * 2 of the License, or (at your option) any later version.
12 */
13#include <linux/kernel.h>
14#include <linux/pci.h>
15#include <linux/init.h>
16
17#include <asm/io.h>
18#include <asm/prom.h>
19#include <asm/pci-bridge.h>
20#include <asm/grackle.h>
21
22#define GRACKLE_CFA(b, d, o) (0x80 | ((b) << 8) | ((d) << 16) \
23 | (((o) & ~3) << 24))
24
25#define GRACKLE_PICR1_STG 0x00000040
26#define GRACKLE_PICR1_LOOPSNOOP 0x00000010
27
28/* N.B. this is called before bridges is initialized, so we can't
29 use grackle_pcibios_{read,write}_config_dword. */
30static inline void grackle_set_stg(struct pci_controller* bp, int enable)
31{
32 unsigned int val;
33
34 out_be32(bp->cfg_addr, GRACKLE_CFA(0, 0, 0xa8));
35 val = in_le32(bp->cfg_data);
36 val = enable? (val | GRACKLE_PICR1_STG) :
37 (val & ~GRACKLE_PICR1_STG);
38 out_be32(bp->cfg_addr, GRACKLE_CFA(0, 0, 0xa8));
39 out_le32(bp->cfg_data, val);
40 (void)in_le32(bp->cfg_data);
41}
42
43static inline void grackle_set_loop_snoop(struct pci_controller *bp, int enable)
44{
45 unsigned int val;
46
47 out_be32(bp->cfg_addr, GRACKLE_CFA(0, 0, 0xa8));
48 val = in_le32(bp->cfg_data);
49 val = enable? (val | GRACKLE_PICR1_LOOPSNOOP) :
50 (val & ~GRACKLE_PICR1_LOOPSNOOP);
51 out_be32(bp->cfg_addr, GRACKLE_CFA(0, 0, 0xa8));
52 out_le32(bp->cfg_data, val);
53 (void)in_le32(bp->cfg_data);
54}
55
56void __init setup_grackle(struct pci_controller *hose)
57{
58 setup_indirect_pci(hose, 0xfec00000, 0xfee00000);
59 if (machine_is_compatible("AAPL,PowerBook1998"))
60 grackle_set_loop_snoop(hose, 1);
61#if 0 /* Disabled for now, HW problems ??? */
62 grackle_set_stg(hose, 1);
63#endif
64}
diff --git a/arch/powerpc/sysdev/i8259.c b/arch/powerpc/sysdev/i8259.c
new file mode 100644
index 000000000000..90bce6e0c191
--- /dev/null
+++ b/arch/powerpc/sysdev/i8259.c
@@ -0,0 +1,221 @@
1/*
2 * i8259 interrupt controller driver.
3 *
4 * This program is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU General Public License
6 * as published by the Free Software Foundation; either version
7 * 2 of the License, or (at your option) any later version.
8 */
9#include <linux/init.h>
10#include <linux/ioport.h>
11#include <linux/interrupt.h>
12#include <asm/io.h>
13#include <asm/i8259.h>
14
15static volatile void __iomem *pci_intack; /* RO, gives us the irq vector */
16
17static unsigned char cached_8259[2] = { 0xff, 0xff };
18#define cached_A1 (cached_8259[0])
19#define cached_21 (cached_8259[1])
20
21static DEFINE_SPINLOCK(i8259_lock);
22
23static int i8259_pic_irq_offset;
24
25/*
26 * Acknowledge the IRQ using either the PCI host bridge's interrupt
27 * acknowledge feature or poll. How i8259_init() is called determines
28 * which is called. It should be noted that polling is broken on some
29 * IBM and Motorola PReP boxes so we must use the int-ack feature on them.
30 */
31int i8259_irq(struct pt_regs *regs)
32{
33 int irq;
34
35 spin_lock(&i8259_lock);
36
37 /* Either int-ack or poll for the IRQ */
38 if (pci_intack)
39 irq = readb(pci_intack);
40 else {
41 /* Perform an interrupt acknowledge cycle on controller 1. */
42 outb(0x0C, 0x20); /* prepare for poll */
43 irq = inb(0x20) & 7;
44 if (irq == 2 ) {
45 /*
46 * Interrupt is cascaded so perform interrupt
47 * acknowledge on controller 2.
48 */
49 outb(0x0C, 0xA0); /* prepare for poll */
50 irq = (inb(0xA0) & 7) + 8;
51 }
52 }
53
54 if (irq == 7) {
55 /*
56 * This may be a spurious interrupt.
57 *
58 * Read the interrupt status register (ISR). If the most
59 * significant bit is not set then there is no valid
60 * interrupt.
61 */
62 if (!pci_intack)
63 outb(0x0B, 0x20); /* ISR register */
64 if(~inb(0x20) & 0x80)
65 irq = -1;
66 }
67
68 spin_unlock(&i8259_lock);
69 return irq + i8259_pic_irq_offset;
70}
71
72int i8259_irq_cascade(struct pt_regs *regs, void *unused)
73{
74 return i8259_irq(regs);
75}
76
77static void i8259_mask_and_ack_irq(unsigned int irq_nr)
78{
79 unsigned long flags;
80
81 spin_lock_irqsave(&i8259_lock, flags);
82 irq_nr -= i8259_pic_irq_offset;
83 if (irq_nr > 7) {
84 cached_A1 |= 1 << (irq_nr-8);
85 inb(0xA1); /* DUMMY */
86 outb(cached_A1, 0xA1);
87 outb(0x20, 0xA0); /* Non-specific EOI */
88 outb(0x20, 0x20); /* Non-specific EOI to cascade */
89 } else {
90 cached_21 |= 1 << irq_nr;
91 inb(0x21); /* DUMMY */
92 outb(cached_21, 0x21);
93 outb(0x20, 0x20); /* Non-specific EOI */
94 }
95 spin_unlock_irqrestore(&i8259_lock, flags);
96}
97
98static void i8259_set_irq_mask(int irq_nr)
99{
100 outb(cached_A1,0xA1);
101 outb(cached_21,0x21);
102}
103
104static void i8259_mask_irq(unsigned int irq_nr)
105{
106 unsigned long flags;
107
108 spin_lock_irqsave(&i8259_lock, flags);
109 irq_nr -= i8259_pic_irq_offset;
110 if (irq_nr < 8)
111 cached_21 |= 1 << irq_nr;
112 else
113 cached_A1 |= 1 << (irq_nr-8);
114 i8259_set_irq_mask(irq_nr);
115 spin_unlock_irqrestore(&i8259_lock, flags);
116}
117
118static void i8259_unmask_irq(unsigned int irq_nr)
119{
120 unsigned long flags;
121
122 spin_lock_irqsave(&i8259_lock, flags);
123 irq_nr -= i8259_pic_irq_offset;
124 if (irq_nr < 8)
125 cached_21 &= ~(1 << irq_nr);
126 else
127 cached_A1 &= ~(1 << (irq_nr-8));
128 i8259_set_irq_mask(irq_nr);
129 spin_unlock_irqrestore(&i8259_lock, flags);
130}
131
132static void i8259_end_irq(unsigned int irq)
133{
134 if (!(irq_desc[irq].status & (IRQ_DISABLED|IRQ_INPROGRESS))
135 && irq_desc[irq].action)
136 i8259_unmask_irq(irq);
137}
138
139struct hw_interrupt_type i8259_pic = {
140 .typename = " i8259 ",
141 .enable = i8259_unmask_irq,
142 .disable = i8259_mask_irq,
143 .ack = i8259_mask_and_ack_irq,
144 .end = i8259_end_irq,
145};
146
147static struct resource pic1_iores = {
148 .name = "8259 (master)",
149 .start = 0x20,
150 .end = 0x21,
151 .flags = IORESOURCE_BUSY,
152};
153
154static struct resource pic2_iores = {
155 .name = "8259 (slave)",
156 .start = 0xa0,
157 .end = 0xa1,
158 .flags = IORESOURCE_BUSY,
159};
160
161static struct resource pic_edgectrl_iores = {
162 .name = "8259 edge control",
163 .start = 0x4d0,
164 .end = 0x4d1,
165 .flags = IORESOURCE_BUSY,
166};
167
168static struct irqaction i8259_irqaction = {
169 .handler = no_action,
170 .flags = SA_INTERRUPT,
171 .mask = CPU_MASK_NONE,
172 .name = "82c59 secondary cascade",
173};
174
175/*
176 * i8259_init()
177 * intack_addr - PCI interrupt acknowledge (real) address which will return
178 * the active irq from the 8259
179 */
180void __init i8259_init(unsigned long intack_addr, int offset)
181{
182 unsigned long flags;
183 int i;
184
185 spin_lock_irqsave(&i8259_lock, flags);
186 i8259_pic_irq_offset = offset;
187
188 /* init master interrupt controller */
189 outb(0x11, 0x20); /* Start init sequence */
190 outb(0x00, 0x21); /* Vector base */
191 outb(0x04, 0x21); /* edge tiggered, Cascade (slave) on IRQ2 */
192 outb(0x01, 0x21); /* Select 8086 mode */
193
194 /* init slave interrupt controller */
195 outb(0x11, 0xA0); /* Start init sequence */
196 outb(0x08, 0xA1); /* Vector base */
197 outb(0x02, 0xA1); /* edge triggered, Cascade (slave) on IRQ2 */
198 outb(0x01, 0xA1); /* Select 8086 mode */
199
200 /* always read ISR */
201 outb(0x0B, 0x20);
202 outb(0x0B, 0xA0);
203
204 /* Mask all interrupts */
205 outb(cached_A1, 0xA1);
206 outb(cached_21, 0x21);
207
208 spin_unlock_irqrestore(&i8259_lock, flags);
209
210 /* reserve our resources */
211 setup_irq(offset + 2, &i8259_irqaction);
212 request_resource(&ioport_resource, &pic1_iores);
213 request_resource(&ioport_resource, &pic2_iores);
214 request_resource(&ioport_resource, &pic_edgectrl_iores);
215
216 if (intack_addr != 0)
217 pci_intack = ioremap(intack_addr, 1);
218
219 for (i = 0; i < NUM_ISA_INTERRUPTS; ++i)
220 irq_desc[offset + i].handler = &i8259_pic;
221}
diff --git a/arch/powerpc/sysdev/indirect_pci.c b/arch/powerpc/sysdev/indirect_pci.c
new file mode 100644
index 000000000000..e71488469704
--- /dev/null
+++ b/arch/powerpc/sysdev/indirect_pci.c
@@ -0,0 +1,134 @@
1/*
2 * Support for indirect PCI bridges.
3 *
4 * Copyright (C) 1998 Gabriel Paubert.
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#include <linux/kernel.h>
13#include <linux/pci.h>
14#include <linux/delay.h>
15#include <linux/string.h>
16#include <linux/init.h>
17
18#include <asm/io.h>
19#include <asm/prom.h>
20#include <asm/pci-bridge.h>
21#include <asm/machdep.h>
22
23#ifdef CONFIG_PPC_INDIRECT_PCI_BE
24#define PCI_CFG_OUT out_be32
25#else
26#define PCI_CFG_OUT out_le32
27#endif
28
29static int
30indirect_read_config(struct pci_bus *bus, unsigned int devfn, int offset,
31 int len, u32 *val)
32{
33 struct pci_controller *hose = bus->sysdata;
34 volatile void __iomem *cfg_data;
35 u8 cfg_type = 0;
36
37 if (ppc_md.pci_exclude_device)
38 if (ppc_md.pci_exclude_device(bus->number, devfn))
39 return PCIBIOS_DEVICE_NOT_FOUND;
40
41 if (hose->set_cfg_type)
42 if (bus->number != hose->first_busno)
43 cfg_type = 1;
44
45 PCI_CFG_OUT(hose->cfg_addr,
46 (0x80000000 | ((bus->number - hose->bus_offset) << 16)
47 | (devfn << 8) | ((offset & 0xfc) | cfg_type)));
48
49 /*
50 * Note: the caller has already checked that offset is
51 * suitably aligned and that len is 1, 2 or 4.
52 */
53 cfg_data = hose->cfg_data + (offset & 3);
54 switch (len) {
55 case 1:
56 *val = in_8(cfg_data);
57 break;
58 case 2:
59 *val = in_le16(cfg_data);
60 break;
61 default:
62 *val = in_le32(cfg_data);
63 break;
64 }
65 return PCIBIOS_SUCCESSFUL;
66}
67
68static int
69indirect_write_config(struct pci_bus *bus, unsigned int devfn, int offset,
70 int len, u32 val)
71{
72 struct pci_controller *hose = bus->sysdata;
73 volatile void __iomem *cfg_data;
74 u8 cfg_type = 0;
75
76 if (ppc_md.pci_exclude_device)
77 if (ppc_md.pci_exclude_device(bus->number, devfn))
78 return PCIBIOS_DEVICE_NOT_FOUND;
79
80 if (hose->set_cfg_type)
81 if (bus->number != hose->first_busno)
82 cfg_type = 1;
83
84 PCI_CFG_OUT(hose->cfg_addr,
85 (0x80000000 | ((bus->number - hose->bus_offset) << 16)
86 | (devfn << 8) | ((offset & 0xfc) | cfg_type)));
87
88 /*
89 * Note: the caller has already checked that offset is
90 * suitably aligned and that len is 1, 2 or 4.
91 */
92 cfg_data = hose->cfg_data + (offset & 3);
93 switch (len) {
94 case 1:
95 out_8(cfg_data, val);
96 break;
97 case 2:
98 out_le16(cfg_data, val);
99 break;
100 default:
101 out_le32(cfg_data, val);
102 break;
103 }
104 return PCIBIOS_SUCCESSFUL;
105}
106
107static struct pci_ops indirect_pci_ops =
108{
109 indirect_read_config,
110 indirect_write_config
111};
112
113void __init
114setup_indirect_pci_nomap(struct pci_controller* hose, void __iomem * cfg_addr,
115 void __iomem * cfg_data)
116{
117 hose->cfg_addr = cfg_addr;
118 hose->cfg_data = cfg_data;
119 hose->ops = &indirect_pci_ops;
120}
121
122void __init
123setup_indirect_pci(struct pci_controller* hose, u32 cfg_addr, u32 cfg_data)
124{
125 unsigned long base = cfg_addr & PAGE_MASK;
126 void __iomem *mbase, *addr, *data;
127
128 mbase = ioremap(base, PAGE_SIZE);
129 addr = mbase + (cfg_addr & ~PAGE_MASK);
130 if ((cfg_data & PAGE_MASK) != base)
131 mbase = ioremap(cfg_data & PAGE_MASK, PAGE_SIZE);
132 data = mbase + (cfg_data & ~PAGE_MASK);
133 setup_indirect_pci_nomap(hose, addr, data);
134}
diff --git a/arch/powerpc/sysdev/mpic.c b/arch/powerpc/sysdev/mpic.c
new file mode 100644
index 000000000000..105f05341a41
--- /dev/null
+++ b/arch/powerpc/sysdev/mpic.c
@@ -0,0 +1,927 @@
1/*
2 * arch/powerpc/kernel/mpic.c
3 *
4 * Driver for interrupt controllers following the OpenPIC standard, the
5 * common implementation beeing IBM's MPIC. This driver also can deal
6 * with various broken implementations of this HW.
7 *
8 * Copyright (C) 2004 Benjamin Herrenschmidt, IBM Corp.
9 *
10 * This file is subject to the terms and conditions of the GNU General Public
11 * License. See the file COPYING in the main directory of this archive
12 * for more details.
13 */
14
15#undef DEBUG
16
17#include <linux/config.h>
18#include <linux/types.h>
19#include <linux/kernel.h>
20#include <linux/init.h>
21#include <linux/irq.h>
22#include <linux/smp.h>
23#include <linux/interrupt.h>
24#include <linux/bootmem.h>
25#include <linux/spinlock.h>
26#include <linux/pci.h>
27
28#include <asm/ptrace.h>
29#include <asm/signal.h>
30#include <asm/io.h>
31#include <asm/pgtable.h>
32#include <asm/irq.h>
33#include <asm/machdep.h>
34#include <asm/mpic.h>
35#include <asm/smp.h>
36
37#ifdef DEBUG
38#define DBG(fmt...) printk(fmt)
39#else
40#define DBG(fmt...)
41#endif
42
43static struct mpic *mpics;
44static struct mpic *mpic_primary;
45static DEFINE_SPINLOCK(mpic_lock);
46
47#ifdef CONFIG_PPC32 /* XXX for now */
48#define distribute_irqs CONFIG_IRQ_ALL_CPUS
49#endif
50
51/*
52 * Register accessor functions
53 */
54
55
56static inline u32 _mpic_read(unsigned int be, volatile u32 __iomem *base,
57 unsigned int reg)
58{
59 if (be)
60 return in_be32(base + (reg >> 2));
61 else
62 return in_le32(base + (reg >> 2));
63}
64
65static inline void _mpic_write(unsigned int be, volatile u32 __iomem *base,
66 unsigned int reg, u32 value)
67{
68 if (be)
69 out_be32(base + (reg >> 2), value);
70 else
71 out_le32(base + (reg >> 2), value);
72}
73
74static inline u32 _mpic_ipi_read(struct mpic *mpic, unsigned int ipi)
75{
76 unsigned int be = (mpic->flags & MPIC_BIG_ENDIAN) != 0;
77 unsigned int offset = MPIC_GREG_IPI_VECTOR_PRI_0 + (ipi * 0x10);
78
79 if (mpic->flags & MPIC_BROKEN_IPI)
80 be = !be;
81 return _mpic_read(be, mpic->gregs, offset);
82}
83
84static inline void _mpic_ipi_write(struct mpic *mpic, unsigned int ipi, u32 value)
85{
86 unsigned int offset = MPIC_GREG_IPI_VECTOR_PRI_0 + (ipi * 0x10);
87
88 _mpic_write(mpic->flags & MPIC_BIG_ENDIAN, mpic->gregs, offset, value);
89}
90
91static inline u32 _mpic_cpu_read(struct mpic *mpic, unsigned int reg)
92{
93 unsigned int cpu = 0;
94
95 if (mpic->flags & MPIC_PRIMARY)
96 cpu = hard_smp_processor_id();
97
98 return _mpic_read(mpic->flags & MPIC_BIG_ENDIAN, mpic->cpuregs[cpu], reg);
99}
100
101static inline void _mpic_cpu_write(struct mpic *mpic, unsigned int reg, u32 value)
102{
103 unsigned int cpu = 0;
104
105 if (mpic->flags & MPIC_PRIMARY)
106 cpu = hard_smp_processor_id();
107
108 _mpic_write(mpic->flags & MPIC_BIG_ENDIAN, mpic->cpuregs[cpu], reg, value);
109}
110
111static inline u32 _mpic_irq_read(struct mpic *mpic, unsigned int src_no, unsigned int reg)
112{
113 unsigned int isu = src_no >> mpic->isu_shift;
114 unsigned int idx = src_no & mpic->isu_mask;
115
116 return _mpic_read(mpic->flags & MPIC_BIG_ENDIAN, mpic->isus[isu],
117 reg + (idx * MPIC_IRQ_STRIDE));
118}
119
120static inline void _mpic_irq_write(struct mpic *mpic, unsigned int src_no,
121 unsigned int reg, u32 value)
122{
123 unsigned int isu = src_no >> mpic->isu_shift;
124 unsigned int idx = src_no & mpic->isu_mask;
125
126 _mpic_write(mpic->flags & MPIC_BIG_ENDIAN, mpic->isus[isu],
127 reg + (idx * MPIC_IRQ_STRIDE), value);
128}
129
130#define mpic_read(b,r) _mpic_read(mpic->flags & MPIC_BIG_ENDIAN,(b),(r))
131#define mpic_write(b,r,v) _mpic_write(mpic->flags & MPIC_BIG_ENDIAN,(b),(r),(v))
132#define mpic_ipi_read(i) _mpic_ipi_read(mpic,(i))
133#define mpic_ipi_write(i,v) _mpic_ipi_write(mpic,(i),(v))
134#define mpic_cpu_read(i) _mpic_cpu_read(mpic,(i))
135#define mpic_cpu_write(i,v) _mpic_cpu_write(mpic,(i),(v))
136#define mpic_irq_read(s,r) _mpic_irq_read(mpic,(s),(r))
137#define mpic_irq_write(s,r,v) _mpic_irq_write(mpic,(s),(r),(v))
138
139
140/*
141 * Low level utility functions
142 */
143
144
145
146/* Check if we have one of those nice broken MPICs with a flipped endian on
147 * reads from IPI registers
148 */
149static void __init mpic_test_broken_ipi(struct mpic *mpic)
150{
151 u32 r;
152
153 mpic_write(mpic->gregs, MPIC_GREG_IPI_VECTOR_PRI_0, MPIC_VECPRI_MASK);
154 r = mpic_read(mpic->gregs, MPIC_GREG_IPI_VECTOR_PRI_0);
155
156 if (r == le32_to_cpu(MPIC_VECPRI_MASK)) {
157 printk(KERN_INFO "mpic: Detected reversed IPI registers\n");
158 mpic->flags |= MPIC_BROKEN_IPI;
159 }
160}
161
162#ifdef CONFIG_MPIC_BROKEN_U3
163
164/* Test if an interrupt is sourced from HyperTransport (used on broken U3s)
165 * to force the edge setting on the MPIC and do the ack workaround.
166 */
167static inline int mpic_is_ht_interrupt(struct mpic *mpic, unsigned int source_no)
168{
169 if (source_no >= 128 || !mpic->fixups)
170 return 0;
171 return mpic->fixups[source_no].base != NULL;
172}
173
174static inline void mpic_apic_end_irq(struct mpic *mpic, unsigned int source_no)
175{
176 struct mpic_irq_fixup *fixup = &mpic->fixups[source_no];
177 u32 tmp;
178
179 spin_lock(&mpic->fixup_lock);
180 writeb(0x11 + 2 * fixup->irq, fixup->base);
181 tmp = readl(fixup->base + 2);
182 writel(tmp | 0x80000000ul, fixup->base + 2);
183 /* config writes shouldn't be posted but let's be safe ... */
184 (void)readl(fixup->base + 2);
185 spin_unlock(&mpic->fixup_lock);
186}
187
188
189static void __init mpic_amd8111_read_irq(struct mpic *mpic, u8 __iomem *devbase)
190{
191 int i, irq;
192 u32 tmp;
193
194 printk(KERN_INFO "mpic: - Workarounds on AMD 8111 @ %p\n", devbase);
195
196 for (i=0; i < 24; i++) {
197 writeb(0x10 + 2*i, devbase + 0xf2);
198 tmp = readl(devbase + 0xf4);
199 if ((tmp & 0x1) || !(tmp & 0x20))
200 continue;
201 irq = (tmp >> 16) & 0xff;
202 mpic->fixups[irq].irq = i;
203 mpic->fixups[irq].base = devbase + 0xf2;
204 }
205}
206
207static void __init mpic_amd8131_read_irq(struct mpic *mpic, u8 __iomem *devbase)
208{
209 int i, irq;
210 u32 tmp;
211
212 printk(KERN_INFO "mpic: - Workarounds on AMD 8131 @ %p\n", devbase);
213
214 for (i=0; i < 4; i++) {
215 writeb(0x10 + 2*i, devbase + 0xba);
216 tmp = readl(devbase + 0xbc);
217 if ((tmp & 0x1) || !(tmp & 0x20))
218 continue;
219 irq = (tmp >> 16) & 0xff;
220 mpic->fixups[irq].irq = i;
221 mpic->fixups[irq].base = devbase + 0xba;
222 }
223}
224
225static void __init mpic_scan_ioapics(struct mpic *mpic)
226{
227 unsigned int devfn;
228 u8 __iomem *cfgspace;
229
230 printk(KERN_INFO "mpic: Setting up IO-APICs workarounds for U3\n");
231
232 /* Allocate fixups array */
233 mpic->fixups = alloc_bootmem(128 * sizeof(struct mpic_irq_fixup));
234 BUG_ON(mpic->fixups == NULL);
235 memset(mpic->fixups, 0, 128 * sizeof(struct mpic_irq_fixup));
236
237 /* Init spinlock */
238 spin_lock_init(&mpic->fixup_lock);
239
240 /* Map u3 config space. We assume all IO-APICs are on the primary bus
241 * and slot will never be above "0xf" so we only need to map 32k
242 */
243 cfgspace = (unsigned char __iomem *)ioremap(0xf2000000, 0x8000);
244 BUG_ON(cfgspace == NULL);
245
246 /* Now we scan all slots. We do a very quick scan, we read the header type,
247 * vendor ID and device ID only, that's plenty enough
248 */
249 for (devfn = 0; devfn < PCI_DEVFN(0x10,0); devfn ++) {
250 u8 __iomem *devbase = cfgspace + (devfn << 8);
251 u8 hdr_type = readb(devbase + PCI_HEADER_TYPE);
252 u32 l = readl(devbase + PCI_VENDOR_ID);
253 u16 vendor_id, device_id;
254 int multifunc = 0;
255
256 DBG("devfn %x, l: %x\n", devfn, l);
257
258 /* If no device, skip */
259 if (l == 0xffffffff || l == 0x00000000 ||
260 l == 0x0000ffff || l == 0xffff0000)
261 goto next;
262
263 /* Check if it's a multifunction device (only really used
264 * to function 0 though
265 */
266 multifunc = !!(hdr_type & 0x80);
267 vendor_id = l & 0xffff;
268 device_id = (l >> 16) & 0xffff;
269
270 /* If a known device, go to fixup setup code */
271 if (vendor_id == PCI_VENDOR_ID_AMD && device_id == 0x7460)
272 mpic_amd8111_read_irq(mpic, devbase);
273 if (vendor_id == PCI_VENDOR_ID_AMD && device_id == 0x7450)
274 mpic_amd8131_read_irq(mpic, devbase);
275 next:
276 /* next device, if function 0 */
277 if ((PCI_FUNC(devfn) == 0) && !multifunc)
278 devfn += 7;
279 }
280}
281
282#endif /* CONFIG_MPIC_BROKEN_U3 */
283
284
285/* Find an mpic associated with a given linux interrupt */
286static struct mpic *mpic_find(unsigned int irq, unsigned int *is_ipi)
287{
288 struct mpic *mpic = mpics;
289
290 while(mpic) {
291 /* search IPIs first since they may override the main interrupts */
292 if (irq >= mpic->ipi_offset && irq < (mpic->ipi_offset + 4)) {
293 if (is_ipi)
294 *is_ipi = 1;
295 return mpic;
296 }
297 if (irq >= mpic->irq_offset &&
298 irq < (mpic->irq_offset + mpic->irq_count)) {
299 if (is_ipi)
300 *is_ipi = 0;
301 return mpic;
302 }
303 mpic = mpic -> next;
304 }
305 return NULL;
306}
307
308/* Convert a cpu mask from logical to physical cpu numbers. */
309static inline u32 mpic_physmask(u32 cpumask)
310{
311 int i;
312 u32 mask = 0;
313
314 for (i = 0; i < NR_CPUS; ++i, cpumask >>= 1)
315 mask |= (cpumask & 1) << get_hard_smp_processor_id(i);
316 return mask;
317}
318
319#ifdef CONFIG_SMP
320/* Get the mpic structure from the IPI number */
321static inline struct mpic * mpic_from_ipi(unsigned int ipi)
322{
323 return container_of(irq_desc[ipi].handler, struct mpic, hc_ipi);
324}
325#endif
326
327/* Get the mpic structure from the irq number */
328static inline struct mpic * mpic_from_irq(unsigned int irq)
329{
330 return container_of(irq_desc[irq].handler, struct mpic, hc_irq);
331}
332
333/* Send an EOI */
334static inline void mpic_eoi(struct mpic *mpic)
335{
336 mpic_cpu_write(MPIC_CPU_EOI, 0);
337 (void)mpic_cpu_read(MPIC_CPU_WHOAMI);
338}
339
340#ifdef CONFIG_SMP
341static irqreturn_t mpic_ipi_action(int irq, void *dev_id, struct pt_regs *regs)
342{
343 struct mpic *mpic = dev_id;
344
345 smp_message_recv(irq - mpic->ipi_offset, regs);
346 return IRQ_HANDLED;
347}
348#endif /* CONFIG_SMP */
349
350/*
351 * Linux descriptor level callbacks
352 */
353
354
355static void mpic_enable_irq(unsigned int irq)
356{
357 unsigned int loops = 100000;
358 struct mpic *mpic = mpic_from_irq(irq);
359 unsigned int src = irq - mpic->irq_offset;
360
361 DBG("%p: %s: enable_irq: %d (src %d)\n", mpic, mpic->name, irq, src);
362
363 mpic_irq_write(src, MPIC_IRQ_VECTOR_PRI,
364 mpic_irq_read(src, MPIC_IRQ_VECTOR_PRI) & ~MPIC_VECPRI_MASK);
365
366 /* make sure mask gets to controller before we return to user */
367 do {
368 if (!loops--) {
369 printk(KERN_ERR "mpic_enable_irq timeout\n");
370 break;
371 }
372 } while(mpic_irq_read(src, MPIC_IRQ_VECTOR_PRI) & MPIC_VECPRI_MASK);
373}
374
375static void mpic_disable_irq(unsigned int irq)
376{
377 unsigned int loops = 100000;
378 struct mpic *mpic = mpic_from_irq(irq);
379 unsigned int src = irq - mpic->irq_offset;
380
381 DBG("%s: disable_irq: %d (src %d)\n", mpic->name, irq, src);
382
383 mpic_irq_write(src, MPIC_IRQ_VECTOR_PRI,
384 mpic_irq_read(src, MPIC_IRQ_VECTOR_PRI) | MPIC_VECPRI_MASK);
385
386 /* make sure mask gets to controller before we return to user */
387 do {
388 if (!loops--) {
389 printk(KERN_ERR "mpic_enable_irq timeout\n");
390 break;
391 }
392 } while(!(mpic_irq_read(src, MPIC_IRQ_VECTOR_PRI) & MPIC_VECPRI_MASK));
393}
394
395static void mpic_end_irq(unsigned int irq)
396{
397 struct mpic *mpic = mpic_from_irq(irq);
398
399 DBG("%s: end_irq: %d\n", mpic->name, irq);
400
401 /* We always EOI on end_irq() even for edge interrupts since that
402 * should only lower the priority, the MPIC should have properly
403 * latched another edge interrupt coming in anyway
404 */
405
406#ifdef CONFIG_MPIC_BROKEN_U3
407 if (mpic->flags & MPIC_BROKEN_U3) {
408 unsigned int src = irq - mpic->irq_offset;
409 if (mpic_is_ht_interrupt(mpic, src))
410 mpic_apic_end_irq(mpic, src);
411 }
412#endif /* CONFIG_MPIC_BROKEN_U3 */
413
414 mpic_eoi(mpic);
415}
416
417#ifdef CONFIG_SMP
418
419static void mpic_enable_ipi(unsigned int irq)
420{
421 struct mpic *mpic = mpic_from_ipi(irq);
422 unsigned int src = irq - mpic->ipi_offset;
423
424 DBG("%s: enable_ipi: %d (ipi %d)\n", mpic->name, irq, src);
425 mpic_ipi_write(src, mpic_ipi_read(src) & ~MPIC_VECPRI_MASK);
426}
427
428static void mpic_disable_ipi(unsigned int irq)
429{
430 /* NEVER disable an IPI... that's just plain wrong! */
431}
432
433static void mpic_end_ipi(unsigned int irq)
434{
435 struct mpic *mpic = mpic_from_ipi(irq);
436
437 /*
438 * IPIs are marked IRQ_PER_CPU. This has the side effect of
439 * preventing the IRQ_PENDING/IRQ_INPROGRESS logic from
440 * applying to them. We EOI them late to avoid re-entering.
441 * We mark IPI's with SA_INTERRUPT as they must run with
442 * irqs disabled.
443 */
444 mpic_eoi(mpic);
445}
446
447#endif /* CONFIG_SMP */
448
449static void mpic_set_affinity(unsigned int irq, cpumask_t cpumask)
450{
451 struct mpic *mpic = mpic_from_irq(irq);
452
453 cpumask_t tmp;
454
455 cpus_and(tmp, cpumask, cpu_online_map);
456
457 mpic_irq_write(irq - mpic->irq_offset, MPIC_IRQ_DESTINATION,
458 mpic_physmask(cpus_addr(tmp)[0]));
459}
460
461
462/*
463 * Exported functions
464 */
465
466
467struct mpic * __init mpic_alloc(unsigned long phys_addr,
468 unsigned int flags,
469 unsigned int isu_size,
470 unsigned int irq_offset,
471 unsigned int irq_count,
472 unsigned int ipi_offset,
473 unsigned char *senses,
474 unsigned int senses_count,
475 const char *name)
476{
477 struct mpic *mpic;
478 u32 reg;
479 const char *vers;
480 int i;
481
482 mpic = alloc_bootmem(sizeof(struct mpic));
483 if (mpic == NULL)
484 return NULL;
485
486
487 memset(mpic, 0, sizeof(struct mpic));
488 mpic->name = name;
489
490 mpic->hc_irq.typename = name;
491 mpic->hc_irq.enable = mpic_enable_irq;
492 mpic->hc_irq.disable = mpic_disable_irq;
493 mpic->hc_irq.end = mpic_end_irq;
494 if (flags & MPIC_PRIMARY)
495 mpic->hc_irq.set_affinity = mpic_set_affinity;
496#ifdef CONFIG_SMP
497 mpic->hc_ipi.typename = name;
498 mpic->hc_ipi.enable = mpic_enable_ipi;
499 mpic->hc_ipi.disable = mpic_disable_ipi;
500 mpic->hc_ipi.end = mpic_end_ipi;
501#endif /* CONFIG_SMP */
502
503 mpic->flags = flags;
504 mpic->isu_size = isu_size;
505 mpic->irq_offset = irq_offset;
506 mpic->irq_count = irq_count;
507 mpic->ipi_offset = ipi_offset;
508 mpic->num_sources = 0; /* so far */
509 mpic->senses = senses;
510 mpic->senses_count = senses_count;
511
512 /* Map the global registers */
513 mpic->gregs = ioremap(phys_addr + MPIC_GREG_BASE, 0x1000);
514 mpic->tmregs = mpic->gregs + ((MPIC_TIMER_BASE - MPIC_GREG_BASE) >> 2);
515 BUG_ON(mpic->gregs == NULL);
516
517 /* Reset */
518 if (flags & MPIC_WANTS_RESET) {
519 mpic_write(mpic->gregs, MPIC_GREG_GLOBAL_CONF_0,
520 mpic_read(mpic->gregs, MPIC_GREG_GLOBAL_CONF_0)
521 | MPIC_GREG_GCONF_RESET);
522 while( mpic_read(mpic->gregs, MPIC_GREG_GLOBAL_CONF_0)
523 & MPIC_GREG_GCONF_RESET)
524 mb();
525 }
526
527 /* Read feature register, calculate num CPUs and, for non-ISU
528 * MPICs, num sources as well. On ISU MPICs, sources are counted
529 * as ISUs are added
530 */
531 reg = mpic_read(mpic->gregs, MPIC_GREG_FEATURE_0);
532 mpic->num_cpus = ((reg & MPIC_GREG_FEATURE_LAST_CPU_MASK)
533 >> MPIC_GREG_FEATURE_LAST_CPU_SHIFT) + 1;
534 if (isu_size == 0)
535 mpic->num_sources = ((reg & MPIC_GREG_FEATURE_LAST_SRC_MASK)
536 >> MPIC_GREG_FEATURE_LAST_SRC_SHIFT) + 1;
537
538 /* Map the per-CPU registers */
539 for (i = 0; i < mpic->num_cpus; i++) {
540 mpic->cpuregs[i] = ioremap(phys_addr + MPIC_CPU_BASE +
541 i * MPIC_CPU_STRIDE, 0x1000);
542 BUG_ON(mpic->cpuregs[i] == NULL);
543 }
544
545 /* Initialize main ISU if none provided */
546 if (mpic->isu_size == 0) {
547 mpic->isu_size = mpic->num_sources;
548 mpic->isus[0] = ioremap(phys_addr + MPIC_IRQ_BASE,
549 MPIC_IRQ_STRIDE * mpic->isu_size);
550 BUG_ON(mpic->isus[0] == NULL);
551 }
552 mpic->isu_shift = 1 + __ilog2(mpic->isu_size - 1);
553 mpic->isu_mask = (1 << mpic->isu_shift) - 1;
554
555 /* Display version */
556 switch (reg & MPIC_GREG_FEATURE_VERSION_MASK) {
557 case 1:
558 vers = "1.0";
559 break;
560 case 2:
561 vers = "1.2";
562 break;
563 case 3:
564 vers = "1.3";
565 break;
566 default:
567 vers = "<unknown>";
568 break;
569 }
570 printk(KERN_INFO "mpic: Setting up MPIC \"%s\" version %s at %lx, max %d CPUs\n",
571 name, vers, phys_addr, mpic->num_cpus);
572 printk(KERN_INFO "mpic: ISU size: %d, shift: %d, mask: %x\n", mpic->isu_size,
573 mpic->isu_shift, mpic->isu_mask);
574
575 mpic->next = mpics;
576 mpics = mpic;
577
578 if (flags & MPIC_PRIMARY)
579 mpic_primary = mpic;
580
581 return mpic;
582}
583
584void __init mpic_assign_isu(struct mpic *mpic, unsigned int isu_num,
585 unsigned long phys_addr)
586{
587 unsigned int isu_first = isu_num * mpic->isu_size;
588
589 BUG_ON(isu_num >= MPIC_MAX_ISU);
590
591 mpic->isus[isu_num] = ioremap(phys_addr, MPIC_IRQ_STRIDE * mpic->isu_size);
592 if ((isu_first + mpic->isu_size) > mpic->num_sources)
593 mpic->num_sources = isu_first + mpic->isu_size;
594}
595
596void __init mpic_setup_cascade(unsigned int irq, mpic_cascade_t handler,
597 void *data)
598{
599 struct mpic *mpic = mpic_find(irq, NULL);
600 unsigned long flags;
601
602 /* Synchronization here is a bit dodgy, so don't try to replace cascade
603 * interrupts on the fly too often ... but normally it's set up at boot.
604 */
605 spin_lock_irqsave(&mpic_lock, flags);
606 if (mpic->cascade)
607 mpic_disable_irq(mpic->cascade_vec + mpic->irq_offset);
608 mpic->cascade = NULL;
609 wmb();
610 mpic->cascade_vec = irq - mpic->irq_offset;
611 mpic->cascade_data = data;
612 wmb();
613 mpic->cascade = handler;
614 mpic_enable_irq(irq);
615 spin_unlock_irqrestore(&mpic_lock, flags);
616}
617
618void __init mpic_init(struct mpic *mpic)
619{
620 int i;
621
622 BUG_ON(mpic->num_sources == 0);
623
624 printk(KERN_INFO "mpic: Initializing for %d sources\n", mpic->num_sources);
625
626 /* Set current processor priority to max */
627 mpic_cpu_write(MPIC_CPU_CURRENT_TASK_PRI, 0xf);
628
629 /* Initialize timers: just disable them all */
630 for (i = 0; i < 4; i++) {
631 mpic_write(mpic->tmregs,
632 i * MPIC_TIMER_STRIDE + MPIC_TIMER_DESTINATION, 0);
633 mpic_write(mpic->tmregs,
634 i * MPIC_TIMER_STRIDE + MPIC_TIMER_VECTOR_PRI,
635 MPIC_VECPRI_MASK |
636 (MPIC_VEC_TIMER_0 + i));
637 }
638
639 /* Initialize IPIs to our reserved vectors and mark them disabled for now */
640 mpic_test_broken_ipi(mpic);
641 for (i = 0; i < 4; i++) {
642 mpic_ipi_write(i,
643 MPIC_VECPRI_MASK |
644 (10 << MPIC_VECPRI_PRIORITY_SHIFT) |
645 (MPIC_VEC_IPI_0 + i));
646#ifdef CONFIG_SMP
647 if (!(mpic->flags & MPIC_PRIMARY))
648 continue;
649 irq_desc[mpic->ipi_offset+i].status |= IRQ_PER_CPU;
650 irq_desc[mpic->ipi_offset+i].handler = &mpic->hc_ipi;
651#endif /* CONFIG_SMP */
652 }
653
654 /* Initialize interrupt sources */
655 if (mpic->irq_count == 0)
656 mpic->irq_count = mpic->num_sources;
657
658#ifdef CONFIG_MPIC_BROKEN_U3
659 /* Do the ioapic fixups on U3 broken mpic */
660 DBG("MPIC flags: %x\n", mpic->flags);
661 if ((mpic->flags & MPIC_BROKEN_U3) && (mpic->flags & MPIC_PRIMARY))
662 mpic_scan_ioapics(mpic);
663#endif /* CONFIG_MPIC_BROKEN_U3 */
664
665 for (i = 0; i < mpic->num_sources; i++) {
666 /* start with vector = source number, and masked */
667 u32 vecpri = MPIC_VECPRI_MASK | i | (8 << MPIC_VECPRI_PRIORITY_SHIFT);
668 int level = 0;
669
670 /* if it's an IPI, we skip it */
671 if ((mpic->irq_offset + i) >= (mpic->ipi_offset + i) &&
672 (mpic->irq_offset + i) < (mpic->ipi_offset + i + 4))
673 continue;
674
675 /* do senses munging */
676 if (mpic->senses && i < mpic->senses_count) {
677 if (mpic->senses[i] & IRQ_SENSE_LEVEL)
678 vecpri |= MPIC_VECPRI_SENSE_LEVEL;
679 if (mpic->senses[i] & IRQ_POLARITY_POSITIVE)
680 vecpri |= MPIC_VECPRI_POLARITY_POSITIVE;
681 } else
682 vecpri |= MPIC_VECPRI_SENSE_LEVEL;
683
684 /* remember if it was a level interrupts */
685 level = (vecpri & MPIC_VECPRI_SENSE_LEVEL);
686
687 /* deal with broken U3 */
688 if (mpic->flags & MPIC_BROKEN_U3) {
689#ifdef CONFIG_MPIC_BROKEN_U3
690 if (mpic_is_ht_interrupt(mpic, i)) {
691 vecpri &= ~(MPIC_VECPRI_SENSE_MASK |
692 MPIC_VECPRI_POLARITY_MASK);
693 vecpri |= MPIC_VECPRI_POLARITY_POSITIVE;
694 }
695#else
696 printk(KERN_ERR "mpic: BROKEN_U3 set, but CONFIG doesn't match\n");
697#endif
698 }
699
700 DBG("setup source %d, vecpri: %08x, level: %d\n", i, vecpri,
701 (level != 0));
702
703 /* init hw */
704 mpic_irq_write(i, MPIC_IRQ_VECTOR_PRI, vecpri);
705 mpic_irq_write(i, MPIC_IRQ_DESTINATION,
706 1 << hard_smp_processor_id());
707
708 /* init linux descriptors */
709 if (i < mpic->irq_count) {
710 irq_desc[mpic->irq_offset+i].status = level ? IRQ_LEVEL : 0;
711 irq_desc[mpic->irq_offset+i].handler = &mpic->hc_irq;
712 }
713 }
714
715 /* Init spurrious vector */
716 mpic_write(mpic->gregs, MPIC_GREG_SPURIOUS, MPIC_VEC_SPURRIOUS);
717
718 /* Disable 8259 passthrough */
719 mpic_write(mpic->gregs, MPIC_GREG_GLOBAL_CONF_0,
720 mpic_read(mpic->gregs, MPIC_GREG_GLOBAL_CONF_0)
721 | MPIC_GREG_GCONF_8259_PTHROU_DIS);
722
723 /* Set current processor priority to 0 */
724 mpic_cpu_write(MPIC_CPU_CURRENT_TASK_PRI, 0);
725}
726
727
728
729void mpic_irq_set_priority(unsigned int irq, unsigned int pri)
730{
731 int is_ipi;
732 struct mpic *mpic = mpic_find(irq, &is_ipi);
733 unsigned long flags;
734 u32 reg;
735
736 spin_lock_irqsave(&mpic_lock, flags);
737 if (is_ipi) {
738 reg = mpic_ipi_read(irq - mpic->ipi_offset) & MPIC_VECPRI_PRIORITY_MASK;
739 mpic_ipi_write(irq - mpic->ipi_offset,
740 reg | (pri << MPIC_VECPRI_PRIORITY_SHIFT));
741 } else {
742 reg = mpic_irq_read(irq - mpic->irq_offset, MPIC_IRQ_VECTOR_PRI)
743 & MPIC_VECPRI_PRIORITY_MASK;
744 mpic_irq_write(irq - mpic->irq_offset, MPIC_IRQ_VECTOR_PRI,
745 reg | (pri << MPIC_VECPRI_PRIORITY_SHIFT));
746 }
747 spin_unlock_irqrestore(&mpic_lock, flags);
748}
749
750unsigned int mpic_irq_get_priority(unsigned int irq)
751{
752 int is_ipi;
753 struct mpic *mpic = mpic_find(irq, &is_ipi);
754 unsigned long flags;
755 u32 reg;
756
757 spin_lock_irqsave(&mpic_lock, flags);
758 if (is_ipi)
759 reg = mpic_ipi_read(irq - mpic->ipi_offset);
760 else
761 reg = mpic_irq_read(irq - mpic->irq_offset, MPIC_IRQ_VECTOR_PRI);
762 spin_unlock_irqrestore(&mpic_lock, flags);
763 return (reg & MPIC_VECPRI_PRIORITY_MASK) >> MPIC_VECPRI_PRIORITY_SHIFT;
764}
765
766void mpic_setup_this_cpu(void)
767{
768#ifdef CONFIG_SMP
769 struct mpic *mpic = mpic_primary;
770 unsigned long flags;
771 u32 msk = 1 << hard_smp_processor_id();
772 unsigned int i;
773
774 BUG_ON(mpic == NULL);
775
776 DBG("%s: setup_this_cpu(%d)\n", mpic->name, hard_smp_processor_id());
777
778 spin_lock_irqsave(&mpic_lock, flags);
779
780 /* let the mpic know we want intrs. default affinity is 0xffffffff
781 * until changed via /proc. That's how it's done on x86. If we want
782 * it differently, then we should make sure we also change the default
783 * values of irq_affinity in irq.c.
784 */
785 if (distribute_irqs) {
786 for (i = 0; i < mpic->num_sources ; i++)
787 mpic_irq_write(i, MPIC_IRQ_DESTINATION,
788 mpic_irq_read(i, MPIC_IRQ_DESTINATION) | msk);
789 }
790
791 /* Set current processor priority to 0 */
792 mpic_cpu_write(MPIC_CPU_CURRENT_TASK_PRI, 0);
793
794 spin_unlock_irqrestore(&mpic_lock, flags);
795#endif /* CONFIG_SMP */
796}
797
798int mpic_cpu_get_priority(void)
799{
800 struct mpic *mpic = mpic_primary;
801
802 return mpic_cpu_read(MPIC_CPU_CURRENT_TASK_PRI);
803}
804
805void mpic_cpu_set_priority(int prio)
806{
807 struct mpic *mpic = mpic_primary;
808
809 prio &= MPIC_CPU_TASKPRI_MASK;
810 mpic_cpu_write(MPIC_CPU_CURRENT_TASK_PRI, prio);
811}
812
813/*
814 * XXX: someone who knows mpic should check this.
815 * do we need to eoi the ipi including for kexec cpu here (see xics comments)?
816 * or can we reset the mpic in the new kernel?
817 */
818void mpic_teardown_this_cpu(int secondary)
819{
820 struct mpic *mpic = mpic_primary;
821 unsigned long flags;
822 u32 msk = 1 << hard_smp_processor_id();
823 unsigned int i;
824
825 BUG_ON(mpic == NULL);
826
827 DBG("%s: teardown_this_cpu(%d)\n", mpic->name, hard_smp_processor_id());
828 spin_lock_irqsave(&mpic_lock, flags);
829
830 /* let the mpic know we don't want intrs. */
831 for (i = 0; i < mpic->num_sources ; i++)
832 mpic_irq_write(i, MPIC_IRQ_DESTINATION,
833 mpic_irq_read(i, MPIC_IRQ_DESTINATION) & ~msk);
834
835 /* Set current processor priority to max */
836 mpic_cpu_write(MPIC_CPU_CURRENT_TASK_PRI, 0xf);
837
838 spin_unlock_irqrestore(&mpic_lock, flags);
839}
840
841
842void mpic_send_ipi(unsigned int ipi_no, unsigned int cpu_mask)
843{
844 struct mpic *mpic = mpic_primary;
845
846 BUG_ON(mpic == NULL);
847
848 DBG("%s: send_ipi(ipi_no: %d)\n", mpic->name, ipi_no);
849
850 mpic_cpu_write(MPIC_CPU_IPI_DISPATCH_0 + ipi_no * 0x10,
851 mpic_physmask(cpu_mask & cpus_addr(cpu_online_map)[0]));
852}
853
854int mpic_get_one_irq(struct mpic *mpic, struct pt_regs *regs)
855{
856 u32 irq;
857
858 irq = mpic_cpu_read(MPIC_CPU_INTACK) & MPIC_VECPRI_VECTOR_MASK;
859 DBG("%s: get_one_irq(): %d\n", mpic->name, irq);
860
861 if (mpic->cascade && irq == mpic->cascade_vec) {
862 DBG("%s: cascading ...\n", mpic->name);
863 irq = mpic->cascade(regs, mpic->cascade_data);
864 mpic_eoi(mpic);
865 return irq;
866 }
867 if (unlikely(irq == MPIC_VEC_SPURRIOUS))
868 return -1;
869 if (irq < MPIC_VEC_IPI_0)
870 return irq + mpic->irq_offset;
871 DBG("%s: ipi %d !\n", mpic->name, irq - MPIC_VEC_IPI_0);
872 return irq - MPIC_VEC_IPI_0 + mpic->ipi_offset;
873}
874
875int mpic_get_irq(struct pt_regs *regs)
876{
877 struct mpic *mpic = mpic_primary;
878
879 BUG_ON(mpic == NULL);
880
881 return mpic_get_one_irq(mpic, regs);
882}
883
884
885#ifdef CONFIG_SMP
886void mpic_request_ipis(void)
887{
888 struct mpic *mpic = mpic_primary;
889
890 BUG_ON(mpic == NULL);
891
892 printk("requesting IPIs ... \n");
893
894 /* IPIs are marked SA_INTERRUPT as they must run with irqs disabled */
895 request_irq(mpic->ipi_offset+0, mpic_ipi_action, SA_INTERRUPT,
896 "IPI0 (call function)", mpic);
897 request_irq(mpic->ipi_offset+1, mpic_ipi_action, SA_INTERRUPT,
898 "IPI1 (reschedule)", mpic);
899 request_irq(mpic->ipi_offset+2, mpic_ipi_action, SA_INTERRUPT,
900 "IPI2 (unused)", mpic);
901 request_irq(mpic->ipi_offset+3, mpic_ipi_action, SA_INTERRUPT,
902 "IPI3 (debugger break)", mpic);
903
904 printk("IPIs requested... \n");
905}
906
907void smp_mpic_message_pass(int target, int msg)
908{
909 /* make sure we're sending something that translates to an IPI */
910 if ((unsigned int)msg > 3) {
911 printk("SMP %d: smp_message_pass: unknown msg %d\n",
912 smp_processor_id(), msg);
913 return;
914 }
915 switch (target) {
916 case MSG_ALL:
917 mpic_send_ipi(msg, 0xffffffff);
918 break;
919 case MSG_ALL_BUT_SELF:
920 mpic_send_ipi(msg, 0xffffffff & ~(1 << smp_processor_id()));
921 break;
922 default:
923 mpic_send_ipi(msg, 1 << target);
924 break;
925 }
926}
927#endif /* CONFIG_SMP */
diff --git a/arch/powerpc/sysdev/u3_iommu.c b/arch/powerpc/sysdev/u3_iommu.c
new file mode 100644
index 000000000000..fba871a1bda5
--- /dev/null
+++ b/arch/powerpc/sysdev/u3_iommu.c
@@ -0,0 +1,327 @@
1/*
2 * arch/ppc64/kernel/u3_iommu.c
3 *
4 * Copyright (C) 2004 Olof Johansson <olof@austin.ibm.com>, IBM Corporation
5 *
6 * Based on pSeries_iommu.c:
7 * Copyright (C) 2001 Mike Corrigan & Dave Engebretsen, IBM Corporation
8 * Copyright (C) 2004 Olof Johansson <olof@austin.ibm.com>, IBM Corporation
9 *
10 * Dynamic DMA mapping support, Apple U3 & IBM CPC925 "DART" iommu.
11 *
12 *
13 * This program is free software; you can redistribute it and/or modify
14 * it under the terms of the GNU General Public License as published by
15 * the Free Software Foundation; either version 2 of the License, or
16 * (at your option) any later version.
17 *
18 * This program is distributed in the hope that it will be useful,
19 * but WITHOUT ANY WARRANTY; without even the implied warranty of
20 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
21 * GNU General Public License for more details.
22 *
23 * You should have received a copy of the GNU General Public License
24 * along with this program; if not, write to the Free Software
25 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
26 */
27
28#include <linux/config.h>
29#include <linux/init.h>
30#include <linux/types.h>
31#include <linux/slab.h>
32#include <linux/mm.h>
33#include <linux/spinlock.h>
34#include <linux/string.h>
35#include <linux/pci.h>
36#include <linux/dma-mapping.h>
37#include <linux/vmalloc.h>
38#include <asm/io.h>
39#include <asm/prom.h>
40#include <asm/ppcdebug.h>
41#include <asm/iommu.h>
42#include <asm/pci-bridge.h>
43#include <asm/machdep.h>
44#include <asm/abs_addr.h>
45#include <asm/cacheflush.h>
46#include <asm/lmb.h>
47#include <asm/dart.h>
48#include <asm/ppc-pci.h>
49
50extern int iommu_force_on;
51
52/* Physical base address and size of the DART table */
53unsigned long dart_tablebase; /* exported to htab_initialize */
54static unsigned long dart_tablesize;
55
56/* Virtual base address of the DART table */
57static u32 *dart_vbase;
58
59/* Mapped base address for the dart */
60static unsigned int *dart;
61
62/* Dummy val that entries are set to when unused */
63static unsigned int dart_emptyval;
64
65static struct iommu_table iommu_table_u3;
66static int iommu_table_u3_inited;
67static int dart_dirty;
68
69#define DBG(...)
70
71static inline void dart_tlb_invalidate_all(void)
72{
73 unsigned long l = 0;
74 unsigned int reg;
75 unsigned long limit;
76
77 DBG("dart: flush\n");
78
79 /* To invalidate the DART, set the DARTCNTL_FLUSHTLB bit in the
80 * control register and wait for it to clear.
81 *
82 * Gotcha: Sometimes, the DART won't detect that the bit gets
83 * set. If so, clear it and set it again.
84 */
85
86 limit = 0;
87
88retry:
89 reg = in_be32((unsigned int *)dart+DARTCNTL);
90 reg |= DARTCNTL_FLUSHTLB;
91 out_be32((unsigned int *)dart+DARTCNTL, reg);
92
93 l = 0;
94 while ((in_be32((unsigned int *)dart+DARTCNTL) & DARTCNTL_FLUSHTLB) &&
95 l < (1L<<limit)) {
96 l++;
97 }
98 if (l == (1L<<limit)) {
99 if (limit < 4) {
100 limit++;
101 reg = in_be32((unsigned int *)dart+DARTCNTL);
102 reg &= ~DARTCNTL_FLUSHTLB;
103 out_be32((unsigned int *)dart+DARTCNTL, reg);
104 goto retry;
105 } else
106 panic("U3-DART: TLB did not flush after waiting a long "
107 "time. Buggy U3 ?");
108 }
109}
110
111static void dart_flush(struct iommu_table *tbl)
112{
113 if (dart_dirty)
114 dart_tlb_invalidate_all();
115 dart_dirty = 0;
116}
117
118static void dart_build(struct iommu_table *tbl, long index,
119 long npages, unsigned long uaddr,
120 enum dma_data_direction direction)
121{
122 unsigned int *dp;
123 unsigned int rpn;
124
125 DBG("dart: build at: %lx, %lx, addr: %x\n", index, npages, uaddr);
126
127 index <<= DART_PAGE_FACTOR;
128 npages <<= DART_PAGE_FACTOR;
129
130 dp = ((unsigned int*)tbl->it_base) + index;
131
132 /* On U3, all memory is contigous, so we can move this
133 * out of the loop.
134 */
135 while (npages--) {
136 rpn = virt_to_abs(uaddr) >> DART_PAGE_SHIFT;
137
138 *(dp++) = DARTMAP_VALID | (rpn & DARTMAP_RPNMASK);
139
140 rpn++;
141 uaddr += DART_PAGE_SIZE;
142 }
143
144 dart_dirty = 1;
145}
146
147
148static void dart_free(struct iommu_table *tbl, long index, long npages)
149{
150 unsigned int *dp;
151
152 /* We don't worry about flushing the TLB cache. The only drawback of
153 * not doing it is that we won't catch buggy device drivers doing
154 * bad DMAs, but then no 32-bit architecture ever does either.
155 */
156
157 DBG("dart: free at: %lx, %lx\n", index, npages);
158
159 index <<= DART_PAGE_FACTOR;
160 npages <<= DART_PAGE_FACTOR;
161
162 dp = ((unsigned int *)tbl->it_base) + index;
163
164 while (npages--)
165 *(dp++) = dart_emptyval;
166}
167
168
169static int dart_init(struct device_node *dart_node)
170{
171 unsigned int regword;
172 unsigned int i;
173 unsigned long tmp;
174
175 if (dart_tablebase == 0 || dart_tablesize == 0) {
176 printk(KERN_INFO "U3-DART: table not allocated, using direct DMA\n");
177 return -ENODEV;
178 }
179
180 /* Make sure nothing from the DART range remains in the CPU cache
181 * from a previous mapping that existed before the kernel took
182 * over
183 */
184 flush_dcache_phys_range(dart_tablebase, dart_tablebase + dart_tablesize);
185
186 /* Allocate a spare page to map all invalid DART pages. We need to do
187 * that to work around what looks like a problem with the HT bridge
188 * prefetching into invalid pages and corrupting data
189 */
190 tmp = lmb_alloc(DART_PAGE_SIZE, DART_PAGE_SIZE);
191 if (!tmp)
192 panic("U3-DART: Cannot allocate spare page!");
193 dart_emptyval = DARTMAP_VALID | ((tmp >> DART_PAGE_SHIFT) & DARTMAP_RPNMASK);
194
195 /* Map in DART registers. FIXME: Use device node to get base address */
196 dart = ioremap(DART_BASE, 0x7000);
197 if (dart == NULL)
198 panic("U3-DART: Cannot map registers!");
199
200 /* Set initial control register contents: table base,
201 * table size and enable bit
202 */
203 regword = DARTCNTL_ENABLE |
204 ((dart_tablebase >> DART_PAGE_SHIFT) << DARTCNTL_BASE_SHIFT) |
205 (((dart_tablesize >> DART_PAGE_SHIFT) & DARTCNTL_SIZE_MASK)
206 << DARTCNTL_SIZE_SHIFT);
207 dart_vbase = ioremap(virt_to_abs(dart_tablebase), dart_tablesize);
208
209 /* Fill initial table */
210 for (i = 0; i < dart_tablesize/4; i++)
211 dart_vbase[i] = dart_emptyval;
212
213 /* Initialize DART with table base and enable it. */
214 out_be32((unsigned int *)dart, regword);
215
216 /* Invalidate DART to get rid of possible stale TLBs */
217 dart_tlb_invalidate_all();
218
219 printk(KERN_INFO "U3/CPC925 DART IOMMU initialized\n");
220
221 return 0;
222}
223
224static void iommu_table_u3_setup(void)
225{
226 iommu_table_u3.it_busno = 0;
227 iommu_table_u3.it_offset = 0;
228 /* it_size is in number of entries */
229 iommu_table_u3.it_size = dart_tablesize / sizeof(u32);
230
231 /* Initialize the common IOMMU code */
232 iommu_table_u3.it_base = (unsigned long)dart_vbase;
233 iommu_table_u3.it_index = 0;
234 iommu_table_u3.it_blocksize = 1;
235 iommu_init_table(&iommu_table_u3);
236
237 /* Reserve the last page of the DART to avoid possible prefetch
238 * past the DART mapped area
239 */
240 set_bit(iommu_table_u3.it_size - 1, iommu_table_u3.it_map);
241}
242
243static void iommu_dev_setup_u3(struct pci_dev *dev)
244{
245 struct device_node *dn;
246
247 /* We only have one iommu table on the mac for now, which makes
248 * things simple. Setup all PCI devices to point to this table
249 *
250 * We must use pci_device_to_OF_node() to make sure that
251 * we get the real "final" pointer to the device in the
252 * pci_dev sysdata and not the temporary PHB one
253 */
254 dn = pci_device_to_OF_node(dev);
255
256 if (dn)
257 PCI_DN(dn)->iommu_table = &iommu_table_u3;
258}
259
260static void iommu_bus_setup_u3(struct pci_bus *bus)
261{
262 struct device_node *dn;
263
264 if (!iommu_table_u3_inited) {
265 iommu_table_u3_inited = 1;
266 iommu_table_u3_setup();
267 }
268
269 dn = pci_bus_to_OF_node(bus);
270
271 if (dn)
272 PCI_DN(dn)->iommu_table = &iommu_table_u3;
273}
274
275static void iommu_dev_setup_null(struct pci_dev *dev) { }
276static void iommu_bus_setup_null(struct pci_bus *bus) { }
277
278void iommu_init_early_u3(void)
279{
280 struct device_node *dn;
281
282 /* Find the DART in the device-tree */
283 dn = of_find_compatible_node(NULL, "dart", "u3-dart");
284 if (dn == NULL)
285 return;
286
287 /* Setup low level TCE operations for the core IOMMU code */
288 ppc_md.tce_build = dart_build;
289 ppc_md.tce_free = dart_free;
290 ppc_md.tce_flush = dart_flush;
291
292 /* Initialize the DART HW */
293 if (dart_init(dn)) {
294 /* If init failed, use direct iommu and null setup functions */
295 ppc_md.iommu_dev_setup = iommu_dev_setup_null;
296 ppc_md.iommu_bus_setup = iommu_bus_setup_null;
297
298 /* Setup pci_dma ops */
299 pci_direct_iommu_init();
300 } else {
301 ppc_md.iommu_dev_setup = iommu_dev_setup_u3;
302 ppc_md.iommu_bus_setup = iommu_bus_setup_u3;
303
304 /* Setup pci_dma ops */
305 pci_iommu_init();
306 }
307}
308
309
310void __init alloc_u3_dart_table(void)
311{
312 /* Only reserve DART space if machine has more than 2GB of RAM
313 * or if requested with iommu=on on cmdline.
314 */
315 if (lmb_end_of_DRAM() <= 0x80000000ull && !iommu_force_on)
316 return;
317
318 /* 512 pages (2MB) is max DART tablesize. */
319 dart_tablesize = 1UL << 21;
320 /* 16MB (1 << 24) alignment. We allocate a full 16Mb chuck since we
321 * will blow up an entire large page anyway in the kernel mapping
322 */
323 dart_tablebase = (unsigned long)
324 abs_to_virt(lmb_alloc_base(1UL<<24, 1UL<<24, 0x80000000L));
325
326 printk(KERN_INFO "U3-DART allocated at: %lx\n", dart_tablebase);
327}