aboutsummaryrefslogtreecommitdiffstats
path: root/arch/ia64
diff options
context:
space:
mode:
Diffstat (limited to 'arch/ia64')
-rw-r--r--arch/ia64/kernel/iosapic.c20
-rw-r--r--arch/ia64/kernel/irq_ia64.c42
-rw-r--r--arch/ia64/kernel/msi_ia64.c20
3 files changed, 70 insertions, 12 deletions
diff --git a/arch/ia64/kernel/iosapic.c b/arch/ia64/kernel/iosapic.c
index e647254c2707..c101c8bff27b 100644
--- a/arch/ia64/kernel/iosapic.c
+++ b/arch/ia64/kernel/iosapic.c
@@ -354,11 +354,13 @@ iosapic_set_affinity (unsigned int irq, cpumask_t mask)
354 354
355 irq &= (~IA64_IRQ_REDIRECTED); 355 irq &= (~IA64_IRQ_REDIRECTED);
356 356
357 /* IRQ migration across domain is not supported yet */ 357 cpus_and(mask, mask, cpu_online_map);
358 cpus_and(mask, mask, irq_to_domain(irq));
359 if (cpus_empty(mask)) 358 if (cpus_empty(mask))
360 return; 359 return;
361 360
361 if (reassign_irq_vector(irq, first_cpu(mask)))
362 return;
363
362 dest = cpu_physical_id(first_cpu(mask)); 364 dest = cpu_physical_id(first_cpu(mask));
363 365
364 if (list_empty(&iosapic_intr_info[irq].rtes)) 366 if (list_empty(&iosapic_intr_info[irq].rtes))
@@ -376,6 +378,8 @@ iosapic_set_affinity (unsigned int irq, cpumask_t mask)
376 else 378 else
377 /* change delivery mode to fixed */ 379 /* change delivery mode to fixed */
378 low32 |= (IOSAPIC_FIXED << IOSAPIC_DELIVERY_SHIFT); 380 low32 |= (IOSAPIC_FIXED << IOSAPIC_DELIVERY_SHIFT);
381 low32 &= IOSAPIC_VECTOR_MASK;
382 low32 |= irq_to_vector(irq);
379 383
380 iosapic_intr_info[irq].low32 = low32; 384 iosapic_intr_info[irq].low32 = low32;
381 iosapic_intr_info[irq].dest = dest; 385 iosapic_intr_info[irq].dest = dest;
@@ -404,10 +408,20 @@ iosapic_end_level_irq (unsigned int irq)
404{ 408{
405 ia64_vector vec = irq_to_vector(irq); 409 ia64_vector vec = irq_to_vector(irq);
406 struct iosapic_rte_info *rte; 410 struct iosapic_rte_info *rte;
411 int do_unmask_irq = 0;
412
413 if (unlikely(irq_desc[irq].status & IRQ_MOVE_PENDING)) {
414 do_unmask_irq = 1;
415 mask_irq(irq);
416 }
407 417
408 move_native_irq(irq);
409 list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list) 418 list_for_each_entry(rte, &iosapic_intr_info[irq].rtes, rte_list)
410 iosapic_eoi(rte->iosapic->addr, vec); 419 iosapic_eoi(rte->iosapic->addr, vec);
420
421 if (unlikely(do_unmask_irq)) {
422 move_masked_irq(irq);
423 unmask_irq(irq);
424 }
411} 425}
412 426
413#define iosapic_shutdown_level_irq mask_irq 427#define iosapic_shutdown_level_irq mask_irq
diff --git a/arch/ia64/kernel/irq_ia64.c b/arch/ia64/kernel/irq_ia64.c
index a3667631ed80..22806b94025a 100644
--- a/arch/ia64/kernel/irq_ia64.c
+++ b/arch/ia64/kernel/irq_ia64.c
@@ -172,15 +172,13 @@ int bind_irq_vector(int irq, int vector, cpumask_t domain)
172 return ret; 172 return ret;
173} 173}
174 174
175static void clear_irq_vector(int irq) 175static void __clear_irq_vector(int irq)
176{ 176{
177 unsigned long flags;
178 int vector, cpu, pos; 177 int vector, cpu, pos;
179 cpumask_t mask; 178 cpumask_t mask;
180 cpumask_t domain; 179 cpumask_t domain;
181 struct irq_cfg *cfg = &irq_cfg[irq]; 180 struct irq_cfg *cfg = &irq_cfg[irq];
182 181
183 spin_lock_irqsave(&vector_lock, flags);
184 BUG_ON((unsigned)irq >= NR_IRQS); 182 BUG_ON((unsigned)irq >= NR_IRQS);
185 BUG_ON(cfg->vector == IRQ_VECTOR_UNASSIGNED); 183 BUG_ON(cfg->vector == IRQ_VECTOR_UNASSIGNED);
186 vector = cfg->vector; 184 vector = cfg->vector;
@@ -193,6 +191,14 @@ static void clear_irq_vector(int irq)
193 irq_status[irq] = IRQ_UNUSED; 191 irq_status[irq] = IRQ_UNUSED;
194 pos = vector - IA64_FIRST_DEVICE_VECTOR; 192 pos = vector - IA64_FIRST_DEVICE_VECTOR;
195 cpus_andnot(vector_table[pos], vector_table[pos], domain); 193 cpus_andnot(vector_table[pos], vector_table[pos], domain);
194}
195
196static void clear_irq_vector(int irq)
197{
198 unsigned long flags;
199
200 spin_lock_irqsave(&vector_lock, flags);
201 __clear_irq_vector(irq);
196 spin_unlock_irqrestore(&vector_lock, flags); 202 spin_unlock_irqrestore(&vector_lock, flags);
197} 203}
198 204
@@ -275,6 +281,36 @@ void destroy_and_reserve_irq(unsigned int irq)
275 reserve_irq(irq); 281 reserve_irq(irq);
276} 282}
277 283
284static int __reassign_irq_vector(int irq, int cpu)
285{
286 struct irq_cfg *cfg = &irq_cfg[irq];
287 int vector;
288 cpumask_t domain;
289
290 if (cfg->vector == IRQ_VECTOR_UNASSIGNED || !cpu_online(cpu))
291 return -EINVAL;
292 if (cpu_isset(cpu, cfg->domain))
293 return 0;
294 domain = vector_allocation_domain(cpu);
295 vector = find_unassigned_vector(domain);
296 if (vector < 0)
297 return -ENOSPC;
298 __clear_irq_vector(irq);
299 BUG_ON(__bind_irq_vector(irq, vector, domain));
300 return 0;
301}
302
303int reassign_irq_vector(int irq, int cpu)
304{
305 unsigned long flags;
306 int ret;
307
308 spin_lock_irqsave(&vector_lock, flags);
309 ret = __reassign_irq_vector(irq, cpu);
310 spin_unlock_irqrestore(&vector_lock, flags);
311 return ret;
312}
313
278/* 314/*
279 * Dynamic irq allocate and deallocation for MSI 315 * Dynamic irq allocate and deallocation for MSI
280 */ 316 */
diff --git a/arch/ia64/kernel/msi_ia64.c b/arch/ia64/kernel/msi_ia64.c
index 1d22670cc88b..2fdbd5c3f213 100644
--- a/arch/ia64/kernel/msi_ia64.c
+++ b/arch/ia64/kernel/msi_ia64.c
@@ -13,6 +13,7 @@
13 13
14#define MSI_DATA_VECTOR_SHIFT 0 14#define MSI_DATA_VECTOR_SHIFT 0
15#define MSI_DATA_VECTOR(v) (((u8)v) << MSI_DATA_VECTOR_SHIFT) 15#define MSI_DATA_VECTOR(v) (((u8)v) << MSI_DATA_VECTOR_SHIFT)
16#define MSI_DATA_VECTOR_MASK 0xffffff00
16 17
17#define MSI_DATA_DELIVERY_SHIFT 8 18#define MSI_DATA_DELIVERY_SHIFT 8
18#define MSI_DATA_DELIVERY_FIXED (0 << MSI_DATA_DELIVERY_SHIFT) 19#define MSI_DATA_DELIVERY_FIXED (0 << MSI_DATA_DELIVERY_SHIFT)
@@ -50,22 +51,29 @@ static struct irq_chip ia64_msi_chip;
50static void ia64_set_msi_irq_affinity(unsigned int irq, cpumask_t cpu_mask) 51static void ia64_set_msi_irq_affinity(unsigned int irq, cpumask_t cpu_mask)
51{ 52{
52 struct msi_msg msg; 53 struct msi_msg msg;
53 u32 addr; 54 u32 addr, data;
55 int cpu = first_cpu(cpu_mask);
54 56
55 /* IRQ migration across domain is not supported yet */ 57 if (!cpu_online(cpu))
56 cpus_and(cpu_mask, cpu_mask, irq_to_domain(irq)); 58 return;
57 if (cpus_empty(cpu_mask)) 59
60 if (reassign_irq_vector(irq, cpu))
58 return; 61 return;
59 62
60 read_msi_msg(irq, &msg); 63 read_msi_msg(irq, &msg);
61 64
62 addr = msg.address_lo; 65 addr = msg.address_lo;
63 addr &= MSI_ADDR_DESTID_MASK; 66 addr &= MSI_ADDR_DESTID_MASK;
64 addr |= MSI_ADDR_DESTID_CPU(cpu_physical_id(first_cpu(cpu_mask))); 67 addr |= MSI_ADDR_DESTID_CPU(cpu_physical_id(cpu));
65 msg.address_lo = addr; 68 msg.address_lo = addr;
66 69
70 data = msg.data;
71 data &= MSI_DATA_VECTOR_MASK;
72 data |= MSI_DATA_VECTOR(irq_to_vector(irq));
73 msg.data = data;
74
67 write_msi_msg(irq, &msg); 75 write_msi_msg(irq, &msg);
68 irq_desc[irq].affinity = cpu_mask; 76 irq_desc[irq].affinity = cpumask_of_cpu(cpu);
69} 77}
70#endif /* CONFIG_SMP */ 78#endif /* CONFIG_SMP */
71 79