diff options
author | Thomas Gleixner <tglx@linutronix.de> | 2009-11-17 10:46:45 -0500 |
---|---|---|
committer | Thomas Gleixner <tglx@linutronix.de> | 2009-12-14 17:55:33 -0500 |
commit | 239007b8440abff689632f50cdf0f2b9e895b534 (patch) | |
tree | 569cab843af4a999d6d868ec9a824530d2bfa733 /kernel/irq/handle.c | |
parent | 9f5a5621e78cf48d86682a71ceb3fcdbde38b222 (diff) |
genirq: Convert irq_desc.lock to raw_spinlock
Convert locks which cannot be sleeping locks in preempt-rt to
raw_spinlocks.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/irq/handle.c')
-rw-r--r-- | kernel/irq/handle.c | 22 |
1 files changed, 11 insertions, 11 deletions
diff --git a/kernel/irq/handle.c b/kernel/irq/handle.c index 17c71bb565c6..814940e7f485 100644 --- a/kernel/irq/handle.c +++ b/kernel/irq/handle.c | |||
@@ -80,7 +80,7 @@ static struct irq_desc irq_desc_init = { | |||
80 | .chip = &no_irq_chip, | 80 | .chip = &no_irq_chip, |
81 | .handle_irq = handle_bad_irq, | 81 | .handle_irq = handle_bad_irq, |
82 | .depth = 1, | 82 | .depth = 1, |
83 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc_init.lock), | 83 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc_init.lock), |
84 | }; | 84 | }; |
85 | 85 | ||
86 | void __ref init_kstat_irqs(struct irq_desc *desc, int node, int nr) | 86 | void __ref init_kstat_irqs(struct irq_desc *desc, int node, int nr) |
@@ -108,7 +108,7 @@ static void init_one_irq_desc(int irq, struct irq_desc *desc, int node) | |||
108 | { | 108 | { |
109 | memcpy(desc, &irq_desc_init, sizeof(struct irq_desc)); | 109 | memcpy(desc, &irq_desc_init, sizeof(struct irq_desc)); |
110 | 110 | ||
111 | spin_lock_init(&desc->lock); | 111 | raw_spin_lock_init(&desc->lock); |
112 | desc->irq = irq; | 112 | desc->irq = irq; |
113 | #ifdef CONFIG_SMP | 113 | #ifdef CONFIG_SMP |
114 | desc->node = node; | 114 | desc->node = node; |
@@ -130,7 +130,7 @@ static void init_one_irq_desc(int irq, struct irq_desc *desc, int node) | |||
130 | /* | 130 | /* |
131 | * Protect the sparse_irqs: | 131 | * Protect the sparse_irqs: |
132 | */ | 132 | */ |
133 | DEFINE_SPINLOCK(sparse_irq_lock); | 133 | DEFINE_RAW_SPINLOCK(sparse_irq_lock); |
134 | 134 | ||
135 | struct irq_desc **irq_desc_ptrs __read_mostly; | 135 | struct irq_desc **irq_desc_ptrs __read_mostly; |
136 | 136 | ||
@@ -141,7 +141,7 @@ static struct irq_desc irq_desc_legacy[NR_IRQS_LEGACY] __cacheline_aligned_in_sm | |||
141 | .chip = &no_irq_chip, | 141 | .chip = &no_irq_chip, |
142 | .handle_irq = handle_bad_irq, | 142 | .handle_irq = handle_bad_irq, |
143 | .depth = 1, | 143 | .depth = 1, |
144 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc_init.lock), | 144 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc_init.lock), |
145 | } | 145 | } |
146 | }; | 146 | }; |
147 | 147 | ||
@@ -212,7 +212,7 @@ struct irq_desc * __ref irq_to_desc_alloc_node(unsigned int irq, int node) | |||
212 | if (desc) | 212 | if (desc) |
213 | return desc; | 213 | return desc; |
214 | 214 | ||
215 | spin_lock_irqsave(&sparse_irq_lock, flags); | 215 | raw_spin_lock_irqsave(&sparse_irq_lock, flags); |
216 | 216 | ||
217 | /* We have to check it to avoid races with another CPU */ | 217 | /* We have to check it to avoid races with another CPU */ |
218 | desc = irq_desc_ptrs[irq]; | 218 | desc = irq_desc_ptrs[irq]; |
@@ -234,7 +234,7 @@ struct irq_desc * __ref irq_to_desc_alloc_node(unsigned int irq, int node) | |||
234 | irq_desc_ptrs[irq] = desc; | 234 | irq_desc_ptrs[irq] = desc; |
235 | 235 | ||
236 | out_unlock: | 236 | out_unlock: |
237 | spin_unlock_irqrestore(&sparse_irq_lock, flags); | 237 | raw_spin_unlock_irqrestore(&sparse_irq_lock, flags); |
238 | 238 | ||
239 | return desc; | 239 | return desc; |
240 | } | 240 | } |
@@ -247,7 +247,7 @@ struct irq_desc irq_desc[NR_IRQS] __cacheline_aligned_in_smp = { | |||
247 | .chip = &no_irq_chip, | 247 | .chip = &no_irq_chip, |
248 | .handle_irq = handle_bad_irq, | 248 | .handle_irq = handle_bad_irq, |
249 | .depth = 1, | 249 | .depth = 1, |
250 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc->lock), | 250 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc->lock), |
251 | } | 251 | } |
252 | }; | 252 | }; |
253 | 253 | ||
@@ -473,7 +473,7 @@ unsigned int __do_IRQ(unsigned int irq) | |||
473 | return 1; | 473 | return 1; |
474 | } | 474 | } |
475 | 475 | ||
476 | spin_lock(&desc->lock); | 476 | raw_spin_lock(&desc->lock); |
477 | if (desc->chip->ack) | 477 | if (desc->chip->ack) |
478 | desc->chip->ack(irq); | 478 | desc->chip->ack(irq); |
479 | /* | 479 | /* |
@@ -517,13 +517,13 @@ unsigned int __do_IRQ(unsigned int irq) | |||
517 | for (;;) { | 517 | for (;;) { |
518 | irqreturn_t action_ret; | 518 | irqreturn_t action_ret; |
519 | 519 | ||
520 | spin_unlock(&desc->lock); | 520 | raw_spin_unlock(&desc->lock); |
521 | 521 | ||
522 | action_ret = handle_IRQ_event(irq, action); | 522 | action_ret = handle_IRQ_event(irq, action); |
523 | if (!noirqdebug) | 523 | if (!noirqdebug) |
524 | note_interrupt(irq, desc, action_ret); | 524 | note_interrupt(irq, desc, action_ret); |
525 | 525 | ||
526 | spin_lock(&desc->lock); | 526 | raw_spin_lock(&desc->lock); |
527 | if (likely(!(desc->status & IRQ_PENDING))) | 527 | if (likely(!(desc->status & IRQ_PENDING))) |
528 | break; | 528 | break; |
529 | desc->status &= ~IRQ_PENDING; | 529 | desc->status &= ~IRQ_PENDING; |
@@ -536,7 +536,7 @@ out: | |||
536 | * disabled while the handler was running. | 536 | * disabled while the handler was running. |
537 | */ | 537 | */ |
538 | desc->chip->end(irq); | 538 | desc->chip->end(irq); |
539 | spin_unlock(&desc->lock); | 539 | raw_spin_unlock(&desc->lock); |
540 | 540 | ||
541 | return 1; | 541 | return 1; |
542 | } | 542 | } |