diff options
author | Thomas Gleixner <tglx@linutronix.de> | 2009-11-17 10:46:45 -0500 |
---|---|---|
committer | Thomas Gleixner <tglx@linutronix.de> | 2009-12-14 17:55:33 -0500 |
commit | 239007b8440abff689632f50cdf0f2b9e895b534 (patch) | |
tree | 569cab843af4a999d6d868ec9a824530d2bfa733 /kernel/irq | |
parent | 9f5a5621e78cf48d86682a71ceb3fcdbde38b222 (diff) |
genirq: Convert irq_desc.lock to raw_spinlock
Convert locks which cannot be sleeping locks in preempt-rt to
raw_spinlocks.
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Acked-by: Peter Zijlstra <peterz@infradead.org>
Acked-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/irq')
-rw-r--r-- | kernel/irq/autoprobe.c | 20 | ||||
-rw-r--r-- | kernel/irq/chip.c | 86 | ||||
-rw-r--r-- | kernel/irq/handle.c | 22 | ||||
-rw-r--r-- | kernel/irq/internals.h | 2 | ||||
-rw-r--r-- | kernel/irq/manage.c | 50 | ||||
-rw-r--r-- | kernel/irq/migration.c | 2 | ||||
-rw-r--r-- | kernel/irq/numa_migrate.c | 8 | ||||
-rw-r--r-- | kernel/irq/pm.c | 8 | ||||
-rw-r--r-- | kernel/irq/proc.c | 4 | ||||
-rw-r--r-- | kernel/irq/spurious.c | 14 |
10 files changed, 108 insertions, 108 deletions
diff --git a/kernel/irq/autoprobe.c b/kernel/irq/autoprobe.c index 1de9700f416..2295a31ef11 100644 --- a/kernel/irq/autoprobe.c +++ b/kernel/irq/autoprobe.c | |||
@@ -45,7 +45,7 @@ unsigned long probe_irq_on(void) | |||
45 | * flush such a longstanding irq before considering it as spurious. | 45 | * flush such a longstanding irq before considering it as spurious. |
46 | */ | 46 | */ |
47 | for_each_irq_desc_reverse(i, desc) { | 47 | for_each_irq_desc_reverse(i, desc) { |
48 | spin_lock_irq(&desc->lock); | 48 | raw_spin_lock_irq(&desc->lock); |
49 | if (!desc->action && !(desc->status & IRQ_NOPROBE)) { | 49 | if (!desc->action && !(desc->status & IRQ_NOPROBE)) { |
50 | /* | 50 | /* |
51 | * An old-style architecture might still have | 51 | * An old-style architecture might still have |
@@ -61,7 +61,7 @@ unsigned long probe_irq_on(void) | |||
61 | desc->chip->set_type(i, IRQ_TYPE_PROBE); | 61 | desc->chip->set_type(i, IRQ_TYPE_PROBE); |
62 | desc->chip->startup(i); | 62 | desc->chip->startup(i); |
63 | } | 63 | } |
64 | spin_unlock_irq(&desc->lock); | 64 | raw_spin_unlock_irq(&desc->lock); |
65 | } | 65 | } |
66 | 66 | ||
67 | /* Wait for longstanding interrupts to trigger. */ | 67 | /* Wait for longstanding interrupts to trigger. */ |
@@ -73,13 +73,13 @@ unsigned long probe_irq_on(void) | |||
73 | * happened in the previous stage, it may have masked itself) | 73 | * happened in the previous stage, it may have masked itself) |
74 | */ | 74 | */ |
75 | for_each_irq_desc_reverse(i, desc) { | 75 | for_each_irq_desc_reverse(i, desc) { |
76 | spin_lock_irq(&desc->lock); | 76 | raw_spin_lock_irq(&desc->lock); |
77 | if (!desc->action && !(desc->status & IRQ_NOPROBE)) { | 77 | if (!desc->action && !(desc->status & IRQ_NOPROBE)) { |
78 | desc->status |= IRQ_AUTODETECT | IRQ_WAITING; | 78 | desc->status |= IRQ_AUTODETECT | IRQ_WAITING; |
79 | if (desc->chip->startup(i)) | 79 | if (desc->chip->startup(i)) |
80 | desc->status |= IRQ_PENDING; | 80 | desc->status |= IRQ_PENDING; |
81 | } | 81 | } |
82 | spin_unlock_irq(&desc->lock); | 82 | raw_spin_unlock_irq(&desc->lock); |
83 | } | 83 | } |
84 | 84 | ||
85 | /* | 85 | /* |
@@ -91,7 +91,7 @@ unsigned long probe_irq_on(void) | |||
91 | * Now filter out any obviously spurious interrupts | 91 | * Now filter out any obviously spurious interrupts |
92 | */ | 92 | */ |
93 | for_each_irq_desc(i, desc) { | 93 | for_each_irq_desc(i, desc) { |
94 | spin_lock_irq(&desc->lock); | 94 | raw_spin_lock_irq(&desc->lock); |
95 | status = desc->status; | 95 | status = desc->status; |
96 | 96 | ||
97 | if (status & IRQ_AUTODETECT) { | 97 | if (status & IRQ_AUTODETECT) { |
@@ -103,7 +103,7 @@ unsigned long probe_irq_on(void) | |||
103 | if (i < 32) | 103 | if (i < 32) |
104 | mask |= 1 << i; | 104 | mask |= 1 << i; |
105 | } | 105 | } |
106 | spin_unlock_irq(&desc->lock); | 106 | raw_spin_unlock_irq(&desc->lock); |
107 | } | 107 | } |
108 | 108 | ||
109 | return mask; | 109 | return mask; |
@@ -129,7 +129,7 @@ unsigned int probe_irq_mask(unsigned long val) | |||
129 | int i; | 129 | int i; |
130 | 130 | ||
131 | for_each_irq_desc(i, desc) { | 131 | for_each_irq_desc(i, desc) { |
132 | spin_lock_irq(&desc->lock); | 132 | raw_spin_lock_irq(&desc->lock); |
133 | status = desc->status; | 133 | status = desc->status; |
134 | 134 | ||
135 | if (status & IRQ_AUTODETECT) { | 135 | if (status & IRQ_AUTODETECT) { |
@@ -139,7 +139,7 @@ unsigned int probe_irq_mask(unsigned long val) | |||
139 | desc->status = status & ~IRQ_AUTODETECT; | 139 | desc->status = status & ~IRQ_AUTODETECT; |
140 | desc->chip->shutdown(i); | 140 | desc->chip->shutdown(i); |
141 | } | 141 | } |
142 | spin_unlock_irq(&desc->lock); | 142 | raw_spin_unlock_irq(&desc->lock); |
143 | } | 143 | } |
144 | mutex_unlock(&probing_active); | 144 | mutex_unlock(&probing_active); |
145 | 145 | ||
@@ -171,7 +171,7 @@ int probe_irq_off(unsigned long val) | |||
171 | unsigned int status; | 171 | unsigned int status; |
172 | 172 | ||
173 | for_each_irq_desc(i, desc) { | 173 | for_each_irq_desc(i, desc) { |
174 | spin_lock_irq(&desc->lock); | 174 | raw_spin_lock_irq(&desc->lock); |
175 | status = desc->status; | 175 | status = desc->status; |
176 | 176 | ||
177 | if (status & IRQ_AUTODETECT) { | 177 | if (status & IRQ_AUTODETECT) { |
@@ -183,7 +183,7 @@ int probe_irq_off(unsigned long val) | |||
183 | desc->status = status & ~IRQ_AUTODETECT; | 183 | desc->status = status & ~IRQ_AUTODETECT; |
184 | desc->chip->shutdown(i); | 184 | desc->chip->shutdown(i); |
185 | } | 185 | } |
186 | spin_unlock_irq(&desc->lock); | 186 | raw_spin_unlock_irq(&desc->lock); |
187 | } | 187 | } |
188 | mutex_unlock(&probing_active); | 188 | mutex_unlock(&probing_active); |
189 | 189 | ||
diff --git a/kernel/irq/chip.c b/kernel/irq/chip.c index ba566c261ad..ecc3fa28f66 100644 --- a/kernel/irq/chip.c +++ b/kernel/irq/chip.c | |||
@@ -34,7 +34,7 @@ void dynamic_irq_init(unsigned int irq) | |||
34 | } | 34 | } |
35 | 35 | ||
36 | /* Ensure we don't have left over values from a previous use of this irq */ | 36 | /* Ensure we don't have left over values from a previous use of this irq */ |
37 | spin_lock_irqsave(&desc->lock, flags); | 37 | raw_spin_lock_irqsave(&desc->lock, flags); |
38 | desc->status = IRQ_DISABLED; | 38 | desc->status = IRQ_DISABLED; |
39 | desc->chip = &no_irq_chip; | 39 | desc->chip = &no_irq_chip; |
40 | desc->handle_irq = handle_bad_irq; | 40 | desc->handle_irq = handle_bad_irq; |
@@ -51,7 +51,7 @@ void dynamic_irq_init(unsigned int irq) | |||
51 | cpumask_clear(desc->pending_mask); | 51 | cpumask_clear(desc->pending_mask); |
52 | #endif | 52 | #endif |
53 | #endif | 53 | #endif |
54 | spin_unlock_irqrestore(&desc->lock, flags); | 54 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
55 | } | 55 | } |
56 | 56 | ||
57 | /** | 57 | /** |
@@ -68,9 +68,9 @@ void dynamic_irq_cleanup(unsigned int irq) | |||
68 | return; | 68 | return; |
69 | } | 69 | } |
70 | 70 | ||
71 | spin_lock_irqsave(&desc->lock, flags); | 71 | raw_spin_lock_irqsave(&desc->lock, flags); |
72 | if (desc->action) { | 72 | if (desc->action) { |
73 | spin_unlock_irqrestore(&desc->lock, flags); | 73 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
74 | WARN(1, KERN_ERR "Destroying IRQ%d without calling free_irq\n", | 74 | WARN(1, KERN_ERR "Destroying IRQ%d without calling free_irq\n", |
75 | irq); | 75 | irq); |
76 | return; | 76 | return; |
@@ -82,7 +82,7 @@ void dynamic_irq_cleanup(unsigned int irq) | |||
82 | desc->chip = &no_irq_chip; | 82 | desc->chip = &no_irq_chip; |
83 | desc->name = NULL; | 83 | desc->name = NULL; |
84 | clear_kstat_irqs(desc); | 84 | clear_kstat_irqs(desc); |
85 | spin_unlock_irqrestore(&desc->lock, flags); | 85 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
86 | } | 86 | } |
87 | 87 | ||
88 | 88 | ||
@@ -104,10 +104,10 @@ int set_irq_chip(unsigned int irq, struct irq_chip *chip) | |||
104 | if (!chip) | 104 | if (!chip) |
105 | chip = &no_irq_chip; | 105 | chip = &no_irq_chip; |
106 | 106 | ||
107 | spin_lock_irqsave(&desc->lock, flags); | 107 | raw_spin_lock_irqsave(&desc->lock, flags); |
108 | irq_chip_set_defaults(chip); | 108 | irq_chip_set_defaults(chip); |
109 | desc->chip = chip; | 109 | desc->chip = chip; |
110 | spin_unlock_irqrestore(&desc->lock, flags); | 110 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
111 | 111 | ||
112 | return 0; | 112 | return 0; |
113 | } | 113 | } |
@@ -133,9 +133,9 @@ int set_irq_type(unsigned int irq, unsigned int type) | |||
133 | if (type == IRQ_TYPE_NONE) | 133 | if (type == IRQ_TYPE_NONE) |
134 | return 0; | 134 | return 0; |
135 | 135 | ||
136 | spin_lock_irqsave(&desc->lock, flags); | 136 | raw_spin_lock_irqsave(&desc->lock, flags); |
137 | ret = __irq_set_trigger(desc, irq, type); | 137 | ret = __irq_set_trigger(desc, irq, type); |
138 | spin_unlock_irqrestore(&desc->lock, flags); | 138 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
139 | return ret; | 139 | return ret; |
140 | } | 140 | } |
141 | EXPORT_SYMBOL(set_irq_type); | 141 | EXPORT_SYMBOL(set_irq_type); |
@@ -158,9 +158,9 @@ int set_irq_data(unsigned int irq, void *data) | |||
158 | return -EINVAL; | 158 | return -EINVAL; |
159 | } | 159 | } |
160 | 160 | ||
161 | spin_lock_irqsave(&desc->lock, flags); | 161 | raw_spin_lock_irqsave(&desc->lock, flags); |
162 | desc->handler_data = data; | 162 | desc->handler_data = data; |
163 | spin_unlock_irqrestore(&desc->lock, flags); | 163 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
164 | return 0; | 164 | return 0; |
165 | } | 165 | } |
166 | EXPORT_SYMBOL(set_irq_data); | 166 | EXPORT_SYMBOL(set_irq_data); |
@@ -183,11 +183,11 @@ int set_irq_msi(unsigned int irq, struct msi_desc *entry) | |||
183 | return -EINVAL; | 183 | return -EINVAL; |
184 | } | 184 | } |
185 | 185 | ||
186 | spin_lock_irqsave(&desc->lock, flags); | 186 | raw_spin_lock_irqsave(&desc->lock, flags); |
187 | desc->msi_desc = entry; | 187 | desc->msi_desc = entry; |
188 | if (entry) | 188 | if (entry) |
189 | entry->irq = irq; | 189 | entry->irq = irq; |
190 | spin_unlock_irqrestore(&desc->lock, flags); | 190 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
191 | return 0; | 191 | return 0; |
192 | } | 192 | } |
193 | 193 | ||
@@ -214,9 +214,9 @@ int set_irq_chip_data(unsigned int irq, void *data) | |||
214 | return -EINVAL; | 214 | return -EINVAL; |
215 | } | 215 | } |
216 | 216 | ||
217 | spin_lock_irqsave(&desc->lock, flags); | 217 | raw_spin_lock_irqsave(&desc->lock, flags); |
218 | desc->chip_data = data; | 218 | desc->chip_data = data; |
219 | spin_unlock_irqrestore(&desc->lock, flags); | 219 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
220 | 220 | ||
221 | return 0; | 221 | return 0; |
222 | } | 222 | } |
@@ -241,12 +241,12 @@ void set_irq_nested_thread(unsigned int irq, int nest) | |||
241 | if (!desc) | 241 | if (!desc) |
242 | return; | 242 | return; |
243 | 243 | ||
244 | spin_lock_irqsave(&desc->lock, flags); | 244 | raw_spin_lock_irqsave(&desc->lock, flags); |
245 | if (nest) | 245 | if (nest) |
246 | desc->status |= IRQ_NESTED_THREAD; | 246 | desc->status |= IRQ_NESTED_THREAD; |
247 | else | 247 | else |
248 | desc->status &= ~IRQ_NESTED_THREAD; | 248 | desc->status &= ~IRQ_NESTED_THREAD; |
249 | spin_unlock_irqrestore(&desc->lock, flags); | 249 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
250 | } | 250 | } |
251 | EXPORT_SYMBOL_GPL(set_irq_nested_thread); | 251 | EXPORT_SYMBOL_GPL(set_irq_nested_thread); |
252 | 252 | ||
@@ -343,7 +343,7 @@ void handle_nested_irq(unsigned int irq) | |||
343 | 343 | ||
344 | might_sleep(); | 344 | might_sleep(); |
345 | 345 | ||
346 | spin_lock_irq(&desc->lock); | 346 | raw_spin_lock_irq(&desc->lock); |
347 | 347 | ||
348 | kstat_incr_irqs_this_cpu(irq, desc); | 348 | kstat_incr_irqs_this_cpu(irq, desc); |
349 | 349 | ||
@@ -352,17 +352,17 @@ void handle_nested_irq(unsigned int irq) | |||
352 | goto out_unlock; | 352 | goto out_unlock; |
353 | 353 | ||
354 | desc->status |= IRQ_INPROGRESS; | 354 | desc->status |= IRQ_INPROGRESS; |
355 | spin_unlock_irq(&desc->lock); | 355 | raw_spin_unlock_irq(&desc->lock); |
356 | 356 | ||
357 | action_ret = action->thread_fn(action->irq, action->dev_id); | 357 | action_ret = action->thread_fn(action->irq, action->dev_id); |
358 | if (!noirqdebug) | 358 | if (!noirqdebug) |
359 | note_interrupt(irq, desc, action_ret); | 359 | note_interrupt(irq, desc, action_ret); |
360 | 360 | ||
361 | spin_lock_irq(&desc->lock); | 361 | raw_spin_lock_irq(&desc->lock); |
362 | desc->status &= ~IRQ_INPROGRESS; | 362 | desc->status &= ~IRQ_INPROGRESS; |
363 | 363 | ||
364 | out_unlock: | 364 | out_unlock: |
365 | spin_unlock_irq(&desc->lock); | 365 | raw_spin_unlock_irq(&desc->lock); |
366 | } | 366 | } |
367 | EXPORT_SYMBOL_GPL(handle_nested_irq); | 367 | EXPORT_SYMBOL_GPL(handle_nested_irq); |
368 | 368 | ||
@@ -384,7 +384,7 @@ handle_simple_irq(unsigned int irq, struct irq_desc *desc) | |||
384 | struct irqaction *action; | 384 | struct irqaction *action; |
385 | irqreturn_t action_ret; | 385 | irqreturn_t action_ret; |
386 | 386 | ||
387 | spin_lock(&desc->lock); | 387 | raw_spin_lock(&desc->lock); |
388 | 388 | ||
389 | if (unlikely(desc->status & IRQ_INPROGRESS)) | 389 | if (unlikely(desc->status & IRQ_INPROGRESS)) |
390 | goto out_unlock; | 390 | goto out_unlock; |
@@ -396,16 +396,16 @@ handle_simple_irq(unsigned int irq, struct irq_desc *desc) | |||
396 | goto out_unlock; | 396 | goto out_unlock; |
397 | 397 | ||
398 | desc->status |= IRQ_INPROGRESS; | 398 | desc->status |= IRQ_INPROGRESS; |
399 | spin_unlock(&desc->lock); | 399 | raw_spin_unlock(&desc->lock); |
400 | 400 | ||
401 | action_ret = handle_IRQ_event(irq, action); | 401 | action_ret = handle_IRQ_event(irq, action); |
402 | if (!noirqdebug) | 402 | if (!noirqdebug) |
403 | note_interrupt(irq, desc, action_ret); | 403 | note_interrupt(irq, desc, action_ret); |
404 | 404 | ||
405 | spin_lock(&desc->lock); | 405 | raw_spin_lock(&desc->lock); |
406 | desc->status &= ~IRQ_INPROGRESS; | 406 | desc->status &= ~IRQ_INPROGRESS; |
407 | out_unlock: | 407 | out_unlock: |
408 | spin_unlock(&desc->lock); | 408 | raw_spin_unlock(&desc->lock); |
409 | } | 409 | } |
410 | 410 | ||
411 | /** | 411 | /** |
@@ -424,7 +424,7 @@ handle_level_irq(unsigned int irq, struct irq_desc *desc) | |||
424 | struct irqaction *action; | 424 | struct irqaction *action; |
425 | irqreturn_t action_ret; | 425 | irqreturn_t action_ret; |
426 | 426 | ||
427 | spin_lock(&desc->lock); | 427 | raw_spin_lock(&desc->lock); |
428 | mask_ack_irq(desc, irq); | 428 | mask_ack_irq(desc, irq); |
429 | 429 | ||
430 | if (unlikely(desc->status & IRQ_INPROGRESS)) | 430 | if (unlikely(desc->status & IRQ_INPROGRESS)) |
@@ -441,13 +441,13 @@ handle_level_irq(unsigned int irq, struct irq_desc *desc) | |||
441 | goto out_unlock; | 441 | goto out_unlock; |
442 | 442 | ||
443 | desc->status |= IRQ_INPROGRESS; | 443 | desc->status |= IRQ_INPROGRESS; |
444 | spin_unlock(&desc->lock); | 444 | raw_spin_unlock(&desc->lock); |
445 | 445 | ||
446 | action_ret = handle_IRQ_event(irq, action); | 446 | action_ret = handle_IRQ_event(irq, action); |
447 | if (!noirqdebug) | 447 | if (!noirqdebug) |
448 | note_interrupt(irq, desc, action_ret); | 448 | note_interrupt(irq, desc, action_ret); |
449 | 449 | ||
450 | spin_lock(&desc->lock); | 450 | raw_spin_lock(&desc->lock); |
451 | desc->status &= ~IRQ_INPROGRESS; | 451 | desc->status &= ~IRQ_INPROGRESS; |
452 | 452 | ||
453 | if (unlikely(desc->status & IRQ_ONESHOT)) | 453 | if (unlikely(desc->status & IRQ_ONESHOT)) |
@@ -455,7 +455,7 @@ handle_level_irq(unsigned int irq, struct irq_desc *desc) | |||
455 | else if (!(desc->status & IRQ_DISABLED) && desc->chip->unmask) | 455 | else if (!(desc->status & IRQ_DISABLED) && desc->chip->unmask) |
456 | desc->chip->unmask(irq); | 456 | desc->chip->unmask(irq); |
457 | out_unlock: | 457 | out_unlock: |
458 | spin_unlock(&desc->lock); | 458 | raw_spin_unlock(&desc->lock); |
459 | } | 459 | } |
460 | EXPORT_SYMBOL_GPL(handle_level_irq); | 460 | EXPORT_SYMBOL_GPL(handle_level_irq); |
461 | 461 | ||
@@ -475,7 +475,7 @@ handle_fasteoi_irq(unsigned int irq, struct irq_desc *desc) | |||
475 | struct irqaction *action; | 475 | struct irqaction *action; |
476 | irqreturn_t action_ret; | 476 | irqreturn_t action_ret; |
477 | 477 | ||
478 | spin_lock(&desc->lock); | 478 | raw_spin_lock(&desc->lock); |
479 | 479 | ||
480 | if (unlikely(desc->status & IRQ_INPROGRESS)) | 480 | if (unlikely(desc->status & IRQ_INPROGRESS)) |
481 | goto out; | 481 | goto out; |
@@ -497,18 +497,18 @@ handle_fasteoi_irq(unsigned int irq, struct irq_desc *desc) | |||
497 | 497 | ||
498 | desc->status |= IRQ_INPROGRESS; | 498 | desc->status |= IRQ_INPROGRESS; |
499 | desc->status &= ~IRQ_PENDING; | 499 | desc->status &= ~IRQ_PENDING; |
500 | spin_unlock(&desc->lock); | 500 | raw_spin_unlock(&desc->lock); |
501 | 501 | ||
502 | action_ret = handle_IRQ_event(irq, action); | 502 | action_ret = handle_IRQ_event(irq, action); |
503 | if (!noirqdebug) | 503 | if (!noirqdebug) |
504 | note_interrupt(irq, desc, action_ret); | 504 | note_interrupt(irq, desc, action_ret); |
505 | 505 | ||
506 | spin_lock(&desc->lock); | 506 | raw_spin_lock(&desc->lock); |
507 | desc->status &= ~IRQ_INPROGRESS; | 507 | desc->status &= ~IRQ_INPROGRESS; |
508 | out: | 508 | out: |
509 | desc->chip->eoi(irq); | 509 | desc->chip->eoi(irq); |
510 | 510 | ||
511 | spin_unlock(&desc->lock); | 511 | raw_spin_unlock(&desc->lock); |
512 | } | 512 | } |
513 | 513 | ||
514 | /** | 514 | /** |
@@ -530,7 +530,7 @@ out: | |||
530 | void | 530 | void |
531 | handle_edge_irq(unsigned int irq, struct irq_desc *desc) | 531 | handle_edge_irq(unsigned int irq, struct irq_desc *desc) |
532 | { | 532 | { |
533 | spin_lock(&desc->lock); | 533 | raw_spin_lock(&desc->lock); |
534 | 534 | ||
535 | desc->status &= ~(IRQ_REPLAY | IRQ_WAITING); | 535 | desc->status &= ~(IRQ_REPLAY | IRQ_WAITING); |
536 | 536 | ||
@@ -576,17 +576,17 @@ handle_edge_irq(unsigned int irq, struct irq_desc *desc) | |||
576 | } | 576 | } |
577 | 577 | ||
578 | desc->status &= ~IRQ_PENDING; | 578 | desc->status &= ~IRQ_PENDING; |
579 | spin_unlock(&desc->lock); | 579 | raw_spin_unlock(&desc->lock); |
580 | action_ret = handle_IRQ_event(irq, action); | 580 | action_ret = handle_IRQ_event(irq, action); |
581 | if (!noirqdebug) | 581 | if (!noirqdebug) |
582 | note_interrupt(irq, desc, action_ret); | 582 | note_interrupt(irq, desc, action_ret); |
583 | spin_lock(&desc->lock); | 583 | raw_spin_lock(&desc->lock); |
584 | 584 | ||
585 | } while ((desc->status & (IRQ_PENDING | IRQ_DISABLED)) == IRQ_PENDING); | 585 | } while ((desc->status & (IRQ_PENDING | IRQ_DISABLED)) == IRQ_PENDING); |
586 | 586 | ||
587 | desc->status &= ~IRQ_INPROGRESS; | 587 | desc->status &= ~IRQ_INPROGRESS; |
588 | out_unlock: | 588 | out_unlock: |
589 | spin_unlock(&desc->lock); | 589 | raw_spin_unlock(&desc->lock); |
590 | } | 590 | } |
591 | 591 | ||
592 | /** | 592 | /** |
@@ -643,7 +643,7 @@ __set_irq_handler(unsigned int irq, irq_flow_handler_t handle, int is_chained, | |||
643 | } | 643 | } |
644 | 644 | ||
645 | chip_bus_lock(irq, desc); | 645 | chip_bus_lock(irq, desc); |
646 | spin_lock_irqsave(&desc->lock, flags); | 646 | raw_spin_lock_irqsave(&desc->lock, flags); |
647 | 647 | ||
648 | /* Uninstall? */ | 648 | /* Uninstall? */ |
649 | if (handle == handle_bad_irq) { | 649 | if (handle == handle_bad_irq) { |
@@ -661,7 +661,7 @@ __set_irq_handler(unsigned int irq, irq_flow_handler_t handle, int is_chained, | |||
661 | desc->depth = 0; | 661 | desc->depth = 0; |
662 | desc->chip->startup(irq); | 662 | desc->chip->startup(irq); |
663 | } | 663 | } |
664 | spin_unlock_irqrestore(&desc->lock, flags); | 664 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
665 | chip_bus_sync_unlock(irq, desc); | 665 | chip_bus_sync_unlock(irq, desc); |
666 | } | 666 | } |
667 | EXPORT_SYMBOL_GPL(__set_irq_handler); | 667 | EXPORT_SYMBOL_GPL(__set_irq_handler); |
@@ -692,9 +692,9 @@ void __init set_irq_noprobe(unsigned int irq) | |||
692 | return; | 692 | return; |
693 | } | 693 | } |
694 | 694 | ||
695 | spin_lock_irqsave(&desc->lock, flags); | 695 | raw_spin_lock_irqsave(&desc->lock, flags); |
696 | desc->status |= IRQ_NOPROBE; | 696 | desc->status |= IRQ_NOPROBE; |
697 | spin_unlock_irqrestore(&desc->lock, flags); | 697 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
698 | } | 698 | } |
699 | 699 | ||
700 | void __init set_irq_probe(unsigned int irq) | 700 | void __init set_irq_probe(unsigned int irq) |
@@ -707,7 +707,7 @@ void __init set_irq_probe(unsigned int irq) | |||
707 | return; | 707 | return; |
708 | } | 708 | } |
709 | 709 | ||
710 | spin_lock_irqsave(&desc->lock, flags); | 710 | raw_spin_lock_irqsave(&desc->lock, flags); |
711 | desc->status &= ~IRQ_NOPROBE; | 711 | desc->status &= ~IRQ_NOPROBE; |
712 | spin_unlock_irqrestore(&desc->lock, flags); | 712 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
713 | } | 713 | } |
diff --git a/kernel/irq/handle.c b/kernel/irq/handle.c index 17c71bb565c..814940e7f48 100644 --- a/kernel/irq/handle.c +++ b/kernel/irq/handle.c | |||
@@ -80,7 +80,7 @@ static struct irq_desc irq_desc_init = { | |||
80 | .chip = &no_irq_chip, | 80 | .chip = &no_irq_chip, |
81 | .handle_irq = handle_bad_irq, | 81 | .handle_irq = handle_bad_irq, |
82 | .depth = 1, | 82 | .depth = 1, |
83 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc_init.lock), | 83 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc_init.lock), |
84 | }; | 84 | }; |
85 | 85 | ||
86 | void __ref init_kstat_irqs(struct irq_desc *desc, int node, int nr) | 86 | void __ref init_kstat_irqs(struct irq_desc *desc, int node, int nr) |
@@ -108,7 +108,7 @@ static void init_one_irq_desc(int irq, struct irq_desc *desc, int node) | |||
108 | { | 108 | { |
109 | memcpy(desc, &irq_desc_init, sizeof(struct irq_desc)); | 109 | memcpy(desc, &irq_desc_init, sizeof(struct irq_desc)); |
110 | 110 | ||
111 | spin_lock_init(&desc->lock); | 111 | raw_spin_lock_init(&desc->lock); |
112 | desc->irq = irq; | 112 | desc->irq = irq; |
113 | #ifdef CONFIG_SMP | 113 | #ifdef CONFIG_SMP |
114 | desc->node = node; | 114 | desc->node = node; |
@@ -130,7 +130,7 @@ static void init_one_irq_desc(int irq, struct irq_desc *desc, int node) | |||
130 | /* | 130 | /* |
131 | * Protect the sparse_irqs: | 131 | * Protect the sparse_irqs: |
132 | */ | 132 | */ |
133 | DEFINE_SPINLOCK(sparse_irq_lock); | 133 | DEFINE_RAW_SPINLOCK(sparse_irq_lock); |
134 | 134 | ||
135 | struct irq_desc **irq_desc_ptrs __read_mostly; | 135 | struct irq_desc **irq_desc_ptrs __read_mostly; |
136 | 136 | ||
@@ -141,7 +141,7 @@ static struct irq_desc irq_desc_legacy[NR_IRQS_LEGACY] __cacheline_aligned_in_sm | |||
141 | .chip = &no_irq_chip, | 141 | .chip = &no_irq_chip, |
142 | .handle_irq = handle_bad_irq, | 142 | .handle_irq = handle_bad_irq, |
143 | .depth = 1, | 143 | .depth = 1, |
144 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc_init.lock), | 144 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc_init.lock), |
145 | } | 145 | } |
146 | }; | 146 | }; |
147 | 147 | ||
@@ -212,7 +212,7 @@ struct irq_desc * __ref irq_to_desc_alloc_node(unsigned int irq, int node) | |||
212 | if (desc) | 212 | if (desc) |
213 | return desc; | 213 | return desc; |
214 | 214 | ||
215 | spin_lock_irqsave(&sparse_irq_lock, flags); | 215 | raw_spin_lock_irqsave(&sparse_irq_lock, flags); |
216 | 216 | ||
217 | /* We have to check it to avoid races with another CPU */ | 217 | /* We have to check it to avoid races with another CPU */ |
218 | desc = irq_desc_ptrs[irq]; | 218 | desc = irq_desc_ptrs[irq]; |
@@ -234,7 +234,7 @@ struct irq_desc * __ref irq_to_desc_alloc_node(unsigned int irq, int node) | |||
234 | irq_desc_ptrs[irq] = desc; | 234 | irq_desc_ptrs[irq] = desc; |
235 | 235 | ||
236 | out_unlock: | 236 | out_unlock: |
237 | spin_unlock_irqrestore(&sparse_irq_lock, flags); | 237 | raw_spin_unlock_irqrestore(&sparse_irq_lock, flags); |
238 | 238 | ||
239 | return desc; | 239 | return desc; |
240 | } | 240 | } |
@@ -247,7 +247,7 @@ struct irq_desc irq_desc[NR_IRQS] __cacheline_aligned_in_smp = { | |||
247 | .chip = &no_irq_chip, | 247 | .chip = &no_irq_chip, |
248 | .handle_irq = handle_bad_irq, | 248 | .handle_irq = handle_bad_irq, |
249 | .depth = 1, | 249 | .depth = 1, |
250 | .lock = __SPIN_LOCK_UNLOCKED(irq_desc->lock), | 250 | .lock = __RAW_SPIN_LOCK_UNLOCKED(irq_desc->lock), |
251 | } | 251 | } |
252 | }; | 252 | }; |
253 | 253 | ||
@@ -473,7 +473,7 @@ unsigned int __do_IRQ(unsigned int irq) | |||
473 | return 1; | 473 | return 1; |
474 | } | 474 | } |
475 | 475 | ||
476 | spin_lock(&desc->lock); | 476 | raw_spin_lock(&desc->lock); |
477 | if (desc->chip->ack) | 477 | if (desc->chip->ack) |
478 | desc->chip->ack(irq); | 478 | desc->chip->ack(irq); |
479 | /* | 479 | /* |
@@ -517,13 +517,13 @@ unsigned int __do_IRQ(unsigned int irq) | |||
517 | for (;;) { | 517 | for (;;) { |
518 | irqreturn_t action_ret; | 518 | irqreturn_t action_ret; |
519 | 519 | ||
520 | spin_unlock(&desc->lock); | 520 | raw_spin_unlock(&desc->lock); |
521 | 521 | ||
522 | action_ret = handle_IRQ_event(irq, action); | 522 | action_ret = handle_IRQ_event(irq, action); |
523 | if (!noirqdebug) | 523 | if (!noirqdebug) |
524 | note_interrupt(irq, desc, action_ret); | 524 | note_interrupt(irq, desc, action_ret); |
525 | 525 | ||
526 | spin_lock(&desc->lock); | 526 | raw_spin_lock(&desc->lock); |
527 | if (likely(!(desc->status & IRQ_PENDING))) | 527 | if (likely(!(desc->status & IRQ_PENDING))) |
528 | break; | 528 | break; |
529 | desc->status &= ~IRQ_PENDING; | 529 | desc->status &= ~IRQ_PENDING; |
@@ -536,7 +536,7 @@ out: | |||
536 | * disabled while the handler was running. | 536 | * disabled while the handler was running. |
537 | */ | 537 | */ |
538 | desc->chip->end(irq); | 538 | desc->chip->end(irq); |
539 | spin_unlock(&desc->lock); | 539 | raw_spin_unlock(&desc->lock); |
540 | 540 | ||
541 | return 1; | 541 | return 1; |
542 | } | 542 | } |
diff --git a/kernel/irq/internals.h b/kernel/irq/internals.h index 1b5d742c6a7..b2821f070a3 100644 --- a/kernel/irq/internals.h +++ b/kernel/irq/internals.h | |||
@@ -18,7 +18,7 @@ extern void __enable_irq(struct irq_desc *desc, unsigned int irq, bool resume); | |||
18 | extern struct lock_class_key irq_desc_lock_class; | 18 | extern struct lock_class_key irq_desc_lock_class; |
19 | extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr); | 19 | extern void init_kstat_irqs(struct irq_desc *desc, int node, int nr); |
20 | extern void clear_kstat_irqs(struct irq_desc *desc); | 20 | extern void clear_kstat_irqs(struct irq_desc *desc); |
21 | extern spinlock_t sparse_irq_lock; | 21 | extern raw_spinlock_t sparse_irq_lock; |
22 | 22 | ||
23 | #ifdef CONFIG_SPARSE_IRQ | 23 | #ifdef CONFIG_SPARSE_IRQ |
24 | /* irq_desc_ptrs allocated at boot time */ | 24 | /* irq_desc_ptrs allocated at boot time */ |
diff --git a/kernel/irq/manage.c b/kernel/irq/manage.c index 7305b297d1e..eb6078ca60c 100644 --- a/kernel/irq/manage.c +++ b/kernel/irq/manage.c | |||
@@ -46,9 +46,9 @@ void synchronize_irq(unsigned int irq) | |||
46 | cpu_relax(); | 46 | cpu_relax(); |
47 | 47 | ||
48 | /* Ok, that indicated we're done: double-check carefully. */ | 48 | /* Ok, that indicated we're done: double-check carefully. */ |
49 | spin_lock_irqsave(&desc->lock, flags); | 49 | raw_spin_lock_irqsave(&desc->lock, flags); |
50 | status = desc->status; | 50 | status = desc->status; |
51 | spin_unlock_irqrestore(&desc->lock, flags); | 51 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
52 | 52 | ||
53 | /* Oops, that failed? */ | 53 | /* Oops, that failed? */ |
54 | } while (status & IRQ_INPROGRESS); | 54 | } while (status & IRQ_INPROGRESS); |
@@ -114,7 +114,7 @@ int irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) | |||
114 | if (!desc->chip->set_affinity) | 114 | if (!desc->chip->set_affinity) |
115 | return -EINVAL; | 115 | return -EINVAL; |
116 | 116 | ||
117 | spin_lock_irqsave(&desc->lock, flags); | 117 | raw_spin_lock_irqsave(&desc->lock, flags); |
118 | 118 | ||
119 | #ifdef CONFIG_GENERIC_PENDING_IRQ | 119 | #ifdef CONFIG_GENERIC_PENDING_IRQ |
120 | if (desc->status & IRQ_MOVE_PCNTXT) { | 120 | if (desc->status & IRQ_MOVE_PCNTXT) { |
@@ -134,7 +134,7 @@ int irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) | |||
134 | } | 134 | } |
135 | #endif | 135 | #endif |
136 | desc->status |= IRQ_AFFINITY_SET; | 136 | desc->status |= IRQ_AFFINITY_SET; |
137 | spin_unlock_irqrestore(&desc->lock, flags); | 137 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
138 | return 0; | 138 | return 0; |
139 | } | 139 | } |
140 | 140 | ||
@@ -181,11 +181,11 @@ int irq_select_affinity_usr(unsigned int irq) | |||
181 | unsigned long flags; | 181 | unsigned long flags; |
182 | int ret; | 182 | int ret; |
183 | 183 | ||
184 | spin_lock_irqsave(&desc->lock, flags); | 184 | raw_spin_lock_irqsave(&desc->lock, flags); |
185 | ret = setup_affinity(irq, desc); | 185 | ret = setup_affinity(irq, desc); |
186 | if (!ret) | 186 | if (!ret) |
187 | irq_set_thread_affinity(desc); | 187 | irq_set_thread_affinity(desc); |
188 | spin_unlock_irqrestore(&desc->lock, flags); | 188 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
189 | 189 | ||
190 | return ret; | 190 | return ret; |
191 | } | 191 | } |
@@ -231,9 +231,9 @@ void disable_irq_nosync(unsigned int irq) | |||
231 | return; | 231 | return; |
232 | 232 | ||
233 | chip_bus_lock(irq, desc); | 233 | chip_bus_lock(irq, desc); |
234 | spin_lock_irqsave(&desc->lock, flags); | 234 | raw_spin_lock_irqsave(&desc->lock, flags); |
235 | __disable_irq(desc, irq, false); | 235 | __disable_irq(desc, irq, false); |
236 | spin_unlock_irqrestore(&desc->lock, flags); | 236 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
237 | chip_bus_sync_unlock(irq, desc); | 237 | chip_bus_sync_unlock(irq, desc); |
238 | } | 238 | } |
239 | EXPORT_SYMBOL(disable_irq_nosync); | 239 | EXPORT_SYMBOL(disable_irq_nosync); |
@@ -308,9 +308,9 @@ void enable_irq(unsigned int irq) | |||
308 | return; | 308 | return; |
309 | 309 | ||
310 | chip_bus_lock(irq, desc); | 310 | chip_bus_lock(irq, desc); |
311 | spin_lock_irqsave(&desc->lock, flags); | 311 | raw_spin_lock_irqsave(&desc->lock, flags); |
312 | __enable_irq(desc, irq, false); | 312 | __enable_irq(desc, irq, false); |
313 | spin_unlock_irqrestore(&desc->lock, flags); | 313 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
314 | chip_bus_sync_unlock(irq, desc); | 314 | chip_bus_sync_unlock(irq, desc); |
315 | } | 315 | } |
316 | EXPORT_SYMBOL(enable_irq); | 316 | EXPORT_SYMBOL(enable_irq); |
@@ -347,7 +347,7 @@ int set_irq_wake(unsigned int irq, unsigned int on) | |||
347 | /* wakeup-capable irqs can be shared between drivers that | 347 | /* wakeup-capable irqs can be shared between drivers that |
348 | * don't need to have the same sleep mode behaviors. | 348 | * don't need to have the same sleep mode behaviors. |
349 | */ | 349 | */ |
350 | spin_lock_irqsave(&desc->lock, flags); | 350 | raw_spin_lock_irqsave(&desc->lock, flags); |
351 | if (on) { | 351 | if (on) { |
352 | if (desc->wake_depth++ == 0) { | 352 | if (desc->wake_depth++ == 0) { |
353 | ret = set_irq_wake_real(irq, on); | 353 | ret = set_irq_wake_real(irq, on); |
@@ -368,7 +368,7 @@ int set_irq_wake(unsigned int irq, unsigned int on) | |||
368 | } | 368 | } |
369 | } | 369 | } |
370 | 370 | ||
371 | spin_unlock_irqrestore(&desc->lock, flags); | 371 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
372 | return ret; | 372 | return ret; |
373 | } | 373 | } |
374 | EXPORT_SYMBOL(set_irq_wake); | 374 | EXPORT_SYMBOL(set_irq_wake); |
@@ -484,12 +484,12 @@ static int irq_wait_for_interrupt(struct irqaction *action) | |||
484 | static void irq_finalize_oneshot(unsigned int irq, struct irq_desc *desc) | 484 | static void irq_finalize_oneshot(unsigned int irq, struct irq_desc *desc) |
485 | { | 485 | { |
486 | chip_bus_lock(irq, desc); | 486 | chip_bus_lock(irq, desc); |
487 | spin_lock_irq(&desc->lock); | 487 | raw_spin_lock_irq(&desc->lock); |
488 | if (!(desc->status & IRQ_DISABLED) && (desc->status & IRQ_MASKED)) { | 488 | if (!(desc->status & IRQ_DISABLED) && (desc->status & IRQ_MASKED)) { |
489 | desc->status &= ~IRQ_MASKED; | 489 | desc->status &= ~IRQ_MASKED; |
490 | desc->chip->unmask(irq); | 490 | desc->chip->unmask(irq); |
491 | } | 491 | } |
492 | spin_unlock_irq(&desc->lock); | 492 | raw_spin_unlock_irq(&desc->lock); |
493 | chip_bus_sync_unlock(irq, desc); | 493 | chip_bus_sync_unlock(irq, desc); |
494 | } | 494 | } |
495 | 495 | ||
@@ -514,9 +514,9 @@ irq_thread_check_affinity(struct irq_desc *desc, struct irqaction *action) | |||
514 | return; | 514 | return; |
515 | } | 515 | } |
516 | 516 | ||
517 | spin_lock_irq(&desc->lock); | 517 | raw_spin_lock_irq(&desc->lock); |
518 | cpumask_copy(mask, desc->affinity); | 518 | cpumask_copy(mask, desc->affinity); |
519 | spin_unlock_irq(&desc->lock); | 519 | raw_spin_unlock_irq(&desc->lock); |
520 | 520 | ||
521 | set_cpus_allowed_ptr(current, mask); | 521 | set_cpus_allowed_ptr(current, mask); |
522 | free_cpumask_var(mask); | 522 | free_cpumask_var(mask); |
@@ -545,7 +545,7 @@ static int irq_thread(void *data) | |||
545 | 545 | ||
546 | atomic_inc(&desc->threads_active); | 546 | atomic_inc(&desc->threads_active); |
547 | 547 | ||
548 | spin_lock_irq(&desc->lock); | 548 | raw_spin_lock_irq(&desc->lock); |
549 | if (unlikely(desc->status & IRQ_DISABLED)) { | 549 | if (unlikely(desc->status & IRQ_DISABLED)) { |
550 | /* | 550 | /* |
551 | * CHECKME: We might need a dedicated | 551 | * CHECKME: We might need a dedicated |
@@ -555,9 +555,9 @@ static int irq_thread(void *data) | |||
555 | * retriggers the interrupt itself --- tglx | 555 | * retriggers the interrupt itself --- tglx |
556 | */ | 556 | */ |
557 | desc->status |= IRQ_PENDING; | 557 | desc->status |= IRQ_PENDING; |
558 | spin_unlock_irq(&desc->lock); | 558 | raw_spin_unlock_irq(&desc->lock); |
559 | } else { | 559 | } else { |
560 | spin_unlock_irq(&desc->lock); | 560 | raw_spin_unlock_irq(&desc->lock); |
561 | 561 | ||
562 | action->thread_fn(action->irq, action->dev_id); | 562 | action->thread_fn(action->irq, action->dev_id); |
563 | 563 | ||
@@ -679,7 +679,7 @@ __setup_irq(unsigned int irq, struct irq_desc *desc, struct irqaction *new) | |||
679 | /* | 679 | /* |
680 | * The following block of code has to be executed atomically | 680 | * The following block of code has to be executed atomically |
681 | */ | 681 | */ |
682 | spin_lock_irqsave(&desc->lock, flags); | 682 | raw_spin_lock_irqsave(&desc->lock, flags); |
683 | old_ptr = &desc->action; | 683 | old_ptr = &desc->action; |
684 | old = *old_ptr; | 684 | old = *old_ptr; |
685 | if (old) { | 685 | if (old) { |
@@ -775,7 +775,7 @@ __setup_irq(unsigned int irq, struct irq_desc *desc, struct irqaction *new) | |||
775 | __enable_irq(desc, irq, false); | 775 | __enable_irq(desc, irq, false); |
776 | } | 776 | } |
777 | 777 | ||
778 | spin_unlock_irqrestore(&desc->lock, flags); | 778 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
779 | 779 | ||
780 | /* | 780 | /* |
781 | * Strictly no need to wake it up, but hung_task complains | 781 | * Strictly no need to wake it up, but hung_task complains |
@@ -802,7 +802,7 @@ mismatch: | |||
802 | ret = -EBUSY; | 802 | ret = -EBUSY; |
803 | 803 | ||
804 | out_thread: | 804 | out_thread: |
805 | spin_unlock_irqrestore(&desc->lock, flags); | 805 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
806 | if (new->thread) { | 806 | if (new->thread) { |
807 | struct task_struct *t = new->thread; | 807 | struct task_struct *t = new->thread; |
808 | 808 | ||
@@ -844,7 +844,7 @@ static struct irqaction *__free_irq(unsigned int irq, void *dev_id) | |||
844 | if (!desc) | 844 | if (!desc) |
845 | return NULL; | 845 | return NULL; |
846 | 846 | ||
847 | spin_lock_irqsave(&desc->lock, flags); | 847 | raw_spin_lock_irqsave(&desc->lock, flags); |
848 | 848 | ||
849 | /* | 849 | /* |
850 | * There can be multiple actions per IRQ descriptor, find the right | 850 | * There can be multiple actions per IRQ descriptor, find the right |
@@ -856,7 +856,7 @@ static struct irqaction *__free_irq(unsigned int irq, void *dev_id) | |||
856 | 856 | ||
857 | if (!action) { | 857 | if (!action) { |
858 | WARN(1, "Trying to free already-free IRQ %d\n", irq); | 858 | WARN(1, "Trying to free already-free IRQ %d\n", irq); |
859 | spin_unlock_irqrestore(&desc->lock, flags); | 859 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
860 | 860 | ||
861 | return NULL; | 861 | return NULL; |
862 | } | 862 | } |
@@ -884,7 +884,7 @@ static struct irqaction *__free_irq(unsigned int irq, void *dev_id) | |||
884 | desc->chip->disable(irq); | 884 | desc->chip->disable(irq); |
885 | } | 885 | } |
886 | 886 | ||
887 | spin_unlock_irqrestore(&desc->lock, flags); | 887 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
888 | 888 | ||
889 | unregister_handler_proc(irq, action); | 889 | unregister_handler_proc(irq, action); |
890 | 890 | ||
diff --git a/kernel/irq/migration.c b/kernel/irq/migration.c index fcb6c96f262..24196228083 100644 --- a/kernel/irq/migration.c +++ b/kernel/irq/migration.c | |||
@@ -27,7 +27,7 @@ void move_masked_irq(int irq) | |||
27 | if (!desc->chip->set_affinity) | 27 | if (!desc->chip->set_affinity) |
28 | return; | 28 | return; |
29 | 29 | ||
30 | assert_spin_locked(&desc->lock); | 30 | assert_raw_spin_locked(&desc->lock); |
31 | 31 | ||
32 | /* | 32 | /* |
33 | * If there was a valid mask to work with, please | 33 | * If there was a valid mask to work with, please |
diff --git a/kernel/irq/numa_migrate.c b/kernel/irq/numa_migrate.c index 3fd30197da2..26bac9d8f86 100644 --- a/kernel/irq/numa_migrate.c +++ b/kernel/irq/numa_migrate.c | |||
@@ -42,7 +42,7 @@ static bool init_copy_one_irq_desc(int irq, struct irq_desc *old_desc, | |||
42 | "for migration.\n", irq); | 42 | "for migration.\n", irq); |
43 | return false; | 43 | return false; |
44 | } | 44 | } |
45 | spin_lock_init(&desc->lock); | 45 | raw_spin_lock_init(&desc->lock); |
46 | desc->node = node; | 46 | desc->node = node; |
47 | lockdep_set_class(&desc->lock, &irq_desc_lock_class); | 47 | lockdep_set_class(&desc->lock, &irq_desc_lock_class); |
48 | init_copy_kstat_irqs(old_desc, desc, node, nr_cpu_ids); | 48 | init_copy_kstat_irqs(old_desc, desc, node, nr_cpu_ids); |
@@ -67,7 +67,7 @@ static struct irq_desc *__real_move_irq_desc(struct irq_desc *old_desc, | |||
67 | 67 | ||
68 | irq = old_desc->irq; | 68 | irq = old_desc->irq; |
69 | 69 | ||
70 | spin_lock_irqsave(&sparse_irq_lock, flags); | 70 | raw_spin_lock_irqsave(&sparse_irq_lock, flags); |
71 | 71 | ||
72 | /* We have to check it to avoid races with another CPU */ | 72 | /* We have to check it to avoid races with another CPU */ |
73 | desc = irq_desc_ptrs[irq]; | 73 | desc = irq_desc_ptrs[irq]; |
@@ -91,7 +91,7 @@ static struct irq_desc *__real_move_irq_desc(struct irq_desc *old_desc, | |||
91 | } | 91 | } |
92 | 92 | ||
93 | irq_desc_ptrs[irq] = desc; | 93 | irq_desc_ptrs[irq] = desc; |
94 | spin_unlock_irqrestore(&sparse_irq_lock, flags); | 94 | raw_spin_unlock_irqrestore(&sparse_irq_lock, flags); |
95 | 95 | ||
96 | /* free the old one */ | 96 | /* free the old one */ |
97 | free_one_irq_desc(old_desc, desc); | 97 | free_one_irq_desc(old_desc, desc); |
@@ -100,7 +100,7 @@ static struct irq_desc *__real_move_irq_desc(struct irq_desc *old_desc, | |||
100 | return desc; | 100 | return desc; |
101 | 101 | ||
102 | out_unlock: | 102 | out_unlock: |
103 | spin_unlock_irqrestore(&sparse_irq_lock, flags); | 103 | raw_spin_unlock_irqrestore(&sparse_irq_lock, flags); |
104 | 104 | ||
105 | return desc; | 105 | return desc; |
106 | } | 106 | } |
diff --git a/kernel/irq/pm.c b/kernel/irq/pm.c index a0bb09e7986..0d4005d85b0 100644 --- a/kernel/irq/pm.c +++ b/kernel/irq/pm.c | |||
@@ -28,9 +28,9 @@ void suspend_device_irqs(void) | |||
28 | for_each_irq_desc(irq, desc) { | 28 | for_each_irq_desc(irq, desc) { |
29 | unsigned long flags; | 29 | unsigned long flags; |
30 | 30 | ||
31 | spin_lock_irqsave(&desc->lock, flags); | 31 | raw_spin_lock_irqsave(&desc->lock, flags); |
32 | __disable_irq(desc, irq, true); | 32 | __disable_irq(desc, irq, true); |
33 | spin_unlock_irqrestore(&desc->lock, flags); | 33 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
34 | } | 34 | } |
35 | 35 | ||
36 | for_each_irq_desc(irq, desc) | 36 | for_each_irq_desc(irq, desc) |
@@ -56,9 +56,9 @@ void resume_device_irqs(void) | |||
56 | if (!(desc->status & IRQ_SUSPENDED)) | 56 | if (!(desc->status & IRQ_SUSPENDED)) |
57 | continue; | 57 | continue; |
58 | 58 | ||
59 | spin_lock_irqsave(&desc->lock, flags); | 59 | raw_spin_lock_irqsave(&desc->lock, flags); |
60 | __enable_irq(desc, irq, true); | 60 | __enable_irq(desc, irq, true); |
61 | spin_unlock_irqrestore(&desc->lock, flags); | 61 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
62 | } | 62 | } |
63 | } | 63 | } |
64 | EXPORT_SYMBOL_GPL(resume_device_irqs); | 64 | EXPORT_SYMBOL_GPL(resume_device_irqs); |
diff --git a/kernel/irq/proc.c b/kernel/irq/proc.c index 0832145fea9..6f50eccc79c 100644 --- a/kernel/irq/proc.c +++ b/kernel/irq/proc.c | |||
@@ -179,7 +179,7 @@ static int name_unique(unsigned int irq, struct irqaction *new_action) | |||
179 | unsigned long flags; | 179 | unsigned long flags; |
180 | int ret = 1; | 180 | int ret = 1; |
181 | 181 | ||
182 | spin_lock_irqsave(&desc->lock, flags); | 182 | raw_spin_lock_irqsave(&desc->lock, flags); |
183 | for (action = desc->action ; action; action = action->next) { | 183 | for (action = desc->action ; action; action = action->next) { |
184 | if ((action != new_action) && action->name && | 184 | if ((action != new_action) && action->name && |
185 | !strcmp(new_action->name, action->name)) { | 185 | !strcmp(new_action->name, action->name)) { |
@@ -187,7 +187,7 @@ static int name_unique(unsigned int irq, struct irqaction *new_action) | |||
187 | break; | 187 | break; |
188 | } | 188 | } |
189 | } | 189 | } |
190 | spin_unlock_irqrestore(&desc->lock, flags); | 190 | raw_spin_unlock_irqrestore(&desc->lock, flags); |
191 | return ret; | 191 | return ret; |
192 | } | 192 | } |
193 | 193 | ||
diff --git a/kernel/irq/spurious.c b/kernel/irq/spurious.c index e49ea1c5232..89fb90ae534 100644 --- a/kernel/irq/spurious.c +++ b/kernel/irq/spurious.c | |||
@@ -28,7 +28,7 @@ static int try_one_irq(int irq, struct irq_desc *desc) | |||
28 | struct irqaction *action; | 28 | struct irqaction *action; |
29 | int ok = 0, work = 0; | 29 | int ok = 0, work = 0; |
30 | 30 | ||
31 | spin_lock(&desc->lock); | 31 | raw_spin_lock(&desc->lock); |
32 | /* Already running on another processor */ | 32 | /* Already running on another processor */ |
33 | if (desc->status & IRQ_INPROGRESS) { | 33 | if (desc->status & IRQ_INPROGRESS) { |
34 | /* | 34 | /* |
@@ -37,13 +37,13 @@ static int try_one_irq(int irq, struct irq_desc *desc) | |||
37 | */ | 37 | */ |
38 | if (desc->action && (desc->action->flags & IRQF_SHARED)) | 38 | if (desc->action && (desc->action->flags & IRQF_SHARED)) |
39 | desc->status |= IRQ_PENDING; | 39 | desc->status |= IRQ_PENDING; |
40 | spin_unlock(&desc->lock); | 40 | raw_spin_unlock(&desc->lock); |
41 | return ok; | 41 | return ok; |
42 | } | 42 | } |
43 | /* Honour the normal IRQ locking */ | 43 | /* Honour the normal IRQ locking */ |
44 | desc->status |= IRQ_INPROGRESS; | 44 | desc->status |= IRQ_INPROGRESS; |
45 | action = desc->action; | 45 | action = desc->action; |
46 | spin_unlock(&desc->lock); | 46 | raw_spin_unlock(&desc->lock); |
47 | 47 | ||
48 | while (action) { | 48 | while (action) { |
49 | /* Only shared IRQ handlers are safe to call */ | 49 | /* Only shared IRQ handlers are safe to call */ |
@@ -56,7 +56,7 @@ static int try_one_irq(int irq, struct irq_desc *desc) | |||
56 | } | 56 | } |
57 | local_irq_disable(); | 57 | local_irq_disable(); |
58 | /* Now clean up the flags */ | 58 | /* Now clean up the flags */ |
59 | spin_lock(&desc->lock); | 59 | raw_spin_lock(&desc->lock); |
60 | action = desc->action; | 60 | action = desc->action; |
61 | 61 | ||
62 | /* | 62 | /* |
@@ -68,9 +68,9 @@ static int try_one_irq(int irq, struct irq_desc *desc) | |||
68 | * Perform real IRQ processing for the IRQ we deferred | 68 | * Perform real IRQ processing for the IRQ we deferred |
69 | */ | 69 | */ |
70 | work = 1; | 70 | work = 1; |
71 | spin_unlock(&desc->lock); | 71 | raw_spin_unlock(&desc->lock); |
72 | handle_IRQ_event(irq, action); | 72 | handle_IRQ_event(irq, action); |
73 | spin_lock(&desc->lock); | 73 | raw_spin_lock(&desc->lock); |
74 | desc->status &= ~IRQ_PENDING; | 74 | desc->status &= ~IRQ_PENDING; |
75 | } | 75 | } |
76 | desc->status &= ~IRQ_INPROGRESS; | 76 | desc->status &= ~IRQ_INPROGRESS; |
@@ -80,7 +80,7 @@ static int try_one_irq(int irq, struct irq_desc *desc) | |||
80 | */ | 80 | */ |
81 | if (work && desc->chip && desc->chip->end) | 81 | if (work && desc->chip && desc->chip->end) |
82 | desc->chip->end(irq); | 82 | desc->chip->end(irq); |
83 | spin_unlock(&desc->lock); | 83 | raw_spin_unlock(&desc->lock); |
84 | 84 | ||
85 | return ok; | 85 | return ok; |
86 | } | 86 | } |