aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/powerpc/kernel/irq.c57
-rw-r--r--include/asm-powerpc/hw_irq.h20
2 files changed, 67 insertions, 10 deletions
diff --git a/arch/powerpc/kernel/irq.c b/arch/powerpc/kernel/irq.c
index e1936952017c..0bd8c7665834 100644
--- a/arch/powerpc/kernel/irq.c
+++ b/arch/powerpc/kernel/irq.c
@@ -97,22 +97,69 @@ EXPORT_SYMBOL(irq_desc);
97 97
98int distribute_irqs = 1; 98int distribute_irqs = 1;
99 99
100static inline unsigned long get_hard_enabled(void)
101{
102 unsigned long enabled;
103
104 __asm__ __volatile__("lbz %0,%1(13)"
105 : "=r" (enabled) : "i" (offsetof(struct paca_struct, hard_enabled)));
106
107 return enabled;
108}
109
110static inline void set_soft_enabled(unsigned long enable)
111{
112 __asm__ __volatile__("stb %0,%1(13)"
113 : : "r" (enable), "i" (offsetof(struct paca_struct, soft_enabled)));
114}
115
100void local_irq_restore(unsigned long en) 116void local_irq_restore(unsigned long en)
101{ 117{
102 get_paca()->soft_enabled = en; 118 /*
119 * get_paca()->soft_enabled = en;
120 * Is it ever valid to use local_irq_restore(0) when soft_enabled is 1?
121 * That was allowed before, and in such a case we do need to take care
122 * that gcc will set soft_enabled directly via r13, not choose to use
123 * an intermediate register, lest we're preempted to a different cpu.
124 */
125 set_soft_enabled(en);
103 if (!en) 126 if (!en)
104 return; 127 return;
105 128
106 if (firmware_has_feature(FW_FEATURE_ISERIES)) { 129 if (firmware_has_feature(FW_FEATURE_ISERIES)) {
107 if (get_paca()->lppaca_ptr->int_dword.any_int) 130 /*
131 * Do we need to disable preemption here? Not really: in the
132 * unlikely event that we're preempted to a different cpu in
133 * between getting r13, loading its lppaca_ptr, and loading
134 * its any_int, we might call iseries_handle_interrupts without
135 * an interrupt pending on the new cpu, but that's no disaster,
136 * is it? And the business of preempting us off the old cpu
137 * would itself involve a local_irq_restore which handles the
138 * interrupt to that cpu.
139 *
140 * But use "local_paca->lppaca_ptr" instead of "get_lppaca()"
141 * to avoid any preemption checking added into get_paca().
142 */
143 if (local_paca->lppaca_ptr->int_dword.any_int)
108 iseries_handle_interrupts(); 144 iseries_handle_interrupts();
109 return; 145 return;
110 } 146 }
111 147
112 if (get_paca()->hard_enabled) 148 /*
149 * if (get_paca()->hard_enabled) return;
150 * But again we need to take care that gcc gets hard_enabled directly
151 * via r13, not choose to use an intermediate register, lest we're
152 * preempted to a different cpu in between the two instructions.
153 */
154 if (get_hard_enabled())
113 return; 155 return;
114 /* need to hard-enable interrupts here */ 156
115 get_paca()->hard_enabled = en; 157 /*
158 * Need to hard-enable interrupts here. Since currently disabled,
159 * no need to take further asm precautions against preemption; but
160 * use local_paca instead of get_paca() to avoid preemption checking.
161 */
162 local_paca->hard_enabled = en;
116 if ((int)mfspr(SPRN_DEC) < 0) 163 if ((int)mfspr(SPRN_DEC) < 0)
117 mtspr(SPRN_DEC, 1); 164 mtspr(SPRN_DEC, 1);
118 hard_irq_enable(); 165 hard_irq_enable();
diff --git a/include/asm-powerpc/hw_irq.h b/include/asm-powerpc/hw_irq.h
index c4a1ab608f6f..fd3f2a206271 100644
--- a/include/asm-powerpc/hw_irq.h
+++ b/include/asm-powerpc/hw_irq.h
@@ -18,15 +18,25 @@ extern void timer_interrupt(struct pt_regs *);
18 18
19static inline unsigned long local_get_flags(void) 19static inline unsigned long local_get_flags(void)
20{ 20{
21 return get_paca()->soft_enabled; 21 unsigned long flags;
22
23 __asm__ __volatile__("lbz %0,%1(13)"
24 : "=r" (flags)
25 : "i" (offsetof(struct paca_struct, soft_enabled)));
26
27 return flags;
22} 28}
23 29
24static inline unsigned long local_irq_disable(void) 30static inline unsigned long local_irq_disable(void)
25{ 31{
26 unsigned long flag = get_paca()->soft_enabled; 32 unsigned long flags, zero;
27 get_paca()->soft_enabled = 0; 33
28 barrier(); 34 __asm__ __volatile__("li %1,0; lbz %0,%2(13); stb %1,%2(13)"
29 return flag; 35 : "=r" (flags), "=&r" (zero)
36 : "i" (offsetof(struct paca_struct, soft_enabled))
37 : "memory");
38
39 return flags;
30} 40}
31 41
32extern void local_irq_restore(unsigned long); 42extern void local_irq_restore(unsigned long);