diff options
Diffstat (limited to 'kernel/irq_work.c')
-rw-r--r-- | kernel/irq_work.c | 112 |
1 files changed, 86 insertions, 26 deletions
diff --git a/kernel/irq_work.c b/kernel/irq_work.c index c9d7478e4889..55fcce6065cf 100644 --- a/kernel/irq_work.c +++ b/kernel/irq_work.c | |||
@@ -12,22 +12,15 @@ | |||
12 | #include <linux/percpu.h> | 12 | #include <linux/percpu.h> |
13 | #include <linux/hardirq.h> | 13 | #include <linux/hardirq.h> |
14 | #include <linux/irqflags.h> | 14 | #include <linux/irqflags.h> |
15 | #include <linux/sched.h> | ||
16 | #include <linux/tick.h> | ||
17 | #include <linux/cpu.h> | ||
18 | #include <linux/notifier.h> | ||
15 | #include <asm/processor.h> | 19 | #include <asm/processor.h> |
16 | 20 | ||
17 | /* | ||
18 | * An entry can be in one of four states: | ||
19 | * | ||
20 | * free NULL, 0 -> {claimed} : free to be used | ||
21 | * claimed NULL, 3 -> {pending} : claimed to be enqueued | ||
22 | * pending next, 3 -> {busy} : queued, pending callback | ||
23 | * busy NULL, 2 -> {free, claimed} : callback in progress, can be claimed | ||
24 | */ | ||
25 | |||
26 | #define IRQ_WORK_PENDING 1UL | ||
27 | #define IRQ_WORK_BUSY 2UL | ||
28 | #define IRQ_WORK_FLAGS 3UL | ||
29 | 21 | ||
30 | static DEFINE_PER_CPU(struct llist_head, irq_work_list); | 22 | static DEFINE_PER_CPU(struct llist_head, irq_work_list); |
23 | static DEFINE_PER_CPU(int, irq_work_raised); | ||
31 | 24 | ||
32 | /* | 25 | /* |
33 | * Claim the entry so that no one else will poke at it. | 26 | * Claim the entry so that no one else will poke at it. |
@@ -70,8 +63,6 @@ void __weak arch_irq_work_raise(void) | |||
70 | */ | 63 | */ |
71 | void irq_work_queue(struct irq_work *work) | 64 | void irq_work_queue(struct irq_work *work) |
72 | { | 65 | { |
73 | bool empty; | ||
74 | |||
75 | /* Only queue if not already pending */ | 66 | /* Only queue if not already pending */ |
76 | if (!irq_work_claim(work)) | 67 | if (!irq_work_claim(work)) |
77 | return; | 68 | return; |
@@ -79,30 +70,55 @@ void irq_work_queue(struct irq_work *work) | |||
79 | /* Queue the entry and raise the IPI if needed. */ | 70 | /* Queue the entry and raise the IPI if needed. */ |
80 | preempt_disable(); | 71 | preempt_disable(); |
81 | 72 | ||
82 | empty = llist_add(&work->llnode, &__get_cpu_var(irq_work_list)); | 73 | llist_add(&work->llnode, &__get_cpu_var(irq_work_list)); |
83 | /* The list was empty, raise self-interrupt to start processing. */ | 74 | |
84 | if (empty) | 75 | /* |
85 | arch_irq_work_raise(); | 76 | * If the work is not "lazy" or the tick is stopped, raise the irq |
77 | * work interrupt (if supported by the arch), otherwise, just wait | ||
78 | * for the next tick. | ||
79 | */ | ||
80 | if (!(work->flags & IRQ_WORK_LAZY) || tick_nohz_tick_stopped()) { | ||
81 | if (!this_cpu_cmpxchg(irq_work_raised, 0, 1)) | ||
82 | arch_irq_work_raise(); | ||
83 | } | ||
86 | 84 | ||
87 | preempt_enable(); | 85 | preempt_enable(); |
88 | } | 86 | } |
89 | EXPORT_SYMBOL_GPL(irq_work_queue); | 87 | EXPORT_SYMBOL_GPL(irq_work_queue); |
90 | 88 | ||
91 | /* | 89 | bool irq_work_needs_cpu(void) |
92 | * Run the irq_work entries on this cpu. Requires to be ran from hardirq | 90 | { |
93 | * context with local IRQs disabled. | 91 | struct llist_head *this_list; |
94 | */ | 92 | |
95 | void irq_work_run(void) | 93 | this_list = &__get_cpu_var(irq_work_list); |
94 | if (llist_empty(this_list)) | ||
95 | return false; | ||
96 | |||
97 | /* All work should have been flushed before going offline */ | ||
98 | WARN_ON_ONCE(cpu_is_offline(smp_processor_id())); | ||
99 | |||
100 | return true; | ||
101 | } | ||
102 | |||
103 | static void __irq_work_run(void) | ||
96 | { | 104 | { |
105 | unsigned long flags; | ||
97 | struct irq_work *work; | 106 | struct irq_work *work; |
98 | struct llist_head *this_list; | 107 | struct llist_head *this_list; |
99 | struct llist_node *llnode; | 108 | struct llist_node *llnode; |
100 | 109 | ||
110 | |||
111 | /* | ||
112 | * Reset the "raised" state right before we check the list because | ||
113 | * an NMI may enqueue after we find the list empty from the runner. | ||
114 | */ | ||
115 | __this_cpu_write(irq_work_raised, 0); | ||
116 | barrier(); | ||
117 | |||
101 | this_list = &__get_cpu_var(irq_work_list); | 118 | this_list = &__get_cpu_var(irq_work_list); |
102 | if (llist_empty(this_list)) | 119 | if (llist_empty(this_list)) |
103 | return; | 120 | return; |
104 | 121 | ||
105 | BUG_ON(!in_irq()); | ||
106 | BUG_ON(!irqs_disabled()); | 122 | BUG_ON(!irqs_disabled()); |
107 | 123 | ||
108 | llnode = llist_del_all(this_list); | 124 | llnode = llist_del_all(this_list); |
@@ -118,15 +134,27 @@ void irq_work_run(void) | |||
118 | * to claim that work don't rely on us to handle their data | 134 | * to claim that work don't rely on us to handle their data |
119 | * while we are in the middle of the func. | 135 | * while we are in the middle of the func. |
120 | */ | 136 | */ |
121 | xchg(&work->flags, IRQ_WORK_BUSY); | 137 | flags = work->flags & ~IRQ_WORK_PENDING; |
138 | xchg(&work->flags, flags); | ||
139 | |||
122 | work->func(work); | 140 | work->func(work); |
123 | /* | 141 | /* |
124 | * Clear the BUSY bit and return to the free state if | 142 | * Clear the BUSY bit and return to the free state if |
125 | * no-one else claimed it meanwhile. | 143 | * no-one else claimed it meanwhile. |
126 | */ | 144 | */ |
127 | (void)cmpxchg(&work->flags, IRQ_WORK_BUSY, 0); | 145 | (void)cmpxchg(&work->flags, flags, flags & ~IRQ_WORK_BUSY); |
128 | } | 146 | } |
129 | } | 147 | } |
148 | |||
149 | /* | ||
150 | * Run the irq_work entries on this cpu. Requires to be ran from hardirq | ||
151 | * context with local IRQs disabled. | ||
152 | */ | ||
153 | void irq_work_run(void) | ||
154 | { | ||
155 | BUG_ON(!in_irq()); | ||
156 | __irq_work_run(); | ||
157 | } | ||
130 | EXPORT_SYMBOL_GPL(irq_work_run); | 158 | EXPORT_SYMBOL_GPL(irq_work_run); |
131 | 159 | ||
132 | /* | 160 | /* |
@@ -141,3 +169,35 @@ void irq_work_sync(struct irq_work *work) | |||
141 | cpu_relax(); | 169 | cpu_relax(); |
142 | } | 170 | } |
143 | EXPORT_SYMBOL_GPL(irq_work_sync); | 171 | EXPORT_SYMBOL_GPL(irq_work_sync); |
172 | |||
173 | #ifdef CONFIG_HOTPLUG_CPU | ||
174 | static int irq_work_cpu_notify(struct notifier_block *self, | ||
175 | unsigned long action, void *hcpu) | ||
176 | { | ||
177 | long cpu = (long)hcpu; | ||
178 | |||
179 | switch (action) { | ||
180 | case CPU_DYING: | ||
181 | /* Called from stop_machine */ | ||
182 | if (WARN_ON_ONCE(cpu != smp_processor_id())) | ||
183 | break; | ||
184 | __irq_work_run(); | ||
185 | break; | ||
186 | default: | ||
187 | break; | ||
188 | } | ||
189 | return NOTIFY_OK; | ||
190 | } | ||
191 | |||
192 | static struct notifier_block cpu_notify; | ||
193 | |||
194 | static __init int irq_work_init_cpu_notifier(void) | ||
195 | { | ||
196 | cpu_notify.notifier_call = irq_work_cpu_notify; | ||
197 | cpu_notify.priority = 0; | ||
198 | register_cpu_notifier(&cpu_notify); | ||
199 | return 0; | ||
200 | } | ||
201 | device_initcall(irq_work_init_cpu_notifier); | ||
202 | |||
203 | #endif /* CONFIG_HOTPLUG_CPU */ | ||