aboutsummaryrefslogtreecommitdiffstats
path: root/virt/kvm/irq_comm.c
diff options
context:
space:
mode:
authorGleb Natapov <gleb@redhat.com>2013-04-28 05:50:07 -0400
committerGleb Natapov <gleb@redhat.com>2013-04-28 05:50:07 -0400
commit064d1afaa5a60fc391d0b4b77599fc8f63f99cd3 (patch)
tree2e640cdfa50b0048c52e021f07a8b24560251b26 /virt/kvm/irq_comm.c
parent730dca42c1d363c939da18c1499c7327c66e2b37 (diff)
parent8b78645c93b5d469e8006d68dbc92edc2640c654 (diff)
Merge git://github.com/agraf/linux-2.6.git kvm-ppc-next into queue
Diffstat (limited to 'virt/kvm/irq_comm.c')
-rw-r--r--virt/kvm/irq_comm.c194
1 files changed, 3 insertions, 191 deletions
diff --git a/virt/kvm/irq_comm.c b/virt/kvm/irq_comm.c
index 25ab48007adb..e2e6b4473a96 100644
--- a/virt/kvm/irq_comm.c
+++ b/virt/kvm/irq_comm.c
@@ -151,59 +151,6 @@ static int kvm_set_msi_inatomic(struct kvm_kernel_irq_routing_entry *e,
151 return -EWOULDBLOCK; 151 return -EWOULDBLOCK;
152} 152}
153 153
154int kvm_send_userspace_msi(struct kvm *kvm, struct kvm_msi *msi)
155{
156 struct kvm_kernel_irq_routing_entry route;
157
158 if (!irqchip_in_kernel(kvm) || msi->flags != 0)
159 return -EINVAL;
160
161 route.msi.address_lo = msi->address_lo;
162 route.msi.address_hi = msi->address_hi;
163 route.msi.data = msi->data;
164
165 return kvm_set_msi(&route, kvm, KVM_USERSPACE_IRQ_SOURCE_ID, 1, false);
166}
167
168/*
169 * Return value:
170 * < 0 Interrupt was ignored (masked or not delivered for other reasons)
171 * = 0 Interrupt was coalesced (previous irq is still pending)
172 * > 0 Number of CPUs interrupt was delivered to
173 */
174int kvm_set_irq(struct kvm *kvm, int irq_source_id, u32 irq, int level,
175 bool line_status)
176{
177 struct kvm_kernel_irq_routing_entry *e, irq_set[KVM_NR_IRQCHIPS];
178 int ret = -1, i = 0;
179 struct kvm_irq_routing_table *irq_rt;
180
181 trace_kvm_set_irq(irq, level, irq_source_id);
182
183 /* Not possible to detect if the guest uses the PIC or the
184 * IOAPIC. So set the bit in both. The guest will ignore
185 * writes to the unused one.
186 */
187 rcu_read_lock();
188 irq_rt = rcu_dereference(kvm->irq_routing);
189 if (irq < irq_rt->nr_rt_entries)
190 hlist_for_each_entry(e, &irq_rt->map[irq], link)
191 irq_set[i++] = *e;
192 rcu_read_unlock();
193
194 while(i--) {
195 int r;
196 r = irq_set[i].set(&irq_set[i], kvm, irq_source_id, level,
197 line_status);
198 if (r < 0)
199 continue;
200
201 ret = r + ((ret < 0) ? 0 : ret);
202 }
203
204 return ret;
205}
206
207/* 154/*
208 * Deliver an IRQ in an atomic context if we can, or return a failure, 155 * Deliver an IRQ in an atomic context if we can, or return a failure,
209 * user can retry in a process context. 156 * user can retry in a process context.
@@ -241,63 +188,6 @@ int kvm_set_irq_inatomic(struct kvm *kvm, int irq_source_id, u32 irq, int level)
241 return ret; 188 return ret;
242} 189}
243 190
244bool kvm_irq_has_notifier(struct kvm *kvm, unsigned irqchip, unsigned pin)
245{
246 struct kvm_irq_ack_notifier *kian;
247 int gsi;
248
249 rcu_read_lock();
250 gsi = rcu_dereference(kvm->irq_routing)->chip[irqchip][pin];
251 if (gsi != -1)
252 hlist_for_each_entry_rcu(kian, &kvm->irq_ack_notifier_list,
253 link)
254 if (kian->gsi == gsi) {
255 rcu_read_unlock();
256 return true;
257 }
258
259 rcu_read_unlock();
260
261 return false;
262}
263EXPORT_SYMBOL_GPL(kvm_irq_has_notifier);
264
265void kvm_notify_acked_irq(struct kvm *kvm, unsigned irqchip, unsigned pin)
266{
267 struct kvm_irq_ack_notifier *kian;
268 int gsi;
269
270 trace_kvm_ack_irq(irqchip, pin);
271
272 rcu_read_lock();
273 gsi = rcu_dereference(kvm->irq_routing)->chip[irqchip][pin];
274 if (gsi != -1)
275 hlist_for_each_entry_rcu(kian, &kvm->irq_ack_notifier_list,
276 link)
277 if (kian->gsi == gsi)
278 kian->irq_acked(kian);
279 rcu_read_unlock();
280}
281
282void kvm_register_irq_ack_notifier(struct kvm *kvm,
283 struct kvm_irq_ack_notifier *kian)
284{
285 mutex_lock(&kvm->irq_lock);
286 hlist_add_head_rcu(&kian->link, &kvm->irq_ack_notifier_list);
287 mutex_unlock(&kvm->irq_lock);
288 kvm_vcpu_request_scan_ioapic(kvm);
289}
290
291void kvm_unregister_irq_ack_notifier(struct kvm *kvm,
292 struct kvm_irq_ack_notifier *kian)
293{
294 mutex_lock(&kvm->irq_lock);
295 hlist_del_init_rcu(&kian->link);
296 mutex_unlock(&kvm->irq_lock);
297 synchronize_rcu();
298 kvm_vcpu_request_scan_ioapic(kvm);
299}
300
301int kvm_request_irq_source_id(struct kvm *kvm) 191int kvm_request_irq_source_id(struct kvm *kvm)
302{ 192{
303 unsigned long *bitmap = &kvm->arch.irq_sources_bitmap; 193 unsigned long *bitmap = &kvm->arch.irq_sources_bitmap;
@@ -381,34 +271,14 @@ void kvm_fire_mask_notifiers(struct kvm *kvm, unsigned irqchip, unsigned pin,
381 rcu_read_unlock(); 271 rcu_read_unlock();
382} 272}
383 273
384void kvm_free_irq_routing(struct kvm *kvm) 274int kvm_set_routing_entry(struct kvm_irq_routing_table *rt,
385{ 275 struct kvm_kernel_irq_routing_entry *e,
386 /* Called only during vm destruction. Nobody can use the pointer 276 const struct kvm_irq_routing_entry *ue)
387 at this stage */
388 kfree(kvm->irq_routing);
389}
390
391static int setup_routing_entry(struct kvm_irq_routing_table *rt,
392 struct kvm_kernel_irq_routing_entry *e,
393 const struct kvm_irq_routing_entry *ue)
394{ 277{
395 int r = -EINVAL; 278 int r = -EINVAL;
396 int delta; 279 int delta;
397 unsigned max_pin; 280 unsigned max_pin;
398 struct kvm_kernel_irq_routing_entry *ei;
399 281
400 /*
401 * Do not allow GSI to be mapped to the same irqchip more than once.
402 * Allow only one to one mapping between GSI and MSI.
403 */
404 hlist_for_each_entry(ei, &rt->map[ue->gsi], link)
405 if (ei->type == KVM_IRQ_ROUTING_MSI ||
406 ue->type == KVM_IRQ_ROUTING_MSI ||
407 ue->u.irqchip.irqchip == ei->irqchip.irqchip)
408 return r;
409
410 e->gsi = ue->gsi;
411 e->type = ue->type;
412 switch (ue->type) { 282 switch (ue->type) {
413 case KVM_IRQ_ROUTING_IRQCHIP: 283 case KVM_IRQ_ROUTING_IRQCHIP:
414 delta = 0; 284 delta = 0;
@@ -445,69 +315,11 @@ static int setup_routing_entry(struct kvm_irq_routing_table *rt,
445 goto out; 315 goto out;
446 } 316 }
447 317
448 hlist_add_head(&e->link, &rt->map[e->gsi]);
449 r = 0; 318 r = 0;
450out: 319out:
451 return r; 320 return r;
452} 321}
453 322
454
455int kvm_set_irq_routing(struct kvm *kvm,
456 const struct kvm_irq_routing_entry *ue,
457 unsigned nr,
458 unsigned flags)
459{
460 struct kvm_irq_routing_table *new, *old;
461 u32 i, j, nr_rt_entries = 0;
462 int r;
463
464 for (i = 0; i < nr; ++i) {
465 if (ue[i].gsi >= KVM_MAX_IRQ_ROUTES)
466 return -EINVAL;
467 nr_rt_entries = max(nr_rt_entries, ue[i].gsi);
468 }
469
470 nr_rt_entries += 1;
471
472 new = kzalloc(sizeof(*new) + (nr_rt_entries * sizeof(struct hlist_head))
473 + (nr * sizeof(struct kvm_kernel_irq_routing_entry)),
474 GFP_KERNEL);
475
476 if (!new)
477 return -ENOMEM;
478
479 new->rt_entries = (void *)&new->map[nr_rt_entries];
480
481 new->nr_rt_entries = nr_rt_entries;
482 for (i = 0; i < 3; i++)
483 for (j = 0; j < KVM_IOAPIC_NUM_PINS; j++)
484 new->chip[i][j] = -1;
485
486 for (i = 0; i < nr; ++i) {
487 r = -EINVAL;
488 if (ue->flags)
489 goto out;
490 r = setup_routing_entry(new, &new->rt_entries[i], ue);
491 if (r)
492 goto out;
493 ++ue;
494 }
495
496 mutex_lock(&kvm->irq_lock);
497 old = kvm->irq_routing;
498 kvm_irq_routing_update(kvm, new);
499 mutex_unlock(&kvm->irq_lock);
500
501 synchronize_rcu();
502
503 new = old;
504 r = 0;
505
506out:
507 kfree(new);
508 return r;
509}
510
511#define IOAPIC_ROUTING_ENTRY(irq) \ 323#define IOAPIC_ROUTING_ENTRY(irq) \
512 { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \ 324 { .gsi = irq, .type = KVM_IRQ_ROUTING_IRQCHIP, \
513 .u.irqchip.irqchip = KVM_IRQCHIP_IOAPIC, .u.irqchip.pin = (irq) } 325 .u.irqchip.irqchip = KVM_IRQCHIP_IOAPIC, .u.irqchip.pin = (irq) }