diff options
author | Michael S. Tsirkin <mst@redhat.com> | 2009-05-14 06:55:41 -0400 |
---|---|---|
committer | Rusty Russell <rusty@rustcorp.com.au> | 2009-06-12 08:46:37 -0400 |
commit | 82af8ce84ed65d2fb6d8c017d3f2bbbf161061fb (patch) | |
tree | aed59c86724c7d3f565d1c60b38d25a7bf20985b /drivers/virtio | |
parent | 77cf524654a886e0fbbf03b16b44f048deef7b0c (diff) |
virtio_pci: optional MSI-X support
This implements optional MSI-X support in virtio_pci.
MSI-X is used whenever the host supports at least 2 MSI-X
vectors: 1 for configuration changes and 1 for virtqueues.
Per-virtqueue vectors are allocated if enough vectors
available.
Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
Acked-by: Anthony Liguori <aliguori@us.ibm.com>
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> (+ whitespace, style)
Diffstat (limited to 'drivers/virtio')
-rw-r--r-- | drivers/virtio/virtio_pci.c | 228 |
1 files changed, 209 insertions, 19 deletions
diff --git a/drivers/virtio/virtio_pci.c b/drivers/virtio/virtio_pci.c index 951e673e50a4..193c8f0e5cc5 100644 --- a/drivers/virtio/virtio_pci.c +++ b/drivers/virtio/virtio_pci.c | |||
@@ -42,6 +42,26 @@ struct virtio_pci_device | |||
42 | /* a list of queues so we can dispatch IRQs */ | 42 | /* a list of queues so we can dispatch IRQs */ |
43 | spinlock_t lock; | 43 | spinlock_t lock; |
44 | struct list_head virtqueues; | 44 | struct list_head virtqueues; |
45 | |||
46 | /* MSI-X support */ | ||
47 | int msix_enabled; | ||
48 | int intx_enabled; | ||
49 | struct msix_entry *msix_entries; | ||
50 | /* Name strings for interrupts. This size should be enough, | ||
51 | * and I'm too lazy to allocate each name separately. */ | ||
52 | char (*msix_names)[256]; | ||
53 | /* Number of available vectors */ | ||
54 | unsigned msix_vectors; | ||
55 | /* Vectors allocated */ | ||
56 | unsigned msix_used_vectors; | ||
57 | }; | ||
58 | |||
59 | /* Constants for MSI-X */ | ||
60 | /* Use first vector for configuration changes, second and the rest for | ||
61 | * virtqueues Thus, we need at least 2 vectors for MSI. */ | ||
62 | enum { | ||
63 | VP_MSIX_CONFIG_VECTOR = 0, | ||
64 | VP_MSIX_VQ_VECTOR = 1, | ||
45 | }; | 65 | }; |
46 | 66 | ||
47 | struct virtio_pci_vq_info | 67 | struct virtio_pci_vq_info |
@@ -60,6 +80,9 @@ struct virtio_pci_vq_info | |||
60 | 80 | ||
61 | /* the list node for the virtqueues list */ | 81 | /* the list node for the virtqueues list */ |
62 | struct list_head node; | 82 | struct list_head node; |
83 | |||
84 | /* MSI-X vector (or none) */ | ||
85 | unsigned vector; | ||
63 | }; | 86 | }; |
64 | 87 | ||
65 | /* Qumranet donated their vendor ID for devices 0x1000 thru 0x10FF. */ | 88 | /* Qumranet donated their vendor ID for devices 0x1000 thru 0x10FF. */ |
@@ -109,7 +132,8 @@ static void vp_get(struct virtio_device *vdev, unsigned offset, | |||
109 | void *buf, unsigned len) | 132 | void *buf, unsigned len) |
110 | { | 133 | { |
111 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); | 134 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); |
112 | void __iomem *ioaddr = vp_dev->ioaddr + VIRTIO_PCI_CONFIG + offset; | 135 | void __iomem *ioaddr = vp_dev->ioaddr + |
136 | VIRTIO_PCI_CONFIG(vp_dev) + offset; | ||
113 | u8 *ptr = buf; | 137 | u8 *ptr = buf; |
114 | int i; | 138 | int i; |
115 | 139 | ||
@@ -123,7 +147,8 @@ static void vp_set(struct virtio_device *vdev, unsigned offset, | |||
123 | const void *buf, unsigned len) | 147 | const void *buf, unsigned len) |
124 | { | 148 | { |
125 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); | 149 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); |
126 | void __iomem *ioaddr = vp_dev->ioaddr + VIRTIO_PCI_CONFIG + offset; | 150 | void __iomem *ioaddr = vp_dev->ioaddr + |
151 | VIRTIO_PCI_CONFIG(vp_dev) + offset; | ||
127 | const u8 *ptr = buf; | 152 | const u8 *ptr = buf; |
128 | int i; | 153 | int i; |
129 | 154 | ||
@@ -221,7 +246,122 @@ static irqreturn_t vp_interrupt(int irq, void *opaque) | |||
221 | return vp_vring_interrupt(irq, opaque); | 246 | return vp_vring_interrupt(irq, opaque); |
222 | } | 247 | } |
223 | 248 | ||
224 | /* the config->find_vq() implementation */ | 249 | static void vp_free_vectors(struct virtio_device *vdev) |
250 | { | ||
251 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); | ||
252 | int i; | ||
253 | |||
254 | if (vp_dev->intx_enabled) { | ||
255 | free_irq(vp_dev->pci_dev->irq, vp_dev); | ||
256 | vp_dev->intx_enabled = 0; | ||
257 | } | ||
258 | |||
259 | for (i = 0; i < vp_dev->msix_used_vectors; ++i) | ||
260 | free_irq(vp_dev->msix_entries[i].vector, vp_dev); | ||
261 | vp_dev->msix_used_vectors = 0; | ||
262 | |||
263 | if (vp_dev->msix_enabled) { | ||
264 | /* Disable the vector used for configuration */ | ||
265 | iowrite16(VIRTIO_MSI_NO_VECTOR, | ||
266 | vp_dev->ioaddr + VIRTIO_MSI_CONFIG_VECTOR); | ||
267 | /* Flush the write out to device */ | ||
268 | ioread16(vp_dev->ioaddr + VIRTIO_MSI_CONFIG_VECTOR); | ||
269 | |||
270 | vp_dev->msix_enabled = 0; | ||
271 | pci_disable_msix(vp_dev->pci_dev); | ||
272 | } | ||
273 | } | ||
274 | |||
275 | static int vp_enable_msix(struct pci_dev *dev, struct msix_entry *entries, | ||
276 | int *options, int noptions) | ||
277 | { | ||
278 | int i; | ||
279 | for (i = 0; i < noptions; ++i) | ||
280 | if (!pci_enable_msix(dev, entries, options[i])) | ||
281 | return options[i]; | ||
282 | return -EBUSY; | ||
283 | } | ||
284 | |||
285 | static int vp_request_vectors(struct virtio_device *vdev, unsigned max_vqs) | ||
286 | { | ||
287 | struct virtio_pci_device *vp_dev = to_vp_device(vdev); | ||
288 | const char *name = dev_name(&vp_dev->vdev.dev); | ||
289 | unsigned i, v; | ||
290 | int err = -ENOMEM; | ||
291 | /* We want at most one vector per queue and one for config changes. | ||
292 | * Fallback to separate vectors for config and a shared for queues. | ||
293 | * Finally fall back to regular interrupts. */ | ||
294 | int options[] = { max_vqs + 1, 2 }; | ||
295 | int nvectors = max(options[0], options[1]); | ||
296 | |||
297 | vp_dev->msix_entries = kmalloc(nvectors * sizeof *vp_dev->msix_entries, | ||
298 | GFP_KERNEL); | ||
299 | if (!vp_dev->msix_entries) | ||
300 | goto error_entries; | ||
301 | vp_dev->msix_names = kmalloc(nvectors * sizeof *vp_dev->msix_names, | ||
302 | GFP_KERNEL); | ||
303 | if (!vp_dev->msix_names) | ||
304 | goto error_names; | ||
305 | |||
306 | for (i = 0; i < nvectors; ++i) | ||
307 | vp_dev->msix_entries[i].entry = i; | ||
308 | |||
309 | err = vp_enable_msix(vp_dev->pci_dev, vp_dev->msix_entries, | ||
310 | options, ARRAY_SIZE(options)); | ||
311 | if (err < 0) { | ||
312 | /* Can't allocate enough MSI-X vectors, use regular interrupt */ | ||
313 | vp_dev->msix_vectors = 0; | ||
314 | err = request_irq(vp_dev->pci_dev->irq, vp_interrupt, | ||
315 | IRQF_SHARED, name, vp_dev); | ||
316 | if (err) | ||
317 | goto error_irq; | ||
318 | vp_dev->intx_enabled = 1; | ||
319 | } else { | ||
320 | vp_dev->msix_vectors = err; | ||
321 | vp_dev->msix_enabled = 1; | ||
322 | |||
323 | /* Set the vector used for configuration */ | ||
324 | v = vp_dev->msix_used_vectors; | ||
325 | snprintf(vp_dev->msix_names[v], sizeof *vp_dev->msix_names, | ||
326 | "%s-config", name); | ||
327 | err = request_irq(vp_dev->msix_entries[v].vector, | ||
328 | vp_config_changed, 0, vp_dev->msix_names[v], | ||
329 | vp_dev); | ||
330 | if (err) | ||
331 | goto error_irq; | ||
332 | ++vp_dev->msix_used_vectors; | ||
333 | |||
334 | iowrite16(v, vp_dev->ioaddr + VIRTIO_MSI_CONFIG_VECTOR); | ||
335 | /* Verify we had enough resources to assign the vector */ | ||
336 | v = ioread16(vp_dev->ioaddr + VIRTIO_MSI_CONFIG_VECTOR); | ||
337 | if (v == VIRTIO_MSI_NO_VECTOR) { | ||
338 | err = -EBUSY; | ||
339 | goto error_irq; | ||
340 | } | ||
341 | } | ||
342 | |||
343 | if (vp_dev->msix_vectors && vp_dev->msix_vectors != max_vqs + 1) { | ||
344 | /* Shared vector for all VQs */ | ||
345 | v = vp_dev->msix_used_vectors; | ||
346 | snprintf(vp_dev->msix_names[v], sizeof *vp_dev->msix_names, | ||
347 | "%s-virtqueues", name); | ||
348 | err = request_irq(vp_dev->msix_entries[v].vector, | ||
349 | vp_vring_interrupt, 0, vp_dev->msix_names[v], | ||
350 | vp_dev); | ||
351 | if (err) | ||
352 | goto error_irq; | ||
353 | ++vp_dev->msix_used_vectors; | ||
354 | } | ||
355 | return 0; | ||
356 | error_irq: | ||
357 | vp_free_vectors(vdev); | ||
358 | kfree(vp_dev->msix_names); | ||
359 | error_names: | ||
360 | kfree(vp_dev->msix_entries); | ||
361 | error_entries: | ||
362 | return err; | ||
363 | } | ||
364 | |||
225 | static struct virtqueue *vp_find_vq(struct virtio_device *vdev, unsigned index, | 365 | static struct virtqueue *vp_find_vq(struct virtio_device *vdev, unsigned index, |
226 | void (*callback)(struct virtqueue *vq), | 366 | void (*callback)(struct virtqueue *vq), |
227 | const char *name) | 367 | const char *name) |
@@ -230,7 +370,7 @@ static struct virtqueue *vp_find_vq(struct virtio_device *vdev, unsigned index, | |||
230 | struct virtio_pci_vq_info *info; | 370 | struct virtio_pci_vq_info *info; |
231 | struct virtqueue *vq; | 371 | struct virtqueue *vq; |
232 | unsigned long flags, size; | 372 | unsigned long flags, size; |
233 | u16 num; | 373 | u16 num, vector; |
234 | int err; | 374 | int err; |
235 | 375 | ||
236 | /* Select the queue we're interested in */ | 376 | /* Select the queue we're interested in */ |
@@ -249,6 +389,7 @@ static struct virtqueue *vp_find_vq(struct virtio_device *vdev, unsigned index, | |||
249 | 389 | ||
250 | info->queue_index = index; | 390 | info->queue_index = index; |
251 | info->num = num; | 391 | info->num = num; |
392 | info->vector = VIRTIO_MSI_NO_VECTOR; | ||
252 | 393 | ||
253 | size = PAGE_ALIGN(vring_size(num, VIRTIO_PCI_VRING_ALIGN)); | 394 | size = PAGE_ALIGN(vring_size(num, VIRTIO_PCI_VRING_ALIGN)); |
254 | info->queue = alloc_pages_exact(size, GFP_KERNEL|__GFP_ZERO); | 395 | info->queue = alloc_pages_exact(size, GFP_KERNEL|__GFP_ZERO); |
@@ -272,12 +413,43 @@ static struct virtqueue *vp_find_vq(struct virtio_device *vdev, unsigned index, | |||
272 | vq->priv = info; | 413 | vq->priv = info; |
273 | info->vq = vq; | 414 | info->vq = vq; |
274 | 415 | ||
416 | /* allocate per-vq vector if available and necessary */ | ||
417 | if (callback && vp_dev->msix_used_vectors < vp_dev->msix_vectors) { | ||
418 | vector = vp_dev->msix_used_vectors; | ||
419 | snprintf(vp_dev->msix_names[vector], sizeof *vp_dev->msix_names, | ||
420 | "%s-%s", dev_name(&vp_dev->vdev.dev), name); | ||
421 | err = request_irq(vp_dev->msix_entries[vector].vector, | ||
422 | vring_interrupt, 0, | ||
423 | vp_dev->msix_names[vector], vq); | ||
424 | if (err) | ||
425 | goto out_request_irq; | ||
426 | info->vector = vector; | ||
427 | ++vp_dev->msix_used_vectors; | ||
428 | } else | ||
429 | vector = VP_MSIX_VQ_VECTOR; | ||
430 | |||
431 | if (callback && vp_dev->msix_enabled) { | ||
432 | iowrite16(vector, vp_dev->ioaddr + VIRTIO_MSI_QUEUE_VECTOR); | ||
433 | vector = ioread16(vp_dev->ioaddr + VIRTIO_MSI_QUEUE_VECTOR); | ||
434 | if (vector == VIRTIO_MSI_NO_VECTOR) { | ||
435 | err = -EBUSY; | ||
436 | goto out_assign; | ||
437 | } | ||
438 | } | ||
439 | |||
275 | spin_lock_irqsave(&vp_dev->lock, flags); | 440 | spin_lock_irqsave(&vp_dev->lock, flags); |
276 | list_add(&info->node, &vp_dev->virtqueues); | 441 | list_add(&info->node, &vp_dev->virtqueues); |
277 | spin_unlock_irqrestore(&vp_dev->lock, flags); | 442 | spin_unlock_irqrestore(&vp_dev->lock, flags); |
278 | 443 | ||
279 | return vq; | 444 | return vq; |
280 | 445 | ||
446 | out_assign: | ||
447 | if (info->vector != VIRTIO_MSI_NO_VECTOR) { | ||
448 | free_irq(vp_dev->msix_entries[info->vector].vector, vq); | ||
449 | --vp_dev->msix_used_vectors; | ||
450 | } | ||
451 | out_request_irq: | ||
452 | vring_del_virtqueue(vq); | ||
281 | out_activate_queue: | 453 | out_activate_queue: |
282 | iowrite32(0, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_PFN); | 454 | iowrite32(0, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_PFN); |
283 | free_pages_exact(info->queue, size); | 455 | free_pages_exact(info->queue, size); |
@@ -286,17 +458,27 @@ out_info: | |||
286 | return ERR_PTR(err); | 458 | return ERR_PTR(err); |
287 | } | 459 | } |
288 | 460 | ||
289 | /* the config->del_vq() implementation */ | ||
290 | static void vp_del_vq(struct virtqueue *vq) | 461 | static void vp_del_vq(struct virtqueue *vq) |
291 | { | 462 | { |
292 | struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev); | 463 | struct virtio_pci_device *vp_dev = to_vp_device(vq->vdev); |
293 | struct virtio_pci_vq_info *info = vq->priv; | 464 | struct virtio_pci_vq_info *info = vq->priv; |
294 | unsigned long size; | 465 | unsigned long size; |
295 | 466 | ||
467 | iowrite16(info->queue_index, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_SEL); | ||
468 | |||
469 | if (info->vector != VIRTIO_MSI_NO_VECTOR) | ||
470 | free_irq(vp_dev->msix_entries[info->vector].vector, vq); | ||
471 | |||
472 | if (vp_dev->msix_enabled) { | ||
473 | iowrite16(VIRTIO_MSI_NO_VECTOR, | ||
474 | vp_dev->ioaddr + VIRTIO_MSI_QUEUE_VECTOR); | ||
475 | /* Flush the write out to device */ | ||
476 | ioread8(vp_dev->ioaddr + VIRTIO_PCI_ISR); | ||
477 | } | ||
478 | |||
296 | vring_del_virtqueue(vq); | 479 | vring_del_virtqueue(vq); |
297 | 480 | ||
298 | /* Select and deactivate the queue */ | 481 | /* Select and deactivate the queue */ |
299 | iowrite16(info->queue_index, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_SEL); | ||
300 | iowrite32(0, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_PFN); | 482 | iowrite32(0, vp_dev->ioaddr + VIRTIO_PCI_QUEUE_PFN); |
301 | 483 | ||
302 | size = PAGE_ALIGN(vring_size(info->num, VIRTIO_PCI_VRING_ALIGN)); | 484 | size = PAGE_ALIGN(vring_size(info->num, VIRTIO_PCI_VRING_ALIGN)); |
@@ -304,30 +486,46 @@ static void vp_del_vq(struct virtqueue *vq) | |||
304 | kfree(info); | 486 | kfree(info); |
305 | } | 487 | } |
306 | 488 | ||
489 | /* the config->del_vqs() implementation */ | ||
307 | static void vp_del_vqs(struct virtio_device *vdev) | 490 | static void vp_del_vqs(struct virtio_device *vdev) |
308 | { | 491 | { |
309 | struct virtqueue *vq, *n; | 492 | struct virtqueue *vq, *n; |
310 | 493 | ||
311 | list_for_each_entry_safe(vq, n, &vdev->vqs, list) | 494 | list_for_each_entry_safe(vq, n, &vdev->vqs, list) |
312 | vp_del_vq(vq); | 495 | vp_del_vq(vq); |
496 | |||
497 | vp_free_vectors(vdev); | ||
313 | } | 498 | } |
314 | 499 | ||
500 | /* the config->find_vqs() implementation */ | ||
315 | static int vp_find_vqs(struct virtio_device *vdev, unsigned nvqs, | 501 | static int vp_find_vqs(struct virtio_device *vdev, unsigned nvqs, |
316 | struct virtqueue *vqs[], | 502 | struct virtqueue *vqs[], |
317 | vq_callback_t *callbacks[], | 503 | vq_callback_t *callbacks[], |
318 | const char *names[]) | 504 | const char *names[]) |
319 | { | 505 | { |
320 | int i; | 506 | int vectors = 0; |
507 | int i, err; | ||
508 | |||
509 | /* How many vectors would we like? */ | ||
510 | for (i = 0; i < nvqs; ++i) | ||
511 | if (callbacks[i]) | ||
512 | ++vectors; | ||
513 | |||
514 | err = vp_request_vectors(vdev, vectors); | ||
515 | if (err) | ||
516 | goto error_request; | ||
321 | 517 | ||
322 | for (i = 0; i < nvqs; ++i) { | 518 | for (i = 0; i < nvqs; ++i) { |
323 | vqs[i] = vp_find_vq(vdev, i, callbacks[i], names[i]); | 519 | vqs[i] = vp_find_vq(vdev, i, callbacks[i], names[i]); |
324 | if (IS_ERR(vqs[i])) | 520 | if (IS_ERR(vqs[i])) |
325 | goto error; | 521 | goto error_find; |
326 | } | 522 | } |
327 | return 0; | 523 | return 0; |
328 | 524 | ||
329 | error: | 525 | error_find: |
330 | vp_del_vqs(vdev); | 526 | vp_del_vqs(vdev); |
527 | |||
528 | error_request: | ||
331 | return PTR_ERR(vqs[i]); | 529 | return PTR_ERR(vqs[i]); |
332 | } | 530 | } |
333 | 531 | ||
@@ -349,7 +547,7 @@ static void virtio_pci_release_dev(struct device *_d) | |||
349 | struct virtio_pci_device *vp_dev = to_vp_device(dev); | 547 | struct virtio_pci_device *vp_dev = to_vp_device(dev); |
350 | struct pci_dev *pci_dev = vp_dev->pci_dev; | 548 | struct pci_dev *pci_dev = vp_dev->pci_dev; |
351 | 549 | ||
352 | free_irq(pci_dev->irq, vp_dev); | 550 | vp_del_vqs(dev); |
353 | pci_set_drvdata(pci_dev, NULL); | 551 | pci_set_drvdata(pci_dev, NULL); |
354 | pci_iounmap(pci_dev, vp_dev->ioaddr); | 552 | pci_iounmap(pci_dev, vp_dev->ioaddr); |
355 | pci_release_regions(pci_dev); | 553 | pci_release_regions(pci_dev); |
@@ -408,21 +606,13 @@ static int __devinit virtio_pci_probe(struct pci_dev *pci_dev, | |||
408 | vp_dev->vdev.id.vendor = pci_dev->subsystem_vendor; | 606 | vp_dev->vdev.id.vendor = pci_dev->subsystem_vendor; |
409 | vp_dev->vdev.id.device = pci_dev->subsystem_device; | 607 | vp_dev->vdev.id.device = pci_dev->subsystem_device; |
410 | 608 | ||
411 | /* register a handler for the queue with the PCI device's interrupt */ | ||
412 | err = request_irq(vp_dev->pci_dev->irq, vp_interrupt, IRQF_SHARED, | ||
413 | dev_name(&vp_dev->vdev.dev), vp_dev); | ||
414 | if (err) | ||
415 | goto out_set_drvdata; | ||
416 | |||
417 | /* finally register the virtio device */ | 609 | /* finally register the virtio device */ |
418 | err = register_virtio_device(&vp_dev->vdev); | 610 | err = register_virtio_device(&vp_dev->vdev); |
419 | if (err) | 611 | if (err) |
420 | goto out_req_irq; | 612 | goto out_set_drvdata; |
421 | 613 | ||
422 | return 0; | 614 | return 0; |
423 | 615 | ||
424 | out_req_irq: | ||
425 | free_irq(pci_dev->irq, vp_dev); | ||
426 | out_set_drvdata: | 616 | out_set_drvdata: |
427 | pci_set_drvdata(pci_dev, NULL); | 617 | pci_set_drvdata(pci_dev, NULL); |
428 | pci_iounmap(pci_dev, vp_dev->ioaddr); | 618 | pci_iounmap(pci_dev, vp_dev->ioaddr); |