aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/block
diff options
context:
space:
mode:
authorMatthew Wilcox <matthew.r.wilcox@intel.com>2013-06-20 10:53:48 -0400
committerMatthew Wilcox <matthew.r.wilcox@intel.com>2013-06-20 11:09:23 -0400
commit063a8096f3dbca7521d5918b3aea7ab46c5d2fe9 (patch)
tree24007515e6df6c7b78a100508884a5889a8ff789 /drivers/block
parent03ea83e9a37e41d436f8348e6eee3d8281bfff3a (diff)
NVMe: Restructure MSI / MSI-X setup
The current code copies 'nr_io_queues' into 'q_count', modifies 'nr_io_queues' during MSI-X setup, then resets 'nr_io_queues' for MSI setup. Instead, copy 'nr_io_queues' into 'vecs' and modify 'vecs' during both MSI-X and MSI setup. This lets us simplify the for-loops that set up MSI-X and MSI, and opens the possibility of using more I/O queues than we have interrupt vectors, should future benchmarking prove that to be a useful feature. Signed-off-by: Matthew Wilcox <matthew.r.wilcox@intel.com>
Diffstat (limited to 'drivers/block')
-rw-r--r--drivers/block/nvme-core.c44
1 files changed, 23 insertions, 21 deletions
diff --git a/drivers/block/nvme-core.c b/drivers/block/nvme-core.c
index ce79a590b45b..de3a75978c56 100644
--- a/drivers/block/nvme-core.c
+++ b/drivers/block/nvme-core.c
@@ -1638,7 +1638,7 @@ static int set_queue_count(struct nvme_dev *dev, int count)
1638static int nvme_setup_io_queues(struct nvme_dev *dev) 1638static int nvme_setup_io_queues(struct nvme_dev *dev)
1639{ 1639{
1640 struct pci_dev *pdev = dev->pci_dev; 1640 struct pci_dev *pdev = dev->pci_dev;
1641 int result, cpu, i, nr_io_queues, db_bar_size, q_depth, q_count; 1641 int result, cpu, i, vecs, nr_io_queues, db_bar_size, q_depth;
1642 1642
1643 nr_io_queues = num_online_cpus(); 1643 nr_io_queues = num_online_cpus();
1644 result = set_queue_count(dev, nr_io_queues); 1644 result = set_queue_count(dev, nr_io_queues);
@@ -1647,7 +1647,6 @@ static int nvme_setup_io_queues(struct nvme_dev *dev)
1647 if (result < nr_io_queues) 1647 if (result < nr_io_queues)
1648 nr_io_queues = result; 1648 nr_io_queues = result;
1649 1649
1650 q_count = nr_io_queues;
1651 /* Deregister the admin queue's interrupt */ 1650 /* Deregister the admin queue's interrupt */
1652 free_irq(dev->entry[0].vector, dev->queues[0]); 1651 free_irq(dev->entry[0].vector, dev->queues[0]);
1653 1652
@@ -1659,39 +1658,42 @@ static int nvme_setup_io_queues(struct nvme_dev *dev)
1659 dev->queues[0]->q_db = dev->dbs; 1658 dev->queues[0]->q_db = dev->dbs;
1660 } 1659 }
1661 1660
1662 for (i = 0; i < nr_io_queues; i++) 1661 vecs = nr_io_queues;
1662 for (i = 0; i < vecs; i++)
1663 dev->entry[i].entry = i; 1663 dev->entry[i].entry = i;
1664 for (;;) { 1664 for (;;) {
1665 result = pci_enable_msix(pdev, dev->entry, nr_io_queues); 1665 result = pci_enable_msix(pdev, dev->entry, vecs);
1666 if (result == 0) { 1666 if (result <= 0)
1667 break; 1667 break;
1668 } else if (result > 0) { 1668 vecs = result;
1669 nr_io_queues = result;
1670 continue;
1671 } else {
1672 nr_io_queues = 0;
1673 break;
1674 }
1675 } 1669 }
1676 1670
1677 if (nr_io_queues == 0) { 1671 if (result < 0) {
1678 nr_io_queues = q_count; 1672 vecs = nr_io_queues;
1673 if (vecs > 32)
1674 vecs = 32;
1679 for (;;) { 1675 for (;;) {
1680 result = pci_enable_msi_block(pdev, nr_io_queues); 1676 result = pci_enable_msi_block(pdev, vecs);
1681 if (result == 0) { 1677 if (result == 0) {
1682 for (i = 0; i < nr_io_queues; i++) 1678 for (i = 0; i < vecs; i++)
1683 dev->entry[i].vector = i + pdev->irq; 1679 dev->entry[i].vector = i + pdev->irq;
1684 break; 1680 break;
1685 } else if (result > 0) { 1681 } else if (result < 0) {
1686 nr_io_queues = result; 1682 vecs = 1;
1687 continue;
1688 } else {
1689 nr_io_queues = 1;
1690 break; 1683 break;
1691 } 1684 }
1685 vecs = result;
1692 } 1686 }
1693 } 1687 }
1694 1688
1689 /*
1690 * Should investigate if there's a performance win from allocating
1691 * more queues than interrupt vectors; it might allow the submission
1692 * path to scale better, even if the receive path is limited by the
1693 * number of interrupts.
1694 */
1695 nr_io_queues = vecs;
1696
1695 result = queue_request_irq(dev, dev->queues[0], "nvme admin"); 1697 result = queue_request_irq(dev, dev->queues[0], "nvme admin");
1696 /* XXX: handle failure here */ 1698 /* XXX: handle failure here */
1697 1699