diff options
author | Matthew Wilcox <matthew.r.wilcox@intel.com> | 2013-06-20 10:53:48 -0400 |
---|---|---|
committer | Matthew Wilcox <matthew.r.wilcox@intel.com> | 2013-06-20 11:09:23 -0400 |
commit | 063a8096f3dbca7521d5918b3aea7ab46c5d2fe9 (patch) | |
tree | 24007515e6df6c7b78a100508884a5889a8ff789 /drivers/block | |
parent | 03ea83e9a37e41d436f8348e6eee3d8281bfff3a (diff) |
NVMe: Restructure MSI / MSI-X setup
The current code copies 'nr_io_queues' into 'q_count', modifies
'nr_io_queues' during MSI-X setup, then resets 'nr_io_queues' for
MSI setup. Instead, copy 'nr_io_queues' into 'vecs' and modify 'vecs'
during both MSI-X and MSI setup.
This lets us simplify the for-loops that set up MSI-X and MSI, and opens
the possibility of using more I/O queues than we have interrupt vectors,
should future benchmarking prove that to be a useful feature.
Signed-off-by: Matthew Wilcox <matthew.r.wilcox@intel.com>
Diffstat (limited to 'drivers/block')
-rw-r--r-- | drivers/block/nvme-core.c | 44 |
1 files changed, 23 insertions, 21 deletions
diff --git a/drivers/block/nvme-core.c b/drivers/block/nvme-core.c index ce79a590b45b..de3a75978c56 100644 --- a/drivers/block/nvme-core.c +++ b/drivers/block/nvme-core.c | |||
@@ -1638,7 +1638,7 @@ static int set_queue_count(struct nvme_dev *dev, int count) | |||
1638 | static int nvme_setup_io_queues(struct nvme_dev *dev) | 1638 | static int nvme_setup_io_queues(struct nvme_dev *dev) |
1639 | { | 1639 | { |
1640 | struct pci_dev *pdev = dev->pci_dev; | 1640 | struct pci_dev *pdev = dev->pci_dev; |
1641 | int result, cpu, i, nr_io_queues, db_bar_size, q_depth, q_count; | 1641 | int result, cpu, i, vecs, nr_io_queues, db_bar_size, q_depth; |
1642 | 1642 | ||
1643 | nr_io_queues = num_online_cpus(); | 1643 | nr_io_queues = num_online_cpus(); |
1644 | result = set_queue_count(dev, nr_io_queues); | 1644 | result = set_queue_count(dev, nr_io_queues); |
@@ -1647,7 +1647,6 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) | |||
1647 | if (result < nr_io_queues) | 1647 | if (result < nr_io_queues) |
1648 | nr_io_queues = result; | 1648 | nr_io_queues = result; |
1649 | 1649 | ||
1650 | q_count = nr_io_queues; | ||
1651 | /* Deregister the admin queue's interrupt */ | 1650 | /* Deregister the admin queue's interrupt */ |
1652 | free_irq(dev->entry[0].vector, dev->queues[0]); | 1651 | free_irq(dev->entry[0].vector, dev->queues[0]); |
1653 | 1652 | ||
@@ -1659,39 +1658,42 @@ static int nvme_setup_io_queues(struct nvme_dev *dev) | |||
1659 | dev->queues[0]->q_db = dev->dbs; | 1658 | dev->queues[0]->q_db = dev->dbs; |
1660 | } | 1659 | } |
1661 | 1660 | ||
1662 | for (i = 0; i < nr_io_queues; i++) | 1661 | vecs = nr_io_queues; |
1662 | for (i = 0; i < vecs; i++) | ||
1663 | dev->entry[i].entry = i; | 1663 | dev->entry[i].entry = i; |
1664 | for (;;) { | 1664 | for (;;) { |
1665 | result = pci_enable_msix(pdev, dev->entry, nr_io_queues); | 1665 | result = pci_enable_msix(pdev, dev->entry, vecs); |
1666 | if (result == 0) { | 1666 | if (result <= 0) |
1667 | break; | 1667 | break; |
1668 | } else if (result > 0) { | 1668 | vecs = result; |
1669 | nr_io_queues = result; | ||
1670 | continue; | ||
1671 | } else { | ||
1672 | nr_io_queues = 0; | ||
1673 | break; | ||
1674 | } | ||
1675 | } | 1669 | } |
1676 | 1670 | ||
1677 | if (nr_io_queues == 0) { | 1671 | if (result < 0) { |
1678 | nr_io_queues = q_count; | 1672 | vecs = nr_io_queues; |
1673 | if (vecs > 32) | ||
1674 | vecs = 32; | ||
1679 | for (;;) { | 1675 | for (;;) { |
1680 | result = pci_enable_msi_block(pdev, nr_io_queues); | 1676 | result = pci_enable_msi_block(pdev, vecs); |
1681 | if (result == 0) { | 1677 | if (result == 0) { |
1682 | for (i = 0; i < nr_io_queues; i++) | 1678 | for (i = 0; i < vecs; i++) |
1683 | dev->entry[i].vector = i + pdev->irq; | 1679 | dev->entry[i].vector = i + pdev->irq; |
1684 | break; | 1680 | break; |
1685 | } else if (result > 0) { | 1681 | } else if (result < 0) { |
1686 | nr_io_queues = result; | 1682 | vecs = 1; |
1687 | continue; | ||
1688 | } else { | ||
1689 | nr_io_queues = 1; | ||
1690 | break; | 1683 | break; |
1691 | } | 1684 | } |
1685 | vecs = result; | ||
1692 | } | 1686 | } |
1693 | } | 1687 | } |
1694 | 1688 | ||
1689 | /* | ||
1690 | * Should investigate if there's a performance win from allocating | ||
1691 | * more queues than interrupt vectors; it might allow the submission | ||
1692 | * path to scale better, even if the receive path is limited by the | ||
1693 | * number of interrupts. | ||
1694 | */ | ||
1695 | nr_io_queues = vecs; | ||
1696 | |||
1695 | result = queue_request_irq(dev, dev->queues[0], "nvme admin"); | 1697 | result = queue_request_irq(dev, dev->queues[0], "nvme admin"); |
1696 | /* XXX: handle failure here */ | 1698 | /* XXX: handle failure here */ |
1697 | 1699 | ||