diff options
author | Ayaz Abdulla <aabdulla@nvidia.com> | 2007-01-23 12:00:56 -0500 |
---|---|---|
committer | Jeff Garzik <jeff@garzik.org> | 2007-02-05 16:58:48 -0500 |
commit | 4e16ed1b0e17a3832310031e2ddaeb0914eb837d (patch) | |
tree | 4871cd0af1bfb95c70498bd3f1e9722b8e264eba /drivers/net | |
parent | f0734ab658390380079369f7090dcf7aa226f394 (diff) |
forcedeth: tx max work
This patch adds a limit to how much tx work can be done in each
iteration of tx processing. If the max limit is reached, remaining tx
completions will be handled by timer interrupt.
Signed-Off-By: Ayaz Abdulla <aabdulla@nvidia.com>
Signed-off-by: Jeff Garzik <jeff@garzik.org>
Diffstat (limited to 'drivers/net')
-rw-r--r-- | drivers/net/forcedeth.c | 18 |
1 files changed, 12 insertions, 6 deletions
diff --git a/drivers/net/forcedeth.c b/drivers/net/forcedeth.c index 7c0f0ccbbb29..ae6cf7827e57 100644 --- a/drivers/net/forcedeth.c +++ b/drivers/net/forcedeth.c | |||
@@ -210,7 +210,7 @@ enum { | |||
210 | * NVREG_POLL_DEFAULT=97 would result in an interval length of 1 ms | 210 | * NVREG_POLL_DEFAULT=97 would result in an interval length of 1 ms |
211 | */ | 211 | */ |
212 | NvRegPollingInterval = 0x00c, | 212 | NvRegPollingInterval = 0x00c, |
213 | #define NVREG_POLL_DEFAULT_THROUGHPUT 970 | 213 | #define NVREG_POLL_DEFAULT_THROUGHPUT 970 /* backup tx cleanup if loop max reached */ |
214 | #define NVREG_POLL_DEFAULT_CPU 13 | 214 | #define NVREG_POLL_DEFAULT_CPU 13 |
215 | NvRegMSIMap0 = 0x020, | 215 | NvRegMSIMap0 = 0x020, |
216 | NvRegMSIMap1 = 0x024, | 216 | NvRegMSIMap1 = 0x024, |
@@ -1859,14 +1859,15 @@ static void nv_tx_done(struct net_device *dev) | |||
1859 | } | 1859 | } |
1860 | } | 1860 | } |
1861 | 1861 | ||
1862 | static void nv_tx_done_optimized(struct net_device *dev) | 1862 | static void nv_tx_done_optimized(struct net_device *dev, int limit) |
1863 | { | 1863 | { |
1864 | struct fe_priv *np = netdev_priv(dev); | 1864 | struct fe_priv *np = netdev_priv(dev); |
1865 | u32 flags; | 1865 | u32 flags; |
1866 | struct ring_desc_ex* orig_get_tx = np->get_tx.ex; | 1866 | struct ring_desc_ex* orig_get_tx = np->get_tx.ex; |
1867 | 1867 | ||
1868 | while ((np->get_tx.ex != np->put_tx.ex) && | 1868 | while ((np->get_tx.ex != np->put_tx.ex) && |
1869 | !((flags = le32_to_cpu(np->get_tx.ex->flaglen)) & NV_TX_VALID)) { | 1869 | !((flags = le32_to_cpu(np->get_tx.ex->flaglen)) & NV_TX_VALID) && |
1870 | (limit-- > 0)) { | ||
1870 | 1871 | ||
1871 | dprintk(KERN_DEBUG "%s: nv_tx_done_optimized: flags 0x%x.\n", | 1872 | dprintk(KERN_DEBUG "%s: nv_tx_done_optimized: flags 0x%x.\n", |
1872 | dev->name, flags); | 1873 | dev->name, flags); |
@@ -1973,7 +1974,7 @@ static void nv_tx_timeout(struct net_device *dev) | |||
1973 | if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) | 1974 | if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) |
1974 | nv_tx_done(dev); | 1975 | nv_tx_done(dev); |
1975 | else | 1976 | else |
1976 | nv_tx_done_optimized(dev); | 1977 | nv_tx_done_optimized(dev, np->tx_ring_size); |
1977 | 1978 | ||
1978 | /* 3) if there are dead entries: clear everything */ | 1979 | /* 3) if there are dead entries: clear everything */ |
1979 | if (np->get_tx_ctx != np->put_tx_ctx) { | 1980 | if (np->get_tx_ctx != np->put_tx_ctx) { |
@@ -2899,7 +2900,7 @@ static irqreturn_t nv_nic_irq_optimized(int foo, void *data) | |||
2899 | break; | 2900 | break; |
2900 | 2901 | ||
2901 | spin_lock(&np->lock); | 2902 | spin_lock(&np->lock); |
2902 | nv_tx_done_optimized(dev); | 2903 | nv_tx_done_optimized(dev, TX_WORK_PER_LOOP); |
2903 | spin_unlock(&np->lock); | 2904 | spin_unlock(&np->lock); |
2904 | 2905 | ||
2905 | #ifdef CONFIG_FORCEDETH_NAPI | 2906 | #ifdef CONFIG_FORCEDETH_NAPI |
@@ -3006,7 +3007,7 @@ static irqreturn_t nv_nic_irq_tx(int foo, void *data) | |||
3006 | break; | 3007 | break; |
3007 | 3008 | ||
3008 | spin_lock_irqsave(&np->lock, flags); | 3009 | spin_lock_irqsave(&np->lock, flags); |
3009 | nv_tx_done_optimized(dev); | 3010 | nv_tx_done_optimized(dev, TX_WORK_PER_LOOP); |
3010 | spin_unlock_irqrestore(&np->lock, flags); | 3011 | spin_unlock_irqrestore(&np->lock, flags); |
3011 | 3012 | ||
3012 | if (unlikely(events & (NVREG_IRQ_TX_ERR))) { | 3013 | if (unlikely(events & (NVREG_IRQ_TX_ERR))) { |
@@ -3163,6 +3164,11 @@ static irqreturn_t nv_nic_irq_other(int foo, void *data) | |||
3163 | if (!(events & np->irqmask)) | 3164 | if (!(events & np->irqmask)) |
3164 | break; | 3165 | break; |
3165 | 3166 | ||
3167 | /* check tx in case we reached max loop limit in tx isr */ | ||
3168 | spin_lock_irqsave(&np->lock, flags); | ||
3169 | nv_tx_done_optimized(dev, TX_WORK_PER_LOOP); | ||
3170 | spin_unlock_irqrestore(&np->lock, flags); | ||
3171 | |||
3166 | if (events & NVREG_IRQ_LINK) { | 3172 | if (events & NVREG_IRQ_LINK) { |
3167 | spin_lock_irqsave(&np->lock, flags); | 3173 | spin_lock_irqsave(&np->lock, flags); |
3168 | nv_link_irq(dev); | 3174 | nv_link_irq(dev); |