diff options
author | David S. Miller <davem@davemloft.net> | 2008-01-16 01:43:24 -0500 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2008-01-17 04:49:29 -0500 |
commit | d2c7ddd6261eb885091cf6ddbcfae01f4216fb8e (patch) | |
tree | fefe1f6e6162e019be7a3f27788a395b1d121688 /drivers/net/e1000e | |
parent | d8c89eb3a12f0da96d049bd515c7fa3702e511c5 (diff) |
[NET]: Fix TX timeout regression in Intel drivers.
This fixes a regression added by changeset
53e52c729cc169db82a6105fac7a166e10c2ec36 ("[NET]: Make ->poll()
breakout consistent in Intel ethernet drivers.")
As pointed out by Jesse Brandeburg, for three of the drivers edited
above there is breakout logic in the *_clean_tx_irq() code to prevent
running TX reclaim forever. If this occurs, we have to elide NAPI
poll completion or else those TX events will never be serviced.
Signed-off-by: David S. Miller <davem@davemloft.net>
Acked-by: Jesse Brandeburg <jesse.brandeburg@intel.com>
Diffstat (limited to 'drivers/net/e1000e')
-rw-r--r-- | drivers/net/e1000e/netdev.c | 7 |
1 files changed, 5 insertions, 2 deletions
diff --git a/drivers/net/e1000e/netdev.c b/drivers/net/e1000e/netdev.c index 4a6fc7453776..2ab3bfbb8a63 100644 --- a/drivers/net/e1000e/netdev.c +++ b/drivers/net/e1000e/netdev.c | |||
@@ -1384,7 +1384,7 @@ static int e1000_clean(struct napi_struct *napi, int budget) | |||
1384 | { | 1384 | { |
1385 | struct e1000_adapter *adapter = container_of(napi, struct e1000_adapter, napi); | 1385 | struct e1000_adapter *adapter = container_of(napi, struct e1000_adapter, napi); |
1386 | struct net_device *poll_dev = adapter->netdev; | 1386 | struct net_device *poll_dev = adapter->netdev; |
1387 | int work_done = 0; | 1387 | int tx_cleaned = 0, work_done = 0; |
1388 | 1388 | ||
1389 | /* Must NOT use netdev_priv macro here. */ | 1389 | /* Must NOT use netdev_priv macro here. */ |
1390 | adapter = poll_dev->priv; | 1390 | adapter = poll_dev->priv; |
@@ -1394,12 +1394,15 @@ static int e1000_clean(struct napi_struct *napi, int budget) | |||
1394 | * simultaneously. A failure obtaining the lock means | 1394 | * simultaneously. A failure obtaining the lock means |
1395 | * tx_ring is currently being cleaned anyway. */ | 1395 | * tx_ring is currently being cleaned anyway. */ |
1396 | if (spin_trylock(&adapter->tx_queue_lock)) { | 1396 | if (spin_trylock(&adapter->tx_queue_lock)) { |
1397 | e1000_clean_tx_irq(adapter); | 1397 | tx_cleaned = e1000_clean_tx_irq(adapter); |
1398 | spin_unlock(&adapter->tx_queue_lock); | 1398 | spin_unlock(&adapter->tx_queue_lock); |
1399 | } | 1399 | } |
1400 | 1400 | ||
1401 | adapter->clean_rx(adapter, &work_done, budget); | 1401 | adapter->clean_rx(adapter, &work_done, budget); |
1402 | 1402 | ||
1403 | if (tx_cleaned) | ||
1404 | work_done = budget; | ||
1405 | |||
1403 | /* If budget not fully consumed, exit the polling mode */ | 1406 | /* If budget not fully consumed, exit the polling mode */ |
1404 | if (work_done < budget) { | 1407 | if (work_done < budget) { |
1405 | if (adapter->itr_setting & 3) | 1408 | if (adapter->itr_setting & 3) |