diff options
author | Stephen Hemminger <shemminger@osdl.org> | 2006-03-23 14:07:28 -0500 |
---|---|---|
committer | Jeff Garzik <jeff@garzik.org> | 2006-03-23 17:13:54 -0500 |
commit | 29b4e886cbba5e0b0e4379fea7fc313e7808b5c2 (patch) | |
tree | 8871821748ca3b4d4a4d96e33d6ada6735107a32 /drivers/net | |
parent | 866b4f3e94a7568a1cb0018c061e19e120de6922 (diff) |
[PATCH] skge: compute available ring buffers
Don't need to keep track of available buffers, it is simpler
to just compute the value (ala e1000). Don't need tes on link up
because should always have available buffers then.
Signed-off-by: Stephen Hemminger <shemminger@osdl.org>
Signed-off-by: Jeff Garzik <jeff@garzik.org>
Diffstat (limited to 'drivers/net')
-rw-r--r-- | drivers/net/skge.c | 20 | ||||
-rw-r--r-- | drivers/net/skge.h | 1 |
2 files changed, 10 insertions, 11 deletions
diff --git a/drivers/net/skge.c b/drivers/net/skge.c index a261766bc052..a6f42fc85738 100644 --- a/drivers/net/skge.c +++ b/drivers/net/skge.c | |||
@@ -847,8 +847,7 @@ static void skge_link_up(struct skge_port *skge) | |||
847 | LED_BLK_OFF|LED_SYNC_OFF|LED_ON); | 847 | LED_BLK_OFF|LED_SYNC_OFF|LED_ON); |
848 | 848 | ||
849 | netif_carrier_on(skge->netdev); | 849 | netif_carrier_on(skge->netdev); |
850 | if (skge->tx_avail > MAX_SKB_FRAGS + 1) | 850 | netif_wake_queue(skge->netdev); |
851 | netif_wake_queue(skge->netdev); | ||
852 | 851 | ||
853 | if (netif_msg_link(skge)) | 852 | if (netif_msg_link(skge)) |
854 | printk(KERN_INFO PFX | 853 | printk(KERN_INFO PFX |
@@ -2190,8 +2189,6 @@ static int skge_up(struct net_device *dev) | |||
2190 | if (err) | 2189 | if (err) |
2191 | goto free_rx_ring; | 2190 | goto free_rx_ring; |
2192 | 2191 | ||
2193 | skge->tx_avail = skge->tx_ring.count - 1; | ||
2194 | |||
2195 | /* Initialize MAC */ | 2192 | /* Initialize MAC */ |
2196 | spin_lock_bh(&hw->phy_lock); | 2193 | spin_lock_bh(&hw->phy_lock); |
2197 | if (hw->chip_id == CHIP_ID_GENESIS) | 2194 | if (hw->chip_id == CHIP_ID_GENESIS) |
@@ -2294,6 +2291,12 @@ static int skge_down(struct net_device *dev) | |||
2294 | return 0; | 2291 | return 0; |
2295 | } | 2292 | } |
2296 | 2293 | ||
2294 | static inline int skge_avail(const struct skge_ring *ring) | ||
2295 | { | ||
2296 | return ((ring->to_clean > ring->to_use) ? 0 : ring->count) | ||
2297 | + (ring->to_clean - ring->to_use) - 1; | ||
2298 | } | ||
2299 | |||
2297 | static int skge_xmit_frame(struct sk_buff *skb, struct net_device *dev) | 2300 | static int skge_xmit_frame(struct sk_buff *skb, struct net_device *dev) |
2298 | { | 2301 | { |
2299 | struct skge_port *skge = netdev_priv(dev); | 2302 | struct skge_port *skge = netdev_priv(dev); |
@@ -2314,7 +2317,7 @@ static int skge_xmit_frame(struct sk_buff *skb, struct net_device *dev) | |||
2314 | return NETDEV_TX_LOCKED; | 2317 | return NETDEV_TX_LOCKED; |
2315 | } | 2318 | } |
2316 | 2319 | ||
2317 | if (unlikely(skge->tx_avail < skb_shinfo(skb)->nr_frags +1)) { | 2320 | if (unlikely(skge_avail(&skge->tx_ring) < skb_shinfo(skb)->nr_frags + 1)) { |
2318 | if (!netif_queue_stopped(dev)) { | 2321 | if (!netif_queue_stopped(dev)) { |
2319 | netif_stop_queue(dev); | 2322 | netif_stop_queue(dev); |
2320 | 2323 | ||
@@ -2390,8 +2393,7 @@ static int skge_xmit_frame(struct sk_buff *skb, struct net_device *dev) | |||
2390 | dev->name, e - ring->start, skb->len); | 2393 | dev->name, e - ring->start, skb->len); |
2391 | 2394 | ||
2392 | ring->to_use = e->next; | 2395 | ring->to_use = e->next; |
2393 | skge->tx_avail -= skb_shinfo(skb)->nr_frags + 1; | 2396 | if (skge_avail(&skge->tx_ring) <= MAX_SKB_FRAGS + 1) { |
2394 | if (skge->tx_avail <= MAX_SKB_FRAGS + 1) { | ||
2395 | pr_debug("%s: transmit queue full\n", dev->name); | 2397 | pr_debug("%s: transmit queue full\n", dev->name); |
2396 | netif_stop_queue(dev); | 2398 | netif_stop_queue(dev); |
2397 | } | 2399 | } |
@@ -2416,14 +2418,12 @@ static void skge_tx_complete(struct skge_port *skge, struct skge_element *last) | |||
2416 | e->skb = NULL; | 2418 | e->skb = NULL; |
2417 | pci_unmap_single(pdev, pci_unmap_addr(e, mapaddr), | 2419 | pci_unmap_single(pdev, pci_unmap_addr(e, mapaddr), |
2418 | skb_headlen(skb), PCI_DMA_TODEVICE); | 2420 | skb_headlen(skb), PCI_DMA_TODEVICE); |
2419 | ++skge->tx_avail; | ||
2420 | 2421 | ||
2421 | for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { | 2422 | for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { |
2422 | e = e->next; | 2423 | e = e->next; |
2423 | pci_unmap_page(pdev, pci_unmap_addr(e, mapaddr), | 2424 | pci_unmap_page(pdev, pci_unmap_addr(e, mapaddr), |
2424 | skb_shinfo(skb)->frags[i].size, | 2425 | skb_shinfo(skb)->frags[i].size, |
2425 | PCI_DMA_TODEVICE); | 2426 | PCI_DMA_TODEVICE); |
2426 | ++skge->tx_avail; | ||
2427 | } | 2427 | } |
2428 | 2428 | ||
2429 | dev_kfree_skb(skb); | 2429 | dev_kfree_skb(skb); |
@@ -2688,7 +2688,7 @@ static void skge_tx_done(struct skge_port *skge) | |||
2688 | 2688 | ||
2689 | skge_write8(skge->hw, Q_ADDR(txqaddr[skge->port], Q_CSR), CSR_IRQ_CL_F); | 2689 | skge_write8(skge->hw, Q_ADDR(txqaddr[skge->port], Q_CSR), CSR_IRQ_CL_F); |
2690 | 2690 | ||
2691 | if (skge->tx_avail > MAX_SKB_FRAGS + 1) | 2691 | if (skge_avail(&skge->tx_ring) > MAX_SKB_FRAGS + 1) |
2692 | netif_wake_queue(skge->netdev); | 2692 | netif_wake_queue(skge->netdev); |
2693 | 2693 | ||
2694 | spin_unlock(&skge->tx_lock); | 2694 | spin_unlock(&skge->tx_lock); |
diff --git a/drivers/net/skge.h b/drivers/net/skge.h index 2efdacc290e5..1f1ce88c8186 100644 --- a/drivers/net/skge.h +++ b/drivers/net/skge.h | |||
@@ -2418,7 +2418,6 @@ struct skge_port { | |||
2418 | int port; | 2418 | int port; |
2419 | 2419 | ||
2420 | spinlock_t tx_lock; | 2420 | spinlock_t tx_lock; |
2421 | u32 tx_avail; | ||
2422 | struct skge_ring tx_ring; | 2421 | struct skge_ring tx_ring; |
2423 | struct skge_ring rx_ring; | 2422 | struct skge_ring rx_ring; |
2424 | 2423 | ||