diff options
author | Stephen Hemminger <shemminger@linux-foundation.org> | 2007-10-03 19:41:36 -0400 |
---|---|---|
committer | David S. Miller <davem@sunset.davemloft.net> | 2007-10-10 19:47:45 -0400 |
commit | bea3348eef27e6044b6161fd04c3152215f96411 (patch) | |
tree | f0990b263e5ce42505d290a4c346fe990bcd4c33 /drivers/net | |
parent | dde4e47e8fe333a5649a3fa0e7db1fa7c08d6158 (diff) |
[NET]: Make NAPI polling independent of struct net_device objects.
Several devices have multiple independant RX queues per net
device, and some have a single interrupt doorbell for several
queues.
In either case, it's easier to support layouts like that if the
structure representing the poll is independant from the net
device itself.
The signature of the ->poll() call back goes from:
int foo_poll(struct net_device *dev, int *budget)
to
int foo_poll(struct napi_struct *napi, int budget)
The caller is returned the number of RX packets processed (or
the number of "NAPI credits" consumed if you want to get
abstract). The callee no longer messes around bumping
dev->quota, *budget, etc. because that is all handled in the
caller upon return.
The napi_struct is to be embedded in the device driver private data
structures.
Furthermore, it is the driver's responsibility to disable all NAPI
instances in it's ->stop() device close handler. Since the
napi_struct is privatized into the driver's private data structures,
only the driver knows how to get at all of the napi_struct instances
it may have per-device.
With lots of help and suggestions from Rusty Russell, Roland Dreier,
Michael Chan, Jeff Garzik, and Jamal Hadi Salim.
Bug fixes from Thomas Graf, Roland Dreier, Peter Zijlstra,
Joseph Fannin, Scott Wood, Hans J. Koch, and Michael Chan.
[ Ported to current tree and all drivers converted. Integrated
Stephen's follow-on kerneldoc additions, and restored poll_list
handling to the old style to fix mutual exclusion issues. -DaveM ]
Signed-off-by: Stephen Hemminger <shemminger@linux-foundation.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'drivers/net')
78 files changed, 1241 insertions, 1380 deletions
diff --git a/drivers/net/8139cp.c b/drivers/net/8139cp.c index a79f28c7a100..7f18ca23d9f8 100644 --- a/drivers/net/8139cp.c +++ b/drivers/net/8139cp.c | |||
@@ -334,6 +334,8 @@ struct cp_private { | |||
334 | spinlock_t lock; | 334 | spinlock_t lock; |
335 | u32 msg_enable; | 335 | u32 msg_enable; |
336 | 336 | ||
337 | struct napi_struct napi; | ||
338 | |||
337 | struct pci_dev *pdev; | 339 | struct pci_dev *pdev; |
338 | u32 rx_config; | 340 | u32 rx_config; |
339 | u16 cpcmd; | 341 | u16 cpcmd; |
@@ -501,12 +503,12 @@ static inline unsigned int cp_rx_csum_ok (u32 status) | |||
501 | return 0; | 503 | return 0; |
502 | } | 504 | } |
503 | 505 | ||
504 | static int cp_rx_poll (struct net_device *dev, int *budget) | 506 | static int cp_rx_poll(struct napi_struct *napi, int budget) |
505 | { | 507 | { |
506 | struct cp_private *cp = netdev_priv(dev); | 508 | struct cp_private *cp = container_of(napi, struct cp_private, napi); |
507 | unsigned rx_tail = cp->rx_tail; | 509 | struct net_device *dev = cp->dev; |
508 | unsigned rx_work = dev->quota; | 510 | unsigned int rx_tail = cp->rx_tail; |
509 | unsigned rx; | 511 | int rx; |
510 | 512 | ||
511 | rx_status_loop: | 513 | rx_status_loop: |
512 | rx = 0; | 514 | rx = 0; |
@@ -588,33 +590,28 @@ rx_next: | |||
588 | desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); | 590 | desc->opts1 = cpu_to_le32(DescOwn | cp->rx_buf_sz); |
589 | rx_tail = NEXT_RX(rx_tail); | 591 | rx_tail = NEXT_RX(rx_tail); |
590 | 592 | ||
591 | if (!rx_work--) | 593 | if (rx >= budget) |
592 | break; | 594 | break; |
593 | } | 595 | } |
594 | 596 | ||
595 | cp->rx_tail = rx_tail; | 597 | cp->rx_tail = rx_tail; |
596 | 598 | ||
597 | dev->quota -= rx; | ||
598 | *budget -= rx; | ||
599 | |||
600 | /* if we did not reach work limit, then we're done with | 599 | /* if we did not reach work limit, then we're done with |
601 | * this round of polling | 600 | * this round of polling |
602 | */ | 601 | */ |
603 | if (rx_work) { | 602 | if (rx < budget) { |
604 | unsigned long flags; | 603 | unsigned long flags; |
605 | 604 | ||
606 | if (cpr16(IntrStatus) & cp_rx_intr_mask) | 605 | if (cpr16(IntrStatus) & cp_rx_intr_mask) |
607 | goto rx_status_loop; | 606 | goto rx_status_loop; |
608 | 607 | ||
609 | local_irq_save(flags); | 608 | spin_lock_irqsave(&cp->lock, flags); |
610 | cpw16_f(IntrMask, cp_intr_mask); | 609 | cpw16_f(IntrMask, cp_intr_mask); |
611 | __netif_rx_complete(dev); | 610 | __netif_rx_complete(dev, napi); |
612 | local_irq_restore(flags); | 611 | spin_unlock_irqrestore(&cp->lock, flags); |
613 | |||
614 | return 0; /* done */ | ||
615 | } | 612 | } |
616 | 613 | ||
617 | return 1; /* not done */ | 614 | return rx; |
618 | } | 615 | } |
619 | 616 | ||
620 | static irqreturn_t cp_interrupt (int irq, void *dev_instance) | 617 | static irqreturn_t cp_interrupt (int irq, void *dev_instance) |
@@ -647,9 +644,9 @@ static irqreturn_t cp_interrupt (int irq, void *dev_instance) | |||
647 | } | 644 | } |
648 | 645 | ||
649 | if (status & (RxOK | RxErr | RxEmpty | RxFIFOOvr)) | 646 | if (status & (RxOK | RxErr | RxEmpty | RxFIFOOvr)) |
650 | if (netif_rx_schedule_prep(dev)) { | 647 | if (netif_rx_schedule_prep(dev, &cp->napi)) { |
651 | cpw16_f(IntrMask, cp_norx_intr_mask); | 648 | cpw16_f(IntrMask, cp_norx_intr_mask); |
652 | __netif_rx_schedule(dev); | 649 | __netif_rx_schedule(dev, &cp->napi); |
653 | } | 650 | } |
654 | 651 | ||
655 | if (status & (TxOK | TxErr | TxEmpty | SWInt)) | 652 | if (status & (TxOK | TxErr | TxEmpty | SWInt)) |
@@ -1175,6 +1172,8 @@ static int cp_open (struct net_device *dev) | |||
1175 | if (rc) | 1172 | if (rc) |
1176 | return rc; | 1173 | return rc; |
1177 | 1174 | ||
1175 | napi_enable(&cp->napi); | ||
1176 | |||
1178 | cp_init_hw(cp); | 1177 | cp_init_hw(cp); |
1179 | 1178 | ||
1180 | rc = request_irq(dev->irq, cp_interrupt, IRQF_SHARED, dev->name, dev); | 1179 | rc = request_irq(dev->irq, cp_interrupt, IRQF_SHARED, dev->name, dev); |
@@ -1188,6 +1187,7 @@ static int cp_open (struct net_device *dev) | |||
1188 | return 0; | 1187 | return 0; |
1189 | 1188 | ||
1190 | err_out_hw: | 1189 | err_out_hw: |
1190 | napi_disable(&cp->napi); | ||
1191 | cp_stop_hw(cp); | 1191 | cp_stop_hw(cp); |
1192 | cp_free_rings(cp); | 1192 | cp_free_rings(cp); |
1193 | return rc; | 1193 | return rc; |
@@ -1198,6 +1198,8 @@ static int cp_close (struct net_device *dev) | |||
1198 | struct cp_private *cp = netdev_priv(dev); | 1198 | struct cp_private *cp = netdev_priv(dev); |
1199 | unsigned long flags; | 1199 | unsigned long flags; |
1200 | 1200 | ||
1201 | napi_disable(&cp->napi); | ||
1202 | |||
1201 | if (netif_msg_ifdown(cp)) | 1203 | if (netif_msg_ifdown(cp)) |
1202 | printk(KERN_DEBUG "%s: disabling interface\n", dev->name); | 1204 | printk(KERN_DEBUG "%s: disabling interface\n", dev->name); |
1203 | 1205 | ||
@@ -1933,11 +1935,10 @@ static int cp_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1933 | dev->hard_start_xmit = cp_start_xmit; | 1935 | dev->hard_start_xmit = cp_start_xmit; |
1934 | dev->get_stats = cp_get_stats; | 1936 | dev->get_stats = cp_get_stats; |
1935 | dev->do_ioctl = cp_ioctl; | 1937 | dev->do_ioctl = cp_ioctl; |
1936 | dev->poll = cp_rx_poll; | ||
1937 | #ifdef CONFIG_NET_POLL_CONTROLLER | 1938 | #ifdef CONFIG_NET_POLL_CONTROLLER |
1938 | dev->poll_controller = cp_poll_controller; | 1939 | dev->poll_controller = cp_poll_controller; |
1939 | #endif | 1940 | #endif |
1940 | dev->weight = 16; /* arbitrary? from NAPI_HOWTO.txt. */ | 1941 | netif_napi_add(dev, &cp->napi, cp_rx_poll, 16); |
1941 | #ifdef BROKEN | 1942 | #ifdef BROKEN |
1942 | dev->change_mtu = cp_change_mtu; | 1943 | dev->change_mtu = cp_change_mtu; |
1943 | #endif | 1944 | #endif |
diff --git a/drivers/net/8139too.c b/drivers/net/8139too.c index f4e4298d24b9..20af6baecfcb 100644 --- a/drivers/net/8139too.c +++ b/drivers/net/8139too.c | |||
@@ -573,6 +573,8 @@ struct rtl8139_private { | |||
573 | int drv_flags; | 573 | int drv_flags; |
574 | struct pci_dev *pci_dev; | 574 | struct pci_dev *pci_dev; |
575 | u32 msg_enable; | 575 | u32 msg_enable; |
576 | struct napi_struct napi; | ||
577 | struct net_device *dev; | ||
576 | struct net_device_stats stats; | 578 | struct net_device_stats stats; |
577 | unsigned char *rx_ring; | 579 | unsigned char *rx_ring; |
578 | unsigned int cur_rx; /* Index into the Rx buffer of next Rx pkt. */ | 580 | unsigned int cur_rx; /* Index into the Rx buffer of next Rx pkt. */ |
@@ -625,10 +627,10 @@ static void rtl8139_tx_timeout (struct net_device *dev); | |||
625 | static void rtl8139_init_ring (struct net_device *dev); | 627 | static void rtl8139_init_ring (struct net_device *dev); |
626 | static int rtl8139_start_xmit (struct sk_buff *skb, | 628 | static int rtl8139_start_xmit (struct sk_buff *skb, |
627 | struct net_device *dev); | 629 | struct net_device *dev); |
628 | static int rtl8139_poll(struct net_device *dev, int *budget); | ||
629 | #ifdef CONFIG_NET_POLL_CONTROLLER | 630 | #ifdef CONFIG_NET_POLL_CONTROLLER |
630 | static void rtl8139_poll_controller(struct net_device *dev); | 631 | static void rtl8139_poll_controller(struct net_device *dev); |
631 | #endif | 632 | #endif |
633 | static int rtl8139_poll(struct napi_struct *napi, int budget); | ||
632 | static irqreturn_t rtl8139_interrupt (int irq, void *dev_instance); | 634 | static irqreturn_t rtl8139_interrupt (int irq, void *dev_instance); |
633 | static int rtl8139_close (struct net_device *dev); | 635 | static int rtl8139_close (struct net_device *dev); |
634 | static int netdev_ioctl (struct net_device *dev, struct ifreq *rq, int cmd); | 636 | static int netdev_ioctl (struct net_device *dev, struct ifreq *rq, int cmd); |
@@ -963,6 +965,7 @@ static int __devinit rtl8139_init_one (struct pci_dev *pdev, | |||
963 | 965 | ||
964 | assert (dev != NULL); | 966 | assert (dev != NULL); |
965 | tp = netdev_priv(dev); | 967 | tp = netdev_priv(dev); |
968 | tp->dev = dev; | ||
966 | 969 | ||
967 | ioaddr = tp->mmio_addr; | 970 | ioaddr = tp->mmio_addr; |
968 | assert (ioaddr != NULL); | 971 | assert (ioaddr != NULL); |
@@ -976,8 +979,7 @@ static int __devinit rtl8139_init_one (struct pci_dev *pdev, | |||
976 | /* The Rtl8139-specific entries in the device structure. */ | 979 | /* The Rtl8139-specific entries in the device structure. */ |
977 | dev->open = rtl8139_open; | 980 | dev->open = rtl8139_open; |
978 | dev->hard_start_xmit = rtl8139_start_xmit; | 981 | dev->hard_start_xmit = rtl8139_start_xmit; |
979 | dev->poll = rtl8139_poll; | 982 | netif_napi_add(dev, &tp->napi, rtl8139_poll, 64); |
980 | dev->weight = 64; | ||
981 | dev->stop = rtl8139_close; | 983 | dev->stop = rtl8139_close; |
982 | dev->get_stats = rtl8139_get_stats; | 984 | dev->get_stats = rtl8139_get_stats; |
983 | dev->set_multicast_list = rtl8139_set_rx_mode; | 985 | dev->set_multicast_list = rtl8139_set_rx_mode; |
@@ -1332,6 +1334,8 @@ static int rtl8139_open (struct net_device *dev) | |||
1332 | 1334 | ||
1333 | } | 1335 | } |
1334 | 1336 | ||
1337 | napi_enable(&tp->napi); | ||
1338 | |||
1335 | tp->mii.full_duplex = tp->mii.force_media; | 1339 | tp->mii.full_duplex = tp->mii.force_media; |
1336 | tp->tx_flag = (TX_FIFO_THRESH << 11) & 0x003f0000; | 1340 | tp->tx_flag = (TX_FIFO_THRESH << 11) & 0x003f0000; |
1337 | 1341 | ||
@@ -2103,39 +2107,32 @@ static void rtl8139_weird_interrupt (struct net_device *dev, | |||
2103 | } | 2107 | } |
2104 | } | 2108 | } |
2105 | 2109 | ||
2106 | static int rtl8139_poll(struct net_device *dev, int *budget) | 2110 | static int rtl8139_poll(struct napi_struct *napi, int budget) |
2107 | { | 2111 | { |
2108 | struct rtl8139_private *tp = netdev_priv(dev); | 2112 | struct rtl8139_private *tp = container_of(napi, struct rtl8139_private, napi); |
2113 | struct net_device *dev = tp->dev; | ||
2109 | void __iomem *ioaddr = tp->mmio_addr; | 2114 | void __iomem *ioaddr = tp->mmio_addr; |
2110 | int orig_budget = min(*budget, dev->quota); | 2115 | int work_done; |
2111 | int done = 1; | ||
2112 | 2116 | ||
2113 | spin_lock(&tp->rx_lock); | 2117 | spin_lock(&tp->rx_lock); |
2114 | if (likely(RTL_R16(IntrStatus) & RxAckBits)) { | 2118 | work_done = 0; |
2115 | int work_done; | 2119 | if (likely(RTL_R16(IntrStatus) & RxAckBits)) |
2116 | 2120 | work_done += rtl8139_rx(dev, tp, budget); | |
2117 | work_done = rtl8139_rx(dev, tp, orig_budget); | ||
2118 | if (likely(work_done > 0)) { | ||
2119 | *budget -= work_done; | ||
2120 | dev->quota -= work_done; | ||
2121 | done = (work_done < orig_budget); | ||
2122 | } | ||
2123 | } | ||
2124 | 2121 | ||
2125 | if (done) { | 2122 | if (work_done < budget) { |
2126 | unsigned long flags; | 2123 | unsigned long flags; |
2127 | /* | 2124 | /* |
2128 | * Order is important since data can get interrupted | 2125 | * Order is important since data can get interrupted |
2129 | * again when we think we are done. | 2126 | * again when we think we are done. |
2130 | */ | 2127 | */ |
2131 | local_irq_save(flags); | 2128 | spin_lock_irqsave(&tp->lock, flags); |
2132 | RTL_W16_F(IntrMask, rtl8139_intr_mask); | 2129 | RTL_W16_F(IntrMask, rtl8139_intr_mask); |
2133 | __netif_rx_complete(dev); | 2130 | __netif_rx_complete(dev, napi); |
2134 | local_irq_restore(flags); | 2131 | spin_unlock_irqrestore(&tp->lock, flags); |
2135 | } | 2132 | } |
2136 | spin_unlock(&tp->rx_lock); | 2133 | spin_unlock(&tp->rx_lock); |
2137 | 2134 | ||
2138 | return !done; | 2135 | return work_done; |
2139 | } | 2136 | } |
2140 | 2137 | ||
2141 | /* The interrupt handler does all of the Rx thread work and cleans up | 2138 | /* The interrupt handler does all of the Rx thread work and cleans up |
@@ -2180,9 +2177,9 @@ static irqreturn_t rtl8139_interrupt (int irq, void *dev_instance) | |||
2180 | /* Receive packets are processed by poll routine. | 2177 | /* Receive packets are processed by poll routine. |
2181 | If not running start it now. */ | 2178 | If not running start it now. */ |
2182 | if (status & RxAckBits){ | 2179 | if (status & RxAckBits){ |
2183 | if (netif_rx_schedule_prep(dev)) { | 2180 | if (netif_rx_schedule_prep(dev, &tp->napi)) { |
2184 | RTL_W16_F (IntrMask, rtl8139_norx_intr_mask); | 2181 | RTL_W16_F (IntrMask, rtl8139_norx_intr_mask); |
2185 | __netif_rx_schedule (dev); | 2182 | __netif_rx_schedule(dev, &tp->napi); |
2186 | } | 2183 | } |
2187 | } | 2184 | } |
2188 | 2185 | ||
@@ -2223,7 +2220,8 @@ static int rtl8139_close (struct net_device *dev) | |||
2223 | void __iomem *ioaddr = tp->mmio_addr; | 2220 | void __iomem *ioaddr = tp->mmio_addr; |
2224 | unsigned long flags; | 2221 | unsigned long flags; |
2225 | 2222 | ||
2226 | netif_stop_queue (dev); | 2223 | netif_stop_queue(dev); |
2224 | napi_disable(&tp->napi); | ||
2227 | 2225 | ||
2228 | if (netif_msg_ifdown(tp)) | 2226 | if (netif_msg_ifdown(tp)) |
2229 | printk(KERN_DEBUG "%s: Shutting down ethercard, status was 0x%4.4x.\n", | 2227 | printk(KERN_DEBUG "%s: Shutting down ethercard, status was 0x%4.4x.\n", |
diff --git a/drivers/net/amd8111e.c b/drivers/net/amd8111e.c index a61b2f89fc33..cf06fc067e92 100644 --- a/drivers/net/amd8111e.c +++ b/drivers/net/amd8111e.c | |||
@@ -723,9 +723,10 @@ static int amd8111e_tx(struct net_device *dev) | |||
723 | 723 | ||
724 | #ifdef CONFIG_AMD8111E_NAPI | 724 | #ifdef CONFIG_AMD8111E_NAPI |
725 | /* This function handles the driver receive operation in polling mode */ | 725 | /* This function handles the driver receive operation in polling mode */ |
726 | static int amd8111e_rx_poll(struct net_device *dev, int * budget) | 726 | static int amd8111e_rx_poll(struct napi_struct *napi, int budget) |
727 | { | 727 | { |
728 | struct amd8111e_priv *lp = netdev_priv(dev); | 728 | struct amd8111e_priv *lp = container_of(napi, struct amd8111e_priv, napi); |
729 | struct net_device *dev = lp->amd8111e_net_dev; | ||
729 | int rx_index = lp->rx_idx & RX_RING_DR_MOD_MASK; | 730 | int rx_index = lp->rx_idx & RX_RING_DR_MOD_MASK; |
730 | void __iomem *mmio = lp->mmio; | 731 | void __iomem *mmio = lp->mmio; |
731 | struct sk_buff *skb,*new_skb; | 732 | struct sk_buff *skb,*new_skb; |
@@ -737,7 +738,7 @@ static int amd8111e_rx_poll(struct net_device *dev, int * budget) | |||
737 | #if AMD8111E_VLAN_TAG_USED | 738 | #if AMD8111E_VLAN_TAG_USED |
738 | short vtag; | 739 | short vtag; |
739 | #endif | 740 | #endif |
740 | int rx_pkt_limit = dev->quota; | 741 | int rx_pkt_limit = budget; |
741 | unsigned long flags; | 742 | unsigned long flags; |
742 | 743 | ||
743 | do{ | 744 | do{ |
@@ -838,21 +839,14 @@ static int amd8111e_rx_poll(struct net_device *dev, int * budget) | |||
838 | } while(intr0 & RINT0); | 839 | } while(intr0 & RINT0); |
839 | 840 | ||
840 | /* Receive descriptor is empty now */ | 841 | /* Receive descriptor is empty now */ |
841 | dev->quota -= num_rx_pkt; | ||
842 | *budget -= num_rx_pkt; | ||
843 | |||
844 | spin_lock_irqsave(&lp->lock, flags); | 842 | spin_lock_irqsave(&lp->lock, flags); |
845 | netif_rx_complete(dev); | 843 | __netif_rx_complete(dev, napi); |
846 | writel(VAL0|RINTEN0, mmio + INTEN0); | 844 | writel(VAL0|RINTEN0, mmio + INTEN0); |
847 | writel(VAL2 | RDMD0, mmio + CMD0); | 845 | writel(VAL2 | RDMD0, mmio + CMD0); |
848 | spin_unlock_irqrestore(&lp->lock, flags); | 846 | spin_unlock_irqrestore(&lp->lock, flags); |
849 | return 0; | ||
850 | 847 | ||
851 | rx_not_empty: | 848 | rx_not_empty: |
852 | /* Do not call a netif_rx_complete */ | 849 | return num_rx_pkt; |
853 | dev->quota -= num_rx_pkt; | ||
854 | *budget -= num_rx_pkt; | ||
855 | return 1; | ||
856 | } | 850 | } |
857 | 851 | ||
858 | #else | 852 | #else |
@@ -1287,11 +1281,11 @@ static irqreturn_t amd8111e_interrupt(int irq, void *dev_id) | |||
1287 | /* Check if Receive Interrupt has occurred. */ | 1281 | /* Check if Receive Interrupt has occurred. */ |
1288 | #ifdef CONFIG_AMD8111E_NAPI | 1282 | #ifdef CONFIG_AMD8111E_NAPI |
1289 | if(intr0 & RINT0){ | 1283 | if(intr0 & RINT0){ |
1290 | if(netif_rx_schedule_prep(dev)){ | 1284 | if(netif_rx_schedule_prep(dev, &lp->napi)){ |
1291 | /* Disable receive interupts */ | 1285 | /* Disable receive interupts */ |
1292 | writel(RINTEN0, mmio + INTEN0); | 1286 | writel(RINTEN0, mmio + INTEN0); |
1293 | /* Schedule a polling routine */ | 1287 | /* Schedule a polling routine */ |
1294 | __netif_rx_schedule(dev); | 1288 | __netif_rx_schedule(dev, &lp->napi); |
1295 | } | 1289 | } |
1296 | else if (intren0 & RINTEN0) { | 1290 | else if (intren0 & RINTEN0) { |
1297 | printk("************Driver bug! \ | 1291 | printk("************Driver bug! \ |
@@ -1345,6 +1339,8 @@ static int amd8111e_close(struct net_device * dev) | |||
1345 | struct amd8111e_priv *lp = netdev_priv(dev); | 1339 | struct amd8111e_priv *lp = netdev_priv(dev); |
1346 | netif_stop_queue(dev); | 1340 | netif_stop_queue(dev); |
1347 | 1341 | ||
1342 | napi_disable(&lp->napi); | ||
1343 | |||
1348 | spin_lock_irq(&lp->lock); | 1344 | spin_lock_irq(&lp->lock); |
1349 | 1345 | ||
1350 | amd8111e_disable_interrupt(lp); | 1346 | amd8111e_disable_interrupt(lp); |
@@ -1375,12 +1371,15 @@ static int amd8111e_open(struct net_device * dev ) | |||
1375 | dev->name, dev)) | 1371 | dev->name, dev)) |
1376 | return -EAGAIN; | 1372 | return -EAGAIN; |
1377 | 1373 | ||
1374 | napi_enable(&lp->napi); | ||
1375 | |||
1378 | spin_lock_irq(&lp->lock); | 1376 | spin_lock_irq(&lp->lock); |
1379 | 1377 | ||
1380 | amd8111e_init_hw_default(lp); | 1378 | amd8111e_init_hw_default(lp); |
1381 | 1379 | ||
1382 | if(amd8111e_restart(dev)){ | 1380 | if(amd8111e_restart(dev)){ |
1383 | spin_unlock_irq(&lp->lock); | 1381 | spin_unlock_irq(&lp->lock); |
1382 | napi_disable(&lp->napi); | ||
1384 | if (dev->irq) | 1383 | if (dev->irq) |
1385 | free_irq(dev->irq, dev); | 1384 | free_irq(dev->irq, dev); |
1386 | return -ENOMEM; | 1385 | return -ENOMEM; |
@@ -2031,8 +2030,7 @@ static int __devinit amd8111e_probe_one(struct pci_dev *pdev, | |||
2031 | dev->tx_timeout = amd8111e_tx_timeout; | 2030 | dev->tx_timeout = amd8111e_tx_timeout; |
2032 | dev->watchdog_timeo = AMD8111E_TX_TIMEOUT; | 2031 | dev->watchdog_timeo = AMD8111E_TX_TIMEOUT; |
2033 | #ifdef CONFIG_AMD8111E_NAPI | 2032 | #ifdef CONFIG_AMD8111E_NAPI |
2034 | dev->poll = amd8111e_rx_poll; | 2033 | netif_napi_add(dev, &lp->napi, amd8111e_rx_poll, 32); |
2035 | dev->weight = 32; | ||
2036 | #endif | 2034 | #endif |
2037 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2035 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2038 | dev->poll_controller = amd8111e_poll; | 2036 | dev->poll_controller = amd8111e_poll; |
diff --git a/drivers/net/amd8111e.h b/drivers/net/amd8111e.h index e65080a5994a..612e653610e1 100644 --- a/drivers/net/amd8111e.h +++ b/drivers/net/amd8111e.h | |||
@@ -763,6 +763,8 @@ struct amd8111e_priv{ | |||
763 | /* Reg memory mapped address */ | 763 | /* Reg memory mapped address */ |
764 | void __iomem *mmio; | 764 | void __iomem *mmio; |
765 | 765 | ||
766 | struct napi_struct napi; | ||
767 | |||
766 | spinlock_t lock; /* Guard lock */ | 768 | spinlock_t lock; /* Guard lock */ |
767 | unsigned long rx_idx, tx_idx; /* The next free ring entry */ | 769 | unsigned long rx_idx, tx_idx; /* The next free ring entry */ |
768 | unsigned long tx_complete_idx; | 770 | unsigned long tx_complete_idx; |
diff --git a/drivers/net/arm/ep93xx_eth.c b/drivers/net/arm/ep93xx_eth.c index f6ece1d43f6e..7f016f3d5bf0 100644 --- a/drivers/net/arm/ep93xx_eth.c +++ b/drivers/net/arm/ep93xx_eth.c | |||
@@ -169,6 +169,9 @@ struct ep93xx_priv | |||
169 | spinlock_t tx_pending_lock; | 169 | spinlock_t tx_pending_lock; |
170 | unsigned int tx_pending; | 170 | unsigned int tx_pending; |
171 | 171 | ||
172 | struct net_device *dev; | ||
173 | struct napi_struct napi; | ||
174 | |||
172 | struct net_device_stats stats; | 175 | struct net_device_stats stats; |
173 | 176 | ||
174 | struct mii_if_info mii; | 177 | struct mii_if_info mii; |
@@ -190,15 +193,11 @@ static struct net_device_stats *ep93xx_get_stats(struct net_device *dev) | |||
190 | return &(ep->stats); | 193 | return &(ep->stats); |
191 | } | 194 | } |
192 | 195 | ||
193 | static int ep93xx_rx(struct net_device *dev, int *budget) | 196 | static int ep93xx_rx(struct net_device *dev, int processed, int budget) |
194 | { | 197 | { |
195 | struct ep93xx_priv *ep = netdev_priv(dev); | 198 | struct ep93xx_priv *ep = netdev_priv(dev); |
196 | int rx_done; | ||
197 | int processed; | ||
198 | 199 | ||
199 | rx_done = 0; | 200 | while (processed < budget) { |
200 | processed = 0; | ||
201 | while (*budget > 0) { | ||
202 | int entry; | 201 | int entry; |
203 | struct ep93xx_rstat *rstat; | 202 | struct ep93xx_rstat *rstat; |
204 | u32 rstat0; | 203 | u32 rstat0; |
@@ -211,10 +210,8 @@ static int ep93xx_rx(struct net_device *dev, int *budget) | |||
211 | 210 | ||
212 | rstat0 = rstat->rstat0; | 211 | rstat0 = rstat->rstat0; |
213 | rstat1 = rstat->rstat1; | 212 | rstat1 = rstat->rstat1; |
214 | if (!(rstat0 & RSTAT0_RFP) || !(rstat1 & RSTAT1_RFP)) { | 213 | if (!(rstat0 & RSTAT0_RFP) || !(rstat1 & RSTAT1_RFP)) |
215 | rx_done = 1; | ||
216 | break; | 214 | break; |
217 | } | ||
218 | 215 | ||
219 | rstat->rstat0 = 0; | 216 | rstat->rstat0 = 0; |
220 | rstat->rstat1 = 0; | 217 | rstat->rstat1 = 0; |
@@ -275,8 +272,6 @@ static int ep93xx_rx(struct net_device *dev, int *budget) | |||
275 | err: | 272 | err: |
276 | ep->rx_pointer = (entry + 1) & (RX_QUEUE_ENTRIES - 1); | 273 | ep->rx_pointer = (entry + 1) & (RX_QUEUE_ENTRIES - 1); |
277 | processed++; | 274 | processed++; |
278 | dev->quota--; | ||
279 | (*budget)--; | ||
280 | } | 275 | } |
281 | 276 | ||
282 | if (processed) { | 277 | if (processed) { |
@@ -284,7 +279,7 @@ err: | |||
284 | wrw(ep, REG_RXSTSENQ, processed); | 279 | wrw(ep, REG_RXSTSENQ, processed); |
285 | } | 280 | } |
286 | 281 | ||
287 | return !rx_done; | 282 | return processed; |
288 | } | 283 | } |
289 | 284 | ||
290 | static int ep93xx_have_more_rx(struct ep93xx_priv *ep) | 285 | static int ep93xx_have_more_rx(struct ep93xx_priv *ep) |
@@ -293,36 +288,32 @@ static int ep93xx_have_more_rx(struct ep93xx_priv *ep) | |||
293 | return !!((rstat->rstat0 & RSTAT0_RFP) && (rstat->rstat1 & RSTAT1_RFP)); | 288 | return !!((rstat->rstat0 & RSTAT0_RFP) && (rstat->rstat1 & RSTAT1_RFP)); |
294 | } | 289 | } |
295 | 290 | ||
296 | static int ep93xx_poll(struct net_device *dev, int *budget) | 291 | static int ep93xx_poll(struct napi_struct *napi, int budget) |
297 | { | 292 | { |
298 | struct ep93xx_priv *ep = netdev_priv(dev); | 293 | struct ep93xx_priv *ep = container_of(napi, struct ep93xx_priv, napi); |
299 | 294 | struct net_device *dev = ep->dev; | |
300 | /* | 295 | int rx = 0; |
301 | * @@@ Have to stop polling if device is downed while we | ||
302 | * are polling. | ||
303 | */ | ||
304 | 296 | ||
305 | poll_some_more: | 297 | poll_some_more: |
306 | if (ep93xx_rx(dev, budget)) | 298 | rx = ep93xx_rx(dev, rx, budget); |
307 | return 1; | 299 | if (rx < budget) { |
308 | 300 | int more = 0; | |
309 | netif_rx_complete(dev); | 301 | |
310 | 302 | spin_lock_irq(&ep->rx_lock); | |
311 | spin_lock_irq(&ep->rx_lock); | 303 | __netif_rx_complete(dev, napi); |
312 | wrl(ep, REG_INTEN, REG_INTEN_TX | REG_INTEN_RX); | 304 | wrl(ep, REG_INTEN, REG_INTEN_TX | REG_INTEN_RX); |
313 | if (ep93xx_have_more_rx(ep)) { | 305 | if (ep93xx_have_more_rx(ep)) { |
314 | wrl(ep, REG_INTEN, REG_INTEN_TX); | 306 | wrl(ep, REG_INTEN, REG_INTEN_TX); |
315 | wrl(ep, REG_INTSTSP, REG_INTSTS_RX); | 307 | wrl(ep, REG_INTSTSP, REG_INTSTS_RX); |
308 | more = 1; | ||
309 | } | ||
316 | spin_unlock_irq(&ep->rx_lock); | 310 | spin_unlock_irq(&ep->rx_lock); |
317 | 311 | ||
318 | if (netif_rx_reschedule(dev, 0)) | 312 | if (more && netif_rx_reschedule(dev, napi)) |
319 | goto poll_some_more; | 313 | goto poll_some_more; |
320 | |||
321 | return 0; | ||
322 | } | 314 | } |
323 | spin_unlock_irq(&ep->rx_lock); | ||
324 | 315 | ||
325 | return 0; | 316 | return rx; |
326 | } | 317 | } |
327 | 318 | ||
328 | static int ep93xx_xmit(struct sk_buff *skb, struct net_device *dev) | 319 | static int ep93xx_xmit(struct sk_buff *skb, struct net_device *dev) |
@@ -426,9 +417,9 @@ static irqreturn_t ep93xx_irq(int irq, void *dev_id) | |||
426 | 417 | ||
427 | if (status & REG_INTSTS_RX) { | 418 | if (status & REG_INTSTS_RX) { |
428 | spin_lock(&ep->rx_lock); | 419 | spin_lock(&ep->rx_lock); |
429 | if (likely(__netif_rx_schedule_prep(dev))) { | 420 | if (likely(__netif_rx_schedule_prep(dev, &ep->napi))) { |
430 | wrl(ep, REG_INTEN, REG_INTEN_TX); | 421 | wrl(ep, REG_INTEN, REG_INTEN_TX); |
431 | __netif_rx_schedule(dev); | 422 | __netif_rx_schedule(dev, &ep->napi); |
432 | } | 423 | } |
433 | spin_unlock(&ep->rx_lock); | 424 | spin_unlock(&ep->rx_lock); |
434 | } | 425 | } |
@@ -648,7 +639,10 @@ static int ep93xx_open(struct net_device *dev) | |||
648 | dev->dev_addr[4], dev->dev_addr[5]); | 639 | dev->dev_addr[4], dev->dev_addr[5]); |
649 | } | 640 | } |
650 | 641 | ||
642 | napi_enable(&ep->napi); | ||
643 | |||
651 | if (ep93xx_start_hw(dev)) { | 644 | if (ep93xx_start_hw(dev)) { |
645 | napi_disable(&ep->napi); | ||
652 | ep93xx_free_buffers(ep); | 646 | ep93xx_free_buffers(ep); |
653 | return -EIO; | 647 | return -EIO; |
654 | } | 648 | } |
@@ -662,6 +656,7 @@ static int ep93xx_open(struct net_device *dev) | |||
662 | 656 | ||
663 | err = request_irq(ep->irq, ep93xx_irq, IRQF_SHARED, dev->name, dev); | 657 | err = request_irq(ep->irq, ep93xx_irq, IRQF_SHARED, dev->name, dev); |
664 | if (err) { | 658 | if (err) { |
659 | napi_disable(&ep->napi); | ||
665 | ep93xx_stop_hw(dev); | 660 | ep93xx_stop_hw(dev); |
666 | ep93xx_free_buffers(ep); | 661 | ep93xx_free_buffers(ep); |
667 | return err; | 662 | return err; |
@@ -678,6 +673,7 @@ static int ep93xx_close(struct net_device *dev) | |||
678 | { | 673 | { |
679 | struct ep93xx_priv *ep = netdev_priv(dev); | 674 | struct ep93xx_priv *ep = netdev_priv(dev); |
680 | 675 | ||
676 | napi_disable(&ep->napi); | ||
681 | netif_stop_queue(dev); | 677 | netif_stop_queue(dev); |
682 | 678 | ||
683 | wrl(ep, REG_GIINTMSK, 0); | 679 | wrl(ep, REG_GIINTMSK, 0); |
@@ -788,14 +784,12 @@ struct net_device *ep93xx_dev_alloc(struct ep93xx_eth_data *data) | |||
788 | 784 | ||
789 | dev->get_stats = ep93xx_get_stats; | 785 | dev->get_stats = ep93xx_get_stats; |
790 | dev->ethtool_ops = &ep93xx_ethtool_ops; | 786 | dev->ethtool_ops = &ep93xx_ethtool_ops; |
791 | dev->poll = ep93xx_poll; | ||
792 | dev->hard_start_xmit = ep93xx_xmit; | 787 | dev->hard_start_xmit = ep93xx_xmit; |
793 | dev->open = ep93xx_open; | 788 | dev->open = ep93xx_open; |
794 | dev->stop = ep93xx_close; | 789 | dev->stop = ep93xx_close; |
795 | dev->do_ioctl = ep93xx_ioctl; | 790 | dev->do_ioctl = ep93xx_ioctl; |
796 | 791 | ||
797 | dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM; | 792 | dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM; |
798 | dev->weight = 64; | ||
799 | 793 | ||
800 | return dev; | 794 | return dev; |
801 | } | 795 | } |
@@ -847,6 +841,8 @@ static int ep93xx_eth_probe(struct platform_device *pdev) | |||
847 | goto err_out; | 841 | goto err_out; |
848 | } | 842 | } |
849 | ep = netdev_priv(dev); | 843 | ep = netdev_priv(dev); |
844 | ep->dev = dev; | ||
845 | netif_napi_add(dev, &ep->napi, ep93xx_poll, 64); | ||
850 | 846 | ||
851 | platform_set_drvdata(pdev, dev); | 847 | platform_set_drvdata(pdev, dev); |
852 | 848 | ||
diff --git a/drivers/net/b44.c b/drivers/net/b44.c index 0795df235492..b92b3e25c42a 100644 --- a/drivers/net/b44.c +++ b/drivers/net/b44.c | |||
@@ -848,10 +848,11 @@ static int b44_rx(struct b44 *bp, int budget) | |||
848 | return received; | 848 | return received; |
849 | } | 849 | } |
850 | 850 | ||
851 | static int b44_poll(struct net_device *netdev, int *budget) | 851 | static int b44_poll(struct napi_struct *napi, int budget) |
852 | { | 852 | { |
853 | struct b44 *bp = netdev_priv(netdev); | 853 | struct b44 *bp = container_of(napi, struct b44, napi); |
854 | int done; | 854 | struct net_device *netdev = bp->dev; |
855 | int work_done; | ||
855 | 856 | ||
856 | spin_lock_irq(&bp->lock); | 857 | spin_lock_irq(&bp->lock); |
857 | 858 | ||
@@ -862,22 +863,9 @@ static int b44_poll(struct net_device *netdev, int *budget) | |||
862 | } | 863 | } |
863 | spin_unlock_irq(&bp->lock); | 864 | spin_unlock_irq(&bp->lock); |
864 | 865 | ||
865 | done = 1; | 866 | work_done = 0; |
866 | if (bp->istat & ISTAT_RX) { | 867 | if (bp->istat & ISTAT_RX) |
867 | int orig_budget = *budget; | 868 | work_done += b44_rx(bp, budget); |
868 | int work_done; | ||
869 | |||
870 | if (orig_budget > netdev->quota) | ||
871 | orig_budget = netdev->quota; | ||
872 | |||
873 | work_done = b44_rx(bp, orig_budget); | ||
874 | |||
875 | *budget -= work_done; | ||
876 | netdev->quota -= work_done; | ||
877 | |||
878 | if (work_done >= orig_budget) | ||
879 | done = 0; | ||
880 | } | ||
881 | 869 | ||
882 | if (bp->istat & ISTAT_ERRORS) { | 870 | if (bp->istat & ISTAT_ERRORS) { |
883 | unsigned long flags; | 871 | unsigned long flags; |
@@ -888,15 +876,15 @@ static int b44_poll(struct net_device *netdev, int *budget) | |||
888 | b44_init_hw(bp, B44_FULL_RESET_SKIP_PHY); | 876 | b44_init_hw(bp, B44_FULL_RESET_SKIP_PHY); |
889 | netif_wake_queue(bp->dev); | 877 | netif_wake_queue(bp->dev); |
890 | spin_unlock_irqrestore(&bp->lock, flags); | 878 | spin_unlock_irqrestore(&bp->lock, flags); |
891 | done = 1; | 879 | work_done = 0; |
892 | } | 880 | } |
893 | 881 | ||
894 | if (done) { | 882 | if (work_done < budget) { |
895 | netif_rx_complete(netdev); | 883 | netif_rx_complete(netdev, napi); |
896 | b44_enable_ints(bp); | 884 | b44_enable_ints(bp); |
897 | } | 885 | } |
898 | 886 | ||
899 | return (done ? 0 : 1); | 887 | return work_done; |
900 | } | 888 | } |
901 | 889 | ||
902 | static irqreturn_t b44_interrupt(int irq, void *dev_id) | 890 | static irqreturn_t b44_interrupt(int irq, void *dev_id) |
@@ -924,13 +912,13 @@ static irqreturn_t b44_interrupt(int irq, void *dev_id) | |||
924 | goto irq_ack; | 912 | goto irq_ack; |
925 | } | 913 | } |
926 | 914 | ||
927 | if (netif_rx_schedule_prep(dev)) { | 915 | if (netif_rx_schedule_prep(dev, &bp->napi)) { |
928 | /* NOTE: These writes are posted by the readback of | 916 | /* NOTE: These writes are posted by the readback of |
929 | * the ISTAT register below. | 917 | * the ISTAT register below. |
930 | */ | 918 | */ |
931 | bp->istat = istat; | 919 | bp->istat = istat; |
932 | __b44_disable_ints(bp); | 920 | __b44_disable_ints(bp); |
933 | __netif_rx_schedule(dev); | 921 | __netif_rx_schedule(dev, &bp->napi); |
934 | } else { | 922 | } else { |
935 | printk(KERN_ERR PFX "%s: Error, poll already scheduled\n", | 923 | printk(KERN_ERR PFX "%s: Error, poll already scheduled\n", |
936 | dev->name); | 924 | dev->name); |
@@ -1420,6 +1408,8 @@ static int b44_open(struct net_device *dev) | |||
1420 | if (err) | 1408 | if (err) |
1421 | goto out; | 1409 | goto out; |
1422 | 1410 | ||
1411 | napi_enable(&bp->napi); | ||
1412 | |||
1423 | b44_init_rings(bp); | 1413 | b44_init_rings(bp); |
1424 | b44_init_hw(bp, B44_FULL_RESET); | 1414 | b44_init_hw(bp, B44_FULL_RESET); |
1425 | 1415 | ||
@@ -1427,6 +1417,7 @@ static int b44_open(struct net_device *dev) | |||
1427 | 1417 | ||
1428 | err = request_irq(dev->irq, b44_interrupt, IRQF_SHARED, dev->name, dev); | 1418 | err = request_irq(dev->irq, b44_interrupt, IRQF_SHARED, dev->name, dev); |
1429 | if (unlikely(err < 0)) { | 1419 | if (unlikely(err < 0)) { |
1420 | napi_disable(&bp->napi); | ||
1430 | b44_chip_reset(bp); | 1421 | b44_chip_reset(bp); |
1431 | b44_free_rings(bp); | 1422 | b44_free_rings(bp); |
1432 | b44_free_consistent(bp); | 1423 | b44_free_consistent(bp); |
@@ -1609,7 +1600,7 @@ static int b44_close(struct net_device *dev) | |||
1609 | 1600 | ||
1610 | netif_stop_queue(dev); | 1601 | netif_stop_queue(dev); |
1611 | 1602 | ||
1612 | netif_poll_disable(dev); | 1603 | napi_disable(&bp->napi); |
1613 | 1604 | ||
1614 | del_timer_sync(&bp->timer); | 1605 | del_timer_sync(&bp->timer); |
1615 | 1606 | ||
@@ -1626,8 +1617,6 @@ static int b44_close(struct net_device *dev) | |||
1626 | 1617 | ||
1627 | free_irq(dev->irq, dev); | 1618 | free_irq(dev->irq, dev); |
1628 | 1619 | ||
1629 | netif_poll_enable(dev); | ||
1630 | |||
1631 | if (bp->flags & B44_FLAG_WOL_ENABLE) { | 1620 | if (bp->flags & B44_FLAG_WOL_ENABLE) { |
1632 | b44_init_hw(bp, B44_PARTIAL_RESET); | 1621 | b44_init_hw(bp, B44_PARTIAL_RESET); |
1633 | b44_setup_wol(bp); | 1622 | b44_setup_wol(bp); |
@@ -2194,8 +2183,7 @@ static int __devinit b44_init_one(struct pci_dev *pdev, | |||
2194 | dev->set_mac_address = b44_set_mac_addr; | 2183 | dev->set_mac_address = b44_set_mac_addr; |
2195 | dev->do_ioctl = b44_ioctl; | 2184 | dev->do_ioctl = b44_ioctl; |
2196 | dev->tx_timeout = b44_tx_timeout; | 2185 | dev->tx_timeout = b44_tx_timeout; |
2197 | dev->poll = b44_poll; | 2186 | netif_napi_add(dev, &bp->napi, b44_poll, 64); |
2198 | dev->weight = 64; | ||
2199 | dev->watchdog_timeo = B44_TX_TIMEOUT; | 2187 | dev->watchdog_timeo = B44_TX_TIMEOUT; |
2200 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2188 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2201 | dev->poll_controller = b44_poll_controller; | 2189 | dev->poll_controller = b44_poll_controller; |
diff --git a/drivers/net/b44.h b/drivers/net/b44.h index e537e63f292e..63c55a4ab3cd 100644 --- a/drivers/net/b44.h +++ b/drivers/net/b44.h | |||
@@ -423,6 +423,8 @@ struct b44 { | |||
423 | struct ring_info *rx_buffers; | 423 | struct ring_info *rx_buffers; |
424 | struct ring_info *tx_buffers; | 424 | struct ring_info *tx_buffers; |
425 | 425 | ||
426 | struct napi_struct napi; | ||
427 | |||
426 | u32 dma_offset; | 428 | u32 dma_offset; |
427 | u32 flags; | 429 | u32 flags; |
428 | #define B44_FLAG_B0_ANDLATER 0x00000001 | 430 | #define B44_FLAG_B0_ANDLATER 0x00000001 |
diff --git a/drivers/net/bnx2.c b/drivers/net/bnx2.c index 66eed22cbd21..ab028ad04235 100644 --- a/drivers/net/bnx2.c +++ b/drivers/net/bnx2.c | |||
@@ -428,7 +428,7 @@ bnx2_netif_stop(struct bnx2 *bp) | |||
428 | { | 428 | { |
429 | bnx2_disable_int_sync(bp); | 429 | bnx2_disable_int_sync(bp); |
430 | if (netif_running(bp->dev)) { | 430 | if (netif_running(bp->dev)) { |
431 | netif_poll_disable(bp->dev); | 431 | napi_disable(&bp->napi); |
432 | netif_tx_disable(bp->dev); | 432 | netif_tx_disable(bp->dev); |
433 | bp->dev->trans_start = jiffies; /* prevent tx timeout */ | 433 | bp->dev->trans_start = jiffies; /* prevent tx timeout */ |
434 | } | 434 | } |
@@ -440,7 +440,7 @@ bnx2_netif_start(struct bnx2 *bp) | |||
440 | if (atomic_dec_and_test(&bp->intr_sem)) { | 440 | if (atomic_dec_and_test(&bp->intr_sem)) { |
441 | if (netif_running(bp->dev)) { | 441 | if (netif_running(bp->dev)) { |
442 | netif_wake_queue(bp->dev); | 442 | netif_wake_queue(bp->dev); |
443 | netif_poll_enable(bp->dev); | 443 | napi_enable(&bp->napi); |
444 | bnx2_enable_int(bp); | 444 | bnx2_enable_int(bp); |
445 | } | 445 | } |
446 | } | 446 | } |
@@ -2551,7 +2551,7 @@ bnx2_msi(int irq, void *dev_instance) | |||
2551 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) | 2551 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) |
2552 | return IRQ_HANDLED; | 2552 | return IRQ_HANDLED; |
2553 | 2553 | ||
2554 | netif_rx_schedule(dev); | 2554 | netif_rx_schedule(dev, &bp->napi); |
2555 | 2555 | ||
2556 | return IRQ_HANDLED; | 2556 | return IRQ_HANDLED; |
2557 | } | 2557 | } |
@@ -2568,7 +2568,7 @@ bnx2_msi_1shot(int irq, void *dev_instance) | |||
2568 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) | 2568 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) |
2569 | return IRQ_HANDLED; | 2569 | return IRQ_HANDLED; |
2570 | 2570 | ||
2571 | netif_rx_schedule(dev); | 2571 | netif_rx_schedule(dev, &bp->napi); |
2572 | 2572 | ||
2573 | return IRQ_HANDLED; | 2573 | return IRQ_HANDLED; |
2574 | } | 2574 | } |
@@ -2604,9 +2604,9 @@ bnx2_interrupt(int irq, void *dev_instance) | |||
2604 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) | 2604 | if (unlikely(atomic_read(&bp->intr_sem) != 0)) |
2605 | return IRQ_HANDLED; | 2605 | return IRQ_HANDLED; |
2606 | 2606 | ||
2607 | if (netif_rx_schedule_prep(dev)) { | 2607 | if (netif_rx_schedule_prep(dev, &bp->napi)) { |
2608 | bp->last_status_idx = sblk->status_idx; | 2608 | bp->last_status_idx = sblk->status_idx; |
2609 | __netif_rx_schedule(dev); | 2609 | __netif_rx_schedule(dev, &bp->napi); |
2610 | } | 2610 | } |
2611 | 2611 | ||
2612 | return IRQ_HANDLED; | 2612 | return IRQ_HANDLED; |
@@ -2632,12 +2632,14 @@ bnx2_has_work(struct bnx2 *bp) | |||
2632 | } | 2632 | } |
2633 | 2633 | ||
2634 | static int | 2634 | static int |
2635 | bnx2_poll(struct net_device *dev, int *budget) | 2635 | bnx2_poll(struct napi_struct *napi, int budget) |
2636 | { | 2636 | { |
2637 | struct bnx2 *bp = netdev_priv(dev); | 2637 | struct bnx2 *bp = container_of(napi, struct bnx2, napi); |
2638 | struct net_device *dev = bp->dev; | ||
2638 | struct status_block *sblk = bp->status_blk; | 2639 | struct status_block *sblk = bp->status_blk; |
2639 | u32 status_attn_bits = sblk->status_attn_bits; | 2640 | u32 status_attn_bits = sblk->status_attn_bits; |
2640 | u32 status_attn_bits_ack = sblk->status_attn_bits_ack; | 2641 | u32 status_attn_bits_ack = sblk->status_attn_bits_ack; |
2642 | int work_done = 0; | ||
2641 | 2643 | ||
2642 | if ((status_attn_bits & STATUS_ATTN_EVENTS) != | 2644 | if ((status_attn_bits & STATUS_ATTN_EVENTS) != |
2643 | (status_attn_bits_ack & STATUS_ATTN_EVENTS)) { | 2645 | (status_attn_bits_ack & STATUS_ATTN_EVENTS)) { |
@@ -2655,23 +2657,14 @@ bnx2_poll(struct net_device *dev, int *budget) | |||
2655 | if (bp->status_blk->status_tx_quick_consumer_index0 != bp->hw_tx_cons) | 2657 | if (bp->status_blk->status_tx_quick_consumer_index0 != bp->hw_tx_cons) |
2656 | bnx2_tx_int(bp); | 2658 | bnx2_tx_int(bp); |
2657 | 2659 | ||
2658 | if (bp->status_blk->status_rx_quick_consumer_index0 != bp->hw_rx_cons) { | 2660 | if (bp->status_blk->status_rx_quick_consumer_index0 != bp->hw_rx_cons) |
2659 | int orig_budget = *budget; | 2661 | work_done = bnx2_rx_int(bp, budget); |
2660 | int work_done; | ||
2661 | |||
2662 | if (orig_budget > dev->quota) | ||
2663 | orig_budget = dev->quota; | ||
2664 | |||
2665 | work_done = bnx2_rx_int(bp, orig_budget); | ||
2666 | *budget -= work_done; | ||
2667 | dev->quota -= work_done; | ||
2668 | } | ||
2669 | 2662 | ||
2670 | bp->last_status_idx = bp->status_blk->status_idx; | 2663 | bp->last_status_idx = bp->status_blk->status_idx; |
2671 | rmb(); | 2664 | rmb(); |
2672 | 2665 | ||
2673 | if (!bnx2_has_work(bp)) { | 2666 | if (!bnx2_has_work(bp)) { |
2674 | netif_rx_complete(dev); | 2667 | netif_rx_complete(dev, napi); |
2675 | if (likely(bp->flags & USING_MSI_FLAG)) { | 2668 | if (likely(bp->flags & USING_MSI_FLAG)) { |
2676 | REG_WR(bp, BNX2_PCICFG_INT_ACK_CMD, | 2669 | REG_WR(bp, BNX2_PCICFG_INT_ACK_CMD, |
2677 | BNX2_PCICFG_INT_ACK_CMD_INDEX_VALID | | 2670 | BNX2_PCICFG_INT_ACK_CMD_INDEX_VALID | |
@@ -2686,10 +2679,9 @@ bnx2_poll(struct net_device *dev, int *budget) | |||
2686 | REG_WR(bp, BNX2_PCICFG_INT_ACK_CMD, | 2679 | REG_WR(bp, BNX2_PCICFG_INT_ACK_CMD, |
2687 | BNX2_PCICFG_INT_ACK_CMD_INDEX_VALID | | 2680 | BNX2_PCICFG_INT_ACK_CMD_INDEX_VALID | |
2688 | bp->last_status_idx); | 2681 | bp->last_status_idx); |
2689 | return 0; | ||
2690 | } | 2682 | } |
2691 | 2683 | ||
2692 | return 1; | 2684 | return work_done; |
2693 | } | 2685 | } |
2694 | 2686 | ||
2695 | /* Called with rtnl_lock from vlan functions and also netif_tx_lock | 2687 | /* Called with rtnl_lock from vlan functions and also netif_tx_lock |
@@ -5039,6 +5031,8 @@ bnx2_open(struct net_device *dev) | |||
5039 | if (rc) | 5031 | if (rc) |
5040 | return rc; | 5032 | return rc; |
5041 | 5033 | ||
5034 | napi_enable(&bp->napi); | ||
5035 | |||
5042 | if ((bp->flags & MSI_CAP_FLAG) && !disable_msi) { | 5036 | if ((bp->flags & MSI_CAP_FLAG) && !disable_msi) { |
5043 | if (pci_enable_msi(bp->pdev) == 0) { | 5037 | if (pci_enable_msi(bp->pdev) == 0) { |
5044 | bp->flags |= USING_MSI_FLAG; | 5038 | bp->flags |= USING_MSI_FLAG; |
@@ -5049,6 +5043,7 @@ bnx2_open(struct net_device *dev) | |||
5049 | rc = bnx2_request_irq(bp); | 5043 | rc = bnx2_request_irq(bp); |
5050 | 5044 | ||
5051 | if (rc) { | 5045 | if (rc) { |
5046 | napi_disable(&bp->napi); | ||
5052 | bnx2_free_mem(bp); | 5047 | bnx2_free_mem(bp); |
5053 | return rc; | 5048 | return rc; |
5054 | } | 5049 | } |
@@ -5056,6 +5051,7 @@ bnx2_open(struct net_device *dev) | |||
5056 | rc = bnx2_init_nic(bp); | 5051 | rc = bnx2_init_nic(bp); |
5057 | 5052 | ||
5058 | if (rc) { | 5053 | if (rc) { |
5054 | napi_disable(&bp->napi); | ||
5059 | bnx2_free_irq(bp); | 5055 | bnx2_free_irq(bp); |
5060 | bnx2_free_skbs(bp); | 5056 | bnx2_free_skbs(bp); |
5061 | bnx2_free_mem(bp); | 5057 | bnx2_free_mem(bp); |
@@ -5088,6 +5084,7 @@ bnx2_open(struct net_device *dev) | |||
5088 | rc = bnx2_request_irq(bp); | 5084 | rc = bnx2_request_irq(bp); |
5089 | 5085 | ||
5090 | if (rc) { | 5086 | if (rc) { |
5087 | napi_disable(&bp->napi); | ||
5091 | bnx2_free_skbs(bp); | 5088 | bnx2_free_skbs(bp); |
5092 | bnx2_free_mem(bp); | 5089 | bnx2_free_mem(bp); |
5093 | del_timer_sync(&bp->timer); | 5090 | del_timer_sync(&bp->timer); |
@@ -5301,7 +5298,8 @@ bnx2_close(struct net_device *dev) | |||
5301 | while (bp->in_reset_task) | 5298 | while (bp->in_reset_task) |
5302 | msleep(1); | 5299 | msleep(1); |
5303 | 5300 | ||
5304 | bnx2_netif_stop(bp); | 5301 | bnx2_disable_int_sync(bp); |
5302 | napi_disable(&bp->napi); | ||
5305 | del_timer_sync(&bp->timer); | 5303 | del_timer_sync(&bp->timer); |
5306 | if (bp->flags & NO_WOL_FLAG) | 5304 | if (bp->flags & NO_WOL_FLAG) |
5307 | reset_code = BNX2_DRV_MSG_CODE_UNLOAD_LNK_DN; | 5305 | reset_code = BNX2_DRV_MSG_CODE_UNLOAD_LNK_DN; |
@@ -6858,11 +6856,10 @@ bnx2_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
6858 | #ifdef BCM_VLAN | 6856 | #ifdef BCM_VLAN |
6859 | dev->vlan_rx_register = bnx2_vlan_rx_register; | 6857 | dev->vlan_rx_register = bnx2_vlan_rx_register; |
6860 | #endif | 6858 | #endif |
6861 | dev->poll = bnx2_poll; | ||
6862 | dev->ethtool_ops = &bnx2_ethtool_ops; | 6859 | dev->ethtool_ops = &bnx2_ethtool_ops; |
6863 | dev->weight = 64; | ||
6864 | 6860 | ||
6865 | bp = netdev_priv(dev); | 6861 | bp = netdev_priv(dev); |
6862 | netif_napi_add(dev, &bp->napi, bnx2_poll, 64); | ||
6866 | 6863 | ||
6867 | #if defined(HAVE_POLL_CONTROLLER) || defined(CONFIG_NET_POLL_CONTROLLER) | 6864 | #if defined(HAVE_POLL_CONTROLLER) || defined(CONFIG_NET_POLL_CONTROLLER) |
6868 | dev->poll_controller = poll_bnx2; | 6865 | dev->poll_controller = poll_bnx2; |
diff --git a/drivers/net/bnx2.h b/drivers/net/bnx2.h index 102adfe1e923..fbae439db647 100644 --- a/drivers/net/bnx2.h +++ b/drivers/net/bnx2.h | |||
@@ -6473,6 +6473,8 @@ struct bnx2 { | |||
6473 | struct net_device *dev; | 6473 | struct net_device *dev; |
6474 | struct pci_dev *pdev; | 6474 | struct pci_dev *pdev; |
6475 | 6475 | ||
6476 | struct napi_struct napi; | ||
6477 | |||
6476 | atomic_t intr_sem; | 6478 | atomic_t intr_sem; |
6477 | 6479 | ||
6478 | struct status_block *status_blk; | 6480 | struct status_block *status_blk; |
diff --git a/drivers/net/cassini.c b/drivers/net/cassini.c index f6e4030c73d1..13f14df21e6e 100644 --- a/drivers/net/cassini.c +++ b/drivers/net/cassini.c | |||
@@ -2485,7 +2485,7 @@ static irqreturn_t cas_interruptN(int irq, void *dev_id) | |||
2485 | if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ | 2485 | if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ |
2486 | #ifdef USE_NAPI | 2486 | #ifdef USE_NAPI |
2487 | cas_mask_intr(cp); | 2487 | cas_mask_intr(cp); |
2488 | netif_rx_schedule(dev); | 2488 | netif_rx_schedule(dev, &cp->napi); |
2489 | #else | 2489 | #else |
2490 | cas_rx_ringN(cp, ring, 0); | 2490 | cas_rx_ringN(cp, ring, 0); |
2491 | #endif | 2491 | #endif |
@@ -2536,7 +2536,7 @@ static irqreturn_t cas_interrupt1(int irq, void *dev_id) | |||
2536 | if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ | 2536 | if (status & INTR_RX_DONE_ALT) { /* handle rx separately */ |
2537 | #ifdef USE_NAPI | 2537 | #ifdef USE_NAPI |
2538 | cas_mask_intr(cp); | 2538 | cas_mask_intr(cp); |
2539 | netif_rx_schedule(dev); | 2539 | netif_rx_schedule(dev, &cp->napi); |
2540 | #else | 2540 | #else |
2541 | cas_rx_ringN(cp, 1, 0); | 2541 | cas_rx_ringN(cp, 1, 0); |
2542 | #endif | 2542 | #endif |
@@ -2592,7 +2592,7 @@ static irqreturn_t cas_interrupt(int irq, void *dev_id) | |||
2592 | if (status & INTR_RX_DONE) { | 2592 | if (status & INTR_RX_DONE) { |
2593 | #ifdef USE_NAPI | 2593 | #ifdef USE_NAPI |
2594 | cas_mask_intr(cp); | 2594 | cas_mask_intr(cp); |
2595 | netif_rx_schedule(dev); | 2595 | netif_rx_schedule(dev, &cp->napi); |
2596 | #else | 2596 | #else |
2597 | cas_rx_ringN(cp, 0, 0); | 2597 | cas_rx_ringN(cp, 0, 0); |
2598 | #endif | 2598 | #endif |
@@ -2607,9 +2607,10 @@ static irqreturn_t cas_interrupt(int irq, void *dev_id) | |||
2607 | 2607 | ||
2608 | 2608 | ||
2609 | #ifdef USE_NAPI | 2609 | #ifdef USE_NAPI |
2610 | static int cas_poll(struct net_device *dev, int *budget) | 2610 | static int cas_poll(struct napi_struct *napi, int budget) |
2611 | { | 2611 | { |
2612 | struct cas *cp = netdev_priv(dev); | 2612 | struct cas *cp = container_of(napi, struct cas, napi); |
2613 | struct net_device *dev = cp->dev; | ||
2613 | int i, enable_intr, todo, credits; | 2614 | int i, enable_intr, todo, credits; |
2614 | u32 status = readl(cp->regs + REG_INTR_STATUS); | 2615 | u32 status = readl(cp->regs + REG_INTR_STATUS); |
2615 | unsigned long flags; | 2616 | unsigned long flags; |
@@ -2620,20 +2621,18 @@ static int cas_poll(struct net_device *dev, int *budget) | |||
2620 | 2621 | ||
2621 | /* NAPI rx packets. we spread the credits across all of the | 2622 | /* NAPI rx packets. we spread the credits across all of the |
2622 | * rxc rings | 2623 | * rxc rings |
2623 | */ | 2624 | * |
2624 | todo = min(*budget, dev->quota); | 2625 | * to make sure we're fair with the work we loop through each |
2625 | |||
2626 | /* to make sure we're fair with the work we loop through each | ||
2627 | * ring N_RX_COMP_RING times with a request of | 2626 | * ring N_RX_COMP_RING times with a request of |
2628 | * todo / N_RX_COMP_RINGS | 2627 | * budget / N_RX_COMP_RINGS |
2629 | */ | 2628 | */ |
2630 | enable_intr = 1; | 2629 | enable_intr = 1; |
2631 | credits = 0; | 2630 | credits = 0; |
2632 | for (i = 0; i < N_RX_COMP_RINGS; i++) { | 2631 | for (i = 0; i < N_RX_COMP_RINGS; i++) { |
2633 | int j; | 2632 | int j; |
2634 | for (j = 0; j < N_RX_COMP_RINGS; j++) { | 2633 | for (j = 0; j < N_RX_COMP_RINGS; j++) { |
2635 | credits += cas_rx_ringN(cp, j, todo / N_RX_COMP_RINGS); | 2634 | credits += cas_rx_ringN(cp, j, budget / N_RX_COMP_RINGS); |
2636 | if (credits >= todo) { | 2635 | if (credits >= budget) { |
2637 | enable_intr = 0; | 2636 | enable_intr = 0; |
2638 | goto rx_comp; | 2637 | goto rx_comp; |
2639 | } | 2638 | } |
@@ -2641,9 +2640,6 @@ static int cas_poll(struct net_device *dev, int *budget) | |||
2641 | } | 2640 | } |
2642 | 2641 | ||
2643 | rx_comp: | 2642 | rx_comp: |
2644 | *budget -= credits; | ||
2645 | dev->quota -= credits; | ||
2646 | |||
2647 | /* final rx completion */ | 2643 | /* final rx completion */ |
2648 | spin_lock_irqsave(&cp->lock, flags); | 2644 | spin_lock_irqsave(&cp->lock, flags); |
2649 | if (status) | 2645 | if (status) |
@@ -2674,11 +2670,10 @@ rx_comp: | |||
2674 | #endif | 2670 | #endif |
2675 | spin_unlock_irqrestore(&cp->lock, flags); | 2671 | spin_unlock_irqrestore(&cp->lock, flags); |
2676 | if (enable_intr) { | 2672 | if (enable_intr) { |
2677 | netif_rx_complete(dev); | 2673 | netif_rx_complete(dev, napi); |
2678 | cas_unmask_intr(cp); | 2674 | cas_unmask_intr(cp); |
2679 | return 0; | ||
2680 | } | 2675 | } |
2681 | return 1; | 2676 | return credits; |
2682 | } | 2677 | } |
2683 | #endif | 2678 | #endif |
2684 | 2679 | ||
@@ -4351,6 +4346,9 @@ static int cas_open(struct net_device *dev) | |||
4351 | goto err_spare; | 4346 | goto err_spare; |
4352 | } | 4347 | } |
4353 | 4348 | ||
4349 | #ifdef USE_NAPI | ||
4350 | napi_enable(&cp->napi); | ||
4351 | #endif | ||
4354 | /* init hw */ | 4352 | /* init hw */ |
4355 | cas_lock_all_save(cp, flags); | 4353 | cas_lock_all_save(cp, flags); |
4356 | cas_clean_rings(cp); | 4354 | cas_clean_rings(cp); |
@@ -4376,6 +4374,9 @@ static int cas_close(struct net_device *dev) | |||
4376 | unsigned long flags; | 4374 | unsigned long flags; |
4377 | struct cas *cp = netdev_priv(dev); | 4375 | struct cas *cp = netdev_priv(dev); |
4378 | 4376 | ||
4377 | #ifdef USE_NAPI | ||
4378 | napi_enable(&cp->napi); | ||
4379 | #endif | ||
4379 | /* Make sure we don't get distracted by suspend/resume */ | 4380 | /* Make sure we don't get distracted by suspend/resume */ |
4380 | mutex_lock(&cp->pm_mutex); | 4381 | mutex_lock(&cp->pm_mutex); |
4381 | 4382 | ||
@@ -5062,8 +5063,7 @@ static int __devinit cas_init_one(struct pci_dev *pdev, | |||
5062 | dev->watchdog_timeo = CAS_TX_TIMEOUT; | 5063 | dev->watchdog_timeo = CAS_TX_TIMEOUT; |
5063 | dev->change_mtu = cas_change_mtu; | 5064 | dev->change_mtu = cas_change_mtu; |
5064 | #ifdef USE_NAPI | 5065 | #ifdef USE_NAPI |
5065 | dev->poll = cas_poll; | 5066 | netif_napi_add(dev, &cp->napi, cas_poll, 64); |
5066 | dev->weight = 64; | ||
5067 | #endif | 5067 | #endif |
5068 | #ifdef CONFIG_NET_POLL_CONTROLLER | 5068 | #ifdef CONFIG_NET_POLL_CONTROLLER |
5069 | dev->poll_controller = cas_netpoll; | 5069 | dev->poll_controller = cas_netpoll; |
diff --git a/drivers/net/cassini.h b/drivers/net/cassini.h index a970804487c7..2f93f83342d2 100644 --- a/drivers/net/cassini.h +++ b/drivers/net/cassini.h | |||
@@ -4280,6 +4280,8 @@ struct cas { | |||
4280 | int rx_cur[N_RX_COMP_RINGS], rx_new[N_RX_COMP_RINGS]; | 4280 | int rx_cur[N_RX_COMP_RINGS], rx_new[N_RX_COMP_RINGS]; |
4281 | int rx_last[N_RX_DESC_RINGS]; | 4281 | int rx_last[N_RX_DESC_RINGS]; |
4282 | 4282 | ||
4283 | struct napi_struct napi; | ||
4284 | |||
4283 | /* Set when chip is actually in operational state | 4285 | /* Set when chip is actually in operational state |
4284 | * (ie. not power managed) */ | 4286 | * (ie. not power managed) */ |
4285 | int hw_running; | 4287 | int hw_running; |
diff --git a/drivers/net/chelsio/common.h b/drivers/net/chelsio/common.h index 8ba702c8b560..b5de4452cf24 100644 --- a/drivers/net/chelsio/common.h +++ b/drivers/net/chelsio/common.h | |||
@@ -278,6 +278,7 @@ struct adapter { | |||
278 | struct peespi *espi; | 278 | struct peespi *espi; |
279 | struct petp *tp; | 279 | struct petp *tp; |
280 | 280 | ||
281 | struct napi_struct napi; | ||
281 | struct port_info port[MAX_NPORTS]; | 282 | struct port_info port[MAX_NPORTS]; |
282 | struct delayed_work stats_update_task; | 283 | struct delayed_work stats_update_task; |
283 | struct timer_list stats_update_timer; | 284 | struct timer_list stats_update_timer; |
diff --git a/drivers/net/chelsio/cxgb2.c b/drivers/net/chelsio/cxgb2.c index 231ce43b97cf..593736c7550d 100644 --- a/drivers/net/chelsio/cxgb2.c +++ b/drivers/net/chelsio/cxgb2.c | |||
@@ -255,8 +255,11 @@ static int cxgb_open(struct net_device *dev) | |||
255 | struct adapter *adapter = dev->priv; | 255 | struct adapter *adapter = dev->priv; |
256 | int other_ports = adapter->open_device_map & PORT_MASK; | 256 | int other_ports = adapter->open_device_map & PORT_MASK; |
257 | 257 | ||
258 | if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) | 258 | napi_enable(&adapter->napi); |
259 | if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) { | ||
260 | napi_disable(&adapter->napi); | ||
259 | return err; | 261 | return err; |
262 | } | ||
260 | 263 | ||
261 | __set_bit(dev->if_port, &adapter->open_device_map); | 264 | __set_bit(dev->if_port, &adapter->open_device_map); |
262 | link_start(&adapter->port[dev->if_port]); | 265 | link_start(&adapter->port[dev->if_port]); |
@@ -274,6 +277,7 @@ static int cxgb_close(struct net_device *dev) | |||
274 | struct cmac *mac = p->mac; | 277 | struct cmac *mac = p->mac; |
275 | 278 | ||
276 | netif_stop_queue(dev); | 279 | netif_stop_queue(dev); |
280 | napi_disable(&adapter->napi); | ||
277 | mac->ops->disable(mac, MAC_DIRECTION_TX | MAC_DIRECTION_RX); | 281 | mac->ops->disable(mac, MAC_DIRECTION_TX | MAC_DIRECTION_RX); |
278 | netif_carrier_off(dev); | 282 | netif_carrier_off(dev); |
279 | 283 | ||
@@ -1113,8 +1117,7 @@ static int __devinit init_one(struct pci_dev *pdev, | |||
1113 | netdev->poll_controller = t1_netpoll; | 1117 | netdev->poll_controller = t1_netpoll; |
1114 | #endif | 1118 | #endif |
1115 | #ifdef CONFIG_CHELSIO_T1_NAPI | 1119 | #ifdef CONFIG_CHELSIO_T1_NAPI |
1116 | netdev->weight = 64; | 1120 | netif_napi_add(netdev, &adapter->napi, t1_poll, 64); |
1117 | netdev->poll = t1_poll; | ||
1118 | #endif | 1121 | #endif |
1119 | 1122 | ||
1120 | SET_ETHTOOL_OPS(netdev, &t1_ethtool_ops); | 1123 | SET_ETHTOOL_OPS(netdev, &t1_ethtool_ops); |
diff --git a/drivers/net/chelsio/sge.c b/drivers/net/chelsio/sge.c index e4f874a70fe5..ffa7e649a6ef 100644 --- a/drivers/net/chelsio/sge.c +++ b/drivers/net/chelsio/sge.c | |||
@@ -1620,23 +1620,20 @@ static int process_pure_responses(struct adapter *adapter) | |||
1620 | * or protection from interrupts as data interrupts are off at this point and | 1620 | * or protection from interrupts as data interrupts are off at this point and |
1621 | * other adapter interrupts do not interfere. | 1621 | * other adapter interrupts do not interfere. |
1622 | */ | 1622 | */ |
1623 | int t1_poll(struct net_device *dev, int *budget) | 1623 | int t1_poll(struct napi_struct *napi, int budget) |
1624 | { | 1624 | { |
1625 | struct adapter *adapter = dev->priv; | 1625 | struct adapter *adapter = container_of(napi, struct adapter, napi); |
1626 | struct net_device *dev = adapter->port[0].dev; | ||
1626 | int work_done; | 1627 | int work_done; |
1627 | 1628 | ||
1628 | work_done = process_responses(adapter, min(*budget, dev->quota)); | 1629 | work_done = process_responses(adapter, budget); |
1629 | *budget -= work_done; | ||
1630 | dev->quota -= work_done; | ||
1631 | |||
1632 | if (unlikely(responses_pending(adapter))) | ||
1633 | return 1; | ||
1634 | |||
1635 | netif_rx_complete(dev); | ||
1636 | writel(adapter->sge->respQ.cidx, adapter->regs + A_SG_SLEEPING); | ||
1637 | |||
1638 | return 0; | ||
1639 | 1630 | ||
1631 | if (likely(!responses_pending(adapter))) { | ||
1632 | netif_rx_complete(dev, napi); | ||
1633 | writel(adapter->sge->respQ.cidx, | ||
1634 | adapter->regs + A_SG_SLEEPING); | ||
1635 | } | ||
1636 | return work_done; | ||
1640 | } | 1637 | } |
1641 | 1638 | ||
1642 | /* | 1639 | /* |
@@ -1653,13 +1650,13 @@ irqreturn_t t1_interrupt(int irq, void *data) | |||
1653 | 1650 | ||
1654 | writel(F_PL_INTR_SGE_DATA, adapter->regs + A_PL_CAUSE); | 1651 | writel(F_PL_INTR_SGE_DATA, adapter->regs + A_PL_CAUSE); |
1655 | 1652 | ||
1656 | if (__netif_rx_schedule_prep(dev)) { | 1653 | if (napi_schedule_prep(&adapter->napi)) { |
1657 | if (process_pure_responses(adapter)) | 1654 | if (process_pure_responses(adapter)) |
1658 | __netif_rx_schedule(dev); | 1655 | __netif_rx_schedule(dev, &adapter->napi); |
1659 | else { | 1656 | else { |
1660 | /* no data, no NAPI needed */ | 1657 | /* no data, no NAPI needed */ |
1661 | writel(sge->respQ.cidx, adapter->regs + A_SG_SLEEPING); | 1658 | writel(sge->respQ.cidx, adapter->regs + A_SG_SLEEPING); |
1662 | netif_poll_enable(dev); /* undo schedule_prep */ | 1659 | napi_enable(&adapter->napi); /* undo schedule_prep */ |
1663 | } | 1660 | } |
1664 | } | 1661 | } |
1665 | return IRQ_HANDLED; | 1662 | return IRQ_HANDLED; |
diff --git a/drivers/net/chelsio/sge.h b/drivers/net/chelsio/sge.h index d132a0ef2a22..713d9c55f24d 100644 --- a/drivers/net/chelsio/sge.h +++ b/drivers/net/chelsio/sge.h | |||
@@ -77,7 +77,7 @@ int t1_sge_configure(struct sge *, struct sge_params *); | |||
77 | int t1_sge_set_coalesce_params(struct sge *, struct sge_params *); | 77 | int t1_sge_set_coalesce_params(struct sge *, struct sge_params *); |
78 | void t1_sge_destroy(struct sge *); | 78 | void t1_sge_destroy(struct sge *); |
79 | irqreturn_t t1_interrupt(int irq, void *cookie); | 79 | irqreturn_t t1_interrupt(int irq, void *cookie); |
80 | int t1_poll(struct net_device *, int *); | 80 | int t1_poll(struct napi_struct *, int); |
81 | 81 | ||
82 | int t1_start_xmit(struct sk_buff *skb, struct net_device *dev); | 82 | int t1_start_xmit(struct sk_buff *skb, struct net_device *dev); |
83 | void t1_set_vlan_accel(struct adapter *adapter, int on_off); | 83 | void t1_set_vlan_accel(struct adapter *adapter, int on_off); |
diff --git a/drivers/net/cxgb3/adapter.h b/drivers/net/cxgb3/adapter.h index 20e887de2545..044261703381 100644 --- a/drivers/net/cxgb3/adapter.h +++ b/drivers/net/cxgb3/adapter.h | |||
@@ -49,11 +49,13 @@ | |||
49 | typedef irqreturn_t(*intr_handler_t) (int, void *); | 49 | typedef irqreturn_t(*intr_handler_t) (int, void *); |
50 | 50 | ||
51 | struct vlan_group; | 51 | struct vlan_group; |
52 | |||
53 | struct adapter; | 52 | struct adapter; |
53 | struct sge_qset; | ||
54 | |||
54 | struct port_info { | 55 | struct port_info { |
55 | struct adapter *adapter; | 56 | struct adapter *adapter; |
56 | struct vlan_group *vlan_grp; | 57 | struct vlan_group *vlan_grp; |
58 | struct sge_qset *qs; | ||
57 | const struct port_type_info *port_type; | 59 | const struct port_type_info *port_type; |
58 | u8 port_id; | 60 | u8 port_id; |
59 | u8 rx_csum_offload; | 61 | u8 rx_csum_offload; |
@@ -173,10 +175,12 @@ enum { /* per port SGE statistics */ | |||
173 | }; | 175 | }; |
174 | 176 | ||
175 | struct sge_qset { /* an SGE queue set */ | 177 | struct sge_qset { /* an SGE queue set */ |
178 | struct adapter *adap; | ||
179 | struct napi_struct napi; | ||
176 | struct sge_rspq rspq; | 180 | struct sge_rspq rspq; |
177 | struct sge_fl fl[SGE_RXQ_PER_SET]; | 181 | struct sge_fl fl[SGE_RXQ_PER_SET]; |
178 | struct sge_txq txq[SGE_TXQ_PER_SET]; | 182 | struct sge_txq txq[SGE_TXQ_PER_SET]; |
179 | struct net_device *netdev; /* associated net device */ | 183 | struct net_device *netdev; |
180 | unsigned long txq_stopped; /* which Tx queues are stopped */ | 184 | unsigned long txq_stopped; /* which Tx queues are stopped */ |
181 | struct timer_list tx_reclaim_timer; /* reclaims TX buffers */ | 185 | struct timer_list tx_reclaim_timer; /* reclaims TX buffers */ |
182 | unsigned long port_stats[SGE_PSTAT_MAX]; | 186 | unsigned long port_stats[SGE_PSTAT_MAX]; |
@@ -221,12 +225,6 @@ struct adapter { | |||
221 | struct delayed_work adap_check_task; | 225 | struct delayed_work adap_check_task; |
222 | struct work_struct ext_intr_handler_task; | 226 | struct work_struct ext_intr_handler_task; |
223 | 227 | ||
224 | /* | ||
225 | * Dummy netdevices are needed when using multiple receive queues with | ||
226 | * NAPI as each netdevice can service only one queue. | ||
227 | */ | ||
228 | struct net_device *dummy_netdev[SGE_QSETS - 1]; | ||
229 | |||
230 | struct dentry *debugfs_root; | 228 | struct dentry *debugfs_root; |
231 | 229 | ||
232 | struct mutex mdio_lock; | 230 | struct mutex mdio_lock; |
@@ -253,12 +251,6 @@ static inline struct port_info *adap2pinfo(struct adapter *adap, int idx) | |||
253 | return netdev_priv(adap->port[idx]); | 251 | return netdev_priv(adap->port[idx]); |
254 | } | 252 | } |
255 | 253 | ||
256 | /* | ||
257 | * We use the spare atalk_ptr to map a net device to its SGE queue set. | ||
258 | * This is a macro so it can be used as l-value. | ||
259 | */ | ||
260 | #define dev2qset(netdev) ((netdev)->atalk_ptr) | ||
261 | |||
262 | #define OFFLOAD_DEVMAP_BIT 15 | 254 | #define OFFLOAD_DEVMAP_BIT 15 |
263 | 255 | ||
264 | #define tdev2adap(d) container_of(d, struct adapter, tdev) | 256 | #define tdev2adap(d) container_of(d, struct adapter, tdev) |
@@ -284,7 +276,7 @@ int t3_mgmt_tx(struct adapter *adap, struct sk_buff *skb); | |||
284 | void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p); | 276 | void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p); |
285 | int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports, | 277 | int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports, |
286 | int irq_vec_idx, const struct qset_params *p, | 278 | int irq_vec_idx, const struct qset_params *p, |
287 | int ntxq, struct net_device *netdev); | 279 | int ntxq, struct net_device *dev); |
288 | int t3_get_desc(const struct sge_qset *qs, unsigned int qnum, unsigned int idx, | 280 | int t3_get_desc(const struct sge_qset *qs, unsigned int qnum, unsigned int idx, |
289 | unsigned char *data); | 281 | unsigned char *data); |
290 | irqreturn_t t3_sge_intr_msix(int irq, void *cookie); | 282 | irqreturn_t t3_sge_intr_msix(int irq, void *cookie); |
diff --git a/drivers/net/cxgb3/cxgb3_main.c b/drivers/net/cxgb3/cxgb3_main.c index 5ab319cfe5de..5db7d4e27ec0 100644 --- a/drivers/net/cxgb3/cxgb3_main.c +++ b/drivers/net/cxgb3/cxgb3_main.c | |||
@@ -339,49 +339,17 @@ static void setup_rss(struct adapter *adap) | |||
339 | V_RRCPLCPUSIZE(6), cpus, rspq_map); | 339 | V_RRCPLCPUSIZE(6), cpus, rspq_map); |
340 | } | 340 | } |
341 | 341 | ||
342 | /* | 342 | static void init_napi(struct adapter *adap) |
343 | * If we have multiple receive queues per port serviced by NAPI we need one | ||
344 | * netdevice per queue as NAPI operates on netdevices. We already have one | ||
345 | * netdevice, namely the one associated with the interface, so we use dummy | ||
346 | * ones for any additional queues. Note that these netdevices exist purely | ||
347 | * so that NAPI has something to work with, they do not represent network | ||
348 | * ports and are not registered. | ||
349 | */ | ||
350 | static int init_dummy_netdevs(struct adapter *adap) | ||
351 | { | 343 | { |
352 | int i, j, dummy_idx = 0; | 344 | int i; |
353 | struct net_device *nd; | ||
354 | |||
355 | for_each_port(adap, i) { | ||
356 | struct net_device *dev = adap->port[i]; | ||
357 | const struct port_info *pi = netdev_priv(dev); | ||
358 | |||
359 | for (j = 0; j < pi->nqsets - 1; j++) { | ||
360 | if (!adap->dummy_netdev[dummy_idx]) { | ||
361 | struct port_info *p; | ||
362 | |||
363 | nd = alloc_netdev(sizeof(*p), "", ether_setup); | ||
364 | if (!nd) | ||
365 | goto free_all; | ||
366 | 345 | ||
367 | p = netdev_priv(nd); | 346 | for (i = 0; i < SGE_QSETS; i++) { |
368 | p->adapter = adap; | 347 | struct sge_qset *qs = &adap->sge.qs[i]; |
369 | nd->weight = 64; | ||
370 | set_bit(__LINK_STATE_START, &nd->state); | ||
371 | adap->dummy_netdev[dummy_idx] = nd; | ||
372 | } | ||
373 | strcpy(adap->dummy_netdev[dummy_idx]->name, dev->name); | ||
374 | dummy_idx++; | ||
375 | } | ||
376 | } | ||
377 | return 0; | ||
378 | 348 | ||
379 | free_all: | 349 | if (qs->adap) |
380 | while (--dummy_idx >= 0) { | 350 | netif_napi_add(qs->netdev, &qs->napi, qs->napi.poll, |
381 | free_netdev(adap->dummy_netdev[dummy_idx]); | 351 | 64); |
382 | adap->dummy_netdev[dummy_idx] = NULL; | ||
383 | } | 352 | } |
384 | return -ENOMEM; | ||
385 | } | 353 | } |
386 | 354 | ||
387 | /* | 355 | /* |
@@ -392,20 +360,18 @@ free_all: | |||
392 | static void quiesce_rx(struct adapter *adap) | 360 | static void quiesce_rx(struct adapter *adap) |
393 | { | 361 | { |
394 | int i; | 362 | int i; |
395 | struct net_device *dev; | ||
396 | 363 | ||
397 | for_each_port(adap, i) { | 364 | for (i = 0; i < SGE_QSETS; i++) |
398 | dev = adap->port[i]; | 365 | if (adap->sge.qs[i].adap) |
399 | while (test_bit(__LINK_STATE_RX_SCHED, &dev->state)) | 366 | napi_disable(&adap->sge.qs[i].napi); |
400 | msleep(1); | 367 | } |
401 | } | ||
402 | 368 | ||
403 | for (i = 0; i < ARRAY_SIZE(adap->dummy_netdev); i++) { | 369 | static void enable_all_napi(struct adapter *adap) |
404 | dev = adap->dummy_netdev[i]; | 370 | { |
405 | if (dev) | 371 | int i; |
406 | while (test_bit(__LINK_STATE_RX_SCHED, &dev->state)) | 372 | for (i = 0; i < SGE_QSETS; i++) |
407 | msleep(1); | 373 | if (adap->sge.qs[i].adap) |
408 | } | 374 | napi_enable(&adap->sge.qs[i].napi); |
409 | } | 375 | } |
410 | 376 | ||
411 | /** | 377 | /** |
@@ -418,7 +384,7 @@ static void quiesce_rx(struct adapter *adap) | |||
418 | */ | 384 | */ |
419 | static int setup_sge_qsets(struct adapter *adap) | 385 | static int setup_sge_qsets(struct adapter *adap) |
420 | { | 386 | { |
421 | int i, j, err, irq_idx = 0, qset_idx = 0, dummy_dev_idx = 0; | 387 | int i, j, err, irq_idx = 0, qset_idx = 0; |
422 | unsigned int ntxq = SGE_TXQ_PER_SET; | 388 | unsigned int ntxq = SGE_TXQ_PER_SET; |
423 | 389 | ||
424 | if (adap->params.rev > 0 && !(adap->flags & USING_MSI)) | 390 | if (adap->params.rev > 0 && !(adap->flags & USING_MSI)) |
@@ -426,15 +392,14 @@ static int setup_sge_qsets(struct adapter *adap) | |||
426 | 392 | ||
427 | for_each_port(adap, i) { | 393 | for_each_port(adap, i) { |
428 | struct net_device *dev = adap->port[i]; | 394 | struct net_device *dev = adap->port[i]; |
429 | const struct port_info *pi = netdev_priv(dev); | 395 | struct port_info *pi = netdev_priv(dev); |
430 | 396 | ||
397 | pi->qs = &adap->sge.qs[pi->first_qset]; | ||
431 | for (j = 0; j < pi->nqsets; ++j, ++qset_idx) { | 398 | for (j = 0; j < pi->nqsets; ++j, ++qset_idx) { |
432 | err = t3_sge_alloc_qset(adap, qset_idx, 1, | 399 | err = t3_sge_alloc_qset(adap, qset_idx, 1, |
433 | (adap->flags & USING_MSIX) ? qset_idx + 1 : | 400 | (adap->flags & USING_MSIX) ? qset_idx + 1 : |
434 | irq_idx, | 401 | irq_idx, |
435 | &adap->params.sge.qset[qset_idx], ntxq, | 402 | &adap->params.sge.qset[qset_idx], ntxq, dev); |
436 | j == 0 ? dev : | ||
437 | adap-> dummy_netdev[dummy_dev_idx++]); | ||
438 | if (err) { | 403 | if (err) { |
439 | t3_free_sge_resources(adap); | 404 | t3_free_sge_resources(adap); |
440 | return err; | 405 | return err; |
@@ -845,21 +810,18 @@ static int cxgb_up(struct adapter *adap) | |||
845 | goto out; | 810 | goto out; |
846 | } | 811 | } |
847 | 812 | ||
848 | err = init_dummy_netdevs(adap); | ||
849 | if (err) | ||
850 | goto out; | ||
851 | |||
852 | err = t3_init_hw(adap, 0); | 813 | err = t3_init_hw(adap, 0); |
853 | if (err) | 814 | if (err) |
854 | goto out; | 815 | goto out; |
855 | 816 | ||
856 | t3_write_reg(adap, A_ULPRX_TDDP_PSZ, V_HPZ0(PAGE_SHIFT - 12)); | 817 | t3_write_reg(adap, A_ULPRX_TDDP_PSZ, V_HPZ0(PAGE_SHIFT - 12)); |
857 | 818 | ||
858 | err = setup_sge_qsets(adap); | 819 | err = setup_sge_qsets(adap); |
859 | if (err) | 820 | if (err) |
860 | goto out; | 821 | goto out; |
861 | 822 | ||
862 | setup_rss(adap); | 823 | setup_rss(adap); |
824 | init_napi(adap); | ||
863 | adap->flags |= FULL_INIT_DONE; | 825 | adap->flags |= FULL_INIT_DONE; |
864 | } | 826 | } |
865 | 827 | ||
@@ -886,6 +848,7 @@ static int cxgb_up(struct adapter *adap) | |||
886 | adap->name, adap))) | 848 | adap->name, adap))) |
887 | goto irq_err; | 849 | goto irq_err; |
888 | 850 | ||
851 | enable_all_napi(adap); | ||
889 | t3_sge_start(adap); | 852 | t3_sge_start(adap); |
890 | t3_intr_enable(adap); | 853 | t3_intr_enable(adap); |
891 | 854 | ||
@@ -1012,8 +975,10 @@ static int cxgb_open(struct net_device *dev) | |||
1012 | int other_ports = adapter->open_device_map & PORT_MASK; | 975 | int other_ports = adapter->open_device_map & PORT_MASK; |
1013 | int err; | 976 | int err; |
1014 | 977 | ||
1015 | if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) | 978 | if (!adapter->open_device_map && (err = cxgb_up(adapter)) < 0) { |
979 | quiesce_rx(adapter); | ||
1016 | return err; | 980 | return err; |
981 | } | ||
1017 | 982 | ||
1018 | set_bit(pi->port_id, &adapter->open_device_map); | 983 | set_bit(pi->port_id, &adapter->open_device_map); |
1019 | if (is_offload(adapter) && !ofld_disable) { | 984 | if (is_offload(adapter) && !ofld_disable) { |
@@ -2524,7 +2489,6 @@ static int __devinit init_one(struct pci_dev *pdev, | |||
2524 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2489 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2525 | netdev->poll_controller = cxgb_netpoll; | 2490 | netdev->poll_controller = cxgb_netpoll; |
2526 | #endif | 2491 | #endif |
2527 | netdev->weight = 64; | ||
2528 | 2492 | ||
2529 | SET_ETHTOOL_OPS(netdev, &cxgb_ethtool_ops); | 2493 | SET_ETHTOOL_OPS(netdev, &cxgb_ethtool_ops); |
2530 | } | 2494 | } |
@@ -2625,12 +2589,6 @@ static void __devexit remove_one(struct pci_dev *pdev) | |||
2625 | t3_free_sge_resources(adapter); | 2589 | t3_free_sge_resources(adapter); |
2626 | cxgb_disable_msi(adapter); | 2590 | cxgb_disable_msi(adapter); |
2627 | 2591 | ||
2628 | for (i = 0; i < ARRAY_SIZE(adapter->dummy_netdev); i++) | ||
2629 | if (adapter->dummy_netdev[i]) { | ||
2630 | free_netdev(adapter->dummy_netdev[i]); | ||
2631 | adapter->dummy_netdev[i] = NULL; | ||
2632 | } | ||
2633 | |||
2634 | for_each_port(adapter, i) | 2592 | for_each_port(adapter, i) |
2635 | if (adapter->port[i]) | 2593 | if (adapter->port[i]) |
2636 | free_netdev(adapter->port[i]); | 2594 | free_netdev(adapter->port[i]); |
diff --git a/drivers/net/cxgb3/sge.c b/drivers/net/cxgb3/sge.c index 58a5f60521ed..069c1aca8a6b 100644 --- a/drivers/net/cxgb3/sge.c +++ b/drivers/net/cxgb3/sge.c | |||
@@ -591,9 +591,6 @@ void t3_free_qset(struct adapter *adapter, struct sge_qset *q) | |||
591 | q->rspq.desc, q->rspq.phys_addr); | 591 | q->rspq.desc, q->rspq.phys_addr); |
592 | } | 592 | } |
593 | 593 | ||
594 | if (q->netdev) | ||
595 | q->netdev->atalk_ptr = NULL; | ||
596 | |||
597 | memset(q, 0, sizeof(*q)); | 594 | memset(q, 0, sizeof(*q)); |
598 | } | 595 | } |
599 | 596 | ||
@@ -1074,7 +1071,7 @@ int t3_eth_xmit(struct sk_buff *skb, struct net_device *dev) | |||
1074 | unsigned int ndesc, pidx, credits, gen, compl; | 1071 | unsigned int ndesc, pidx, credits, gen, compl; |
1075 | const struct port_info *pi = netdev_priv(dev); | 1072 | const struct port_info *pi = netdev_priv(dev); |
1076 | struct adapter *adap = pi->adapter; | 1073 | struct adapter *adap = pi->adapter; |
1077 | struct sge_qset *qs = dev2qset(dev); | 1074 | struct sge_qset *qs = pi->qs; |
1078 | struct sge_txq *q = &qs->txq[TXQ_ETH]; | 1075 | struct sge_txq *q = &qs->txq[TXQ_ETH]; |
1079 | 1076 | ||
1080 | /* | 1077 | /* |
@@ -1326,13 +1323,12 @@ static void restart_ctrlq(unsigned long data) | |||
1326 | struct sk_buff *skb; | 1323 | struct sk_buff *skb; |
1327 | struct sge_qset *qs = (struct sge_qset *)data; | 1324 | struct sge_qset *qs = (struct sge_qset *)data; |
1328 | struct sge_txq *q = &qs->txq[TXQ_CTRL]; | 1325 | struct sge_txq *q = &qs->txq[TXQ_CTRL]; |
1329 | const struct port_info *pi = netdev_priv(qs->netdev); | ||
1330 | struct adapter *adap = pi->adapter; | ||
1331 | 1326 | ||
1332 | spin_lock(&q->lock); | 1327 | spin_lock(&q->lock); |
1333 | again:reclaim_completed_tx_imm(q); | 1328 | again:reclaim_completed_tx_imm(q); |
1334 | 1329 | ||
1335 | while (q->in_use < q->size && (skb = __skb_dequeue(&q->sendq)) != NULL) { | 1330 | while (q->in_use < q->size && |
1331 | (skb = __skb_dequeue(&q->sendq)) != NULL) { | ||
1336 | 1332 | ||
1337 | write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); | 1333 | write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); |
1338 | 1334 | ||
@@ -1354,7 +1350,7 @@ static void restart_ctrlq(unsigned long data) | |||
1354 | } | 1350 | } |
1355 | 1351 | ||
1356 | spin_unlock(&q->lock); | 1352 | spin_unlock(&q->lock); |
1357 | t3_write_reg(adap, A_SG_KDOORBELL, | 1353 | t3_write_reg(qs->adap, A_SG_KDOORBELL, |
1358 | F_SELEGRCNTX | V_EGRCNTX(q->cntxt_id)); | 1354 | F_SELEGRCNTX | V_EGRCNTX(q->cntxt_id)); |
1359 | } | 1355 | } |
1360 | 1356 | ||
@@ -1638,8 +1634,7 @@ static inline void offload_enqueue(struct sge_rspq *q, struct sk_buff *skb) | |||
1638 | else { | 1634 | else { |
1639 | struct sge_qset *qs = rspq_to_qset(q); | 1635 | struct sge_qset *qs = rspq_to_qset(q); |
1640 | 1636 | ||
1641 | if (__netif_rx_schedule_prep(qs->netdev)) | 1637 | napi_schedule(&qs->napi); |
1642 | __netif_rx_schedule(qs->netdev); | ||
1643 | q->rx_head = skb; | 1638 | q->rx_head = skb; |
1644 | } | 1639 | } |
1645 | q->rx_tail = skb; | 1640 | q->rx_tail = skb; |
@@ -1675,34 +1670,30 @@ static inline void deliver_partial_bundle(struct t3cdev *tdev, | |||
1675 | * receive handler. Batches need to be of modest size as we do prefetches | 1670 | * receive handler. Batches need to be of modest size as we do prefetches |
1676 | * on the packets in each. | 1671 | * on the packets in each. |
1677 | */ | 1672 | */ |
1678 | static int ofld_poll(struct net_device *dev, int *budget) | 1673 | static int ofld_poll(struct napi_struct *napi, int budget) |
1679 | { | 1674 | { |
1680 | const struct port_info *pi = netdev_priv(dev); | 1675 | struct sge_qset *qs = container_of(napi, struct sge_qset, napi); |
1681 | struct adapter *adapter = pi->adapter; | ||
1682 | struct sge_qset *qs = dev2qset(dev); | ||
1683 | struct sge_rspq *q = &qs->rspq; | 1676 | struct sge_rspq *q = &qs->rspq; |
1684 | int work_done, limit = min(*budget, dev->quota), avail = limit; | 1677 | struct adapter *adapter = qs->adap; |
1678 | int work_done = 0; | ||
1685 | 1679 | ||
1686 | while (avail) { | 1680 | while (work_done < budget) { |
1687 | struct sk_buff *head, *tail, *skbs[RX_BUNDLE_SIZE]; | 1681 | struct sk_buff *head, *tail, *skbs[RX_BUNDLE_SIZE]; |
1688 | int ngathered; | 1682 | int ngathered; |
1689 | 1683 | ||
1690 | spin_lock_irq(&q->lock); | 1684 | spin_lock_irq(&q->lock); |
1691 | head = q->rx_head; | 1685 | head = q->rx_head; |
1692 | if (!head) { | 1686 | if (!head) { |
1693 | work_done = limit - avail; | 1687 | napi_complete(napi); |
1694 | *budget -= work_done; | ||
1695 | dev->quota -= work_done; | ||
1696 | __netif_rx_complete(dev); | ||
1697 | spin_unlock_irq(&q->lock); | 1688 | spin_unlock_irq(&q->lock); |
1698 | return 0; | 1689 | return work_done; |
1699 | } | 1690 | } |
1700 | 1691 | ||
1701 | tail = q->rx_tail; | 1692 | tail = q->rx_tail; |
1702 | q->rx_head = q->rx_tail = NULL; | 1693 | q->rx_head = q->rx_tail = NULL; |
1703 | spin_unlock_irq(&q->lock); | 1694 | spin_unlock_irq(&q->lock); |
1704 | 1695 | ||
1705 | for (ngathered = 0; avail && head; avail--) { | 1696 | for (ngathered = 0; work_done < budget && head; work_done++) { |
1706 | prefetch(head->data); | 1697 | prefetch(head->data); |
1707 | skbs[ngathered] = head; | 1698 | skbs[ngathered] = head; |
1708 | head = head->next; | 1699 | head = head->next; |
@@ -1724,10 +1715,8 @@ static int ofld_poll(struct net_device *dev, int *budget) | |||
1724 | } | 1715 | } |
1725 | deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); | 1716 | deliver_partial_bundle(&adapter->tdev, q, skbs, ngathered); |
1726 | } | 1717 | } |
1727 | work_done = limit - avail; | 1718 | |
1728 | *budget -= work_done; | 1719 | return work_done; |
1729 | dev->quota -= work_done; | ||
1730 | return 1; | ||
1731 | } | 1720 | } |
1732 | 1721 | ||
1733 | /** | 1722 | /** |
@@ -2071,50 +2060,47 @@ static inline int is_pure_response(const struct rsp_desc *r) | |||
2071 | 2060 | ||
2072 | /** | 2061 | /** |
2073 | * napi_rx_handler - the NAPI handler for Rx processing | 2062 | * napi_rx_handler - the NAPI handler for Rx processing |
2074 | * @dev: the net device | 2063 | * @napi: the napi instance |
2075 | * @budget: how many packets we can process in this round | 2064 | * @budget: how many packets we can process in this round |
2076 | * | 2065 | * |
2077 | * Handler for new data events when using NAPI. | 2066 | * Handler for new data events when using NAPI. |
2078 | */ | 2067 | */ |
2079 | static int napi_rx_handler(struct net_device *dev, int *budget) | 2068 | static int napi_rx_handler(struct napi_struct *napi, int budget) |
2080 | { | 2069 | { |
2081 | const struct port_info *pi = netdev_priv(dev); | 2070 | struct sge_qset *qs = container_of(napi, struct sge_qset, napi); |
2082 | struct adapter *adap = pi->adapter; | 2071 | struct adapter *adap = qs->adap; |
2083 | struct sge_qset *qs = dev2qset(dev); | 2072 | int work_done = process_responses(adap, qs, budget); |
2084 | int effective_budget = min(*budget, dev->quota); | ||
2085 | |||
2086 | int work_done = process_responses(adap, qs, effective_budget); | ||
2087 | *budget -= work_done; | ||
2088 | dev->quota -= work_done; | ||
2089 | 2073 | ||
2090 | if (work_done >= effective_budget) | 2074 | if (likely(work_done < budget)) { |
2091 | return 1; | 2075 | napi_complete(napi); |
2092 | |||
2093 | netif_rx_complete(dev); | ||
2094 | 2076 | ||
2095 | /* | 2077 | /* |
2096 | * Because we don't atomically flush the following write it is | 2078 | * Because we don't atomically flush the following |
2097 | * possible that in very rare cases it can reach the device in a way | 2079 | * write it is possible that in very rare cases it can |
2098 | * that races with a new response being written plus an error interrupt | 2080 | * reach the device in a way that races with a new |
2099 | * causing the NAPI interrupt handler below to return unhandled status | 2081 | * response being written plus an error interrupt |
2100 | * to the OS. To protect against this would require flushing the write | 2082 | * causing the NAPI interrupt handler below to return |
2101 | * and doing both the write and the flush with interrupts off. Way too | 2083 | * unhandled status to the OS. To protect against |
2102 | * expensive and unjustifiable given the rarity of the race. | 2084 | * this would require flushing the write and doing |
2103 | * | 2085 | * both the write and the flush with interrupts off. |
2104 | * The race cannot happen at all with MSI-X. | 2086 | * Way too expensive and unjustifiable given the |
2105 | */ | 2087 | * rarity of the race. |
2106 | t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | | 2088 | * |
2107 | V_NEWTIMER(qs->rspq.next_holdoff) | | 2089 | * The race cannot happen at all with MSI-X. |
2108 | V_NEWINDEX(qs->rspq.cidx)); | 2090 | */ |
2109 | return 0; | 2091 | t3_write_reg(adap, A_SG_GTS, V_RSPQ(qs->rspq.cntxt_id) | |
2092 | V_NEWTIMER(qs->rspq.next_holdoff) | | ||
2093 | V_NEWINDEX(qs->rspq.cidx)); | ||
2094 | } | ||
2095 | return work_done; | ||
2110 | } | 2096 | } |
2111 | 2097 | ||
2112 | /* | 2098 | /* |
2113 | * Returns true if the device is already scheduled for polling. | 2099 | * Returns true if the device is already scheduled for polling. |
2114 | */ | 2100 | */ |
2115 | static inline int napi_is_scheduled(struct net_device *dev) | 2101 | static inline int napi_is_scheduled(struct napi_struct *napi) |
2116 | { | 2102 | { |
2117 | return test_bit(__LINK_STATE_RX_SCHED, &dev->state); | 2103 | return test_bit(NAPI_STATE_SCHED, &napi->state); |
2118 | } | 2104 | } |
2119 | 2105 | ||
2120 | /** | 2106 | /** |
@@ -2197,8 +2183,7 @@ static inline int handle_responses(struct adapter *adap, struct sge_rspq *q) | |||
2197 | V_NEWTIMER(q->holdoff_tmr) | V_NEWINDEX(q->cidx)); | 2183 | V_NEWTIMER(q->holdoff_tmr) | V_NEWINDEX(q->cidx)); |
2198 | return 0; | 2184 | return 0; |
2199 | } | 2185 | } |
2200 | if (likely(__netif_rx_schedule_prep(qs->netdev))) | 2186 | napi_schedule(&qs->napi); |
2201 | __netif_rx_schedule(qs->netdev); | ||
2202 | return 1; | 2187 | return 1; |
2203 | } | 2188 | } |
2204 | 2189 | ||
@@ -2209,8 +2194,7 @@ static inline int handle_responses(struct adapter *adap, struct sge_rspq *q) | |||
2209 | irqreturn_t t3_sge_intr_msix(int irq, void *cookie) | 2194 | irqreturn_t t3_sge_intr_msix(int irq, void *cookie) |
2210 | { | 2195 | { |
2211 | struct sge_qset *qs = cookie; | 2196 | struct sge_qset *qs = cookie; |
2212 | const struct port_info *pi = netdev_priv(qs->netdev); | 2197 | struct adapter *adap = qs->adap; |
2213 | struct adapter *adap = pi->adapter; | ||
2214 | struct sge_rspq *q = &qs->rspq; | 2198 | struct sge_rspq *q = &qs->rspq; |
2215 | 2199 | ||
2216 | spin_lock(&q->lock); | 2200 | spin_lock(&q->lock); |
@@ -2229,13 +2213,11 @@ irqreturn_t t3_sge_intr_msix(int irq, void *cookie) | |||
2229 | irqreturn_t t3_sge_intr_msix_napi(int irq, void *cookie) | 2213 | irqreturn_t t3_sge_intr_msix_napi(int irq, void *cookie) |
2230 | { | 2214 | { |
2231 | struct sge_qset *qs = cookie; | 2215 | struct sge_qset *qs = cookie; |
2232 | const struct port_info *pi = netdev_priv(qs->netdev); | ||
2233 | struct adapter *adap = pi->adapter; | ||
2234 | struct sge_rspq *q = &qs->rspq; | 2216 | struct sge_rspq *q = &qs->rspq; |
2235 | 2217 | ||
2236 | spin_lock(&q->lock); | 2218 | spin_lock(&q->lock); |
2237 | 2219 | ||
2238 | if (handle_responses(adap, q) < 0) | 2220 | if (handle_responses(qs->adap, q) < 0) |
2239 | q->unhandled_irqs++; | 2221 | q->unhandled_irqs++; |
2240 | spin_unlock(&q->lock); | 2222 | spin_unlock(&q->lock); |
2241 | return IRQ_HANDLED; | 2223 | return IRQ_HANDLED; |
@@ -2278,11 +2260,13 @@ static irqreturn_t t3_intr_msi(int irq, void *cookie) | |||
2278 | return IRQ_HANDLED; | 2260 | return IRQ_HANDLED; |
2279 | } | 2261 | } |
2280 | 2262 | ||
2281 | static int rspq_check_napi(struct net_device *dev, struct sge_rspq *q) | 2263 | static int rspq_check_napi(struct sge_qset *qs) |
2282 | { | 2264 | { |
2283 | if (!napi_is_scheduled(dev) && is_new_response(&q->desc[q->cidx], q)) { | 2265 | struct sge_rspq *q = &qs->rspq; |
2284 | if (likely(__netif_rx_schedule_prep(dev))) | 2266 | |
2285 | __netif_rx_schedule(dev); | 2267 | if (!napi_is_scheduled(&qs->napi) && |
2268 | is_new_response(&q->desc[q->cidx], q)) { | ||
2269 | napi_schedule(&qs->napi); | ||
2286 | return 1; | 2270 | return 1; |
2287 | } | 2271 | } |
2288 | return 0; | 2272 | return 0; |
@@ -2303,10 +2287,9 @@ irqreturn_t t3_intr_msi_napi(int irq, void *cookie) | |||
2303 | 2287 | ||
2304 | spin_lock(&q->lock); | 2288 | spin_lock(&q->lock); |
2305 | 2289 | ||
2306 | new_packets = rspq_check_napi(adap->sge.qs[0].netdev, q); | 2290 | new_packets = rspq_check_napi(&adap->sge.qs[0]); |
2307 | if (adap->params.nports == 2) | 2291 | if (adap->params.nports == 2) |
2308 | new_packets += rspq_check_napi(adap->sge.qs[1].netdev, | 2292 | new_packets += rspq_check_napi(&adap->sge.qs[1]); |
2309 | &adap->sge.qs[1].rspq); | ||
2310 | if (!new_packets && t3_slow_intr_handler(adap) == 0) | 2293 | if (!new_packets && t3_slow_intr_handler(adap) == 0) |
2311 | q->unhandled_irqs++; | 2294 | q->unhandled_irqs++; |
2312 | 2295 | ||
@@ -2409,9 +2392,9 @@ static irqreturn_t t3b_intr(int irq, void *cookie) | |||
2409 | static irqreturn_t t3b_intr_napi(int irq, void *cookie) | 2392 | static irqreturn_t t3b_intr_napi(int irq, void *cookie) |
2410 | { | 2393 | { |
2411 | u32 map; | 2394 | u32 map; |
2412 | struct net_device *dev; | ||
2413 | struct adapter *adap = cookie; | 2395 | struct adapter *adap = cookie; |
2414 | struct sge_rspq *q0 = &adap->sge.qs[0].rspq; | 2396 | struct sge_qset *qs0 = &adap->sge.qs[0]; |
2397 | struct sge_rspq *q0 = &qs0->rspq; | ||
2415 | 2398 | ||
2416 | t3_write_reg(adap, A_PL_CLI, 0); | 2399 | t3_write_reg(adap, A_PL_CLI, 0); |
2417 | map = t3_read_reg(adap, A_SG_DATA_INTR); | 2400 | map = t3_read_reg(adap, A_SG_DATA_INTR); |
@@ -2424,18 +2407,11 @@ static irqreturn_t t3b_intr_napi(int irq, void *cookie) | |||
2424 | if (unlikely(map & F_ERRINTR)) | 2407 | if (unlikely(map & F_ERRINTR)) |
2425 | t3_slow_intr_handler(adap); | 2408 | t3_slow_intr_handler(adap); |
2426 | 2409 | ||
2427 | if (likely(map & 1)) { | 2410 | if (likely(map & 1)) |
2428 | dev = adap->sge.qs[0].netdev; | 2411 | napi_schedule(&qs0->napi); |
2429 | |||
2430 | if (likely(__netif_rx_schedule_prep(dev))) | ||
2431 | __netif_rx_schedule(dev); | ||
2432 | } | ||
2433 | if (map & 2) { | ||
2434 | dev = adap->sge.qs[1].netdev; | ||
2435 | 2412 | ||
2436 | if (likely(__netif_rx_schedule_prep(dev))) | 2413 | if (map & 2) |
2437 | __netif_rx_schedule(dev); | 2414 | napi_schedule(&adap->sge.qs[1].napi); |
2438 | } | ||
2439 | 2415 | ||
2440 | spin_unlock(&q0->lock); | 2416 | spin_unlock(&q0->lock); |
2441 | return IRQ_HANDLED; | 2417 | return IRQ_HANDLED; |
@@ -2514,8 +2490,7 @@ static void sge_timer_cb(unsigned long data) | |||
2514 | { | 2490 | { |
2515 | spinlock_t *lock; | 2491 | spinlock_t *lock; |
2516 | struct sge_qset *qs = (struct sge_qset *)data; | 2492 | struct sge_qset *qs = (struct sge_qset *)data; |
2517 | const struct port_info *pi = netdev_priv(qs->netdev); | 2493 | struct adapter *adap = qs->adap; |
2518 | struct adapter *adap = pi->adapter; | ||
2519 | 2494 | ||
2520 | if (spin_trylock(&qs->txq[TXQ_ETH].lock)) { | 2495 | if (spin_trylock(&qs->txq[TXQ_ETH].lock)) { |
2521 | reclaim_completed_tx(adap, &qs->txq[TXQ_ETH]); | 2496 | reclaim_completed_tx(adap, &qs->txq[TXQ_ETH]); |
@@ -2526,9 +2501,9 @@ static void sge_timer_cb(unsigned long data) | |||
2526 | spin_unlock(&qs->txq[TXQ_OFLD].lock); | 2501 | spin_unlock(&qs->txq[TXQ_OFLD].lock); |
2527 | } | 2502 | } |
2528 | lock = (adap->flags & USING_MSIX) ? &qs->rspq.lock : | 2503 | lock = (adap->flags & USING_MSIX) ? &qs->rspq.lock : |
2529 | &adap->sge.qs[0].rspq.lock; | 2504 | &adap->sge.qs[0].rspq.lock; |
2530 | if (spin_trylock_irq(lock)) { | 2505 | if (spin_trylock_irq(lock)) { |
2531 | if (!napi_is_scheduled(qs->netdev)) { | 2506 | if (!napi_is_scheduled(&qs->napi)) { |
2532 | u32 status = t3_read_reg(adap, A_SG_RSPQ_FL_STATUS); | 2507 | u32 status = t3_read_reg(adap, A_SG_RSPQ_FL_STATUS); |
2533 | 2508 | ||
2534 | if (qs->fl[0].credits < qs->fl[0].size) | 2509 | if (qs->fl[0].credits < qs->fl[0].size) |
@@ -2562,12 +2537,9 @@ static void sge_timer_cb(unsigned long data) | |||
2562 | */ | 2537 | */ |
2563 | void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p) | 2538 | void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p) |
2564 | { | 2539 | { |
2565 | if (!qs->netdev) | ||
2566 | return; | ||
2567 | |||
2568 | qs->rspq.holdoff_tmr = max(p->coalesce_usecs * 10, 1U);/* can't be 0 */ | 2540 | qs->rspq.holdoff_tmr = max(p->coalesce_usecs * 10, 1U);/* can't be 0 */ |
2569 | qs->rspq.polling = p->polling; | 2541 | qs->rspq.polling = p->polling; |
2570 | qs->netdev->poll = p->polling ? napi_rx_handler : ofld_poll; | 2542 | qs->napi.poll = p->polling ? napi_rx_handler : ofld_poll; |
2571 | } | 2543 | } |
2572 | 2544 | ||
2573 | /** | 2545 | /** |
@@ -2587,7 +2559,7 @@ void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p) | |||
2587 | */ | 2559 | */ |
2588 | int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports, | 2560 | int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports, |
2589 | int irq_vec_idx, const struct qset_params *p, | 2561 | int irq_vec_idx, const struct qset_params *p, |
2590 | int ntxq, struct net_device *netdev) | 2562 | int ntxq, struct net_device *dev) |
2591 | { | 2563 | { |
2592 | int i, ret = -ENOMEM; | 2564 | int i, ret = -ENOMEM; |
2593 | struct sge_qset *q = &adapter->sge.qs[id]; | 2565 | struct sge_qset *q = &adapter->sge.qs[id]; |
@@ -2708,16 +2680,10 @@ int t3_sge_alloc_qset(struct adapter *adapter, unsigned int id, int nports, | |||
2708 | } | 2680 | } |
2709 | 2681 | ||
2710 | spin_unlock(&adapter->sge.reg_lock); | 2682 | spin_unlock(&adapter->sge.reg_lock); |
2711 | q->netdev = netdev; | ||
2712 | t3_update_qset_coalesce(q, p); | ||
2713 | 2683 | ||
2714 | /* | 2684 | q->adap = adapter; |
2715 | * We use atalk_ptr as a backpointer to a qset. In case a device is | 2685 | q->netdev = dev; |
2716 | * associated with multiple queue sets only the first one sets | 2686 | t3_update_qset_coalesce(q, p); |
2717 | * atalk_ptr. | ||
2718 | */ | ||
2719 | if (netdev->atalk_ptr == NULL) | ||
2720 | netdev->atalk_ptr = q; | ||
2721 | 2687 | ||
2722 | refill_fl(adapter, &q->fl[0], q->fl[0].size, GFP_KERNEL); | 2688 | refill_fl(adapter, &q->fl[0], q->fl[0].size, GFP_KERNEL); |
2723 | refill_fl(adapter, &q->fl[1], q->fl[1].size, GFP_KERNEL); | 2689 | refill_fl(adapter, &q->fl[1], q->fl[1].size, GFP_KERNEL); |
diff --git a/drivers/net/e100.c b/drivers/net/e100.c index 280313b9b069..e25f5ec2b279 100644 --- a/drivers/net/e100.c +++ b/drivers/net/e100.c | |||
@@ -539,6 +539,7 @@ struct nic { | |||
539 | struct csr __iomem *csr; | 539 | struct csr __iomem *csr; |
540 | enum scb_cmd_lo cuc_cmd; | 540 | enum scb_cmd_lo cuc_cmd; |
541 | unsigned int cbs_avail; | 541 | unsigned int cbs_avail; |
542 | struct napi_struct napi; | ||
542 | struct cb *cbs; | 543 | struct cb *cbs; |
543 | struct cb *cb_to_use; | 544 | struct cb *cb_to_use; |
544 | struct cb *cb_to_send; | 545 | struct cb *cb_to_send; |
@@ -1974,35 +1975,31 @@ static irqreturn_t e100_intr(int irq, void *dev_id) | |||
1974 | if(stat_ack & stat_ack_rnr) | 1975 | if(stat_ack & stat_ack_rnr) |
1975 | nic->ru_running = RU_SUSPENDED; | 1976 | nic->ru_running = RU_SUSPENDED; |
1976 | 1977 | ||
1977 | if(likely(netif_rx_schedule_prep(netdev))) { | 1978 | if(likely(netif_rx_schedule_prep(netdev, &nic->napi))) { |
1978 | e100_disable_irq(nic); | 1979 | e100_disable_irq(nic); |
1979 | __netif_rx_schedule(netdev); | 1980 | __netif_rx_schedule(netdev, &nic->napi); |
1980 | } | 1981 | } |
1981 | 1982 | ||
1982 | return IRQ_HANDLED; | 1983 | return IRQ_HANDLED; |
1983 | } | 1984 | } |
1984 | 1985 | ||
1985 | static int e100_poll(struct net_device *netdev, int *budget) | 1986 | static int e100_poll(struct napi_struct *napi, int budget) |
1986 | { | 1987 | { |
1987 | struct nic *nic = netdev_priv(netdev); | 1988 | struct nic *nic = container_of(napi, struct nic, napi); |
1988 | unsigned int work_to_do = min(netdev->quota, *budget); | 1989 | struct net_device *netdev = nic->netdev; |
1989 | unsigned int work_done = 0; | 1990 | int work_done = 0; |
1990 | int tx_cleaned; | 1991 | int tx_cleaned; |
1991 | 1992 | ||
1992 | e100_rx_clean(nic, &work_done, work_to_do); | 1993 | e100_rx_clean(nic, &work_done, budget); |
1993 | tx_cleaned = e100_tx_clean(nic); | 1994 | tx_cleaned = e100_tx_clean(nic); |
1994 | 1995 | ||
1995 | /* If no Rx and Tx cleanup work was done, exit polling mode. */ | 1996 | /* If no Rx and Tx cleanup work was done, exit polling mode. */ |
1996 | if((!tx_cleaned && (work_done == 0)) || !netif_running(netdev)) { | 1997 | if((!tx_cleaned && (work_done == 0)) || !netif_running(netdev)) { |
1997 | netif_rx_complete(netdev); | 1998 | netif_rx_complete(netdev, napi); |
1998 | e100_enable_irq(nic); | 1999 | e100_enable_irq(nic); |
1999 | return 0; | ||
2000 | } | 2000 | } |
2001 | 2001 | ||
2002 | *budget -= work_done; | 2002 | return work_done; |
2003 | netdev->quota -= work_done; | ||
2004 | |||
2005 | return 1; | ||
2006 | } | 2003 | } |
2007 | 2004 | ||
2008 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2005 | #ifdef CONFIG_NET_POLL_CONTROLLER |
@@ -2071,7 +2068,7 @@ static int e100_up(struct nic *nic) | |||
2071 | nic->netdev->name, nic->netdev))) | 2068 | nic->netdev->name, nic->netdev))) |
2072 | goto err_no_irq; | 2069 | goto err_no_irq; |
2073 | netif_wake_queue(nic->netdev); | 2070 | netif_wake_queue(nic->netdev); |
2074 | netif_poll_enable(nic->netdev); | 2071 | napi_enable(&nic->napi); |
2075 | /* enable ints _after_ enabling poll, preventing a race between | 2072 | /* enable ints _after_ enabling poll, preventing a race between |
2076 | * disable ints+schedule */ | 2073 | * disable ints+schedule */ |
2077 | e100_enable_irq(nic); | 2074 | e100_enable_irq(nic); |
@@ -2089,7 +2086,7 @@ err_rx_clean_list: | |||
2089 | static void e100_down(struct nic *nic) | 2086 | static void e100_down(struct nic *nic) |
2090 | { | 2087 | { |
2091 | /* wait here for poll to complete */ | 2088 | /* wait here for poll to complete */ |
2092 | netif_poll_disable(nic->netdev); | 2089 | napi_disable(&nic->napi); |
2093 | netif_stop_queue(nic->netdev); | 2090 | netif_stop_queue(nic->netdev); |
2094 | e100_hw_reset(nic); | 2091 | e100_hw_reset(nic); |
2095 | free_irq(nic->pdev->irq, nic->netdev); | 2092 | free_irq(nic->pdev->irq, nic->netdev); |
@@ -2572,14 +2569,13 @@ static int __devinit e100_probe(struct pci_dev *pdev, | |||
2572 | SET_ETHTOOL_OPS(netdev, &e100_ethtool_ops); | 2569 | SET_ETHTOOL_OPS(netdev, &e100_ethtool_ops); |
2573 | netdev->tx_timeout = e100_tx_timeout; | 2570 | netdev->tx_timeout = e100_tx_timeout; |
2574 | netdev->watchdog_timeo = E100_WATCHDOG_PERIOD; | 2571 | netdev->watchdog_timeo = E100_WATCHDOG_PERIOD; |
2575 | netdev->poll = e100_poll; | ||
2576 | netdev->weight = E100_NAPI_WEIGHT; | ||
2577 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2572 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2578 | netdev->poll_controller = e100_netpoll; | 2573 | netdev->poll_controller = e100_netpoll; |
2579 | #endif | 2574 | #endif |
2580 | strncpy(netdev->name, pci_name(pdev), sizeof(netdev->name) - 1); | 2575 | strncpy(netdev->name, pci_name(pdev), sizeof(netdev->name) - 1); |
2581 | 2576 | ||
2582 | nic = netdev_priv(netdev); | 2577 | nic = netdev_priv(netdev); |
2578 | netif_napi_add(netdev, &nic->napi, e100_poll, E100_NAPI_WEIGHT); | ||
2583 | nic->netdev = netdev; | 2579 | nic->netdev = netdev; |
2584 | nic->pdev = pdev; | 2580 | nic->pdev = pdev; |
2585 | nic->msg_enable = (1 << debug) - 1; | 2581 | nic->msg_enable = (1 << debug) - 1; |
@@ -2733,7 +2729,7 @@ static int e100_suspend(struct pci_dev *pdev, pm_message_t state) | |||
2733 | struct nic *nic = netdev_priv(netdev); | 2729 | struct nic *nic = netdev_priv(netdev); |
2734 | 2730 | ||
2735 | if (netif_running(netdev)) | 2731 | if (netif_running(netdev)) |
2736 | netif_poll_disable(nic->netdev); | 2732 | napi_disable(&nic->napi); |
2737 | del_timer_sync(&nic->watchdog); | 2733 | del_timer_sync(&nic->watchdog); |
2738 | netif_carrier_off(nic->netdev); | 2734 | netif_carrier_off(nic->netdev); |
2739 | netif_device_detach(netdev); | 2735 | netif_device_detach(netdev); |
@@ -2779,7 +2775,7 @@ static void e100_shutdown(struct pci_dev *pdev) | |||
2779 | struct nic *nic = netdev_priv(netdev); | 2775 | struct nic *nic = netdev_priv(netdev); |
2780 | 2776 | ||
2781 | if (netif_running(netdev)) | 2777 | if (netif_running(netdev)) |
2782 | netif_poll_disable(nic->netdev); | 2778 | napi_disable(&nic->napi); |
2783 | del_timer_sync(&nic->watchdog); | 2779 | del_timer_sync(&nic->watchdog); |
2784 | netif_carrier_off(nic->netdev); | 2780 | netif_carrier_off(nic->netdev); |
2785 | 2781 | ||
@@ -2804,12 +2800,13 @@ static void e100_shutdown(struct pci_dev *pdev) | |||
2804 | static pci_ers_result_t e100_io_error_detected(struct pci_dev *pdev, pci_channel_state_t state) | 2800 | static pci_ers_result_t e100_io_error_detected(struct pci_dev *pdev, pci_channel_state_t state) |
2805 | { | 2801 | { |
2806 | struct net_device *netdev = pci_get_drvdata(pdev); | 2802 | struct net_device *netdev = pci_get_drvdata(pdev); |
2803 | struct nic *nic = netdev_priv(netdev); | ||
2807 | 2804 | ||
2808 | /* Similar to calling e100_down(), but avoids adpater I/O. */ | 2805 | /* Similar to calling e100_down(), but avoids adpater I/O. */ |
2809 | netdev->stop(netdev); | 2806 | netdev->stop(netdev); |
2810 | 2807 | ||
2811 | /* Detach; put netif into state similar to hotplug unplug. */ | 2808 | /* Detach; put netif into state similar to hotplug unplug. */ |
2812 | netif_poll_enable(netdev); | 2809 | napi_enable(&nic->napi); |
2813 | netif_device_detach(netdev); | 2810 | netif_device_detach(netdev); |
2814 | pci_disable_device(pdev); | 2811 | pci_disable_device(pdev); |
2815 | 2812 | ||
diff --git a/drivers/net/e1000/e1000.h b/drivers/net/e1000/e1000.h index 16a6edfeba41..781ed9968489 100644 --- a/drivers/net/e1000/e1000.h +++ b/drivers/net/e1000/e1000.h | |||
@@ -300,6 +300,7 @@ struct e1000_adapter { | |||
300 | int cleaned_count); | 300 | int cleaned_count); |
301 | struct e1000_rx_ring *rx_ring; /* One per active queue */ | 301 | struct e1000_rx_ring *rx_ring; /* One per active queue */ |
302 | #ifdef CONFIG_E1000_NAPI | 302 | #ifdef CONFIG_E1000_NAPI |
303 | struct napi_struct napi; | ||
303 | struct net_device *polling_netdev; /* One per active queue */ | 304 | struct net_device *polling_netdev; /* One per active queue */ |
304 | #endif | 305 | #endif |
305 | int num_tx_queues; | 306 | int num_tx_queues; |
diff --git a/drivers/net/e1000/e1000_main.c b/drivers/net/e1000/e1000_main.c index e7c8951f47fa..723568d6e44a 100644 --- a/drivers/net/e1000/e1000_main.c +++ b/drivers/net/e1000/e1000_main.c | |||
@@ -166,7 +166,7 @@ static irqreturn_t e1000_intr_msi(int irq, void *data); | |||
166 | static boolean_t e1000_clean_tx_irq(struct e1000_adapter *adapter, | 166 | static boolean_t e1000_clean_tx_irq(struct e1000_adapter *adapter, |
167 | struct e1000_tx_ring *tx_ring); | 167 | struct e1000_tx_ring *tx_ring); |
168 | #ifdef CONFIG_E1000_NAPI | 168 | #ifdef CONFIG_E1000_NAPI |
169 | static int e1000_clean(struct net_device *poll_dev, int *budget); | 169 | static int e1000_clean(struct napi_struct *napi, int budget); |
170 | static boolean_t e1000_clean_rx_irq(struct e1000_adapter *adapter, | 170 | static boolean_t e1000_clean_rx_irq(struct e1000_adapter *adapter, |
171 | struct e1000_rx_ring *rx_ring, | 171 | struct e1000_rx_ring *rx_ring, |
172 | int *work_done, int work_to_do); | 172 | int *work_done, int work_to_do); |
@@ -545,7 +545,7 @@ int e1000_up(struct e1000_adapter *adapter) | |||
545 | clear_bit(__E1000_DOWN, &adapter->flags); | 545 | clear_bit(__E1000_DOWN, &adapter->flags); |
546 | 546 | ||
547 | #ifdef CONFIG_E1000_NAPI | 547 | #ifdef CONFIG_E1000_NAPI |
548 | netif_poll_enable(adapter->netdev); | 548 | napi_enable(&adapter->napi); |
549 | #endif | 549 | #endif |
550 | e1000_irq_enable(adapter); | 550 | e1000_irq_enable(adapter); |
551 | 551 | ||
@@ -634,7 +634,7 @@ e1000_down(struct e1000_adapter *adapter) | |||
634 | set_bit(__E1000_DOWN, &adapter->flags); | 634 | set_bit(__E1000_DOWN, &adapter->flags); |
635 | 635 | ||
636 | #ifdef CONFIG_E1000_NAPI | 636 | #ifdef CONFIG_E1000_NAPI |
637 | netif_poll_disable(netdev); | 637 | napi_disable(&adapter->napi); |
638 | #endif | 638 | #endif |
639 | e1000_irq_disable(adapter); | 639 | e1000_irq_disable(adapter); |
640 | 640 | ||
@@ -936,8 +936,7 @@ e1000_probe(struct pci_dev *pdev, | |||
936 | netdev->tx_timeout = &e1000_tx_timeout; | 936 | netdev->tx_timeout = &e1000_tx_timeout; |
937 | netdev->watchdog_timeo = 5 * HZ; | 937 | netdev->watchdog_timeo = 5 * HZ; |
938 | #ifdef CONFIG_E1000_NAPI | 938 | #ifdef CONFIG_E1000_NAPI |
939 | netdev->poll = &e1000_clean; | 939 | netif_napi_add(netdev, &adapter->napi, e1000_clean, 64); |
940 | netdev->weight = 64; | ||
941 | #endif | 940 | #endif |
942 | netdev->vlan_rx_register = e1000_vlan_rx_register; | 941 | netdev->vlan_rx_register = e1000_vlan_rx_register; |
943 | netdev->vlan_rx_add_vid = e1000_vlan_rx_add_vid; | 942 | netdev->vlan_rx_add_vid = e1000_vlan_rx_add_vid; |
@@ -1151,9 +1150,6 @@ e1000_probe(struct pci_dev *pdev, | |||
1151 | /* tell the stack to leave us alone until e1000_open() is called */ | 1150 | /* tell the stack to leave us alone until e1000_open() is called */ |
1152 | netif_carrier_off(netdev); | 1151 | netif_carrier_off(netdev); |
1153 | netif_stop_queue(netdev); | 1152 | netif_stop_queue(netdev); |
1154 | #ifdef CONFIG_E1000_NAPI | ||
1155 | netif_poll_disable(netdev); | ||
1156 | #endif | ||
1157 | 1153 | ||
1158 | strcpy(netdev->name, "eth%d"); | 1154 | strcpy(netdev->name, "eth%d"); |
1159 | if ((err = register_netdev(netdev))) | 1155 | if ((err = register_netdev(netdev))) |
@@ -1222,12 +1218,13 @@ e1000_remove(struct pci_dev *pdev) | |||
1222 | * would have already happened in close and is redundant. */ | 1218 | * would have already happened in close and is redundant. */ |
1223 | e1000_release_hw_control(adapter); | 1219 | e1000_release_hw_control(adapter); |
1224 | 1220 | ||
1225 | unregister_netdev(netdev); | ||
1226 | #ifdef CONFIG_E1000_NAPI | 1221 | #ifdef CONFIG_E1000_NAPI |
1227 | for (i = 0; i < adapter->num_rx_queues; i++) | 1222 | for (i = 0; i < adapter->num_rx_queues; i++) |
1228 | dev_put(&adapter->polling_netdev[i]); | 1223 | dev_put(&adapter->polling_netdev[i]); |
1229 | #endif | 1224 | #endif |
1230 | 1225 | ||
1226 | unregister_netdev(netdev); | ||
1227 | |||
1231 | if (!e1000_check_phy_reset_block(&adapter->hw)) | 1228 | if (!e1000_check_phy_reset_block(&adapter->hw)) |
1232 | e1000_phy_hw_reset(&adapter->hw); | 1229 | e1000_phy_hw_reset(&adapter->hw); |
1233 | 1230 | ||
@@ -1325,8 +1322,6 @@ e1000_sw_init(struct e1000_adapter *adapter) | |||
1325 | #ifdef CONFIG_E1000_NAPI | 1322 | #ifdef CONFIG_E1000_NAPI |
1326 | for (i = 0; i < adapter->num_rx_queues; i++) { | 1323 | for (i = 0; i < adapter->num_rx_queues; i++) { |
1327 | adapter->polling_netdev[i].priv = adapter; | 1324 | adapter->polling_netdev[i].priv = adapter; |
1328 | adapter->polling_netdev[i].poll = &e1000_clean; | ||
1329 | adapter->polling_netdev[i].weight = 64; | ||
1330 | dev_hold(&adapter->polling_netdev[i]); | 1325 | dev_hold(&adapter->polling_netdev[i]); |
1331 | set_bit(__LINK_STATE_START, &adapter->polling_netdev[i].state); | 1326 | set_bit(__LINK_STATE_START, &adapter->polling_netdev[i].state); |
1332 | } | 1327 | } |
@@ -1443,7 +1438,7 @@ e1000_open(struct net_device *netdev) | |||
1443 | clear_bit(__E1000_DOWN, &adapter->flags); | 1438 | clear_bit(__E1000_DOWN, &adapter->flags); |
1444 | 1439 | ||
1445 | #ifdef CONFIG_E1000_NAPI | 1440 | #ifdef CONFIG_E1000_NAPI |
1446 | netif_poll_enable(netdev); | 1441 | napi_enable(&adapter->napi); |
1447 | #endif | 1442 | #endif |
1448 | 1443 | ||
1449 | e1000_irq_enable(adapter); | 1444 | e1000_irq_enable(adapter); |
@@ -3786,12 +3781,12 @@ e1000_intr_msi(int irq, void *data) | |||
3786 | } | 3781 | } |
3787 | 3782 | ||
3788 | #ifdef CONFIG_E1000_NAPI | 3783 | #ifdef CONFIG_E1000_NAPI |
3789 | if (likely(netif_rx_schedule_prep(netdev))) { | 3784 | if (likely(netif_rx_schedule_prep(netdev, &adapter->napi))) { |
3790 | adapter->total_tx_bytes = 0; | 3785 | adapter->total_tx_bytes = 0; |
3791 | adapter->total_tx_packets = 0; | 3786 | adapter->total_tx_packets = 0; |
3792 | adapter->total_rx_bytes = 0; | 3787 | adapter->total_rx_bytes = 0; |
3793 | adapter->total_rx_packets = 0; | 3788 | adapter->total_rx_packets = 0; |
3794 | __netif_rx_schedule(netdev); | 3789 | __netif_rx_schedule(netdev, &adapter->napi); |
3795 | } else | 3790 | } else |
3796 | e1000_irq_enable(adapter); | 3791 | e1000_irq_enable(adapter); |
3797 | #else | 3792 | #else |
@@ -3871,12 +3866,12 @@ e1000_intr(int irq, void *data) | |||
3871 | E1000_WRITE_REG(hw, IMC, ~0); | 3866 | E1000_WRITE_REG(hw, IMC, ~0); |
3872 | E1000_WRITE_FLUSH(hw); | 3867 | E1000_WRITE_FLUSH(hw); |
3873 | } | 3868 | } |
3874 | if (likely(netif_rx_schedule_prep(netdev))) { | 3869 | if (likely(netif_rx_schedule_prep(netdev, &adapter->napi))) { |
3875 | adapter->total_tx_bytes = 0; | 3870 | adapter->total_tx_bytes = 0; |
3876 | adapter->total_tx_packets = 0; | 3871 | adapter->total_tx_packets = 0; |
3877 | adapter->total_rx_bytes = 0; | 3872 | adapter->total_rx_bytes = 0; |
3878 | adapter->total_rx_packets = 0; | 3873 | adapter->total_rx_packets = 0; |
3879 | __netif_rx_schedule(netdev); | 3874 | __netif_rx_schedule(netdev, &adapter->napi); |
3880 | } else | 3875 | } else |
3881 | /* this really should not happen! if it does it is basically a | 3876 | /* this really should not happen! if it does it is basically a |
3882 | * bug, but not a hard error, so enable ints and continue */ | 3877 | * bug, but not a hard error, so enable ints and continue */ |
@@ -3924,10 +3919,10 @@ e1000_intr(int irq, void *data) | |||
3924 | **/ | 3919 | **/ |
3925 | 3920 | ||
3926 | static int | 3921 | static int |
3927 | e1000_clean(struct net_device *poll_dev, int *budget) | 3922 | e1000_clean(struct napi_struct *napi, int budget) |
3928 | { | 3923 | { |
3929 | struct e1000_adapter *adapter; | 3924 | struct e1000_adapter *adapter = container_of(napi, struct e1000_adapter, napi); |
3930 | int work_to_do = min(*budget, poll_dev->quota); | 3925 | struct net_device *poll_dev = adapter->netdev; |
3931 | int tx_cleaned = 0, work_done = 0; | 3926 | int tx_cleaned = 0, work_done = 0; |
3932 | 3927 | ||
3933 | /* Must NOT use netdev_priv macro here. */ | 3928 | /* Must NOT use netdev_priv macro here. */ |
@@ -3948,23 +3943,19 @@ e1000_clean(struct net_device *poll_dev, int *budget) | |||
3948 | } | 3943 | } |
3949 | 3944 | ||
3950 | adapter->clean_rx(adapter, &adapter->rx_ring[0], | 3945 | adapter->clean_rx(adapter, &adapter->rx_ring[0], |
3951 | &work_done, work_to_do); | 3946 | &work_done, budget); |
3952 | |||
3953 | *budget -= work_done; | ||
3954 | poll_dev->quota -= work_done; | ||
3955 | 3947 | ||
3956 | /* If no Tx and not enough Rx work done, exit the polling mode */ | 3948 | /* If no Tx and not enough Rx work done, exit the polling mode */ |
3957 | if ((!tx_cleaned && (work_done == 0)) || | 3949 | if ((!tx_cleaned && (work_done < budget)) || |
3958 | !netif_running(poll_dev)) { | 3950 | !netif_running(poll_dev)) { |
3959 | quit_polling: | 3951 | quit_polling: |
3960 | if (likely(adapter->itr_setting & 3)) | 3952 | if (likely(adapter->itr_setting & 3)) |
3961 | e1000_set_itr(adapter); | 3953 | e1000_set_itr(adapter); |
3962 | netif_rx_complete(poll_dev); | 3954 | netif_rx_complete(poll_dev, napi); |
3963 | e1000_irq_enable(adapter); | 3955 | e1000_irq_enable(adapter); |
3964 | return 0; | ||
3965 | } | 3956 | } |
3966 | 3957 | ||
3967 | return 1; | 3958 | return work_done; |
3968 | } | 3959 | } |
3969 | 3960 | ||
3970 | #endif | 3961 | #endif |
diff --git a/drivers/net/ehea/ehea.h b/drivers/net/ehea/ehea.h index 8d58be56f4e3..a154681165b9 100644 --- a/drivers/net/ehea/ehea.h +++ b/drivers/net/ehea/ehea.h | |||
@@ -351,6 +351,7 @@ struct ehea_q_skb_arr { | |||
351 | * Port resources | 351 | * Port resources |
352 | */ | 352 | */ |
353 | struct ehea_port_res { | 353 | struct ehea_port_res { |
354 | struct napi_struct napi; | ||
354 | struct port_stats p_stats; | 355 | struct port_stats p_stats; |
355 | struct ehea_mr send_mr; /* send memory region */ | 356 | struct ehea_mr send_mr; /* send memory region */ |
356 | struct ehea_mr recv_mr; /* receive memory region */ | 357 | struct ehea_mr recv_mr; /* receive memory region */ |
@@ -362,7 +363,6 @@ struct ehea_port_res { | |||
362 | struct ehea_cq *send_cq; | 363 | struct ehea_cq *send_cq; |
363 | struct ehea_cq *recv_cq; | 364 | struct ehea_cq *recv_cq; |
364 | struct ehea_eq *eq; | 365 | struct ehea_eq *eq; |
365 | struct net_device *d_netdev; | ||
366 | struct ehea_q_skb_arr rq1_skba; | 366 | struct ehea_q_skb_arr rq1_skba; |
367 | struct ehea_q_skb_arr rq2_skba; | 367 | struct ehea_q_skb_arr rq2_skba; |
368 | struct ehea_q_skb_arr rq3_skba; | 368 | struct ehea_q_skb_arr rq3_skba; |
diff --git a/drivers/net/ehea/ehea_main.c b/drivers/net/ehea/ehea_main.c index 717b12984d10..5ebd545ab04e 100644 --- a/drivers/net/ehea/ehea_main.c +++ b/drivers/net/ehea/ehea_main.c | |||
@@ -393,9 +393,9 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq, | |||
393 | return 0; | 393 | return 0; |
394 | } | 394 | } |
395 | 395 | ||
396 | static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | 396 | static int ehea_proc_rwqes(struct net_device *dev, |
397 | struct ehea_port_res *pr, | 397 | struct ehea_port_res *pr, |
398 | int *budget) | 398 | int budget) |
399 | { | 399 | { |
400 | struct ehea_port *port = pr->port; | 400 | struct ehea_port *port = pr->port; |
401 | struct ehea_qp *qp = pr->qp; | 401 | struct ehea_qp *qp = pr->qp; |
@@ -408,18 +408,16 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
408 | int skb_arr_rq2_len = pr->rq2_skba.len; | 408 | int skb_arr_rq2_len = pr->rq2_skba.len; |
409 | int skb_arr_rq3_len = pr->rq3_skba.len; | 409 | int skb_arr_rq3_len = pr->rq3_skba.len; |
410 | int processed, processed_rq1, processed_rq2, processed_rq3; | 410 | int processed, processed_rq1, processed_rq2, processed_rq3; |
411 | int wqe_index, last_wqe_index, rq, my_quota, port_reset; | 411 | int wqe_index, last_wqe_index, rq, port_reset; |
412 | 412 | ||
413 | processed = processed_rq1 = processed_rq2 = processed_rq3 = 0; | 413 | processed = processed_rq1 = processed_rq2 = processed_rq3 = 0; |
414 | last_wqe_index = 0; | 414 | last_wqe_index = 0; |
415 | my_quota = min(*budget, dev->quota); | ||
416 | 415 | ||
417 | cqe = ehea_poll_rq1(qp, &wqe_index); | 416 | cqe = ehea_poll_rq1(qp, &wqe_index); |
418 | while ((my_quota > 0) && cqe) { | 417 | while ((processed < budget) && cqe) { |
419 | ehea_inc_rq1(qp); | 418 | ehea_inc_rq1(qp); |
420 | processed_rq1++; | 419 | processed_rq1++; |
421 | processed++; | 420 | processed++; |
422 | my_quota--; | ||
423 | if (netif_msg_rx_status(port)) | 421 | if (netif_msg_rx_status(port)) |
424 | ehea_dump(cqe, sizeof(*cqe), "CQE"); | 422 | ehea_dump(cqe, sizeof(*cqe), "CQE"); |
425 | 423 | ||
@@ -434,14 +432,14 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
434 | if (netif_msg_rx_err(port)) | 432 | if (netif_msg_rx_err(port)) |
435 | ehea_error("LL rq1: skb=NULL"); | 433 | ehea_error("LL rq1: skb=NULL"); |
436 | 434 | ||
437 | skb = netdev_alloc_skb(port->netdev, | 435 | skb = netdev_alloc_skb(dev, |
438 | EHEA_L_PKT_SIZE); | 436 | EHEA_L_PKT_SIZE); |
439 | if (!skb) | 437 | if (!skb) |
440 | break; | 438 | break; |
441 | } | 439 | } |
442 | skb_copy_to_linear_data(skb, ((char*)cqe) + 64, | 440 | skb_copy_to_linear_data(skb, ((char*)cqe) + 64, |
443 | cqe->num_bytes_transfered - 4); | 441 | cqe->num_bytes_transfered - 4); |
444 | ehea_fill_skb(port->netdev, skb, cqe); | 442 | ehea_fill_skb(dev, skb, cqe); |
445 | } else if (rq == 2) { /* RQ2 */ | 443 | } else if (rq == 2) { /* RQ2 */ |
446 | skb = get_skb_by_index(skb_arr_rq2, | 444 | skb = get_skb_by_index(skb_arr_rq2, |
447 | skb_arr_rq2_len, cqe); | 445 | skb_arr_rq2_len, cqe); |
@@ -450,7 +448,7 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
450 | ehea_error("rq2: skb=NULL"); | 448 | ehea_error("rq2: skb=NULL"); |
451 | break; | 449 | break; |
452 | } | 450 | } |
453 | ehea_fill_skb(port->netdev, skb, cqe); | 451 | ehea_fill_skb(dev, skb, cqe); |
454 | processed_rq2++; | 452 | processed_rq2++; |
455 | } else { /* RQ3 */ | 453 | } else { /* RQ3 */ |
456 | skb = get_skb_by_index(skb_arr_rq3, | 454 | skb = get_skb_by_index(skb_arr_rq3, |
@@ -460,7 +458,7 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
460 | ehea_error("rq3: skb=NULL"); | 458 | ehea_error("rq3: skb=NULL"); |
461 | break; | 459 | break; |
462 | } | 460 | } |
463 | ehea_fill_skb(port->netdev, skb, cqe); | 461 | ehea_fill_skb(dev, skb, cqe); |
464 | processed_rq3++; | 462 | processed_rq3++; |
465 | } | 463 | } |
466 | 464 | ||
@@ -471,7 +469,7 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
471 | else | 469 | else |
472 | netif_receive_skb(skb); | 470 | netif_receive_skb(skb); |
473 | 471 | ||
474 | port->netdev->last_rx = jiffies; | 472 | dev->last_rx = jiffies; |
475 | } else { | 473 | } else { |
476 | pr->p_stats.poll_receive_errors++; | 474 | pr->p_stats.poll_receive_errors++; |
477 | port_reset = ehea_treat_poll_error(pr, rq, cqe, | 475 | port_reset = ehea_treat_poll_error(pr, rq, cqe, |
@@ -484,14 +482,12 @@ static struct ehea_cqe *ehea_proc_rwqes(struct net_device *dev, | |||
484 | } | 482 | } |
485 | 483 | ||
486 | pr->rx_packets += processed; | 484 | pr->rx_packets += processed; |
487 | *budget -= processed; | ||
488 | 485 | ||
489 | ehea_refill_rq1(pr, last_wqe_index, processed_rq1); | 486 | ehea_refill_rq1(pr, last_wqe_index, processed_rq1); |
490 | ehea_refill_rq2(pr, processed_rq2); | 487 | ehea_refill_rq2(pr, processed_rq2); |
491 | ehea_refill_rq3(pr, processed_rq3); | 488 | ehea_refill_rq3(pr, processed_rq3); |
492 | 489 | ||
493 | cqe = ehea_poll_rq1(qp, &wqe_index); | 490 | return processed; |
494 | return cqe; | ||
495 | } | 491 | } |
496 | 492 | ||
497 | static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota) | 493 | static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota) |
@@ -554,22 +550,27 @@ static struct ehea_cqe *ehea_proc_cqes(struct ehea_port_res *pr, int my_quota) | |||
554 | } | 550 | } |
555 | 551 | ||
556 | #define EHEA_NAPI_POLL_NUM_BEFORE_IRQ 16 | 552 | #define EHEA_NAPI_POLL_NUM_BEFORE_IRQ 16 |
553 | #define EHEA_POLL_MAX_CQES 65535 | ||
557 | 554 | ||
558 | static int ehea_poll(struct net_device *dev, int *budget) | 555 | static int ehea_poll(struct napi_struct *napi, int budget) |
559 | { | 556 | { |
560 | struct ehea_port_res *pr = dev->priv; | 557 | struct ehea_port_res *pr = container_of(napi, struct ehea_port_res, napi); |
558 | struct net_device *dev = pr->port->netdev; | ||
561 | struct ehea_cqe *cqe; | 559 | struct ehea_cqe *cqe; |
562 | struct ehea_cqe *cqe_skb = NULL; | 560 | struct ehea_cqe *cqe_skb = NULL; |
563 | int force_irq, wqe_index; | 561 | int force_irq, wqe_index; |
564 | 562 | int rx = 0; | |
565 | cqe = ehea_poll_rq1(pr->qp, &wqe_index); | ||
566 | cqe_skb = ehea_poll_cq(pr->send_cq); | ||
567 | 563 | ||
568 | force_irq = (pr->poll_counter > EHEA_NAPI_POLL_NUM_BEFORE_IRQ); | 564 | force_irq = (pr->poll_counter > EHEA_NAPI_POLL_NUM_BEFORE_IRQ); |
565 | cqe_skb = ehea_proc_cqes(pr, EHEA_POLL_MAX_CQES); | ||
566 | |||
567 | if (!force_irq) | ||
568 | rx += ehea_proc_rwqes(dev, pr, budget - rx); | ||
569 | 569 | ||
570 | if ((!cqe && !cqe_skb) || force_irq) { | 570 | while ((rx != budget) || force_irq) { |
571 | pr->poll_counter = 0; | 571 | pr->poll_counter = 0; |
572 | netif_rx_complete(dev); | 572 | force_irq = 0; |
573 | netif_rx_complete(dev, napi); | ||
573 | ehea_reset_cq_ep(pr->recv_cq); | 574 | ehea_reset_cq_ep(pr->recv_cq); |
574 | ehea_reset_cq_ep(pr->send_cq); | 575 | ehea_reset_cq_ep(pr->send_cq); |
575 | ehea_reset_cq_n1(pr->recv_cq); | 576 | ehea_reset_cq_n1(pr->recv_cq); |
@@ -578,43 +579,35 @@ static int ehea_poll(struct net_device *dev, int *budget) | |||
578 | cqe_skb = ehea_poll_cq(pr->send_cq); | 579 | cqe_skb = ehea_poll_cq(pr->send_cq); |
579 | 580 | ||
580 | if (!cqe && !cqe_skb) | 581 | if (!cqe && !cqe_skb) |
581 | return 0; | 582 | return rx; |
582 | 583 | ||
583 | if (!netif_rx_reschedule(dev, dev->quota)) | 584 | if (!netif_rx_reschedule(dev, napi)) |
584 | return 0; | 585 | return rx; |
585 | } | ||
586 | |||
587 | cqe = ehea_proc_rwqes(dev, pr, budget); | ||
588 | cqe_skb = ehea_proc_cqes(pr, 300); | ||
589 | 586 | ||
590 | if (cqe || cqe_skb) | 587 | cqe_skb = ehea_proc_cqes(pr, EHEA_POLL_MAX_CQES); |
591 | pr->poll_counter++; | 588 | rx += ehea_proc_rwqes(dev, pr, budget - rx); |
589 | } | ||
592 | 590 | ||
593 | return 1; | 591 | pr->poll_counter++; |
592 | return rx; | ||
594 | } | 593 | } |
595 | 594 | ||
596 | #ifdef CONFIG_NET_POLL_CONTROLLER | 595 | #ifdef CONFIG_NET_POLL_CONTROLLER |
597 | static void ehea_netpoll(struct net_device *dev) | 596 | static void ehea_netpoll(struct net_device *dev) |
598 | { | 597 | { |
599 | struct ehea_port *port = netdev_priv(dev); | 598 | struct ehea_port *port = netdev_priv(dev); |
599 | int i; | ||
600 | 600 | ||
601 | netif_rx_schedule(port->port_res[0].d_netdev); | 601 | for (i = 0; i < port->num_def_qps; i++) |
602 | netif_rx_schedule(dev, &port->port_res[i].napi); | ||
602 | } | 603 | } |
603 | #endif | 604 | #endif |
604 | 605 | ||
605 | static int ehea_poll_firstqueue(struct net_device *dev, int *budget) | ||
606 | { | ||
607 | struct ehea_port *port = netdev_priv(dev); | ||
608 | struct net_device *d_dev = port->port_res[0].d_netdev; | ||
609 | |||
610 | return ehea_poll(d_dev, budget); | ||
611 | } | ||
612 | |||
613 | static irqreturn_t ehea_recv_irq_handler(int irq, void *param) | 606 | static irqreturn_t ehea_recv_irq_handler(int irq, void *param) |
614 | { | 607 | { |
615 | struct ehea_port_res *pr = param; | 608 | struct ehea_port_res *pr = param; |
616 | 609 | ||
617 | netif_rx_schedule(pr->d_netdev); | 610 | netif_rx_schedule(pr->port->netdev, &pr->napi); |
618 | 611 | ||
619 | return IRQ_HANDLED; | 612 | return IRQ_HANDLED; |
620 | } | 613 | } |
@@ -1236,14 +1229,7 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr, | |||
1236 | 1229 | ||
1237 | kfree(init_attr); | 1230 | kfree(init_attr); |
1238 | 1231 | ||
1239 | pr->d_netdev = alloc_netdev(0, "", ether_setup); | 1232 | netif_napi_add(pr->port->netdev, &pr->napi, ehea_poll, 64); |
1240 | if (!pr->d_netdev) | ||
1241 | goto out_free; | ||
1242 | pr->d_netdev->priv = pr; | ||
1243 | pr->d_netdev->weight = 64; | ||
1244 | pr->d_netdev->poll = ehea_poll; | ||
1245 | set_bit(__LINK_STATE_START, &pr->d_netdev->state); | ||
1246 | strcpy(pr->d_netdev->name, port->netdev->name); | ||
1247 | 1233 | ||
1248 | ret = 0; | 1234 | ret = 0; |
1249 | goto out; | 1235 | goto out; |
@@ -1266,8 +1252,6 @@ static int ehea_clean_portres(struct ehea_port *port, struct ehea_port_res *pr) | |||
1266 | { | 1252 | { |
1267 | int ret, i; | 1253 | int ret, i; |
1268 | 1254 | ||
1269 | free_netdev(pr->d_netdev); | ||
1270 | |||
1271 | ret = ehea_destroy_qp(pr->qp); | 1255 | ret = ehea_destroy_qp(pr->qp); |
1272 | 1256 | ||
1273 | if (!ret) { | 1257 | if (!ret) { |
@@ -2248,6 +2232,22 @@ out: | |||
2248 | return ret; | 2232 | return ret; |
2249 | } | 2233 | } |
2250 | 2234 | ||
2235 | static void port_napi_disable(struct ehea_port *port) | ||
2236 | { | ||
2237 | int i; | ||
2238 | |||
2239 | for (i = 0; i < port->num_def_qps; i++) | ||
2240 | napi_disable(&port->port_res[i].napi); | ||
2241 | } | ||
2242 | |||
2243 | static void port_napi_enable(struct ehea_port *port) | ||
2244 | { | ||
2245 | int i; | ||
2246 | |||
2247 | for (i = 0; i < port->num_def_qps; i++) | ||
2248 | napi_enable(&port->port_res[i].napi); | ||
2249 | } | ||
2250 | |||
2251 | static int ehea_open(struct net_device *dev) | 2251 | static int ehea_open(struct net_device *dev) |
2252 | { | 2252 | { |
2253 | int ret; | 2253 | int ret; |
@@ -2259,8 +2259,10 @@ static int ehea_open(struct net_device *dev) | |||
2259 | ehea_info("enabling port %s", dev->name); | 2259 | ehea_info("enabling port %s", dev->name); |
2260 | 2260 | ||
2261 | ret = ehea_up(dev); | 2261 | ret = ehea_up(dev); |
2262 | if (!ret) | 2262 | if (!ret) { |
2263 | port_napi_enable(port); | ||
2263 | netif_start_queue(dev); | 2264 | netif_start_queue(dev); |
2265 | } | ||
2264 | 2266 | ||
2265 | up(&port->port_lock); | 2267 | up(&port->port_lock); |
2266 | 2268 | ||
@@ -2269,7 +2271,7 @@ static int ehea_open(struct net_device *dev) | |||
2269 | 2271 | ||
2270 | static int ehea_down(struct net_device *dev) | 2272 | static int ehea_down(struct net_device *dev) |
2271 | { | 2273 | { |
2272 | int ret, i; | 2274 | int ret; |
2273 | struct ehea_port *port = netdev_priv(dev); | 2275 | struct ehea_port *port = netdev_priv(dev); |
2274 | 2276 | ||
2275 | if (port->state == EHEA_PORT_DOWN) | 2277 | if (port->state == EHEA_PORT_DOWN) |
@@ -2278,10 +2280,7 @@ static int ehea_down(struct net_device *dev) | |||
2278 | ehea_drop_multicast_list(dev); | 2280 | ehea_drop_multicast_list(dev); |
2279 | ehea_free_interrupts(dev); | 2281 | ehea_free_interrupts(dev); |
2280 | 2282 | ||
2281 | for (i = 0; i < port->num_def_qps; i++) | 2283 | port_napi_disable(port); |
2282 | while (test_bit(__LINK_STATE_RX_SCHED, | ||
2283 | &port->port_res[i].d_netdev->state)) | ||
2284 | msleep(1); | ||
2285 | 2284 | ||
2286 | port->state = EHEA_PORT_DOWN; | 2285 | port->state = EHEA_PORT_DOWN; |
2287 | 2286 | ||
@@ -2319,7 +2318,8 @@ static void ehea_reset_port(struct work_struct *work) | |||
2319 | port->resets++; | 2318 | port->resets++; |
2320 | down(&port->port_lock); | 2319 | down(&port->port_lock); |
2321 | netif_stop_queue(dev); | 2320 | netif_stop_queue(dev); |
2322 | netif_poll_disable(dev); | 2321 | |
2322 | port_napi_disable(port); | ||
2323 | 2323 | ||
2324 | ehea_down(dev); | 2324 | ehea_down(dev); |
2325 | 2325 | ||
@@ -2330,7 +2330,8 @@ static void ehea_reset_port(struct work_struct *work) | |||
2330 | if (netif_msg_timer(port)) | 2330 | if (netif_msg_timer(port)) |
2331 | ehea_info("Device %s resetted successfully", dev->name); | 2331 | ehea_info("Device %s resetted successfully", dev->name); |
2332 | 2332 | ||
2333 | netif_poll_enable(dev); | 2333 | port_napi_enable(port); |
2334 | |||
2334 | netif_wake_queue(dev); | 2335 | netif_wake_queue(dev); |
2335 | out: | 2336 | out: |
2336 | up(&port->port_lock); | 2337 | up(&port->port_lock); |
@@ -2358,7 +2359,9 @@ static void ehea_rereg_mrs(struct work_struct *work) | |||
2358 | dev->name); | 2359 | dev->name); |
2359 | down(&port->port_lock); | 2360 | down(&port->port_lock); |
2360 | netif_stop_queue(dev); | 2361 | netif_stop_queue(dev); |
2361 | netif_poll_disable(dev); | 2362 | |
2363 | port_napi_disable(port); | ||
2364 | |||
2362 | ehea_down(dev); | 2365 | ehea_down(dev); |
2363 | up(&port->port_lock); | 2366 | up(&port->port_lock); |
2364 | } | 2367 | } |
@@ -2406,7 +2409,7 @@ static void ehea_rereg_mrs(struct work_struct *work) | |||
2406 | 2409 | ||
2407 | ret = ehea_up(dev); | 2410 | ret = ehea_up(dev); |
2408 | if (!ret) { | 2411 | if (!ret) { |
2409 | netif_poll_enable(dev); | 2412 | port_napi_enable(port); |
2410 | netif_wake_queue(dev); | 2413 | netif_wake_queue(dev); |
2411 | } | 2414 | } |
2412 | 2415 | ||
@@ -2644,11 +2647,9 @@ struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter, | |||
2644 | memcpy(dev->dev_addr, &port->mac_addr, ETH_ALEN); | 2647 | memcpy(dev->dev_addr, &port->mac_addr, ETH_ALEN); |
2645 | 2648 | ||
2646 | dev->open = ehea_open; | 2649 | dev->open = ehea_open; |
2647 | dev->poll = ehea_poll_firstqueue; | ||
2648 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2650 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2649 | dev->poll_controller = ehea_netpoll; | 2651 | dev->poll_controller = ehea_netpoll; |
2650 | #endif | 2652 | #endif |
2651 | dev->weight = 64; | ||
2652 | dev->stop = ehea_stop; | 2653 | dev->stop = ehea_stop; |
2653 | dev->hard_start_xmit = ehea_start_xmit; | 2654 | dev->hard_start_xmit = ehea_start_xmit; |
2654 | dev->get_stats = ehea_get_stats; | 2655 | dev->get_stats = ehea_get_stats; |
diff --git a/drivers/net/epic100.c b/drivers/net/epic100.c index 119778401e48..f8446e373bdd 100644 --- a/drivers/net/epic100.c +++ b/drivers/net/epic100.c | |||
@@ -262,6 +262,7 @@ struct epic_private { | |||
262 | /* Ring pointers. */ | 262 | /* Ring pointers. */ |
263 | spinlock_t lock; /* Group with Tx control cache line. */ | 263 | spinlock_t lock; /* Group with Tx control cache line. */ |
264 | spinlock_t napi_lock; | 264 | spinlock_t napi_lock; |
265 | struct napi_struct napi; | ||
265 | unsigned int reschedule_in_poll; | 266 | unsigned int reschedule_in_poll; |
266 | unsigned int cur_tx, dirty_tx; | 267 | unsigned int cur_tx, dirty_tx; |
267 | 268 | ||
@@ -294,7 +295,7 @@ static void epic_tx_timeout(struct net_device *dev); | |||
294 | static void epic_init_ring(struct net_device *dev); | 295 | static void epic_init_ring(struct net_device *dev); |
295 | static int epic_start_xmit(struct sk_buff *skb, struct net_device *dev); | 296 | static int epic_start_xmit(struct sk_buff *skb, struct net_device *dev); |
296 | static int epic_rx(struct net_device *dev, int budget); | 297 | static int epic_rx(struct net_device *dev, int budget); |
297 | static int epic_poll(struct net_device *dev, int *budget); | 298 | static int epic_poll(struct napi_struct *napi, int budget); |
298 | static irqreturn_t epic_interrupt(int irq, void *dev_instance); | 299 | static irqreturn_t epic_interrupt(int irq, void *dev_instance); |
299 | static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); | 300 | static int netdev_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); |
300 | static const struct ethtool_ops netdev_ethtool_ops; | 301 | static const struct ethtool_ops netdev_ethtool_ops; |
@@ -487,8 +488,7 @@ static int __devinit epic_init_one (struct pci_dev *pdev, | |||
487 | dev->ethtool_ops = &netdev_ethtool_ops; | 488 | dev->ethtool_ops = &netdev_ethtool_ops; |
488 | dev->watchdog_timeo = TX_TIMEOUT; | 489 | dev->watchdog_timeo = TX_TIMEOUT; |
489 | dev->tx_timeout = &epic_tx_timeout; | 490 | dev->tx_timeout = &epic_tx_timeout; |
490 | dev->poll = epic_poll; | 491 | netif_napi_add(dev, &ep->napi, epic_poll, 64); |
491 | dev->weight = 64; | ||
492 | 492 | ||
493 | ret = register_netdev(dev); | 493 | ret = register_netdev(dev); |
494 | if (ret < 0) | 494 | if (ret < 0) |
@@ -660,8 +660,11 @@ static int epic_open(struct net_device *dev) | |||
660 | /* Soft reset the chip. */ | 660 | /* Soft reset the chip. */ |
661 | outl(0x4001, ioaddr + GENCTL); | 661 | outl(0x4001, ioaddr + GENCTL); |
662 | 662 | ||
663 | if ((retval = request_irq(dev->irq, &epic_interrupt, IRQF_SHARED, dev->name, dev))) | 663 | napi_enable(&ep->napi); |
664 | if ((retval = request_irq(dev->irq, &epic_interrupt, IRQF_SHARED, dev->name, dev))) { | ||
665 | napi_disable(&ep->napi); | ||
664 | return retval; | 666 | return retval; |
667 | } | ||
665 | 668 | ||
666 | epic_init_ring(dev); | 669 | epic_init_ring(dev); |
667 | 670 | ||
@@ -1103,9 +1106,9 @@ static irqreturn_t epic_interrupt(int irq, void *dev_instance) | |||
1103 | 1106 | ||
1104 | if ((status & EpicNapiEvent) && !ep->reschedule_in_poll) { | 1107 | if ((status & EpicNapiEvent) && !ep->reschedule_in_poll) { |
1105 | spin_lock(&ep->napi_lock); | 1108 | spin_lock(&ep->napi_lock); |
1106 | if (netif_rx_schedule_prep(dev)) { | 1109 | if (netif_rx_schedule_prep(dev, &ep->napi)) { |
1107 | epic_napi_irq_off(dev, ep); | 1110 | epic_napi_irq_off(dev, ep); |
1108 | __netif_rx_schedule(dev); | 1111 | __netif_rx_schedule(dev, &ep->napi); |
1109 | } else | 1112 | } else |
1110 | ep->reschedule_in_poll++; | 1113 | ep->reschedule_in_poll++; |
1111 | spin_unlock(&ep->napi_lock); | 1114 | spin_unlock(&ep->napi_lock); |
@@ -1257,26 +1260,22 @@ static void epic_rx_err(struct net_device *dev, struct epic_private *ep) | |||
1257 | outw(RxQueued, ioaddr + COMMAND); | 1260 | outw(RxQueued, ioaddr + COMMAND); |
1258 | } | 1261 | } |
1259 | 1262 | ||
1260 | static int epic_poll(struct net_device *dev, int *budget) | 1263 | static int epic_poll(struct napi_struct *napi, int budget) |
1261 | { | 1264 | { |
1262 | struct epic_private *ep = dev->priv; | 1265 | struct epic_private *ep = container_of(napi, struct epic_private, napi); |
1263 | int work_done = 0, orig_budget; | 1266 | struct net_device *dev = ep->mii.dev; |
1267 | int work_done = 0; | ||
1264 | long ioaddr = dev->base_addr; | 1268 | long ioaddr = dev->base_addr; |
1265 | 1269 | ||
1266 | orig_budget = (*budget > dev->quota) ? dev->quota : *budget; | ||
1267 | |||
1268 | rx_action: | 1270 | rx_action: |
1269 | 1271 | ||
1270 | epic_tx(dev, ep); | 1272 | epic_tx(dev, ep); |
1271 | 1273 | ||
1272 | work_done += epic_rx(dev, *budget); | 1274 | work_done += epic_rx(dev, budget); |
1273 | 1275 | ||
1274 | epic_rx_err(dev, ep); | 1276 | epic_rx_err(dev, ep); |
1275 | 1277 | ||
1276 | *budget -= work_done; | 1278 | if (netif_running(dev) && (work_done < budget)) { |
1277 | dev->quota -= work_done; | ||
1278 | |||
1279 | if (netif_running(dev) && (work_done < orig_budget)) { | ||
1280 | unsigned long flags; | 1279 | unsigned long flags; |
1281 | int more; | 1280 | int more; |
1282 | 1281 | ||
@@ -1286,7 +1285,7 @@ rx_action: | |||
1286 | 1285 | ||
1287 | more = ep->reschedule_in_poll; | 1286 | more = ep->reschedule_in_poll; |
1288 | if (!more) { | 1287 | if (!more) { |
1289 | __netif_rx_complete(dev); | 1288 | __netif_rx_complete(dev, napi); |
1290 | outl(EpicNapiEvent, ioaddr + INTSTAT); | 1289 | outl(EpicNapiEvent, ioaddr + INTSTAT); |
1291 | epic_napi_irq_on(dev, ep); | 1290 | epic_napi_irq_on(dev, ep); |
1292 | } else | 1291 | } else |
@@ -1298,7 +1297,7 @@ rx_action: | |||
1298 | goto rx_action; | 1297 | goto rx_action; |
1299 | } | 1298 | } |
1300 | 1299 | ||
1301 | return (work_done >= orig_budget); | 1300 | return work_done; |
1302 | } | 1301 | } |
1303 | 1302 | ||
1304 | static int epic_close(struct net_device *dev) | 1303 | static int epic_close(struct net_device *dev) |
@@ -1309,6 +1308,7 @@ static int epic_close(struct net_device *dev) | |||
1309 | int i; | 1308 | int i; |
1310 | 1309 | ||
1311 | netif_stop_queue(dev); | 1310 | netif_stop_queue(dev); |
1311 | napi_disable(&ep->napi); | ||
1312 | 1312 | ||
1313 | if (debug > 1) | 1313 | if (debug > 1) |
1314 | printk(KERN_DEBUG "%s: Shutting down ethercard, status was %2.2x.\n", | 1314 | printk(KERN_DEBUG "%s: Shutting down ethercard, status was %2.2x.\n", |
diff --git a/drivers/net/fec_8xx/fec_8xx.h b/drivers/net/fec_8xx/fec_8xx.h index 5af60b0f9208..f3b1c6fbba8b 100644 --- a/drivers/net/fec_8xx/fec_8xx.h +++ b/drivers/net/fec_8xx/fec_8xx.h | |||
@@ -105,6 +105,8 @@ struct fec; | |||
105 | struct fec_enet_private { | 105 | struct fec_enet_private { |
106 | spinlock_t lock; /* during all ops except TX pckt processing */ | 106 | spinlock_t lock; /* during all ops except TX pckt processing */ |
107 | spinlock_t tx_lock; /* during fec_start_xmit and fec_tx */ | 107 | spinlock_t tx_lock; /* during fec_start_xmit and fec_tx */ |
108 | struct net_device *dev; | ||
109 | struct napi_struct napi; | ||
108 | int fecno; | 110 | int fecno; |
109 | struct fec *fecp; | 111 | struct fec *fecp; |
110 | const struct fec_platform_info *fpi; | 112 | const struct fec_platform_info *fpi; |
diff --git a/drivers/net/fec_8xx/fec_main.c b/drivers/net/fec_8xx/fec_main.c index e5502af5b8e2..6348fb93ca9c 100644 --- a/drivers/net/fec_8xx/fec_main.c +++ b/drivers/net/fec_8xx/fec_main.c | |||
@@ -465,9 +465,9 @@ void fec_stop(struct net_device *dev) | |||
465 | } | 465 | } |
466 | 466 | ||
467 | /* common receive function */ | 467 | /* common receive function */ |
468 | static int fec_enet_rx_common(struct net_device *dev, int *budget) | 468 | static int fec_enet_rx_common(struct fec_enet_private *ep, |
469 | struct net_device *dev, int budget) | ||
469 | { | 470 | { |
470 | struct fec_enet_private *fep = netdev_priv(dev); | ||
471 | fec_t *fecp = fep->fecp; | 471 | fec_t *fecp = fep->fecp; |
472 | const struct fec_platform_info *fpi = fep->fpi; | 472 | const struct fec_platform_info *fpi = fep->fpi; |
473 | cbd_t *bdp; | 473 | cbd_t *bdp; |
@@ -475,11 +475,8 @@ static int fec_enet_rx_common(struct net_device *dev, int *budget) | |||
475 | int received = 0; | 475 | int received = 0; |
476 | __u16 pkt_len, sc; | 476 | __u16 pkt_len, sc; |
477 | int curidx; | 477 | int curidx; |
478 | int rx_work_limit; | ||
479 | 478 | ||
480 | if (fpi->use_napi) { | 479 | if (fpi->use_napi) { |
481 | rx_work_limit = min(dev->quota, *budget); | ||
482 | |||
483 | if (!netif_running(dev)) | 480 | if (!netif_running(dev)) |
484 | return 0; | 481 | return 0; |
485 | } | 482 | } |
@@ -530,11 +527,6 @@ static int fec_enet_rx_common(struct net_device *dev, int *budget) | |||
530 | BUG_ON(skbn == NULL); | 527 | BUG_ON(skbn == NULL); |
531 | 528 | ||
532 | } else { | 529 | } else { |
533 | |||
534 | /* napi, got packet but no quota */ | ||
535 | if (fpi->use_napi && --rx_work_limit < 0) | ||
536 | break; | ||
537 | |||
538 | skb = fep->rx_skbuff[curidx]; | 530 | skb = fep->rx_skbuff[curidx]; |
539 | BUG_ON(skb == NULL); | 531 | BUG_ON(skb == NULL); |
540 | 532 | ||
@@ -599,25 +591,24 @@ static int fec_enet_rx_common(struct net_device *dev, int *budget) | |||
599 | * able to keep up at the expense of system resources. | 591 | * able to keep up at the expense of system resources. |
600 | */ | 592 | */ |
601 | FW(fecp, r_des_active, 0x01000000); | 593 | FW(fecp, r_des_active, 0x01000000); |
594 | |||
595 | if (received >= budget) | ||
596 | break; | ||
597 | |||
602 | } | 598 | } |
603 | 599 | ||
604 | fep->cur_rx = bdp; | 600 | fep->cur_rx = bdp; |
605 | 601 | ||
606 | if (fpi->use_napi) { | 602 | if (fpi->use_napi) { |
607 | dev->quota -= received; | 603 | if (received < budget) { |
608 | *budget -= received; | 604 | netif_rx_complete(dev, &fep->napi); |
609 | |||
610 | if (rx_work_limit < 0) | ||
611 | return 1; /* not done */ | ||
612 | 605 | ||
613 | /* done */ | 606 | /* enable RX interrupt bits */ |
614 | netif_rx_complete(dev); | 607 | FS(fecp, imask, FEC_ENET_RXF | FEC_ENET_RXB); |
615 | 608 | } | |
616 | /* enable RX interrupt bits */ | ||
617 | FS(fecp, imask, FEC_ENET_RXF | FEC_ENET_RXB); | ||
618 | } | 609 | } |
619 | 610 | ||
620 | return 0; | 611 | return received; |
621 | } | 612 | } |
622 | 613 | ||
623 | static void fec_enet_tx(struct net_device *dev) | 614 | static void fec_enet_tx(struct net_device *dev) |
@@ -743,12 +734,12 @@ fec_enet_interrupt(int irq, void *dev_id) | |||
743 | 734 | ||
744 | if ((int_events & FEC_ENET_RXF) != 0) { | 735 | if ((int_events & FEC_ENET_RXF) != 0) { |
745 | if (!fpi->use_napi) | 736 | if (!fpi->use_napi) |
746 | fec_enet_rx_common(dev, NULL); | 737 | fec_enet_rx_common(fep, dev, ~0); |
747 | else { | 738 | else { |
748 | if (netif_rx_schedule_prep(dev)) { | 739 | if (netif_rx_schedule_prep(dev, &fep->napi)) { |
749 | /* disable rx interrupts */ | 740 | /* disable rx interrupts */ |
750 | FC(fecp, imask, FEC_ENET_RXF | FEC_ENET_RXB); | 741 | FC(fecp, imask, FEC_ENET_RXF | FEC_ENET_RXB); |
751 | __netif_rx_schedule(dev); | 742 | __netif_rx_schedule(dev, &fep->napi); |
752 | } else { | 743 | } else { |
753 | printk(KERN_ERR DRV_MODULE_NAME | 744 | printk(KERN_ERR DRV_MODULE_NAME |
754 | ": %s driver bug! interrupt while in poll!\n", | 745 | ": %s driver bug! interrupt while in poll!\n", |
@@ -893,10 +884,13 @@ static int fec_enet_open(struct net_device *dev) | |||
893 | const struct fec_platform_info *fpi = fep->fpi; | 884 | const struct fec_platform_info *fpi = fep->fpi; |
894 | unsigned long flags; | 885 | unsigned long flags; |
895 | 886 | ||
887 | napi_enable(&fep->napi); | ||
888 | |||
896 | /* Install our interrupt handler. */ | 889 | /* Install our interrupt handler. */ |
897 | if (request_irq(fpi->fec_irq, fec_enet_interrupt, 0, "fec", dev) != 0) { | 890 | if (request_irq(fpi->fec_irq, fec_enet_interrupt, 0, "fec", dev) != 0) { |
898 | printk(KERN_ERR DRV_MODULE_NAME | 891 | printk(KERN_ERR DRV_MODULE_NAME |
899 | ": %s Could not allocate FEC IRQ!", dev->name); | 892 | ": %s Could not allocate FEC IRQ!", dev->name); |
893 | napi_disable(&fep->napi); | ||
900 | return -EINVAL; | 894 | return -EINVAL; |
901 | } | 895 | } |
902 | 896 | ||
@@ -907,6 +901,7 @@ static int fec_enet_open(struct net_device *dev) | |||
907 | printk(KERN_ERR DRV_MODULE_NAME | 901 | printk(KERN_ERR DRV_MODULE_NAME |
908 | ": %s Could not allocate PHY IRQ!", dev->name); | 902 | ": %s Could not allocate PHY IRQ!", dev->name); |
909 | free_irq(fpi->fec_irq, dev); | 903 | free_irq(fpi->fec_irq, dev); |
904 | napi_disable(&fep->napi); | ||
910 | return -EINVAL; | 905 | return -EINVAL; |
911 | } | 906 | } |
912 | 907 | ||
@@ -932,6 +927,7 @@ static int fec_enet_close(struct net_device *dev) | |||
932 | unsigned long flags; | 927 | unsigned long flags; |
933 | 928 | ||
934 | netif_stop_queue(dev); | 929 | netif_stop_queue(dev); |
930 | napi_disable(&fep->napi); | ||
935 | netif_carrier_off(dev); | 931 | netif_carrier_off(dev); |
936 | 932 | ||
937 | if (fpi->use_mdio) | 933 | if (fpi->use_mdio) |
@@ -955,9 +951,12 @@ static struct net_device_stats *fec_enet_get_stats(struct net_device *dev) | |||
955 | return &fep->stats; | 951 | return &fep->stats; |
956 | } | 952 | } |
957 | 953 | ||
958 | static int fec_enet_poll(struct net_device *dev, int *budget) | 954 | static int fec_enet_poll(struct napi_struct *napi, int budget) |
959 | { | 955 | { |
960 | return fec_enet_rx_common(dev, budget); | 956 | struct fec_enet_private *fep = container_of(napi, struct fec_enet_private, napi); |
957 | struct net_device *dev = fep->dev; | ||
958 | |||
959 | return fec_enet_rx_common(fep, dev, budget); | ||
961 | } | 960 | } |
962 | 961 | ||
963 | /*************************************************************************/ | 962 | /*************************************************************************/ |
@@ -1107,6 +1106,7 @@ int fec_8xx_init_one(const struct fec_platform_info *fpi, | |||
1107 | SET_MODULE_OWNER(dev); | 1106 | SET_MODULE_OWNER(dev); |
1108 | 1107 | ||
1109 | fep = netdev_priv(dev); | 1108 | fep = netdev_priv(dev); |
1109 | fep->dev = dev; | ||
1110 | 1110 | ||
1111 | /* partial reset of FEC */ | 1111 | /* partial reset of FEC */ |
1112 | fec_whack_reset(fecp); | 1112 | fec_whack_reset(fecp); |
@@ -1172,10 +1172,9 @@ int fec_8xx_init_one(const struct fec_platform_info *fpi, | |||
1172 | dev->get_stats = fec_enet_get_stats; | 1172 | dev->get_stats = fec_enet_get_stats; |
1173 | dev->set_multicast_list = fec_set_multicast_list; | 1173 | dev->set_multicast_list = fec_set_multicast_list; |
1174 | dev->set_mac_address = fec_set_mac_address; | 1174 | dev->set_mac_address = fec_set_mac_address; |
1175 | if (fpi->use_napi) { | 1175 | netif_napi_add(dev, &fec->napi, |
1176 | dev->poll = fec_enet_poll; | 1176 | fec_enet_poll, fpi->napi_weight); |
1177 | dev->weight = fpi->napi_weight; | 1177 | |
1178 | } | ||
1179 | dev->ethtool_ops = &fec_ethtool_ops; | 1178 | dev->ethtool_ops = &fec_ethtool_ops; |
1180 | dev->do_ioctl = fec_ioctl; | 1179 | dev->do_ioctl = fec_ioctl; |
1181 | 1180 | ||
diff --git a/drivers/net/forcedeth.c b/drivers/net/forcedeth.c index 1938d6dfc863..24c1294614f2 100644 --- a/drivers/net/forcedeth.c +++ b/drivers/net/forcedeth.c | |||
@@ -159,6 +159,8 @@ | |||
159 | #define dprintk(x...) do { } while (0) | 159 | #define dprintk(x...) do { } while (0) |
160 | #endif | 160 | #endif |
161 | 161 | ||
162 | #define TX_WORK_PER_LOOP 64 | ||
163 | #define RX_WORK_PER_LOOP 64 | ||
162 | 164 | ||
163 | /* | 165 | /* |
164 | * Hardware access: | 166 | * Hardware access: |
@@ -745,6 +747,9 @@ struct nv_skb_map { | |||
745 | struct fe_priv { | 747 | struct fe_priv { |
746 | spinlock_t lock; | 748 | spinlock_t lock; |
747 | 749 | ||
750 | struct net_device *dev; | ||
751 | struct napi_struct napi; | ||
752 | |||
748 | /* General data: | 753 | /* General data: |
749 | * Locking: spin_lock(&np->lock); */ | 754 | * Locking: spin_lock(&np->lock); */ |
750 | struct net_device_stats stats; | 755 | struct net_device_stats stats; |
@@ -1586,9 +1591,10 @@ static int nv_alloc_rx_optimized(struct net_device *dev) | |||
1586 | static void nv_do_rx_refill(unsigned long data) | 1591 | static void nv_do_rx_refill(unsigned long data) |
1587 | { | 1592 | { |
1588 | struct net_device *dev = (struct net_device *) data; | 1593 | struct net_device *dev = (struct net_device *) data; |
1594 | struct fe_priv *np = netdev_priv(dev); | ||
1589 | 1595 | ||
1590 | /* Just reschedule NAPI rx processing */ | 1596 | /* Just reschedule NAPI rx processing */ |
1591 | netif_rx_schedule(dev); | 1597 | netif_rx_schedule(dev, &np->napi); |
1592 | } | 1598 | } |
1593 | #else | 1599 | #else |
1594 | static void nv_do_rx_refill(unsigned long data) | 1600 | static void nv_do_rx_refill(unsigned long data) |
@@ -2997,7 +3003,7 @@ static irqreturn_t nv_nic_irq(int foo, void *data) | |||
2997 | 3003 | ||
2998 | #ifdef CONFIG_FORCEDETH_NAPI | 3004 | #ifdef CONFIG_FORCEDETH_NAPI |
2999 | if (events & NVREG_IRQ_RX_ALL) { | 3005 | if (events & NVREG_IRQ_RX_ALL) { |
3000 | netif_rx_schedule(dev); | 3006 | netif_rx_schedule(dev, &np->napi); |
3001 | 3007 | ||
3002 | /* Disable furthur receive irq's */ | 3008 | /* Disable furthur receive irq's */ |
3003 | spin_lock(&np->lock); | 3009 | spin_lock(&np->lock); |
@@ -3010,7 +3016,7 @@ static irqreturn_t nv_nic_irq(int foo, void *data) | |||
3010 | spin_unlock(&np->lock); | 3016 | spin_unlock(&np->lock); |
3011 | } | 3017 | } |
3012 | #else | 3018 | #else |
3013 | if (nv_rx_process(dev, dev->weight)) { | 3019 | if (nv_rx_process(dev, RX_WORK_PER_LOOP)) { |
3014 | if (unlikely(nv_alloc_rx(dev))) { | 3020 | if (unlikely(nv_alloc_rx(dev))) { |
3015 | spin_lock(&np->lock); | 3021 | spin_lock(&np->lock); |
3016 | if (!np->in_shutdown) | 3022 | if (!np->in_shutdown) |
@@ -3079,8 +3085,6 @@ static irqreturn_t nv_nic_irq(int foo, void *data) | |||
3079 | return IRQ_RETVAL(i); | 3085 | return IRQ_RETVAL(i); |
3080 | } | 3086 | } |
3081 | 3087 | ||
3082 | #define TX_WORK_PER_LOOP 64 | ||
3083 | #define RX_WORK_PER_LOOP 64 | ||
3084 | /** | 3088 | /** |
3085 | * All _optimized functions are used to help increase performance | 3089 | * All _optimized functions are used to help increase performance |
3086 | * (reduce CPU and increase throughput). They use descripter version 3, | 3090 | * (reduce CPU and increase throughput). They use descripter version 3, |
@@ -3114,7 +3118,7 @@ static irqreturn_t nv_nic_irq_optimized(int foo, void *data) | |||
3114 | 3118 | ||
3115 | #ifdef CONFIG_FORCEDETH_NAPI | 3119 | #ifdef CONFIG_FORCEDETH_NAPI |
3116 | if (events & NVREG_IRQ_RX_ALL) { | 3120 | if (events & NVREG_IRQ_RX_ALL) { |
3117 | netif_rx_schedule(dev); | 3121 | netif_rx_schedule(dev, &np->napi); |
3118 | 3122 | ||
3119 | /* Disable furthur receive irq's */ | 3123 | /* Disable furthur receive irq's */ |
3120 | spin_lock(&np->lock); | 3124 | spin_lock(&np->lock); |
@@ -3127,7 +3131,7 @@ static irqreturn_t nv_nic_irq_optimized(int foo, void *data) | |||
3127 | spin_unlock(&np->lock); | 3131 | spin_unlock(&np->lock); |
3128 | } | 3132 | } |
3129 | #else | 3133 | #else |
3130 | if (nv_rx_process_optimized(dev, dev->weight)) { | 3134 | if (nv_rx_process_optimized(dev, RX_WORK_PER_LOOP)) { |
3131 | if (unlikely(nv_alloc_rx_optimized(dev))) { | 3135 | if (unlikely(nv_alloc_rx_optimized(dev))) { |
3132 | spin_lock(&np->lock); | 3136 | spin_lock(&np->lock); |
3133 | if (!np->in_shutdown) | 3137 | if (!np->in_shutdown) |
@@ -3245,19 +3249,19 @@ static irqreturn_t nv_nic_irq_tx(int foo, void *data) | |||
3245 | } | 3249 | } |
3246 | 3250 | ||
3247 | #ifdef CONFIG_FORCEDETH_NAPI | 3251 | #ifdef CONFIG_FORCEDETH_NAPI |
3248 | static int nv_napi_poll(struct net_device *dev, int *budget) | 3252 | static int nv_napi_poll(struct napi_struct *napi, int budget) |
3249 | { | 3253 | { |
3250 | int pkts, limit = min(*budget, dev->quota); | 3254 | struct fe_priv *np = container_of(napi, struct fe_priv, napi); |
3251 | struct fe_priv *np = netdev_priv(dev); | 3255 | struct net_device *dev = np->dev; |
3252 | u8 __iomem *base = get_hwbase(dev); | 3256 | u8 __iomem *base = get_hwbase(dev); |
3253 | unsigned long flags; | 3257 | unsigned long flags; |
3254 | int retcode; | 3258 | int pkts, retcode; |
3255 | 3259 | ||
3256 | if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) { | 3260 | if (np->desc_ver == DESC_VER_1 || np->desc_ver == DESC_VER_2) { |
3257 | pkts = nv_rx_process(dev, limit); | 3261 | pkts = nv_rx_process(dev, budget); |
3258 | retcode = nv_alloc_rx(dev); | 3262 | retcode = nv_alloc_rx(dev); |
3259 | } else { | 3263 | } else { |
3260 | pkts = nv_rx_process_optimized(dev, limit); | 3264 | pkts = nv_rx_process_optimized(dev, budget); |
3261 | retcode = nv_alloc_rx_optimized(dev); | 3265 | retcode = nv_alloc_rx_optimized(dev); |
3262 | } | 3266 | } |
3263 | 3267 | ||
@@ -3268,13 +3272,12 @@ static int nv_napi_poll(struct net_device *dev, int *budget) | |||
3268 | spin_unlock_irqrestore(&np->lock, flags); | 3272 | spin_unlock_irqrestore(&np->lock, flags); |
3269 | } | 3273 | } |
3270 | 3274 | ||
3271 | if (pkts < limit) { | 3275 | if (pkts < budget) { |
3272 | /* all done, no more packets present */ | ||
3273 | netif_rx_complete(dev); | ||
3274 | |||
3275 | /* re-enable receive interrupts */ | 3276 | /* re-enable receive interrupts */ |
3276 | spin_lock_irqsave(&np->lock, flags); | 3277 | spin_lock_irqsave(&np->lock, flags); |
3277 | 3278 | ||
3279 | __netif_rx_complete(dev, napi); | ||
3280 | |||
3278 | np->irqmask |= NVREG_IRQ_RX_ALL; | 3281 | np->irqmask |= NVREG_IRQ_RX_ALL; |
3279 | if (np->msi_flags & NV_MSI_X_ENABLED) | 3282 | if (np->msi_flags & NV_MSI_X_ENABLED) |
3280 | writel(NVREG_IRQ_RX_ALL, base + NvRegIrqMask); | 3283 | writel(NVREG_IRQ_RX_ALL, base + NvRegIrqMask); |
@@ -3282,13 +3285,8 @@ static int nv_napi_poll(struct net_device *dev, int *budget) | |||
3282 | writel(np->irqmask, base + NvRegIrqMask); | 3285 | writel(np->irqmask, base + NvRegIrqMask); |
3283 | 3286 | ||
3284 | spin_unlock_irqrestore(&np->lock, flags); | 3287 | spin_unlock_irqrestore(&np->lock, flags); |
3285 | return 0; | ||
3286 | } else { | ||
3287 | /* used up our quantum, so reschedule */ | ||
3288 | dev->quota -= pkts; | ||
3289 | *budget -= pkts; | ||
3290 | return 1; | ||
3291 | } | 3288 | } |
3289 | return pkts; | ||
3292 | } | 3290 | } |
3293 | #endif | 3291 | #endif |
3294 | 3292 | ||
@@ -3296,6 +3294,7 @@ static int nv_napi_poll(struct net_device *dev, int *budget) | |||
3296 | static irqreturn_t nv_nic_irq_rx(int foo, void *data) | 3294 | static irqreturn_t nv_nic_irq_rx(int foo, void *data) |
3297 | { | 3295 | { |
3298 | struct net_device *dev = (struct net_device *) data; | 3296 | struct net_device *dev = (struct net_device *) data; |
3297 | struct fe_priv *np = netdev_priv(dev); | ||
3299 | u8 __iomem *base = get_hwbase(dev); | 3298 | u8 __iomem *base = get_hwbase(dev); |
3300 | u32 events; | 3299 | u32 events; |
3301 | 3300 | ||
@@ -3303,7 +3302,7 @@ static irqreturn_t nv_nic_irq_rx(int foo, void *data) | |||
3303 | writel(NVREG_IRQ_RX_ALL, base + NvRegMSIXIrqStatus); | 3302 | writel(NVREG_IRQ_RX_ALL, base + NvRegMSIXIrqStatus); |
3304 | 3303 | ||
3305 | if (events) { | 3304 | if (events) { |
3306 | netif_rx_schedule(dev); | 3305 | netif_rx_schedule(dev, &np->napi); |
3307 | /* disable receive interrupts on the nic */ | 3306 | /* disable receive interrupts on the nic */ |
3308 | writel(NVREG_IRQ_RX_ALL, base + NvRegIrqMask); | 3307 | writel(NVREG_IRQ_RX_ALL, base + NvRegIrqMask); |
3309 | pci_push(base); | 3308 | pci_push(base); |
@@ -3329,7 +3328,7 @@ static irqreturn_t nv_nic_irq_rx(int foo, void *data) | |||
3329 | if (!(events & np->irqmask)) | 3328 | if (!(events & np->irqmask)) |
3330 | break; | 3329 | break; |
3331 | 3330 | ||
3332 | if (nv_rx_process_optimized(dev, dev->weight)) { | 3331 | if (nv_rx_process_optimized(dev, RX_WORK_PER_LOOP)) { |
3333 | if (unlikely(nv_alloc_rx_optimized(dev))) { | 3332 | if (unlikely(nv_alloc_rx_optimized(dev))) { |
3334 | spin_lock_irqsave(&np->lock, flags); | 3333 | spin_lock_irqsave(&np->lock, flags); |
3335 | if (!np->in_shutdown) | 3334 | if (!np->in_shutdown) |
@@ -4620,7 +4619,9 @@ static void nv_self_test(struct net_device *dev, struct ethtool_test *test, u64 | |||
4620 | if (test->flags & ETH_TEST_FL_OFFLINE) { | 4619 | if (test->flags & ETH_TEST_FL_OFFLINE) { |
4621 | if (netif_running(dev)) { | 4620 | if (netif_running(dev)) { |
4622 | netif_stop_queue(dev); | 4621 | netif_stop_queue(dev); |
4623 | netif_poll_disable(dev); | 4622 | #ifdef CONFIG_FORCEDETH_NAPI |
4623 | napi_disable(&np->napi); | ||
4624 | #endif | ||
4624 | netif_tx_lock_bh(dev); | 4625 | netif_tx_lock_bh(dev); |
4625 | spin_lock_irq(&np->lock); | 4626 | spin_lock_irq(&np->lock); |
4626 | nv_disable_hw_interrupts(dev, np->irqmask); | 4627 | nv_disable_hw_interrupts(dev, np->irqmask); |
@@ -4679,7 +4680,9 @@ static void nv_self_test(struct net_device *dev, struct ethtool_test *test, u64 | |||
4679 | nv_start_rx(dev); | 4680 | nv_start_rx(dev); |
4680 | nv_start_tx(dev); | 4681 | nv_start_tx(dev); |
4681 | netif_start_queue(dev); | 4682 | netif_start_queue(dev); |
4682 | netif_poll_enable(dev); | 4683 | #ifdef CONFIG_FORCEDETH_NAPI |
4684 | napi_enable(&np->napi); | ||
4685 | #endif | ||
4683 | nv_enable_hw_interrupts(dev, np->irqmask); | 4686 | nv_enable_hw_interrupts(dev, np->irqmask); |
4684 | } | 4687 | } |
4685 | } | 4688 | } |
@@ -4911,7 +4914,9 @@ static int nv_open(struct net_device *dev) | |||
4911 | nv_start_rx(dev); | 4914 | nv_start_rx(dev); |
4912 | nv_start_tx(dev); | 4915 | nv_start_tx(dev); |
4913 | netif_start_queue(dev); | 4916 | netif_start_queue(dev); |
4914 | netif_poll_enable(dev); | 4917 | #ifdef CONFIG_FORCEDETH_NAPI |
4918 | napi_enable(&np->napi); | ||
4919 | #endif | ||
4915 | 4920 | ||
4916 | if (ret) { | 4921 | if (ret) { |
4917 | netif_carrier_on(dev); | 4922 | netif_carrier_on(dev); |
@@ -4942,7 +4947,9 @@ static int nv_close(struct net_device *dev) | |||
4942 | spin_lock_irq(&np->lock); | 4947 | spin_lock_irq(&np->lock); |
4943 | np->in_shutdown = 1; | 4948 | np->in_shutdown = 1; |
4944 | spin_unlock_irq(&np->lock); | 4949 | spin_unlock_irq(&np->lock); |
4945 | netif_poll_disable(dev); | 4950 | #ifdef CONFIG_FORCEDETH_NAPI |
4951 | napi_disable(&np->napi); | ||
4952 | #endif | ||
4946 | synchronize_irq(dev->irq); | 4953 | synchronize_irq(dev->irq); |
4947 | 4954 | ||
4948 | del_timer_sync(&np->oom_kick); | 4955 | del_timer_sync(&np->oom_kick); |
@@ -4994,6 +5001,7 @@ static int __devinit nv_probe(struct pci_dev *pci_dev, const struct pci_device_i | |||
4994 | goto out; | 5001 | goto out; |
4995 | 5002 | ||
4996 | np = netdev_priv(dev); | 5003 | np = netdev_priv(dev); |
5004 | np->dev = dev; | ||
4997 | np->pci_dev = pci_dev; | 5005 | np->pci_dev = pci_dev; |
4998 | spin_lock_init(&np->lock); | 5006 | spin_lock_init(&np->lock); |
4999 | SET_MODULE_OWNER(dev); | 5007 | SET_MODULE_OWNER(dev); |
@@ -5155,9 +5163,8 @@ static int __devinit nv_probe(struct pci_dev *pci_dev, const struct pci_device_i | |||
5155 | #ifdef CONFIG_NET_POLL_CONTROLLER | 5163 | #ifdef CONFIG_NET_POLL_CONTROLLER |
5156 | dev->poll_controller = nv_poll_controller; | 5164 | dev->poll_controller = nv_poll_controller; |
5157 | #endif | 5165 | #endif |
5158 | dev->weight = RX_WORK_PER_LOOP; | ||
5159 | #ifdef CONFIG_FORCEDETH_NAPI | 5166 | #ifdef CONFIG_FORCEDETH_NAPI |
5160 | dev->poll = nv_napi_poll; | 5167 | netif_napi_add(dev, &np->napi, nv_napi_poll, RX_WORK_PER_LOOP); |
5161 | #endif | 5168 | #endif |
5162 | SET_ETHTOOL_OPS(dev, &ops); | 5169 | SET_ETHTOOL_OPS(dev, &ops); |
5163 | dev->tx_timeout = nv_tx_timeout; | 5170 | dev->tx_timeout = nv_tx_timeout; |
diff --git a/drivers/net/fs_enet/fs_enet-main.c b/drivers/net/fs_enet/fs_enet-main.c index a4a2a0ea43d3..c509cb13222d 100644 --- a/drivers/net/fs_enet/fs_enet-main.c +++ b/drivers/net/fs_enet/fs_enet-main.c | |||
@@ -70,18 +70,16 @@ static void fs_set_multicast_list(struct net_device *dev) | |||
70 | } | 70 | } |
71 | 71 | ||
72 | /* NAPI receive function */ | 72 | /* NAPI receive function */ |
73 | static int fs_enet_rx_napi(struct net_device *dev, int *budget) | 73 | static int fs_enet_rx_napi(struct napi_struct *napi, int budget) |
74 | { | 74 | { |
75 | struct fs_enet_private *fep = netdev_priv(dev); | 75 | struct fs_enet_private *fep = container_of(napi, struct fs_enet_private, napi); |
76 | struct net_device *dev = to_net_dev(fep->dev); | ||
76 | const struct fs_platform_info *fpi = fep->fpi; | 77 | const struct fs_platform_info *fpi = fep->fpi; |
77 | cbd_t *bdp; | 78 | cbd_t *bdp; |
78 | struct sk_buff *skb, *skbn, *skbt; | 79 | struct sk_buff *skb, *skbn, *skbt; |
79 | int received = 0; | 80 | int received = 0; |
80 | u16 pkt_len, sc; | 81 | u16 pkt_len, sc; |
81 | int curidx; | 82 | int curidx; |
82 | int rx_work_limit = 0; /* pacify gcc */ | ||
83 | |||
84 | rx_work_limit = min(dev->quota, *budget); | ||
85 | 83 | ||
86 | if (!netif_running(dev)) | 84 | if (!netif_running(dev)) |
87 | return 0; | 85 | return 0; |
@@ -96,7 +94,6 @@ static int fs_enet_rx_napi(struct net_device *dev, int *budget) | |||
96 | (*fep->ops->napi_clear_rx_event)(dev); | 94 | (*fep->ops->napi_clear_rx_event)(dev); |
97 | 95 | ||
98 | while (((sc = CBDR_SC(bdp)) & BD_ENET_RX_EMPTY) == 0) { | 96 | while (((sc = CBDR_SC(bdp)) & BD_ENET_RX_EMPTY) == 0) { |
99 | |||
100 | curidx = bdp - fep->rx_bd_base; | 97 | curidx = bdp - fep->rx_bd_base; |
101 | 98 | ||
102 | /* | 99 | /* |
@@ -136,11 +133,6 @@ static int fs_enet_rx_napi(struct net_device *dev, int *budget) | |||
136 | skbn = skb; | 133 | skbn = skb; |
137 | 134 | ||
138 | } else { | 135 | } else { |
139 | |||
140 | /* napi, got packet but no quota */ | ||
141 | if (--rx_work_limit < 0) | ||
142 | break; | ||
143 | |||
144 | skb = fep->rx_skbuff[curidx]; | 136 | skb = fep->rx_skbuff[curidx]; |
145 | 137 | ||
146 | dma_unmap_single(fep->dev, CBDR_BUFADDR(bdp), | 138 | dma_unmap_single(fep->dev, CBDR_BUFADDR(bdp), |
@@ -199,22 +191,19 @@ static int fs_enet_rx_napi(struct net_device *dev, int *budget) | |||
199 | bdp = fep->rx_bd_base; | 191 | bdp = fep->rx_bd_base; |
200 | 192 | ||
201 | (*fep->ops->rx_bd_done)(dev); | 193 | (*fep->ops->rx_bd_done)(dev); |
194 | |||
195 | if (received >= budget) | ||
196 | break; | ||
202 | } | 197 | } |
203 | 198 | ||
204 | fep->cur_rx = bdp; | 199 | fep->cur_rx = bdp; |
205 | 200 | ||
206 | dev->quota -= received; | 201 | if (received >= budget) { |
207 | *budget -= received; | 202 | /* done */ |
208 | 203 | netif_rx_complete(dev, napi); | |
209 | if (rx_work_limit < 0) | 204 | (*fep->ops->napi_enable_rx)(dev); |
210 | return 1; /* not done */ | 205 | } |
211 | 206 | return received; | |
212 | /* done */ | ||
213 | netif_rx_complete(dev); | ||
214 | |||
215 | (*fep->ops->napi_enable_rx)(dev); | ||
216 | |||
217 | return 0; | ||
218 | } | 207 | } |
219 | 208 | ||
220 | /* non NAPI receive function */ | 209 | /* non NAPI receive function */ |
@@ -470,7 +459,7 @@ fs_enet_interrupt(int irq, void *dev_id) | |||
470 | if (!fpi->use_napi) | 459 | if (!fpi->use_napi) |
471 | fs_enet_rx_non_napi(dev); | 460 | fs_enet_rx_non_napi(dev); |
472 | else { | 461 | else { |
473 | napi_ok = netif_rx_schedule_prep(dev); | 462 | napi_ok = napi_schedule_prep(&fep->napi); |
474 | 463 | ||
475 | (*fep->ops->napi_disable_rx)(dev); | 464 | (*fep->ops->napi_disable_rx)(dev); |
476 | (*fep->ops->clear_int_events)(dev, fep->ev_napi_rx); | 465 | (*fep->ops->clear_int_events)(dev, fep->ev_napi_rx); |
@@ -478,7 +467,7 @@ fs_enet_interrupt(int irq, void *dev_id) | |||
478 | /* NOTE: it is possible for FCCs in NAPI mode */ | 467 | /* NOTE: it is possible for FCCs in NAPI mode */ |
479 | /* to submit a spurious interrupt while in poll */ | 468 | /* to submit a spurious interrupt while in poll */ |
480 | if (napi_ok) | 469 | if (napi_ok) |
481 | __netif_rx_schedule(dev); | 470 | __netif_rx_schedule(dev, &fep->napi); |
482 | } | 471 | } |
483 | } | 472 | } |
484 | 473 | ||
@@ -799,18 +788,22 @@ static int fs_enet_open(struct net_device *dev) | |||
799 | int r; | 788 | int r; |
800 | int err; | 789 | int err; |
801 | 790 | ||
791 | napi_enable(&fep->napi); | ||
792 | |||
802 | /* Install our interrupt handler. */ | 793 | /* Install our interrupt handler. */ |
803 | r = fs_request_irq(dev, fep->interrupt, "fs_enet-mac", fs_enet_interrupt); | 794 | r = fs_request_irq(dev, fep->interrupt, "fs_enet-mac", fs_enet_interrupt); |
804 | if (r != 0) { | 795 | if (r != 0) { |
805 | printk(KERN_ERR DRV_MODULE_NAME | 796 | printk(KERN_ERR DRV_MODULE_NAME |
806 | ": %s Could not allocate FS_ENET IRQ!", dev->name); | 797 | ": %s Could not allocate FS_ENET IRQ!", dev->name); |
798 | napi_disable(&fep->napi); | ||
807 | return -EINVAL; | 799 | return -EINVAL; |
808 | } | 800 | } |
809 | 801 | ||
810 | err = fs_init_phy(dev); | 802 | err = fs_init_phy(dev); |
811 | if(err) | 803 | if(err) { |
804 | napi_disable(&fep->napi); | ||
812 | return err; | 805 | return err; |
813 | 806 | } | |
814 | phy_start(fep->phydev); | 807 | phy_start(fep->phydev); |
815 | 808 | ||
816 | return 0; | 809 | return 0; |
@@ -823,6 +816,7 @@ static int fs_enet_close(struct net_device *dev) | |||
823 | 816 | ||
824 | netif_stop_queue(dev); | 817 | netif_stop_queue(dev); |
825 | netif_carrier_off(dev); | 818 | netif_carrier_off(dev); |
819 | napi_disable(&fep->napi); | ||
826 | phy_stop(fep->phydev); | 820 | phy_stop(fep->phydev); |
827 | 821 | ||
828 | spin_lock_irqsave(&fep->lock, flags); | 822 | spin_lock_irqsave(&fep->lock, flags); |
@@ -1047,10 +1041,9 @@ static struct net_device *fs_init_instance(struct device *dev, | |||
1047 | ndev->stop = fs_enet_close; | 1041 | ndev->stop = fs_enet_close; |
1048 | ndev->get_stats = fs_enet_get_stats; | 1042 | ndev->get_stats = fs_enet_get_stats; |
1049 | ndev->set_multicast_list = fs_set_multicast_list; | 1043 | ndev->set_multicast_list = fs_set_multicast_list; |
1050 | if (fpi->use_napi) { | 1044 | netif_napi_add(ndev, &fep->napi, |
1051 | ndev->poll = fs_enet_rx_napi; | 1045 | fs_enet_rx_napi, fpi->napi_weight); |
1052 | ndev->weight = fpi->napi_weight; | 1046 | |
1053 | } | ||
1054 | ndev->ethtool_ops = &fs_ethtool_ops; | 1047 | ndev->ethtool_ops = &fs_ethtool_ops; |
1055 | ndev->do_ioctl = fs_ioctl; | 1048 | ndev->do_ioctl = fs_ioctl; |
1056 | 1049 | ||
diff --git a/drivers/net/fs_enet/fs_enet.h b/drivers/net/fs_enet/fs_enet.h index 569be225cd05..46d0606b1439 100644 --- a/drivers/net/fs_enet/fs_enet.h +++ b/drivers/net/fs_enet/fs_enet.h | |||
@@ -121,6 +121,7 @@ struct fs_enet_mii_bus { | |||
121 | }; | 121 | }; |
122 | 122 | ||
123 | struct fs_enet_private { | 123 | struct fs_enet_private { |
124 | struct napi_struct napi; | ||
124 | struct device *dev; /* pointer back to the device (must be initialized first) */ | 125 | struct device *dev; /* pointer back to the device (must be initialized first) */ |
125 | spinlock_t lock; /* during all ops except TX pckt processing */ | 126 | spinlock_t lock; /* during all ops except TX pckt processing */ |
126 | spinlock_t tx_lock; /* during fs_start_xmit and fs_tx */ | 127 | spinlock_t tx_lock; /* during fs_start_xmit and fs_tx */ |
diff --git a/drivers/net/gianfar.c b/drivers/net/gianfar.c index f92690555dd9..bd2de325bbdd 100644 --- a/drivers/net/gianfar.c +++ b/drivers/net/gianfar.c | |||
@@ -134,7 +134,7 @@ static void gfar_configure_serdes(struct net_device *dev); | |||
134 | extern int gfar_local_mdio_write(struct gfar_mii *regs, int mii_id, int regnum, u16 value); | 134 | extern int gfar_local_mdio_write(struct gfar_mii *regs, int mii_id, int regnum, u16 value); |
135 | extern int gfar_local_mdio_read(struct gfar_mii *regs, int mii_id, int regnum); | 135 | extern int gfar_local_mdio_read(struct gfar_mii *regs, int mii_id, int regnum); |
136 | #ifdef CONFIG_GFAR_NAPI | 136 | #ifdef CONFIG_GFAR_NAPI |
137 | static int gfar_poll(struct net_device *dev, int *budget); | 137 | static int gfar_poll(struct napi_struct *napi, int budget); |
138 | #endif | 138 | #endif |
139 | #ifdef CONFIG_NET_POLL_CONTROLLER | 139 | #ifdef CONFIG_NET_POLL_CONTROLLER |
140 | static void gfar_netpoll(struct net_device *dev); | 140 | static void gfar_netpoll(struct net_device *dev); |
@@ -188,6 +188,7 @@ static int gfar_probe(struct platform_device *pdev) | |||
188 | return -ENOMEM; | 188 | return -ENOMEM; |
189 | 189 | ||
190 | priv = netdev_priv(dev); | 190 | priv = netdev_priv(dev); |
191 | priv->dev = dev; | ||
191 | 192 | ||
192 | /* Set the info in the priv to the current info */ | 193 | /* Set the info in the priv to the current info */ |
193 | priv->einfo = einfo; | 194 | priv->einfo = einfo; |
@@ -261,10 +262,7 @@ static int gfar_probe(struct platform_device *pdev) | |||
261 | dev->hard_start_xmit = gfar_start_xmit; | 262 | dev->hard_start_xmit = gfar_start_xmit; |
262 | dev->tx_timeout = gfar_timeout; | 263 | dev->tx_timeout = gfar_timeout; |
263 | dev->watchdog_timeo = TX_TIMEOUT; | 264 | dev->watchdog_timeo = TX_TIMEOUT; |
264 | #ifdef CONFIG_GFAR_NAPI | 265 | netif_napi_add(dev, &priv->napi, gfar_poll, GFAR_DEV_WEIGHT); |
265 | dev->poll = gfar_poll; | ||
266 | dev->weight = GFAR_DEV_WEIGHT; | ||
267 | #endif | ||
268 | #ifdef CONFIG_NET_POLL_CONTROLLER | 266 | #ifdef CONFIG_NET_POLL_CONTROLLER |
269 | dev->poll_controller = gfar_netpoll; | 267 | dev->poll_controller = gfar_netpoll; |
270 | #endif | 268 | #endif |
@@ -939,6 +937,8 @@ static int gfar_enet_open(struct net_device *dev) | |||
939 | { | 937 | { |
940 | int err; | 938 | int err; |
941 | 939 | ||
940 | napi_enable(&priv->napi); | ||
941 | |||
942 | /* Initialize a bunch of registers */ | 942 | /* Initialize a bunch of registers */ |
943 | init_registers(dev); | 943 | init_registers(dev); |
944 | 944 | ||
@@ -946,10 +946,14 @@ static int gfar_enet_open(struct net_device *dev) | |||
946 | 946 | ||
947 | err = init_phy(dev); | 947 | err = init_phy(dev); |
948 | 948 | ||
949 | if(err) | 949 | if(err) { |
950 | napi_disable(&priv->napi); | ||
950 | return err; | 951 | return err; |
952 | } | ||
951 | 953 | ||
952 | err = startup_gfar(dev); | 954 | err = startup_gfar(dev); |
955 | if (err) | ||
956 | napi_disable(&priv->napi); | ||
953 | 957 | ||
954 | netif_start_queue(dev); | 958 | netif_start_queue(dev); |
955 | 959 | ||
@@ -1102,6 +1106,9 @@ static int gfar_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
1102 | static int gfar_close(struct net_device *dev) | 1106 | static int gfar_close(struct net_device *dev) |
1103 | { | 1107 | { |
1104 | struct gfar_private *priv = netdev_priv(dev); | 1108 | struct gfar_private *priv = netdev_priv(dev); |
1109 | |||
1110 | napi_disable(&priv->napi); | ||
1111 | |||
1105 | stop_gfar(dev); | 1112 | stop_gfar(dev); |
1106 | 1113 | ||
1107 | /* Disconnect from the PHY */ | 1114 | /* Disconnect from the PHY */ |
@@ -1318,7 +1325,7 @@ struct sk_buff * gfar_new_skb(struct net_device *dev, struct rxbd8 *bdp) | |||
1318 | return NULL; | 1325 | return NULL; |
1319 | 1326 | ||
1320 | alignamount = RXBUF_ALIGNMENT - | 1327 | alignamount = RXBUF_ALIGNMENT - |
1321 | (((unsigned) skb->data) & (RXBUF_ALIGNMENT - 1)); | 1328 | (((unsigned long) skb->data) & (RXBUF_ALIGNMENT - 1)); |
1322 | 1329 | ||
1323 | /* We need the data buffer to be aligned properly. We will reserve | 1330 | /* We need the data buffer to be aligned properly. We will reserve |
1324 | * as many bytes as needed to align the data properly | 1331 | * as many bytes as needed to align the data properly |
@@ -1390,12 +1397,12 @@ irqreturn_t gfar_receive(int irq, void *dev_id) | |||
1390 | 1397 | ||
1391 | /* support NAPI */ | 1398 | /* support NAPI */ |
1392 | #ifdef CONFIG_GFAR_NAPI | 1399 | #ifdef CONFIG_GFAR_NAPI |
1393 | if (netif_rx_schedule_prep(dev)) { | 1400 | if (netif_rx_schedule_prep(dev, &priv->napi)) { |
1394 | tempval = gfar_read(&priv->regs->imask); | 1401 | tempval = gfar_read(&priv->regs->imask); |
1395 | tempval &= IMASK_RX_DISABLED; | 1402 | tempval &= IMASK_RX_DISABLED; |
1396 | gfar_write(&priv->regs->imask, tempval); | 1403 | gfar_write(&priv->regs->imask, tempval); |
1397 | 1404 | ||
1398 | __netif_rx_schedule(dev); | 1405 | __netif_rx_schedule(dev, &priv->napi); |
1399 | } else { | 1406 | } else { |
1400 | if (netif_msg_rx_err(priv)) | 1407 | if (netif_msg_rx_err(priv)) |
1401 | printk(KERN_DEBUG "%s: receive called twice (%x)[%x]\n", | 1408 | printk(KERN_DEBUG "%s: receive called twice (%x)[%x]\n", |
@@ -1569,23 +1576,16 @@ int gfar_clean_rx_ring(struct net_device *dev, int rx_work_limit) | |||
1569 | } | 1576 | } |
1570 | 1577 | ||
1571 | #ifdef CONFIG_GFAR_NAPI | 1578 | #ifdef CONFIG_GFAR_NAPI |
1572 | static int gfar_poll(struct net_device *dev, int *budget) | 1579 | static int gfar_poll(struct napi_struct *napi, int budget) |
1573 | { | 1580 | { |
1581 | struct gfar_private *priv = container_of(napi, struct gfar_private, napi); | ||
1582 | struct net_device *dev = priv->dev; | ||
1574 | int howmany; | 1583 | int howmany; |
1575 | struct gfar_private *priv = netdev_priv(dev); | ||
1576 | int rx_work_limit = *budget; | ||
1577 | |||
1578 | if (rx_work_limit > dev->quota) | ||
1579 | rx_work_limit = dev->quota; | ||
1580 | 1584 | ||
1581 | howmany = gfar_clean_rx_ring(dev, rx_work_limit); | 1585 | howmany = gfar_clean_rx_ring(dev, budget); |
1582 | 1586 | ||
1583 | dev->quota -= howmany; | 1587 | if (howmany < budget) { |
1584 | rx_work_limit -= howmany; | 1588 | netif_rx_complete(dev, napi); |
1585 | *budget -= howmany; | ||
1586 | |||
1587 | if (rx_work_limit > 0) { | ||
1588 | netif_rx_complete(dev); | ||
1589 | 1589 | ||
1590 | /* Clear the halt bit in RSTAT */ | 1590 | /* Clear the halt bit in RSTAT */ |
1591 | gfar_write(&priv->regs->rstat, RSTAT_CLEAR_RHALT); | 1591 | gfar_write(&priv->regs->rstat, RSTAT_CLEAR_RHALT); |
@@ -1601,8 +1601,7 @@ static int gfar_poll(struct net_device *dev, int *budget) | |||
1601 | gfar_write(&priv->regs->rxic, 0); | 1601 | gfar_write(&priv->regs->rxic, 0); |
1602 | } | 1602 | } |
1603 | 1603 | ||
1604 | /* Return 1 if there's more work to do */ | 1604 | return howmany; |
1605 | return (rx_work_limit > 0) ? 0 : 1; | ||
1606 | } | 1605 | } |
1607 | #endif | 1606 | #endif |
1608 | 1607 | ||
diff --git a/drivers/net/gianfar.h b/drivers/net/gianfar.h index d8e779c102fa..b8714e00482d 100644 --- a/drivers/net/gianfar.h +++ b/drivers/net/gianfar.h | |||
@@ -691,6 +691,9 @@ struct gfar_private { | |||
691 | /* RX Locked fields */ | 691 | /* RX Locked fields */ |
692 | spinlock_t rxlock; | 692 | spinlock_t rxlock; |
693 | 693 | ||
694 | struct net_device *dev; | ||
695 | struct napi_struct napi; | ||
696 | |||
694 | /* skb array and index */ | 697 | /* skb array and index */ |
695 | struct sk_buff ** rx_skbuff; | 698 | struct sk_buff ** rx_skbuff; |
696 | u16 skb_currx; | 699 | u16 skb_currx; |
diff --git a/drivers/net/ibmveth.c b/drivers/net/ibmveth.c index acba90f1638e..78e28ada1e21 100644 --- a/drivers/net/ibmveth.c +++ b/drivers/net/ibmveth.c | |||
@@ -83,7 +83,7 @@ | |||
83 | static int ibmveth_open(struct net_device *dev); | 83 | static int ibmveth_open(struct net_device *dev); |
84 | static int ibmveth_close(struct net_device *dev); | 84 | static int ibmveth_close(struct net_device *dev); |
85 | static int ibmveth_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd); | 85 | static int ibmveth_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd); |
86 | static int ibmveth_poll(struct net_device *dev, int *budget); | 86 | static int ibmveth_poll(struct napi_struct *napi, int budget); |
87 | static int ibmveth_start_xmit(struct sk_buff *skb, struct net_device *dev); | 87 | static int ibmveth_start_xmit(struct sk_buff *skb, struct net_device *dev); |
88 | static struct net_device_stats *ibmveth_get_stats(struct net_device *dev); | 88 | static struct net_device_stats *ibmveth_get_stats(struct net_device *dev); |
89 | static void ibmveth_set_multicast_list(struct net_device *dev); | 89 | static void ibmveth_set_multicast_list(struct net_device *dev); |
@@ -480,6 +480,8 @@ static int ibmveth_open(struct net_device *netdev) | |||
480 | 480 | ||
481 | ibmveth_debug_printk("open starting\n"); | 481 | ibmveth_debug_printk("open starting\n"); |
482 | 482 | ||
483 | napi_enable(&adapter->napi); | ||
484 | |||
483 | for(i = 0; i<IbmVethNumBufferPools; i++) | 485 | for(i = 0; i<IbmVethNumBufferPools; i++) |
484 | rxq_entries += adapter->rx_buff_pool[i].size; | 486 | rxq_entries += adapter->rx_buff_pool[i].size; |
485 | 487 | ||
@@ -489,6 +491,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
489 | if(!adapter->buffer_list_addr || !adapter->filter_list_addr) { | 491 | if(!adapter->buffer_list_addr || !adapter->filter_list_addr) { |
490 | ibmveth_error_printk("unable to allocate filter or buffer list pages\n"); | 492 | ibmveth_error_printk("unable to allocate filter or buffer list pages\n"); |
491 | ibmveth_cleanup(adapter); | 493 | ibmveth_cleanup(adapter); |
494 | napi_disable(&adapter->napi); | ||
492 | return -ENOMEM; | 495 | return -ENOMEM; |
493 | } | 496 | } |
494 | 497 | ||
@@ -498,6 +501,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
498 | if(!adapter->rx_queue.queue_addr) { | 501 | if(!adapter->rx_queue.queue_addr) { |
499 | ibmveth_error_printk("unable to allocate rx queue pages\n"); | 502 | ibmveth_error_printk("unable to allocate rx queue pages\n"); |
500 | ibmveth_cleanup(adapter); | 503 | ibmveth_cleanup(adapter); |
504 | napi_disable(&adapter->napi); | ||
501 | return -ENOMEM; | 505 | return -ENOMEM; |
502 | } | 506 | } |
503 | 507 | ||
@@ -514,6 +518,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
514 | (dma_mapping_error(adapter->rx_queue.queue_dma))) { | 518 | (dma_mapping_error(adapter->rx_queue.queue_dma))) { |
515 | ibmveth_error_printk("unable to map filter or buffer list pages\n"); | 519 | ibmveth_error_printk("unable to map filter or buffer list pages\n"); |
516 | ibmveth_cleanup(adapter); | 520 | ibmveth_cleanup(adapter); |
521 | napi_disable(&adapter->napi); | ||
517 | return -ENOMEM; | 522 | return -ENOMEM; |
518 | } | 523 | } |
519 | 524 | ||
@@ -545,6 +550,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
545 | rxq_desc.desc, | 550 | rxq_desc.desc, |
546 | mac_address); | 551 | mac_address); |
547 | ibmveth_cleanup(adapter); | 552 | ibmveth_cleanup(adapter); |
553 | napi_disable(&adapter->napi); | ||
548 | return -ENONET; | 554 | return -ENONET; |
549 | } | 555 | } |
550 | 556 | ||
@@ -555,6 +561,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
555 | ibmveth_error_printk("unable to alloc pool\n"); | 561 | ibmveth_error_printk("unable to alloc pool\n"); |
556 | adapter->rx_buff_pool[i].active = 0; | 562 | adapter->rx_buff_pool[i].active = 0; |
557 | ibmveth_cleanup(adapter); | 563 | ibmveth_cleanup(adapter); |
564 | napi_disable(&adapter->napi); | ||
558 | return -ENOMEM ; | 565 | return -ENOMEM ; |
559 | } | 566 | } |
560 | } | 567 | } |
@@ -567,6 +574,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
567 | } while (H_IS_LONG_BUSY(rc) || (rc == H_BUSY)); | 574 | } while (H_IS_LONG_BUSY(rc) || (rc == H_BUSY)); |
568 | 575 | ||
569 | ibmveth_cleanup(adapter); | 576 | ibmveth_cleanup(adapter); |
577 | napi_disable(&adapter->napi); | ||
570 | return rc; | 578 | return rc; |
571 | } | 579 | } |
572 | 580 | ||
@@ -587,6 +595,8 @@ static int ibmveth_close(struct net_device *netdev) | |||
587 | 595 | ||
588 | ibmveth_debug_printk("close starting\n"); | 596 | ibmveth_debug_printk("close starting\n"); |
589 | 597 | ||
598 | napi_disable(&adapter->napi); | ||
599 | |||
590 | if (!adapter->pool_config) | 600 | if (!adapter->pool_config) |
591 | netif_stop_queue(netdev); | 601 | netif_stop_queue(netdev); |
592 | 602 | ||
@@ -767,80 +777,68 @@ out: spin_lock_irqsave(&adapter->stats_lock, flags); | |||
767 | return 0; | 777 | return 0; |
768 | } | 778 | } |
769 | 779 | ||
770 | static int ibmveth_poll(struct net_device *netdev, int *budget) | 780 | static int ibmveth_poll(struct napi_struct *napi, int budget) |
771 | { | 781 | { |
772 | struct ibmveth_adapter *adapter = netdev->priv; | 782 | struct ibmveth_adapter *adapter = container_of(napi, struct ibmveth_adapter, napi); |
773 | int max_frames_to_process = netdev->quota; | 783 | struct net_device *netdev = adapter->netdev; |
774 | int frames_processed = 0; | 784 | int frames_processed = 0; |
775 | int more_work = 1; | ||
776 | unsigned long lpar_rc; | 785 | unsigned long lpar_rc; |
777 | 786 | ||
778 | restart_poll: | 787 | restart_poll: |
779 | do { | 788 | do { |
780 | struct net_device *netdev = adapter->netdev; | 789 | struct sk_buff *skb; |
781 | |||
782 | if(ibmveth_rxq_pending_buffer(adapter)) { | ||
783 | struct sk_buff *skb; | ||
784 | 790 | ||
785 | rmb(); | 791 | if (!ibmveth_rxq_pending_buffer(adapter)) |
792 | break; | ||
786 | 793 | ||
787 | if(!ibmveth_rxq_buffer_valid(adapter)) { | 794 | rmb(); |
788 | wmb(); /* suggested by larson1 */ | 795 | if (!ibmveth_rxq_buffer_valid(adapter)) { |
789 | adapter->rx_invalid_buffer++; | 796 | wmb(); /* suggested by larson1 */ |
790 | ibmveth_debug_printk("recycling invalid buffer\n"); | 797 | adapter->rx_invalid_buffer++; |
791 | ibmveth_rxq_recycle_buffer(adapter); | 798 | ibmveth_debug_printk("recycling invalid buffer\n"); |
792 | } else { | 799 | ibmveth_rxq_recycle_buffer(adapter); |
793 | int length = ibmveth_rxq_frame_length(adapter); | 800 | } else { |
794 | int offset = ibmveth_rxq_frame_offset(adapter); | 801 | int length = ibmveth_rxq_frame_length(adapter); |
795 | skb = ibmveth_rxq_get_buffer(adapter); | 802 | int offset = ibmveth_rxq_frame_offset(adapter); |
803 | skb = ibmveth_rxq_get_buffer(adapter); | ||
796 | 804 | ||
797 | ibmveth_rxq_harvest_buffer(adapter); | 805 | ibmveth_rxq_harvest_buffer(adapter); |
798 | 806 | ||
799 | skb_reserve(skb, offset); | 807 | skb_reserve(skb, offset); |
800 | skb_put(skb, length); | 808 | skb_put(skb, length); |
801 | skb->protocol = eth_type_trans(skb, netdev); | 809 | skb->protocol = eth_type_trans(skb, netdev); |
802 | 810 | ||
803 | netif_receive_skb(skb); /* send it up */ | 811 | netif_receive_skb(skb); /* send it up */ |
804 | 812 | ||
805 | adapter->stats.rx_packets++; | 813 | adapter->stats.rx_packets++; |
806 | adapter->stats.rx_bytes += length; | 814 | adapter->stats.rx_bytes += length; |
807 | frames_processed++; | 815 | frames_processed++; |
808 | netdev->last_rx = jiffies; | 816 | netdev->last_rx = jiffies; |
809 | } | ||
810 | } else { | ||
811 | more_work = 0; | ||
812 | } | 817 | } |
813 | } while(more_work && (frames_processed < max_frames_to_process)); | 818 | } while (frames_processed < budget); |
814 | 819 | ||
815 | ibmveth_replenish_task(adapter); | 820 | ibmveth_replenish_task(adapter); |
816 | 821 | ||
817 | if(more_work) { | 822 | if (frames_processed < budget) { |
818 | /* more work to do - return that we are not done yet */ | 823 | /* We think we are done - reenable interrupts, |
819 | netdev->quota -= frames_processed; | 824 | * then check once more to make sure we are done. |
820 | *budget -= frames_processed; | 825 | */ |
821 | return 1; | 826 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, |
822 | } | 827 | VIO_IRQ_ENABLE); |
823 | |||
824 | /* we think we are done - reenable interrupts, then check once more to make sure we are done */ | ||
825 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_ENABLE); | ||
826 | 828 | ||
827 | ibmveth_assert(lpar_rc == H_SUCCESS); | 829 | ibmveth_assert(lpar_rc == H_SUCCESS); |
828 | 830 | ||
829 | netif_rx_complete(netdev); | 831 | netif_rx_complete(netdev, napi); |
830 | 832 | ||
831 | if(ibmveth_rxq_pending_buffer(adapter) && netif_rx_reschedule(netdev, frames_processed)) | 833 | if (ibmveth_rxq_pending_buffer(adapter) && |
832 | { | 834 | netif_rx_reschedule(netdev, napi)) { |
833 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); | 835 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, |
834 | ibmveth_assert(lpar_rc == H_SUCCESS); | 836 | VIO_IRQ_DISABLE); |
835 | more_work = 1; | 837 | goto restart_poll; |
836 | goto restart_poll; | 838 | } |
837 | } | 839 | } |
838 | 840 | ||
839 | netdev->quota -= frames_processed; | 841 | return frames_processed; |
840 | *budget -= frames_processed; | ||
841 | |||
842 | /* we really are done */ | ||
843 | return 0; | ||
844 | } | 842 | } |
845 | 843 | ||
846 | static irqreturn_t ibmveth_interrupt(int irq, void *dev_instance) | 844 | static irqreturn_t ibmveth_interrupt(int irq, void *dev_instance) |
@@ -849,10 +847,11 @@ static irqreturn_t ibmveth_interrupt(int irq, void *dev_instance) | |||
849 | struct ibmveth_adapter *adapter = netdev->priv; | 847 | struct ibmveth_adapter *adapter = netdev->priv; |
850 | unsigned long lpar_rc; | 848 | unsigned long lpar_rc; |
851 | 849 | ||
852 | if(netif_rx_schedule_prep(netdev)) { | 850 | if (netif_rx_schedule_prep(netdev, &adapter->napi)) { |
853 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); | 851 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, |
852 | VIO_IRQ_DISABLE); | ||
854 | ibmveth_assert(lpar_rc == H_SUCCESS); | 853 | ibmveth_assert(lpar_rc == H_SUCCESS); |
855 | __netif_rx_schedule(netdev); | 854 | __netif_rx_schedule(netdev, &adapter->napi); |
856 | } | 855 | } |
857 | return IRQ_HANDLED; | 856 | return IRQ_HANDLED; |
858 | } | 857 | } |
@@ -1004,6 +1003,8 @@ static int __devinit ibmveth_probe(struct vio_dev *dev, const struct vio_device_ | |||
1004 | adapter->mcastFilterSize= *mcastFilterSize_p; | 1003 | adapter->mcastFilterSize= *mcastFilterSize_p; |
1005 | adapter->pool_config = 0; | 1004 | adapter->pool_config = 0; |
1006 | 1005 | ||
1006 | netif_napi_add(netdev, &adapter->napi, ibmveth_poll, 16); | ||
1007 | |||
1007 | /* Some older boxes running PHYP non-natively have an OF that | 1008 | /* Some older boxes running PHYP non-natively have an OF that |
1008 | returns a 8-byte local-mac-address field (and the first | 1009 | returns a 8-byte local-mac-address field (and the first |
1009 | 2 bytes have to be ignored) while newer boxes' OF return | 1010 | 2 bytes have to be ignored) while newer boxes' OF return |
@@ -1020,8 +1021,6 @@ static int __devinit ibmveth_probe(struct vio_dev *dev, const struct vio_device_ | |||
1020 | 1021 | ||
1021 | netdev->irq = dev->irq; | 1022 | netdev->irq = dev->irq; |
1022 | netdev->open = ibmveth_open; | 1023 | netdev->open = ibmveth_open; |
1023 | netdev->poll = ibmveth_poll; | ||
1024 | netdev->weight = 16; | ||
1025 | netdev->stop = ibmveth_close; | 1024 | netdev->stop = ibmveth_close; |
1026 | netdev->hard_start_xmit = ibmveth_start_xmit; | 1025 | netdev->hard_start_xmit = ibmveth_start_xmit; |
1027 | netdev->get_stats = ibmveth_get_stats; | 1026 | netdev->get_stats = ibmveth_get_stats; |
diff --git a/drivers/net/ibmveth.h b/drivers/net/ibmveth.h index 72cc15a6cab7..e05694126f85 100644 --- a/drivers/net/ibmveth.h +++ b/drivers/net/ibmveth.h | |||
@@ -112,6 +112,7 @@ struct ibmveth_rx_q { | |||
112 | struct ibmveth_adapter { | 112 | struct ibmveth_adapter { |
113 | struct vio_dev *vdev; | 113 | struct vio_dev *vdev; |
114 | struct net_device *netdev; | 114 | struct net_device *netdev; |
115 | struct napi_struct napi; | ||
115 | struct net_device_stats stats; | 116 | struct net_device_stats stats; |
116 | unsigned int mcastFilterSize; | 117 | unsigned int mcastFilterSize; |
117 | unsigned long mac_addr; | 118 | unsigned long mac_addr; |
diff --git a/drivers/net/ixgb/ixgb.h b/drivers/net/ixgb/ixgb.h index 3569d5b03388..1eee8894c732 100644 --- a/drivers/net/ixgb/ixgb.h +++ b/drivers/net/ixgb/ixgb.h | |||
@@ -184,6 +184,7 @@ struct ixgb_adapter { | |||
184 | boolean_t rx_csum; | 184 | boolean_t rx_csum; |
185 | 185 | ||
186 | /* OS defined structs */ | 186 | /* OS defined structs */ |
187 | struct napi_struct napi; | ||
187 | struct net_device *netdev; | 188 | struct net_device *netdev; |
188 | struct pci_dev *pdev; | 189 | struct pci_dev *pdev; |
189 | struct net_device_stats net_stats; | 190 | struct net_device_stats net_stats; |
diff --git a/drivers/net/ixgb/ixgb_main.c b/drivers/net/ixgb/ixgb_main.c index 991c8833e23c..e3f27c67fb28 100644 --- a/drivers/net/ixgb/ixgb_main.c +++ b/drivers/net/ixgb/ixgb_main.c | |||
@@ -97,7 +97,7 @@ static irqreturn_t ixgb_intr(int irq, void *data); | |||
97 | static boolean_t ixgb_clean_tx_irq(struct ixgb_adapter *adapter); | 97 | static boolean_t ixgb_clean_tx_irq(struct ixgb_adapter *adapter); |
98 | 98 | ||
99 | #ifdef CONFIG_IXGB_NAPI | 99 | #ifdef CONFIG_IXGB_NAPI |
100 | static int ixgb_clean(struct net_device *netdev, int *budget); | 100 | static int ixgb_clean(struct napi_struct *napi, int budget); |
101 | static boolean_t ixgb_clean_rx_irq(struct ixgb_adapter *adapter, | 101 | static boolean_t ixgb_clean_rx_irq(struct ixgb_adapter *adapter, |
102 | int *work_done, int work_to_do); | 102 | int *work_done, int work_to_do); |
103 | #else | 103 | #else |
@@ -288,7 +288,7 @@ ixgb_up(struct ixgb_adapter *adapter) | |||
288 | mod_timer(&adapter->watchdog_timer, jiffies); | 288 | mod_timer(&adapter->watchdog_timer, jiffies); |
289 | 289 | ||
290 | #ifdef CONFIG_IXGB_NAPI | 290 | #ifdef CONFIG_IXGB_NAPI |
291 | netif_poll_enable(netdev); | 291 | napi_enable(&adapter->napi); |
292 | #endif | 292 | #endif |
293 | ixgb_irq_enable(adapter); | 293 | ixgb_irq_enable(adapter); |
294 | 294 | ||
@@ -309,7 +309,7 @@ ixgb_down(struct ixgb_adapter *adapter, boolean_t kill_watchdog) | |||
309 | if(kill_watchdog) | 309 | if(kill_watchdog) |
310 | del_timer_sync(&adapter->watchdog_timer); | 310 | del_timer_sync(&adapter->watchdog_timer); |
311 | #ifdef CONFIG_IXGB_NAPI | 311 | #ifdef CONFIG_IXGB_NAPI |
312 | netif_poll_disable(netdev); | 312 | napi_disable(&adapter->napi); |
313 | #endif | 313 | #endif |
314 | adapter->link_speed = 0; | 314 | adapter->link_speed = 0; |
315 | adapter->link_duplex = 0; | 315 | adapter->link_duplex = 0; |
@@ -421,8 +421,7 @@ ixgb_probe(struct pci_dev *pdev, | |||
421 | netdev->tx_timeout = &ixgb_tx_timeout; | 421 | netdev->tx_timeout = &ixgb_tx_timeout; |
422 | netdev->watchdog_timeo = 5 * HZ; | 422 | netdev->watchdog_timeo = 5 * HZ; |
423 | #ifdef CONFIG_IXGB_NAPI | 423 | #ifdef CONFIG_IXGB_NAPI |
424 | netdev->poll = &ixgb_clean; | 424 | netif_napi_add(netdev, &adapter->napi, ixgb_clean, 64); |
425 | netdev->weight = 64; | ||
426 | #endif | 425 | #endif |
427 | netdev->vlan_rx_register = ixgb_vlan_rx_register; | 426 | netdev->vlan_rx_register = ixgb_vlan_rx_register; |
428 | netdev->vlan_rx_add_vid = ixgb_vlan_rx_add_vid; | 427 | netdev->vlan_rx_add_vid = ixgb_vlan_rx_add_vid; |
@@ -1746,7 +1745,7 @@ ixgb_intr(int irq, void *data) | |||
1746 | } | 1745 | } |
1747 | 1746 | ||
1748 | #ifdef CONFIG_IXGB_NAPI | 1747 | #ifdef CONFIG_IXGB_NAPI |
1749 | if(netif_rx_schedule_prep(netdev)) { | 1748 | if (netif_rx_schedule_prep(netdev, &adapter->napi)) { |
1750 | 1749 | ||
1751 | /* Disable interrupts and register for poll. The flush | 1750 | /* Disable interrupts and register for poll. The flush |
1752 | of the posted write is intentionally left out. | 1751 | of the posted write is intentionally left out. |
@@ -1754,7 +1753,7 @@ ixgb_intr(int irq, void *data) | |||
1754 | 1753 | ||
1755 | atomic_inc(&adapter->irq_sem); | 1754 | atomic_inc(&adapter->irq_sem); |
1756 | IXGB_WRITE_REG(&adapter->hw, IMC, ~0); | 1755 | IXGB_WRITE_REG(&adapter->hw, IMC, ~0); |
1757 | __netif_rx_schedule(netdev); | 1756 | __netif_rx_schedule(netdev, &adapter->napi); |
1758 | } | 1757 | } |
1759 | #else | 1758 | #else |
1760 | /* yes, that is actually a & and it is meant to make sure that | 1759 | /* yes, that is actually a & and it is meant to make sure that |
@@ -1776,27 +1775,23 @@ ixgb_intr(int irq, void *data) | |||
1776 | **/ | 1775 | **/ |
1777 | 1776 | ||
1778 | static int | 1777 | static int |
1779 | ixgb_clean(struct net_device *netdev, int *budget) | 1778 | ixgb_clean(struct napi_struct *napi, int budget) |
1780 | { | 1779 | { |
1781 | struct ixgb_adapter *adapter = netdev_priv(netdev); | 1780 | struct ixgb_adapter *adapter = container_of(napi, struct ixgb_adapter, napi); |
1782 | int work_to_do = min(*budget, netdev->quota); | 1781 | struct net_device *netdev = adapter->netdev; |
1783 | int tx_cleaned; | 1782 | int tx_cleaned; |
1784 | int work_done = 0; | 1783 | int work_done = 0; |
1785 | 1784 | ||
1786 | tx_cleaned = ixgb_clean_tx_irq(adapter); | 1785 | tx_cleaned = ixgb_clean_tx_irq(adapter); |
1787 | ixgb_clean_rx_irq(adapter, &work_done, work_to_do); | 1786 | ixgb_clean_rx_irq(adapter, &work_done, budget); |
1788 | |||
1789 | *budget -= work_done; | ||
1790 | netdev->quota -= work_done; | ||
1791 | 1787 | ||
1792 | /* if no Tx and not enough Rx work done, exit the polling mode */ | 1788 | /* if no Tx and not enough Rx work done, exit the polling mode */ |
1793 | if((!tx_cleaned && (work_done == 0)) || !netif_running(netdev)) { | 1789 | if((!tx_cleaned && (work_done == 0)) || !netif_running(netdev)) { |
1794 | netif_rx_complete(netdev); | 1790 | netif_rx_complete(netdev, napi); |
1795 | ixgb_irq_enable(adapter); | 1791 | ixgb_irq_enable(adapter); |
1796 | return 0; | ||
1797 | } | 1792 | } |
1798 | 1793 | ||
1799 | return 1; | 1794 | return work_done; |
1800 | } | 1795 | } |
1801 | #endif | 1796 | #endif |
1802 | 1797 | ||
diff --git a/drivers/net/ixp2000/ixpdev.c b/drivers/net/ixp2000/ixpdev.c index d9ce1aef148a..6c0dd49149d0 100644 --- a/drivers/net/ixp2000/ixpdev.c +++ b/drivers/net/ixp2000/ixpdev.c | |||
@@ -74,9 +74,9 @@ static int ixpdev_xmit(struct sk_buff *skb, struct net_device *dev) | |||
74 | } | 74 | } |
75 | 75 | ||
76 | 76 | ||
77 | static int ixpdev_rx(struct net_device *dev, int *budget) | 77 | static int ixpdev_rx(struct net_device *dev, int processed, int budget) |
78 | { | 78 | { |
79 | while (*budget > 0) { | 79 | while (processed < budget) { |
80 | struct ixpdev_rx_desc *desc; | 80 | struct ixpdev_rx_desc *desc; |
81 | struct sk_buff *skb; | 81 | struct sk_buff *skb; |
82 | void *buf; | 82 | void *buf; |
@@ -122,29 +122,34 @@ static int ixpdev_rx(struct net_device *dev, int *budget) | |||
122 | 122 | ||
123 | err: | 123 | err: |
124 | ixp2000_reg_write(RING_RX_PENDING, _desc); | 124 | ixp2000_reg_write(RING_RX_PENDING, _desc); |
125 | dev->quota--; | 125 | processed++; |
126 | (*budget)--; | ||
127 | } | 126 | } |
128 | 127 | ||
129 | return 1; | 128 | return processed; |
130 | } | 129 | } |
131 | 130 | ||
132 | /* dev always points to nds[0]. */ | 131 | /* dev always points to nds[0]. */ |
133 | static int ixpdev_poll(struct net_device *dev, int *budget) | 132 | static int ixpdev_poll(struct napi_struct *napi, int budget) |
134 | { | 133 | { |
134 | struct ixpdev_priv *ip = container_of(napi, struct ixpdev_priv, napi); | ||
135 | struct net_device *dev = ip->dev; | ||
136 | int rx; | ||
137 | |||
135 | /* @@@ Have to stop polling when nds[0] is administratively | 138 | /* @@@ Have to stop polling when nds[0] is administratively |
136 | * downed while we are polling. */ | 139 | * downed while we are polling. */ |
140 | rx = 0; | ||
137 | do { | 141 | do { |
138 | ixp2000_reg_write(IXP2000_IRQ_THD_RAW_STATUS_A_0, 0x00ff); | 142 | ixp2000_reg_write(IXP2000_IRQ_THD_RAW_STATUS_A_0, 0x00ff); |
139 | 143 | ||
140 | if (ixpdev_rx(dev, budget)) | 144 | rx = ixpdev_rx(dev, rx, budget); |
141 | return 1; | 145 | if (rx >= budget) |
146 | break; | ||
142 | } while (ixp2000_reg_read(IXP2000_IRQ_THD_RAW_STATUS_A_0) & 0x00ff); | 147 | } while (ixp2000_reg_read(IXP2000_IRQ_THD_RAW_STATUS_A_0) & 0x00ff); |
143 | 148 | ||
144 | netif_rx_complete(dev); | 149 | netif_rx_complete(dev, napi); |
145 | ixp2000_reg_write(IXP2000_IRQ_THD_ENABLE_SET_A_0, 0x00ff); | 150 | ixp2000_reg_write(IXP2000_IRQ_THD_ENABLE_SET_A_0, 0x00ff); |
146 | 151 | ||
147 | return 0; | 152 | return rx; |
148 | } | 153 | } |
149 | 154 | ||
150 | static void ixpdev_tx_complete(void) | 155 | static void ixpdev_tx_complete(void) |
@@ -199,9 +204,12 @@ static irqreturn_t ixpdev_interrupt(int irq, void *dev_id) | |||
199 | * Any of the eight receive units signaled RX? | 204 | * Any of the eight receive units signaled RX? |
200 | */ | 205 | */ |
201 | if (status & 0x00ff) { | 206 | if (status & 0x00ff) { |
207 | struct net_device *dev = nds[0]; | ||
208 | struct ixpdev_priv *ip = netdev_priv(dev); | ||
209 | |||
202 | ixp2000_reg_wrb(IXP2000_IRQ_THD_ENABLE_CLEAR_A_0, 0x00ff); | 210 | ixp2000_reg_wrb(IXP2000_IRQ_THD_ENABLE_CLEAR_A_0, 0x00ff); |
203 | if (likely(__netif_rx_schedule_prep(nds[0]))) { | 211 | if (likely(napi_schedule_prep(&ip->napi))) { |
204 | __netif_rx_schedule(nds[0]); | 212 | __netif_rx_schedule(dev, &ip->napi); |
205 | } else { | 213 | } else { |
206 | printk(KERN_CRIT "ixp2000: irq while polling!!\n"); | 214 | printk(KERN_CRIT "ixp2000: irq while polling!!\n"); |
207 | } | 215 | } |
@@ -232,11 +240,13 @@ static int ixpdev_open(struct net_device *dev) | |||
232 | struct ixpdev_priv *ip = netdev_priv(dev); | 240 | struct ixpdev_priv *ip = netdev_priv(dev); |
233 | int err; | 241 | int err; |
234 | 242 | ||
243 | napi_enable(&ip->napi); | ||
235 | if (!nds_open++) { | 244 | if (!nds_open++) { |
236 | err = request_irq(IRQ_IXP2000_THDA0, ixpdev_interrupt, | 245 | err = request_irq(IRQ_IXP2000_THDA0, ixpdev_interrupt, |
237 | IRQF_SHARED, "ixp2000_eth", nds); | 246 | IRQF_SHARED, "ixp2000_eth", nds); |
238 | if (err) { | 247 | if (err) { |
239 | nds_open--; | 248 | nds_open--; |
249 | napi_disable(&ip->napi); | ||
240 | return err; | 250 | return err; |
241 | } | 251 | } |
242 | 252 | ||
@@ -254,6 +264,7 @@ static int ixpdev_close(struct net_device *dev) | |||
254 | struct ixpdev_priv *ip = netdev_priv(dev); | 264 | struct ixpdev_priv *ip = netdev_priv(dev); |
255 | 265 | ||
256 | netif_stop_queue(dev); | 266 | netif_stop_queue(dev); |
267 | napi_disable(&ip->napi); | ||
257 | set_port_admin_status(ip->channel, 0); | 268 | set_port_admin_status(ip->channel, 0); |
258 | 269 | ||
259 | if (!--nds_open) { | 270 | if (!--nds_open) { |
@@ -274,7 +285,6 @@ struct net_device *ixpdev_alloc(int channel, int sizeof_priv) | |||
274 | return NULL; | 285 | return NULL; |
275 | 286 | ||
276 | dev->hard_start_xmit = ixpdev_xmit; | 287 | dev->hard_start_xmit = ixpdev_xmit; |
277 | dev->poll = ixpdev_poll; | ||
278 | dev->open = ixpdev_open; | 288 | dev->open = ixpdev_open; |
279 | dev->stop = ixpdev_close; | 289 | dev->stop = ixpdev_close; |
280 | #ifdef CONFIG_NET_POLL_CONTROLLER | 290 | #ifdef CONFIG_NET_POLL_CONTROLLER |
@@ -282,9 +292,10 @@ struct net_device *ixpdev_alloc(int channel, int sizeof_priv) | |||
282 | #endif | 292 | #endif |
283 | 293 | ||
284 | dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM; | 294 | dev->features |= NETIF_F_SG | NETIF_F_HW_CSUM; |
285 | dev->weight = 64; | ||
286 | 295 | ||
287 | ip = netdev_priv(dev); | 296 | ip = netdev_priv(dev); |
297 | ip->dev = dev; | ||
298 | netif_napi_add(dev, &ip->napi, ixpdev_poll, 64); | ||
288 | ip->channel = channel; | 299 | ip->channel = channel; |
289 | ip->tx_queue_entries = 0; | 300 | ip->tx_queue_entries = 0; |
290 | 301 | ||
diff --git a/drivers/net/ixp2000/ixpdev.h b/drivers/net/ixp2000/ixpdev.h index bd686cb63058..391ece623243 100644 --- a/drivers/net/ixp2000/ixpdev.h +++ b/drivers/net/ixp2000/ixpdev.h | |||
@@ -14,6 +14,8 @@ | |||
14 | 14 | ||
15 | struct ixpdev_priv | 15 | struct ixpdev_priv |
16 | { | 16 | { |
17 | struct net_device *dev; | ||
18 | struct napi_struct napi; | ||
17 | int channel; | 19 | int channel; |
18 | int tx_queue_entries; | 20 | int tx_queue_entries; |
19 | }; | 21 | }; |
diff --git a/drivers/net/macb.c b/drivers/net/macb.c index a4bb0264180a..74c3f7a7ae4a 100644 --- a/drivers/net/macb.c +++ b/drivers/net/macb.c | |||
@@ -470,47 +470,41 @@ static int macb_rx(struct macb *bp, int budget) | |||
470 | return received; | 470 | return received; |
471 | } | 471 | } |
472 | 472 | ||
473 | static int macb_poll(struct net_device *dev, int *budget) | 473 | static int macb_poll(struct napi_struct *napi, int budget) |
474 | { | 474 | { |
475 | struct macb *bp = netdev_priv(dev); | 475 | struct macb *bp = container_of(napi, struct macb, napi); |
476 | int orig_budget, work_done, retval = 0; | 476 | struct net_device *dev = bp->dev; |
477 | int work_done; | ||
477 | u32 status; | 478 | u32 status; |
478 | 479 | ||
479 | status = macb_readl(bp, RSR); | 480 | status = macb_readl(bp, RSR); |
480 | macb_writel(bp, RSR, status); | 481 | macb_writel(bp, RSR, status); |
481 | 482 | ||
483 | work_done = 0; | ||
482 | if (!status) { | 484 | if (!status) { |
483 | /* | 485 | /* |
484 | * This may happen if an interrupt was pending before | 486 | * This may happen if an interrupt was pending before |
485 | * this function was called last time, and no packets | 487 | * this function was called last time, and no packets |
486 | * have been received since. | 488 | * have been received since. |
487 | */ | 489 | */ |
488 | netif_rx_complete(dev); | 490 | netif_rx_complete(dev, napi); |
489 | goto out; | 491 | goto out; |
490 | } | 492 | } |
491 | 493 | ||
492 | dev_dbg(&bp->pdev->dev, "poll: status = %08lx, budget = %d\n", | 494 | dev_dbg(&bp->pdev->dev, "poll: status = %08lx, budget = %d\n", |
493 | (unsigned long)status, *budget); | 495 | (unsigned long)status, budget); |
494 | 496 | ||
495 | if (!(status & MACB_BIT(REC))) { | 497 | if (!(status & MACB_BIT(REC))) { |
496 | dev_warn(&bp->pdev->dev, | 498 | dev_warn(&bp->pdev->dev, |
497 | "No RX buffers complete, status = %02lx\n", | 499 | "No RX buffers complete, status = %02lx\n", |
498 | (unsigned long)status); | 500 | (unsigned long)status); |
499 | netif_rx_complete(dev); | 501 | netif_rx_complete(dev, napi); |
500 | goto out; | 502 | goto out; |
501 | } | 503 | } |
502 | 504 | ||
503 | orig_budget = *budget; | 505 | work_done = macb_rx(bp, budget); |
504 | if (orig_budget > dev->quota) | 506 | if (work_done < budget) |
505 | orig_budget = dev->quota; | 507 | netif_rx_complete(dev, napi); |
506 | |||
507 | work_done = macb_rx(bp, orig_budget); | ||
508 | if (work_done < orig_budget) { | ||
509 | netif_rx_complete(dev); | ||
510 | retval = 0; | ||
511 | } else { | ||
512 | retval = 1; | ||
513 | } | ||
514 | 508 | ||
515 | /* | 509 | /* |
516 | * We've done what we can to clean the buffers. Make sure we | 510 | * We've done what we can to clean the buffers. Make sure we |
@@ -521,7 +515,7 @@ out: | |||
521 | 515 | ||
522 | /* TODO: Handle errors */ | 516 | /* TODO: Handle errors */ |
523 | 517 | ||
524 | return retval; | 518 | return work_done; |
525 | } | 519 | } |
526 | 520 | ||
527 | static irqreturn_t macb_interrupt(int irq, void *dev_id) | 521 | static irqreturn_t macb_interrupt(int irq, void *dev_id) |
@@ -545,7 +539,7 @@ static irqreturn_t macb_interrupt(int irq, void *dev_id) | |||
545 | } | 539 | } |
546 | 540 | ||
547 | if (status & MACB_RX_INT_FLAGS) { | 541 | if (status & MACB_RX_INT_FLAGS) { |
548 | if (netif_rx_schedule_prep(dev)) { | 542 | if (netif_rx_schedule_prep(dev, &bp->napi)) { |
549 | /* | 543 | /* |
550 | * There's no point taking any more interrupts | 544 | * There's no point taking any more interrupts |
551 | * until we have processed the buffers | 545 | * until we have processed the buffers |
@@ -553,7 +547,7 @@ static irqreturn_t macb_interrupt(int irq, void *dev_id) | |||
553 | macb_writel(bp, IDR, MACB_RX_INT_FLAGS); | 547 | macb_writel(bp, IDR, MACB_RX_INT_FLAGS); |
554 | dev_dbg(&bp->pdev->dev, | 548 | dev_dbg(&bp->pdev->dev, |
555 | "scheduling RX softirq\n"); | 549 | "scheduling RX softirq\n"); |
556 | __netif_rx_schedule(dev); | 550 | __netif_rx_schedule(dev, &bp->napi); |
557 | } | 551 | } |
558 | } | 552 | } |
559 | 553 | ||
@@ -937,6 +931,8 @@ static int macb_open(struct net_device *dev) | |||
937 | return err; | 931 | return err; |
938 | } | 932 | } |
939 | 933 | ||
934 | napi_enable(&bp->napi); | ||
935 | |||
940 | macb_init_rings(bp); | 936 | macb_init_rings(bp); |
941 | macb_init_hw(bp); | 937 | macb_init_hw(bp); |
942 | 938 | ||
@@ -954,6 +950,7 @@ static int macb_close(struct net_device *dev) | |||
954 | unsigned long flags; | 950 | unsigned long flags; |
955 | 951 | ||
956 | netif_stop_queue(dev); | 952 | netif_stop_queue(dev); |
953 | napi_disable(&bp->napi); | ||
957 | 954 | ||
958 | if (bp->phy_dev) | 955 | if (bp->phy_dev) |
959 | phy_stop(bp->phy_dev); | 956 | phy_stop(bp->phy_dev); |
@@ -1146,8 +1143,7 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
1146 | dev->get_stats = macb_get_stats; | 1143 | dev->get_stats = macb_get_stats; |
1147 | dev->set_multicast_list = macb_set_rx_mode; | 1144 | dev->set_multicast_list = macb_set_rx_mode; |
1148 | dev->do_ioctl = macb_ioctl; | 1145 | dev->do_ioctl = macb_ioctl; |
1149 | dev->poll = macb_poll; | 1146 | netif_napi_add(dev, &bp->napi, macb_poll, 64); |
1150 | dev->weight = 64; | ||
1151 | dev->ethtool_ops = &macb_ethtool_ops; | 1147 | dev->ethtool_ops = &macb_ethtool_ops; |
1152 | 1148 | ||
1153 | dev->base_addr = regs->start; | 1149 | dev->base_addr = regs->start; |
diff --git a/drivers/net/macb.h b/drivers/net/macb.h index 4e3283ebd97c..57b85acf0d16 100644 --- a/drivers/net/macb.h +++ b/drivers/net/macb.h | |||
@@ -374,6 +374,7 @@ struct macb { | |||
374 | struct clk *pclk; | 374 | struct clk *pclk; |
375 | struct clk *hclk; | 375 | struct clk *hclk; |
376 | struct net_device *dev; | 376 | struct net_device *dev; |
377 | struct napi_struct napi; | ||
377 | struct net_device_stats stats; | 378 | struct net_device_stats stats; |
378 | struct macb_stats hw_stats; | 379 | struct macb_stats hw_stats; |
379 | 380 | ||
diff --git a/drivers/net/mv643xx_eth.c b/drivers/net/mv643xx_eth.c index 315335671f0f..702eba549161 100644 --- a/drivers/net/mv643xx_eth.c +++ b/drivers/net/mv643xx_eth.c | |||
@@ -66,7 +66,7 @@ static int mv643xx_eth_change_mtu(struct net_device *, int); | |||
66 | static struct net_device_stats *mv643xx_eth_get_stats(struct net_device *); | 66 | static struct net_device_stats *mv643xx_eth_get_stats(struct net_device *); |
67 | static void eth_port_init_mac_tables(unsigned int eth_port_num); | 67 | static void eth_port_init_mac_tables(unsigned int eth_port_num); |
68 | #ifdef MV643XX_NAPI | 68 | #ifdef MV643XX_NAPI |
69 | static int mv643xx_poll(struct net_device *dev, int *budget); | 69 | static int mv643xx_poll(struct napi_struct *napi, int budget); |
70 | #endif | 70 | #endif |
71 | static int ethernet_phy_get(unsigned int eth_port_num); | 71 | static int ethernet_phy_get(unsigned int eth_port_num); |
72 | static void ethernet_phy_set(unsigned int eth_port_num, int phy_addr); | 72 | static void ethernet_phy_set(unsigned int eth_port_num, int phy_addr); |
@@ -562,7 +562,7 @@ static irqreturn_t mv643xx_eth_int_handler(int irq, void *dev_id) | |||
562 | /* wait for previous write to complete */ | 562 | /* wait for previous write to complete */ |
563 | mv_read(MV643XX_ETH_INTERRUPT_MASK_REG(port_num)); | 563 | mv_read(MV643XX_ETH_INTERRUPT_MASK_REG(port_num)); |
564 | 564 | ||
565 | netif_rx_schedule(dev); | 565 | netif_rx_schedule(dev, &mp->napi); |
566 | } | 566 | } |
567 | #else | 567 | #else |
568 | if (eth_int_cause & ETH_INT_CAUSE_RX) | 568 | if (eth_int_cause & ETH_INT_CAUSE_RX) |
@@ -880,6 +880,10 @@ static int mv643xx_eth_open(struct net_device *dev) | |||
880 | 880 | ||
881 | mv643xx_eth_rx_refill_descs(dev); /* Fill RX ring with skb's */ | 881 | mv643xx_eth_rx_refill_descs(dev); /* Fill RX ring with skb's */ |
882 | 882 | ||
883 | #ifdef MV643XX_NAPI | ||
884 | napi_enable(&mp->napi); | ||
885 | #endif | ||
886 | |||
883 | eth_port_start(dev); | 887 | eth_port_start(dev); |
884 | 888 | ||
885 | /* Interrupt Coalescing */ | 889 | /* Interrupt Coalescing */ |
@@ -982,7 +986,7 @@ static int mv643xx_eth_stop(struct net_device *dev) | |||
982 | mv_read(MV643XX_ETH_INTERRUPT_MASK_REG(port_num)); | 986 | mv_read(MV643XX_ETH_INTERRUPT_MASK_REG(port_num)); |
983 | 987 | ||
984 | #ifdef MV643XX_NAPI | 988 | #ifdef MV643XX_NAPI |
985 | netif_poll_disable(dev); | 989 | napi_disable(&mp->napi); |
986 | #endif | 990 | #endif |
987 | netif_carrier_off(dev); | 991 | netif_carrier_off(dev); |
988 | netif_stop_queue(dev); | 992 | netif_stop_queue(dev); |
@@ -992,10 +996,6 @@ static int mv643xx_eth_stop(struct net_device *dev) | |||
992 | mv643xx_eth_free_tx_rings(dev); | 996 | mv643xx_eth_free_tx_rings(dev); |
993 | mv643xx_eth_free_rx_rings(dev); | 997 | mv643xx_eth_free_rx_rings(dev); |
994 | 998 | ||
995 | #ifdef MV643XX_NAPI | ||
996 | netif_poll_enable(dev); | ||
997 | #endif | ||
998 | |||
999 | free_irq(dev->irq, dev); | 999 | free_irq(dev->irq, dev); |
1000 | 1000 | ||
1001 | return 0; | 1001 | return 0; |
@@ -1007,11 +1007,12 @@ static int mv643xx_eth_stop(struct net_device *dev) | |||
1007 | * | 1007 | * |
1008 | * This function is used in case of NAPI | 1008 | * This function is used in case of NAPI |
1009 | */ | 1009 | */ |
1010 | static int mv643xx_poll(struct net_device *dev, int *budget) | 1010 | static int mv643xx_poll(struct napi_struct *napi, int budget) |
1011 | { | 1011 | { |
1012 | struct mv643xx_private *mp = netdev_priv(dev); | 1012 | struct mv643xx_private *mp = container_of(napi, struct mv643xx_private, napi); |
1013 | int done = 1, orig_budget, work_done; | 1013 | struct net_device *dev = mp->dev; |
1014 | unsigned int port_num = mp->port_num; | 1014 | unsigned int port_num = mp->port_num; |
1015 | int work_done; | ||
1015 | 1016 | ||
1016 | #ifdef MV643XX_TX_FAST_REFILL | 1017 | #ifdef MV643XX_TX_FAST_REFILL |
1017 | if (++mp->tx_clean_threshold > 5) { | 1018 | if (++mp->tx_clean_threshold > 5) { |
@@ -1020,27 +1021,20 @@ static int mv643xx_poll(struct net_device *dev, int *budget) | |||
1020 | } | 1021 | } |
1021 | #endif | 1022 | #endif |
1022 | 1023 | ||
1024 | work_done = 0; | ||
1023 | if ((mv_read(MV643XX_ETH_RX_CURRENT_QUEUE_DESC_PTR_0(port_num))) | 1025 | if ((mv_read(MV643XX_ETH_RX_CURRENT_QUEUE_DESC_PTR_0(port_num))) |
1024 | != (u32) mp->rx_used_desc_q) { | 1026 | != (u32) mp->rx_used_desc_q) |
1025 | orig_budget = *budget; | 1027 | work_done = mv643xx_eth_receive_queue(dev, budget); |
1026 | if (orig_budget > dev->quota) | ||
1027 | orig_budget = dev->quota; | ||
1028 | work_done = mv643xx_eth_receive_queue(dev, orig_budget); | ||
1029 | *budget -= work_done; | ||
1030 | dev->quota -= work_done; | ||
1031 | if (work_done >= orig_budget) | ||
1032 | done = 0; | ||
1033 | } | ||
1034 | 1028 | ||
1035 | if (done) { | 1029 | if (work_done < budget) { |
1036 | netif_rx_complete(dev); | 1030 | netif_rx_complete(dev, napi); |
1037 | mv_write(MV643XX_ETH_INTERRUPT_CAUSE_REG(port_num), 0); | 1031 | mv_write(MV643XX_ETH_INTERRUPT_CAUSE_REG(port_num), 0); |
1038 | mv_write(MV643XX_ETH_INTERRUPT_CAUSE_EXTEND_REG(port_num), 0); | 1032 | mv_write(MV643XX_ETH_INTERRUPT_CAUSE_EXTEND_REG(port_num), 0); |
1039 | mv_write(MV643XX_ETH_INTERRUPT_MASK_REG(port_num), | 1033 | mv_write(MV643XX_ETH_INTERRUPT_MASK_REG(port_num), |
1040 | ETH_INT_UNMASK_ALL); | 1034 | ETH_INT_UNMASK_ALL); |
1041 | } | 1035 | } |
1042 | 1036 | ||
1043 | return done ? 0 : 1; | 1037 | return work_done; |
1044 | } | 1038 | } |
1045 | #endif | 1039 | #endif |
1046 | 1040 | ||
@@ -1333,6 +1327,10 @@ static int mv643xx_eth_probe(struct platform_device *pdev) | |||
1333 | platform_set_drvdata(pdev, dev); | 1327 | platform_set_drvdata(pdev, dev); |
1334 | 1328 | ||
1335 | mp = netdev_priv(dev); | 1329 | mp = netdev_priv(dev); |
1330 | mp->dev = dev; | ||
1331 | #ifdef MV643XX_NAPI | ||
1332 | netif_napi_add(dev, &mp->napi, mv643xx_poll, 64); | ||
1333 | #endif | ||
1336 | 1334 | ||
1337 | res = platform_get_resource(pdev, IORESOURCE_IRQ, 0); | 1335 | res = platform_get_resource(pdev, IORESOURCE_IRQ, 0); |
1338 | BUG_ON(!res); | 1336 | BUG_ON(!res); |
@@ -1347,10 +1345,6 @@ static int mv643xx_eth_probe(struct platform_device *pdev) | |||
1347 | 1345 | ||
1348 | /* No need to Tx Timeout */ | 1346 | /* No need to Tx Timeout */ |
1349 | dev->tx_timeout = mv643xx_eth_tx_timeout; | 1347 | dev->tx_timeout = mv643xx_eth_tx_timeout; |
1350 | #ifdef MV643XX_NAPI | ||
1351 | dev->poll = mv643xx_poll; | ||
1352 | dev->weight = 64; | ||
1353 | #endif | ||
1354 | 1348 | ||
1355 | #ifdef CONFIG_NET_POLL_CONTROLLER | 1349 | #ifdef CONFIG_NET_POLL_CONTROLLER |
1356 | dev->poll_controller = mv643xx_netpoll; | 1350 | dev->poll_controller = mv643xx_netpoll; |
diff --git a/drivers/net/mv643xx_eth.h b/drivers/net/mv643xx_eth.h index 565b96696aca..be669eb23788 100644 --- a/drivers/net/mv643xx_eth.h +++ b/drivers/net/mv643xx_eth.h | |||
@@ -320,6 +320,8 @@ struct mv643xx_private { | |||
320 | 320 | ||
321 | struct work_struct tx_timeout_task; | 321 | struct work_struct tx_timeout_task; |
322 | 322 | ||
323 | struct net_device *dev; | ||
324 | struct napi_struct napi; | ||
323 | struct net_device_stats stats; | 325 | struct net_device_stats stats; |
324 | struct mv643xx_mib_counters mib_counters; | 326 | struct mv643xx_mib_counters mib_counters; |
325 | spinlock_t lock; | 327 | spinlock_t lock; |
diff --git a/drivers/net/myri10ge/myri10ge.c b/drivers/net/myri10ge/myri10ge.c index 556962f9612d..a30146ea51f0 100644 --- a/drivers/net/myri10ge/myri10ge.c +++ b/drivers/net/myri10ge/myri10ge.c | |||
@@ -163,6 +163,7 @@ struct myri10ge_priv { | |||
163 | int small_bytes; | 163 | int small_bytes; |
164 | int big_bytes; | 164 | int big_bytes; |
165 | struct net_device *dev; | 165 | struct net_device *dev; |
166 | struct napi_struct napi; | ||
166 | struct net_device_stats stats; | 167 | struct net_device_stats stats; |
167 | u8 __iomem *sram; | 168 | u8 __iomem *sram; |
168 | int sram_size; | 169 | int sram_size; |
@@ -1100,7 +1101,7 @@ static inline void myri10ge_tx_done(struct myri10ge_priv *mgp, int mcp_index) | |||
1100 | } | 1101 | } |
1101 | } | 1102 | } |
1102 | 1103 | ||
1103 | static inline void myri10ge_clean_rx_done(struct myri10ge_priv *mgp, int *limit) | 1104 | static inline int myri10ge_clean_rx_done(struct myri10ge_priv *mgp, int budget) |
1104 | { | 1105 | { |
1105 | struct myri10ge_rx_done *rx_done = &mgp->rx_done; | 1106 | struct myri10ge_rx_done *rx_done = &mgp->rx_done; |
1106 | unsigned long rx_bytes = 0; | 1107 | unsigned long rx_bytes = 0; |
@@ -1109,10 +1110,11 @@ static inline void myri10ge_clean_rx_done(struct myri10ge_priv *mgp, int *limit) | |||
1109 | 1110 | ||
1110 | int idx = rx_done->idx; | 1111 | int idx = rx_done->idx; |
1111 | int cnt = rx_done->cnt; | 1112 | int cnt = rx_done->cnt; |
1113 | int work_done = 0; | ||
1112 | u16 length; | 1114 | u16 length; |
1113 | __wsum checksum; | 1115 | __wsum checksum; |
1114 | 1116 | ||
1115 | while (rx_done->entry[idx].length != 0 && *limit != 0) { | 1117 | while (rx_done->entry[idx].length != 0 && work_done++ < budget) { |
1116 | length = ntohs(rx_done->entry[idx].length); | 1118 | length = ntohs(rx_done->entry[idx].length); |
1117 | rx_done->entry[idx].length = 0; | 1119 | rx_done->entry[idx].length = 0; |
1118 | checksum = csum_unfold(rx_done->entry[idx].checksum); | 1120 | checksum = csum_unfold(rx_done->entry[idx].checksum); |
@@ -1128,10 +1130,6 @@ static inline void myri10ge_clean_rx_done(struct myri10ge_priv *mgp, int *limit) | |||
1128 | rx_bytes += rx_ok * (unsigned long)length; | 1130 | rx_bytes += rx_ok * (unsigned long)length; |
1129 | cnt++; | 1131 | cnt++; |
1130 | idx = cnt & (myri10ge_max_intr_slots - 1); | 1132 | idx = cnt & (myri10ge_max_intr_slots - 1); |
1131 | |||
1132 | /* limit potential for livelock by only handling a | ||
1133 | * limited number of frames. */ | ||
1134 | (*limit)--; | ||
1135 | } | 1133 | } |
1136 | rx_done->idx = idx; | 1134 | rx_done->idx = idx; |
1137 | rx_done->cnt = cnt; | 1135 | rx_done->cnt = cnt; |
@@ -1145,6 +1143,7 @@ static inline void myri10ge_clean_rx_done(struct myri10ge_priv *mgp, int *limit) | |||
1145 | if (mgp->rx_big.fill_cnt - mgp->rx_big.cnt < myri10ge_fill_thresh) | 1143 | if (mgp->rx_big.fill_cnt - mgp->rx_big.cnt < myri10ge_fill_thresh) |
1146 | myri10ge_alloc_rx_pages(mgp, &mgp->rx_big, mgp->big_bytes, 0); | 1144 | myri10ge_alloc_rx_pages(mgp, &mgp->rx_big, mgp->big_bytes, 0); |
1147 | 1145 | ||
1146 | return work_done; | ||
1148 | } | 1147 | } |
1149 | 1148 | ||
1150 | static inline void myri10ge_check_statblock(struct myri10ge_priv *mgp) | 1149 | static inline void myri10ge_check_statblock(struct myri10ge_priv *mgp) |
@@ -1189,26 +1188,21 @@ static inline void myri10ge_check_statblock(struct myri10ge_priv *mgp) | |||
1189 | } | 1188 | } |
1190 | } | 1189 | } |
1191 | 1190 | ||
1192 | static int myri10ge_poll(struct net_device *netdev, int *budget) | 1191 | static int myri10ge_poll(struct napi_struct *napi, int budget) |
1193 | { | 1192 | { |
1194 | struct myri10ge_priv *mgp = netdev_priv(netdev); | 1193 | struct myri10ge_priv *mgp = container_of(napi, struct myri10ge_priv, napi); |
1194 | struct net_device *netdev = mgp->dev; | ||
1195 | struct myri10ge_rx_done *rx_done = &mgp->rx_done; | 1195 | struct myri10ge_rx_done *rx_done = &mgp->rx_done; |
1196 | int limit, orig_limit, work_done; | 1196 | int work_done; |
1197 | 1197 | ||
1198 | /* process as many rx events as NAPI will allow */ | 1198 | /* process as many rx events as NAPI will allow */ |
1199 | limit = min(*budget, netdev->quota); | 1199 | work_done = myri10ge_clean_rx_done(mgp, budget); |
1200 | orig_limit = limit; | ||
1201 | myri10ge_clean_rx_done(mgp, &limit); | ||
1202 | work_done = orig_limit - limit; | ||
1203 | *budget -= work_done; | ||
1204 | netdev->quota -= work_done; | ||
1205 | 1200 | ||
1206 | if (rx_done->entry[rx_done->idx].length == 0 || !netif_running(netdev)) { | 1201 | if (rx_done->entry[rx_done->idx].length == 0 || !netif_running(netdev)) { |
1207 | netif_rx_complete(netdev); | 1202 | netif_rx_complete(netdev, napi); |
1208 | put_be32(htonl(3), mgp->irq_claim); | 1203 | put_be32(htonl(3), mgp->irq_claim); |
1209 | return 0; | ||
1210 | } | 1204 | } |
1211 | return 1; | 1205 | return work_done; |
1212 | } | 1206 | } |
1213 | 1207 | ||
1214 | static irqreturn_t myri10ge_intr(int irq, void *arg) | 1208 | static irqreturn_t myri10ge_intr(int irq, void *arg) |
@@ -1226,7 +1220,7 @@ static irqreturn_t myri10ge_intr(int irq, void *arg) | |||
1226 | /* low bit indicates receives are present, so schedule | 1220 | /* low bit indicates receives are present, so schedule |
1227 | * napi poll handler */ | 1221 | * napi poll handler */ |
1228 | if (stats->valid & 1) | 1222 | if (stats->valid & 1) |
1229 | netif_rx_schedule(mgp->dev); | 1223 | netif_rx_schedule(mgp->dev, &mgp->napi); |
1230 | 1224 | ||
1231 | if (!mgp->msi_enabled) { | 1225 | if (!mgp->msi_enabled) { |
1232 | put_be32(0, mgp->irq_deassert); | 1226 | put_be32(0, mgp->irq_deassert); |
@@ -1853,7 +1847,7 @@ static int myri10ge_open(struct net_device *dev) | |||
1853 | mgp->link_state = htonl(~0U); | 1847 | mgp->link_state = htonl(~0U); |
1854 | mgp->rdma_tags_available = 15; | 1848 | mgp->rdma_tags_available = 15; |
1855 | 1849 | ||
1856 | netif_poll_enable(mgp->dev); /* must happen prior to any irq */ | 1850 | napi_enable(&mgp->napi); /* must happen prior to any irq */ |
1857 | 1851 | ||
1858 | status = myri10ge_send_cmd(mgp, MXGEFW_CMD_ETHERNET_UP, &cmd, 0); | 1852 | status = myri10ge_send_cmd(mgp, MXGEFW_CMD_ETHERNET_UP, &cmd, 0); |
1859 | if (status) { | 1853 | if (status) { |
@@ -1897,7 +1891,7 @@ static int myri10ge_close(struct net_device *dev) | |||
1897 | 1891 | ||
1898 | del_timer_sync(&mgp->watchdog_timer); | 1892 | del_timer_sync(&mgp->watchdog_timer); |
1899 | mgp->running = MYRI10GE_ETH_STOPPING; | 1893 | mgp->running = MYRI10GE_ETH_STOPPING; |
1900 | netif_poll_disable(mgp->dev); | 1894 | napi_disable(&mgp->napi); |
1901 | netif_carrier_off(dev); | 1895 | netif_carrier_off(dev); |
1902 | netif_stop_queue(dev); | 1896 | netif_stop_queue(dev); |
1903 | old_down_cnt = mgp->down_cnt; | 1897 | old_down_cnt = mgp->down_cnt; |
@@ -2857,6 +2851,8 @@ static int myri10ge_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2857 | mgp = netdev_priv(netdev); | 2851 | mgp = netdev_priv(netdev); |
2858 | memset(mgp, 0, sizeof(*mgp)); | 2852 | memset(mgp, 0, sizeof(*mgp)); |
2859 | mgp->dev = netdev; | 2853 | mgp->dev = netdev; |
2854 | netif_napi_add(netdev, &mgp->napi, | ||
2855 | myri10ge_poll, myri10ge_napi_weight); | ||
2860 | mgp->pdev = pdev; | 2856 | mgp->pdev = pdev; |
2861 | mgp->csum_flag = MXGEFW_FLAGS_CKSUM; | 2857 | mgp->csum_flag = MXGEFW_FLAGS_CKSUM; |
2862 | mgp->pause = myri10ge_flow_control; | 2858 | mgp->pause = myri10ge_flow_control; |
@@ -2981,8 +2977,6 @@ static int myri10ge_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2981 | netdev->features = NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_TSO; | 2977 | netdev->features = NETIF_F_SG | NETIF_F_HW_CSUM | NETIF_F_TSO; |
2982 | if (dac_enabled) | 2978 | if (dac_enabled) |
2983 | netdev->features |= NETIF_F_HIGHDMA; | 2979 | netdev->features |= NETIF_F_HIGHDMA; |
2984 | netdev->poll = myri10ge_poll; | ||
2985 | netdev->weight = myri10ge_napi_weight; | ||
2986 | 2980 | ||
2987 | /* make sure we can get an irq, and that MSI can be | 2981 | /* make sure we can get an irq, and that MSI can be |
2988 | * setup (if available). Also ensure netdev->irq | 2982 | * setup (if available). Also ensure netdev->irq |
diff --git a/drivers/net/natsemi.c b/drivers/net/natsemi.c index b47a12d684f9..43cfa4b3e294 100644 --- a/drivers/net/natsemi.c +++ b/drivers/net/natsemi.c | |||
@@ -560,6 +560,8 @@ struct netdev_private { | |||
560 | /* address of a sent-in-place packet/buffer, for later free() */ | 560 | /* address of a sent-in-place packet/buffer, for later free() */ |
561 | struct sk_buff *tx_skbuff[TX_RING_SIZE]; | 561 | struct sk_buff *tx_skbuff[TX_RING_SIZE]; |
562 | dma_addr_t tx_dma[TX_RING_SIZE]; | 562 | dma_addr_t tx_dma[TX_RING_SIZE]; |
563 | struct net_device *dev; | ||
564 | struct napi_struct napi; | ||
563 | struct net_device_stats stats; | 565 | struct net_device_stats stats; |
564 | /* Media monitoring timer */ | 566 | /* Media monitoring timer */ |
565 | struct timer_list timer; | 567 | struct timer_list timer; |
@@ -636,7 +638,7 @@ static void init_registers(struct net_device *dev); | |||
636 | static int start_tx(struct sk_buff *skb, struct net_device *dev); | 638 | static int start_tx(struct sk_buff *skb, struct net_device *dev); |
637 | static irqreturn_t intr_handler(int irq, void *dev_instance); | 639 | static irqreturn_t intr_handler(int irq, void *dev_instance); |
638 | static void netdev_error(struct net_device *dev, int intr_status); | 640 | static void netdev_error(struct net_device *dev, int intr_status); |
639 | static int natsemi_poll(struct net_device *dev, int *budget); | 641 | static int natsemi_poll(struct napi_struct *napi, int budget); |
640 | static void netdev_rx(struct net_device *dev, int *work_done, int work_to_do); | 642 | static void netdev_rx(struct net_device *dev, int *work_done, int work_to_do); |
641 | static void netdev_tx_done(struct net_device *dev); | 643 | static void netdev_tx_done(struct net_device *dev); |
642 | static int natsemi_change_mtu(struct net_device *dev, int new_mtu); | 644 | static int natsemi_change_mtu(struct net_device *dev, int new_mtu); |
@@ -861,6 +863,7 @@ static int __devinit natsemi_probe1 (struct pci_dev *pdev, | |||
861 | dev->irq = irq; | 863 | dev->irq = irq; |
862 | 864 | ||
863 | np = netdev_priv(dev); | 865 | np = netdev_priv(dev); |
866 | netif_napi_add(dev, &np->napi, natsemi_poll, 64); | ||
864 | 867 | ||
865 | np->pci_dev = pdev; | 868 | np->pci_dev = pdev; |
866 | pci_set_drvdata(pdev, dev); | 869 | pci_set_drvdata(pdev, dev); |
@@ -931,8 +934,6 @@ static int __devinit natsemi_probe1 (struct pci_dev *pdev, | |||
931 | dev->do_ioctl = &netdev_ioctl; | 934 | dev->do_ioctl = &netdev_ioctl; |
932 | dev->tx_timeout = &tx_timeout; | 935 | dev->tx_timeout = &tx_timeout; |
933 | dev->watchdog_timeo = TX_TIMEOUT; | 936 | dev->watchdog_timeo = TX_TIMEOUT; |
934 | dev->poll = natsemi_poll; | ||
935 | dev->weight = 64; | ||
936 | 937 | ||
937 | #ifdef CONFIG_NET_POLL_CONTROLLER | 938 | #ifdef CONFIG_NET_POLL_CONTROLLER |
938 | dev->poll_controller = &natsemi_poll_controller; | 939 | dev->poll_controller = &natsemi_poll_controller; |
@@ -1554,6 +1555,8 @@ static int netdev_open(struct net_device *dev) | |||
1554 | free_irq(dev->irq, dev); | 1555 | free_irq(dev->irq, dev); |
1555 | return i; | 1556 | return i; |
1556 | } | 1557 | } |
1558 | napi_enable(&np->napi); | ||
1559 | |||
1557 | init_ring(dev); | 1560 | init_ring(dev); |
1558 | spin_lock_irq(&np->lock); | 1561 | spin_lock_irq(&np->lock); |
1559 | init_registers(dev); | 1562 | init_registers(dev); |
@@ -2200,10 +2203,10 @@ static irqreturn_t intr_handler(int irq, void *dev_instance) | |||
2200 | 2203 | ||
2201 | prefetch(&np->rx_skbuff[np->cur_rx % RX_RING_SIZE]); | 2204 | prefetch(&np->rx_skbuff[np->cur_rx % RX_RING_SIZE]); |
2202 | 2205 | ||
2203 | if (netif_rx_schedule_prep(dev)) { | 2206 | if (netif_rx_schedule_prep(dev, &np->napi)) { |
2204 | /* Disable interrupts and register for poll */ | 2207 | /* Disable interrupts and register for poll */ |
2205 | natsemi_irq_disable(dev); | 2208 | natsemi_irq_disable(dev); |
2206 | __netif_rx_schedule(dev); | 2209 | __netif_rx_schedule(dev, &np->napi); |
2207 | } else | 2210 | } else |
2208 | printk(KERN_WARNING | 2211 | printk(KERN_WARNING |
2209 | "%s: Ignoring interrupt, status %#08x, mask %#08x.\n", | 2212 | "%s: Ignoring interrupt, status %#08x, mask %#08x.\n", |
@@ -2216,12 +2219,11 @@ static irqreturn_t intr_handler(int irq, void *dev_instance) | |||
2216 | /* This is the NAPI poll routine. As well as the standard RX handling | 2219 | /* This is the NAPI poll routine. As well as the standard RX handling |
2217 | * it also handles all other interrupts that the chip might raise. | 2220 | * it also handles all other interrupts that the chip might raise. |
2218 | */ | 2221 | */ |
2219 | static int natsemi_poll(struct net_device *dev, int *budget) | 2222 | static int natsemi_poll(struct napi_struct *napi, int budget) |
2220 | { | 2223 | { |
2221 | struct netdev_private *np = netdev_priv(dev); | 2224 | struct netdev_private *np = container_of(napi, struct netdev_private, napi); |
2225 | struct net_device *dev = np->dev; | ||
2222 | void __iomem * ioaddr = ns_ioaddr(dev); | 2226 | void __iomem * ioaddr = ns_ioaddr(dev); |
2223 | |||
2224 | int work_to_do = min(*budget, dev->quota); | ||
2225 | int work_done = 0; | 2227 | int work_done = 0; |
2226 | 2228 | ||
2227 | do { | 2229 | do { |
@@ -2236,7 +2238,7 @@ static int natsemi_poll(struct net_device *dev, int *budget) | |||
2236 | if (np->intr_status & | 2238 | if (np->intr_status & |
2237 | (IntrRxDone | IntrRxIntr | RxStatusFIFOOver | | 2239 | (IntrRxDone | IntrRxIntr | RxStatusFIFOOver | |
2238 | IntrRxErr | IntrRxOverrun)) { | 2240 | IntrRxErr | IntrRxOverrun)) { |
2239 | netdev_rx(dev, &work_done, work_to_do); | 2241 | netdev_rx(dev, &work_done, budget); |
2240 | } | 2242 | } |
2241 | 2243 | ||
2242 | if (np->intr_status & | 2244 | if (np->intr_status & |
@@ -2250,16 +2252,13 @@ static int natsemi_poll(struct net_device *dev, int *budget) | |||
2250 | if (np->intr_status & IntrAbnormalSummary) | 2252 | if (np->intr_status & IntrAbnormalSummary) |
2251 | netdev_error(dev, np->intr_status); | 2253 | netdev_error(dev, np->intr_status); |
2252 | 2254 | ||
2253 | *budget -= work_done; | 2255 | if (work_done >= budget) |
2254 | dev->quota -= work_done; | 2256 | return work_done; |
2255 | |||
2256 | if (work_done >= work_to_do) | ||
2257 | return 1; | ||
2258 | 2257 | ||
2259 | np->intr_status = readl(ioaddr + IntrStatus); | 2258 | np->intr_status = readl(ioaddr + IntrStatus); |
2260 | } while (np->intr_status); | 2259 | } while (np->intr_status); |
2261 | 2260 | ||
2262 | netif_rx_complete(dev); | 2261 | netif_rx_complete(dev, napi); |
2263 | 2262 | ||
2264 | /* Reenable interrupts providing nothing is trying to shut | 2263 | /* Reenable interrupts providing nothing is trying to shut |
2265 | * the chip down. */ | 2264 | * the chip down. */ |
@@ -2268,7 +2267,7 @@ static int natsemi_poll(struct net_device *dev, int *budget) | |||
2268 | natsemi_irq_enable(dev); | 2267 | natsemi_irq_enable(dev); |
2269 | spin_unlock(&np->lock); | 2268 | spin_unlock(&np->lock); |
2270 | 2269 | ||
2271 | return 0; | 2270 | return work_done; |
2272 | } | 2271 | } |
2273 | 2272 | ||
2274 | /* This routine is logically part of the interrupt handler, but separated | 2273 | /* This routine is logically part of the interrupt handler, but separated |
@@ -3158,6 +3157,8 @@ static int netdev_close(struct net_device *dev) | |||
3158 | dev->name, np->cur_tx, np->dirty_tx, | 3157 | dev->name, np->cur_tx, np->dirty_tx, |
3159 | np->cur_rx, np->dirty_rx); | 3158 | np->cur_rx, np->dirty_rx); |
3160 | 3159 | ||
3160 | napi_disable(&np->napi); | ||
3161 | |||
3161 | /* | 3162 | /* |
3162 | * FIXME: what if someone tries to close a device | 3163 | * FIXME: what if someone tries to close a device |
3163 | * that is suspended? | 3164 | * that is suspended? |
@@ -3253,7 +3254,7 @@ static void __devexit natsemi_remove1 (struct pci_dev *pdev) | |||
3253 | * disable_irq() to enforce synchronization. | 3254 | * disable_irq() to enforce synchronization. |
3254 | * * natsemi_poll: checks before reenabling interrupts. suspend | 3255 | * * natsemi_poll: checks before reenabling interrupts. suspend |
3255 | * sets hands_off, disables interrupts and then waits with | 3256 | * sets hands_off, disables interrupts and then waits with |
3256 | * netif_poll_disable(). | 3257 | * napi_disable(). |
3257 | * | 3258 | * |
3258 | * Interrupts must be disabled, otherwise hands_off can cause irq storms. | 3259 | * Interrupts must be disabled, otherwise hands_off can cause irq storms. |
3259 | */ | 3260 | */ |
@@ -3279,7 +3280,7 @@ static int natsemi_suspend (struct pci_dev *pdev, pm_message_t state) | |||
3279 | spin_unlock_irq(&np->lock); | 3280 | spin_unlock_irq(&np->lock); |
3280 | enable_irq(dev->irq); | 3281 | enable_irq(dev->irq); |
3281 | 3282 | ||
3282 | netif_poll_disable(dev); | 3283 | napi_disable(&np->napi); |
3283 | 3284 | ||
3284 | /* Update the error counts. */ | 3285 | /* Update the error counts. */ |
3285 | __get_stats(dev); | 3286 | __get_stats(dev); |
@@ -3320,6 +3321,8 @@ static int natsemi_resume (struct pci_dev *pdev) | |||
3320 | pci_enable_device(pdev); | 3321 | pci_enable_device(pdev); |
3321 | /* pci_power_on(pdev); */ | 3322 | /* pci_power_on(pdev); */ |
3322 | 3323 | ||
3324 | napi_enable(&np->napi); | ||
3325 | |||
3323 | natsemi_reset(dev); | 3326 | natsemi_reset(dev); |
3324 | init_ring(dev); | 3327 | init_ring(dev); |
3325 | disable_irq(dev->irq); | 3328 | disable_irq(dev->irq); |
@@ -3333,7 +3336,6 @@ static int natsemi_resume (struct pci_dev *pdev) | |||
3333 | mod_timer(&np->timer, jiffies + 1*HZ); | 3336 | mod_timer(&np->timer, jiffies + 1*HZ); |
3334 | } | 3337 | } |
3335 | netif_device_attach(dev); | 3338 | netif_device_attach(dev); |
3336 | netif_poll_enable(dev); | ||
3337 | out: | 3339 | out: |
3338 | rtnl_unlock(); | 3340 | rtnl_unlock(); |
3339 | return 0; | 3341 | return 0; |
diff --git a/drivers/net/netxen/netxen_nic.h b/drivers/net/netxen/netxen_nic.h index d4c92cc879d4..aaa34939485b 100644 --- a/drivers/net/netxen/netxen_nic.h +++ b/drivers/net/netxen/netxen_nic.h | |||
@@ -880,6 +880,7 @@ struct netxen_adapter { | |||
880 | struct netxen_adapter *master; | 880 | struct netxen_adapter *master; |
881 | struct net_device *netdev; | 881 | struct net_device *netdev; |
882 | struct pci_dev *pdev; | 882 | struct pci_dev *pdev; |
883 | struct napi_struct napi; | ||
883 | struct net_device_stats net_stats; | 884 | struct net_device_stats net_stats; |
884 | unsigned char mac_addr[ETH_ALEN]; | 885 | unsigned char mac_addr[ETH_ALEN]; |
885 | int mtu; | 886 | int mtu; |
diff --git a/drivers/net/netxen/netxen_nic_main.c b/drivers/net/netxen/netxen_nic_main.c index 3122d0101638..a10bbefbdadd 100644 --- a/drivers/net/netxen/netxen_nic_main.c +++ b/drivers/net/netxen/netxen_nic_main.c | |||
@@ -68,7 +68,7 @@ static void netxen_tx_timeout(struct net_device *netdev); | |||
68 | static void netxen_tx_timeout_task(struct work_struct *work); | 68 | static void netxen_tx_timeout_task(struct work_struct *work); |
69 | static void netxen_watchdog(unsigned long); | 69 | static void netxen_watchdog(unsigned long); |
70 | static int netxen_handle_int(struct netxen_adapter *, struct net_device *); | 70 | static int netxen_handle_int(struct netxen_adapter *, struct net_device *); |
71 | static int netxen_nic_poll(struct net_device *dev, int *budget); | 71 | static int netxen_nic_poll(struct napi_struct *napi, int budget); |
72 | #ifdef CONFIG_NET_POLL_CONTROLLER | 72 | #ifdef CONFIG_NET_POLL_CONTROLLER |
73 | static void netxen_nic_poll_controller(struct net_device *netdev); | 73 | static void netxen_nic_poll_controller(struct net_device *netdev); |
74 | #endif | 74 | #endif |
@@ -402,6 +402,9 @@ netxen_nic_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
402 | adapter->netdev = netdev; | 402 | adapter->netdev = netdev; |
403 | adapter->pdev = pdev; | 403 | adapter->pdev = pdev; |
404 | 404 | ||
405 | netif_napi_add(netdev, &adapter->napi, | ||
406 | netxen_nic_poll, NETXEN_NETDEV_WEIGHT); | ||
407 | |||
405 | /* this will be read from FW later */ | 408 | /* this will be read from FW later */ |
406 | adapter->intr_scheme = -1; | 409 | adapter->intr_scheme = -1; |
407 | 410 | ||
@@ -422,8 +425,6 @@ netxen_nic_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
422 | netxen_nic_change_mtu(netdev, netdev->mtu); | 425 | netxen_nic_change_mtu(netdev, netdev->mtu); |
423 | 426 | ||
424 | SET_ETHTOOL_OPS(netdev, &netxen_nic_ethtool_ops); | 427 | SET_ETHTOOL_OPS(netdev, &netxen_nic_ethtool_ops); |
425 | netdev->poll = netxen_nic_poll; | ||
426 | netdev->weight = NETXEN_NETDEV_WEIGHT; | ||
427 | #ifdef CONFIG_NET_POLL_CONTROLLER | 428 | #ifdef CONFIG_NET_POLL_CONTROLLER |
428 | netdev->poll_controller = netxen_nic_poll_controller; | 429 | netdev->poll_controller = netxen_nic_poll_controller; |
429 | #endif | 430 | #endif |
@@ -885,6 +886,8 @@ static int netxen_nic_open(struct net_device *netdev) | |||
885 | if (!adapter->driver_mismatch) | 886 | if (!adapter->driver_mismatch) |
886 | mod_timer(&adapter->watchdog_timer, jiffies); | 887 | mod_timer(&adapter->watchdog_timer, jiffies); |
887 | 888 | ||
889 | napi_enable(&adapter->napi); | ||
890 | |||
888 | netxen_nic_enable_int(adapter); | 891 | netxen_nic_enable_int(adapter); |
889 | 892 | ||
890 | /* Done here again so that even if phantom sw overwrote it, | 893 | /* Done here again so that even if phantom sw overwrote it, |
@@ -894,6 +897,7 @@ static int netxen_nic_open(struct net_device *netdev) | |||
894 | del_timer_sync(&adapter->watchdog_timer); | 897 | del_timer_sync(&adapter->watchdog_timer); |
895 | printk(KERN_ERR "%s: Failed to initialize port %d\n", | 898 | printk(KERN_ERR "%s: Failed to initialize port %d\n", |
896 | netxen_nic_driver_name, adapter->portnum); | 899 | netxen_nic_driver_name, adapter->portnum); |
900 | napi_disable(&adapter->napi); | ||
897 | return -EIO; | 901 | return -EIO; |
898 | } | 902 | } |
899 | if (adapter->macaddr_set) | 903 | if (adapter->macaddr_set) |
@@ -923,6 +927,7 @@ static int netxen_nic_close(struct net_device *netdev) | |||
923 | 927 | ||
924 | netif_carrier_off(netdev); | 928 | netif_carrier_off(netdev); |
925 | netif_stop_queue(netdev); | 929 | netif_stop_queue(netdev); |
930 | napi_disable(&adapter->napi); | ||
926 | 931 | ||
927 | netxen_nic_disable_int(adapter); | 932 | netxen_nic_disable_int(adapter); |
928 | 933 | ||
@@ -1243,11 +1248,11 @@ netxen_handle_int(struct netxen_adapter *adapter, struct net_device *netdev) | |||
1243 | netxen_nic_disable_int(adapter); | 1248 | netxen_nic_disable_int(adapter); |
1244 | 1249 | ||
1245 | if (netxen_nic_rx_has_work(adapter) || netxen_nic_tx_has_work(adapter)) { | 1250 | if (netxen_nic_rx_has_work(adapter) || netxen_nic_tx_has_work(adapter)) { |
1246 | if (netif_rx_schedule_prep(netdev)) { | 1251 | if (netif_rx_schedule_prep(netdev, &adapter->napi)) { |
1247 | /* | 1252 | /* |
1248 | * Interrupts are already disabled. | 1253 | * Interrupts are already disabled. |
1249 | */ | 1254 | */ |
1250 | __netif_rx_schedule(netdev); | 1255 | __netif_rx_schedule(netdev, &adapter->napi); |
1251 | } else { | 1256 | } else { |
1252 | static unsigned int intcount = 0; | 1257 | static unsigned int intcount = 0; |
1253 | if ((++intcount & 0xfff) == 0xfff) | 1258 | if ((++intcount & 0xfff) == 0xfff) |
@@ -1305,14 +1310,13 @@ irqreturn_t netxen_intr(int irq, void *data) | |||
1305 | return IRQ_HANDLED; | 1310 | return IRQ_HANDLED; |
1306 | } | 1311 | } |
1307 | 1312 | ||
1308 | static int netxen_nic_poll(struct net_device *netdev, int *budget) | 1313 | static int netxen_nic_poll(struct napi_struct *napi, int budget) |
1309 | { | 1314 | { |
1310 | struct netxen_adapter *adapter = netdev_priv(netdev); | 1315 | struct netxen_adapter *adapter = container_of(napi, struct netxen_adapter, napi); |
1311 | int work_to_do = min(*budget, netdev->quota); | 1316 | struct net_device *netdev = adapter->netdev; |
1312 | int done = 1; | 1317 | int done = 1; |
1313 | int ctx; | 1318 | int ctx; |
1314 | int this_work_done; | 1319 | int work_done; |
1315 | int work_done = 0; | ||
1316 | 1320 | ||
1317 | DPRINTK(INFO, "polling for %d descriptors\n", *budget); | 1321 | DPRINTK(INFO, "polling for %d descriptors\n", *budget); |
1318 | 1322 | ||
@@ -1330,16 +1334,11 @@ static int netxen_nic_poll(struct net_device *netdev, int *budget) | |||
1330 | * packets are on one context, it gets only half of the quota, | 1334 | * packets are on one context, it gets only half of the quota, |
1331 | * and ends up not processing it. | 1335 | * and ends up not processing it. |
1332 | */ | 1336 | */ |
1333 | this_work_done = netxen_process_rcv_ring(adapter, ctx, | 1337 | work_done += netxen_process_rcv_ring(adapter, ctx, |
1334 | work_to_do / | 1338 | budget / MAX_RCV_CTX); |
1335 | MAX_RCV_CTX); | ||
1336 | work_done += this_work_done; | ||
1337 | } | 1339 | } |
1338 | 1340 | ||
1339 | netdev->quota -= work_done; | 1341 | if (work_done >= budget && netxen_nic_rx_has_work(adapter) != 0) |
1340 | *budget -= work_done; | ||
1341 | |||
1342 | if (work_done >= work_to_do && netxen_nic_rx_has_work(adapter) != 0) | ||
1343 | done = 0; | 1342 | done = 0; |
1344 | 1343 | ||
1345 | if (netxen_process_cmd_ring((unsigned long)adapter) == 0) | 1344 | if (netxen_process_cmd_ring((unsigned long)adapter) == 0) |
@@ -1348,11 +1347,11 @@ static int netxen_nic_poll(struct net_device *netdev, int *budget) | |||
1348 | DPRINTK(INFO, "new work_done: %d work_to_do: %d\n", | 1347 | DPRINTK(INFO, "new work_done: %d work_to_do: %d\n", |
1349 | work_done, work_to_do); | 1348 | work_done, work_to_do); |
1350 | if (done) { | 1349 | if (done) { |
1351 | netif_rx_complete(netdev); | 1350 | netif_rx_complete(netdev, napi); |
1352 | netxen_nic_enable_int(adapter); | 1351 | netxen_nic_enable_int(adapter); |
1353 | } | 1352 | } |
1354 | 1353 | ||
1355 | return !done; | 1354 | return work_done; |
1356 | } | 1355 | } |
1357 | 1356 | ||
1358 | #ifdef CONFIG_NET_POLL_CONTROLLER | 1357 | #ifdef CONFIG_NET_POLL_CONTROLLER |
diff --git a/drivers/net/pasemi_mac.c b/drivers/net/pasemi_mac.c index 0b3066a6fe40..e63cc335a4ba 100644 --- a/drivers/net/pasemi_mac.c +++ b/drivers/net/pasemi_mac.c | |||
@@ -584,7 +584,7 @@ static irqreturn_t pasemi_mac_rx_intr(int irq, void *data) | |||
584 | if (*mac->rx_status & PAS_STATUS_TIMER) | 584 | if (*mac->rx_status & PAS_STATUS_TIMER) |
585 | reg |= PAS_IOB_DMA_RXCH_RESET_TINTC; | 585 | reg |= PAS_IOB_DMA_RXCH_RESET_TINTC; |
586 | 586 | ||
587 | netif_rx_schedule(dev); | 587 | netif_rx_schedule(dev, &mac->napi); |
588 | 588 | ||
589 | pci_write_config_dword(mac->iob_pdev, | 589 | pci_write_config_dword(mac->iob_pdev, |
590 | PAS_IOB_DMA_RXCH_RESET(mac->dma_rxch), reg); | 590 | PAS_IOB_DMA_RXCH_RESET(mac->dma_rxch), reg); |
@@ -808,7 +808,7 @@ static int pasemi_mac_open(struct net_device *dev) | |||
808 | dev_warn(&mac->pdev->dev, "phy init failed: %d\n", ret); | 808 | dev_warn(&mac->pdev->dev, "phy init failed: %d\n", ret); |
809 | 809 | ||
810 | netif_start_queue(dev); | 810 | netif_start_queue(dev); |
811 | netif_poll_enable(dev); | 811 | napi_enable(&mac->napi); |
812 | 812 | ||
813 | /* Interrupts are a bit different for our DMA controller: While | 813 | /* Interrupts are a bit different for our DMA controller: While |
814 | * it's got one a regular PCI device header, the interrupt there | 814 | * it's got one a regular PCI device header, the interrupt there |
@@ -845,7 +845,7 @@ static int pasemi_mac_open(struct net_device *dev) | |||
845 | out_rx_int: | 845 | out_rx_int: |
846 | free_irq(mac->tx_irq, dev); | 846 | free_irq(mac->tx_irq, dev); |
847 | out_tx_int: | 847 | out_tx_int: |
848 | netif_poll_disable(dev); | 848 | napi_disable(&mac->napi); |
849 | netif_stop_queue(dev); | 849 | netif_stop_queue(dev); |
850 | pasemi_mac_free_tx_resources(dev); | 850 | pasemi_mac_free_tx_resources(dev); |
851 | out_tx_resources: | 851 | out_tx_resources: |
@@ -869,6 +869,7 @@ static int pasemi_mac_close(struct net_device *dev) | |||
869 | } | 869 | } |
870 | 870 | ||
871 | netif_stop_queue(dev); | 871 | netif_stop_queue(dev); |
872 | napi_disable(&mac->napi); | ||
872 | 873 | ||
873 | /* Clean out any pending buffers */ | 874 | /* Clean out any pending buffers */ |
874 | pasemi_mac_clean_tx(mac); | 875 | pasemi_mac_clean_tx(mac); |
@@ -1047,26 +1048,20 @@ static void pasemi_mac_set_rx_mode(struct net_device *dev) | |||
1047 | } | 1048 | } |
1048 | 1049 | ||
1049 | 1050 | ||
1050 | static int pasemi_mac_poll(struct net_device *dev, int *budget) | 1051 | static int pasemi_mac_poll(struct napi_struct *napi, int budget) |
1051 | { | 1052 | { |
1052 | int pkts, limit = min(*budget, dev->quota); | 1053 | struct pasemi_mac *mac = container_of(napi, struct pasemi_mac, napi); |
1053 | struct pasemi_mac *mac = netdev_priv(dev); | 1054 | struct net_device *dev = mac->netdev; |
1054 | 1055 | int pkts; | |
1055 | pkts = pasemi_mac_clean_rx(mac, limit); | ||
1056 | 1056 | ||
1057 | dev->quota -= pkts; | 1057 | pkts = pasemi_mac_clean_rx(mac, budget); |
1058 | *budget -= pkts; | 1058 | if (pkts < budget) { |
1059 | |||
1060 | if (pkts < limit) { | ||
1061 | /* all done, no more packets present */ | 1059 | /* all done, no more packets present */ |
1062 | netif_rx_complete(dev); | 1060 | netif_rx_complete(dev, napi); |
1063 | 1061 | ||
1064 | pasemi_mac_restart_rx_intr(mac); | 1062 | pasemi_mac_restart_rx_intr(mac); |
1065 | return 0; | ||
1066 | } else { | ||
1067 | /* used up our quantum, so reschedule */ | ||
1068 | return 1; | ||
1069 | } | 1063 | } |
1064 | return pkts; | ||
1070 | } | 1065 | } |
1071 | 1066 | ||
1072 | static int __devinit | 1067 | static int __devinit |
@@ -1099,6 +1094,10 @@ pasemi_mac_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1099 | mac->netdev = dev; | 1094 | mac->netdev = dev; |
1100 | mac->dma_pdev = pci_get_device(PCI_VENDOR_ID_PASEMI, 0xa007, NULL); | 1095 | mac->dma_pdev = pci_get_device(PCI_VENDOR_ID_PASEMI, 0xa007, NULL); |
1101 | 1096 | ||
1097 | netif_napi_add(dev, &mac->napi, pasemi_mac_poll, 64); | ||
1098 | |||
1099 | dev->features = NETIF_F_HW_CSUM; | ||
1100 | |||
1102 | if (!mac->dma_pdev) { | 1101 | if (!mac->dma_pdev) { |
1103 | dev_err(&pdev->dev, "Can't find DMA Controller\n"); | 1102 | dev_err(&pdev->dev, "Can't find DMA Controller\n"); |
1104 | err = -ENODEV; | 1103 | err = -ENODEV; |
@@ -1150,9 +1149,6 @@ pasemi_mac_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1150 | dev->hard_start_xmit = pasemi_mac_start_tx; | 1149 | dev->hard_start_xmit = pasemi_mac_start_tx; |
1151 | dev->get_stats = pasemi_mac_get_stats; | 1150 | dev->get_stats = pasemi_mac_get_stats; |
1152 | dev->set_multicast_list = pasemi_mac_set_rx_mode; | 1151 | dev->set_multicast_list = pasemi_mac_set_rx_mode; |
1153 | dev->weight = 64; | ||
1154 | dev->poll = pasemi_mac_poll; | ||
1155 | dev->features = NETIF_F_HW_CSUM; | ||
1156 | 1152 | ||
1157 | /* The dma status structure is located in the I/O bridge, and | 1153 | /* The dma status structure is located in the I/O bridge, and |
1158 | * is cache coherent. | 1154 | * is cache coherent. |
diff --git a/drivers/net/pasemi_mac.h b/drivers/net/pasemi_mac.h index c29ee159c33d..85d3b7856e5f 100644 --- a/drivers/net/pasemi_mac.h +++ b/drivers/net/pasemi_mac.h | |||
@@ -56,6 +56,7 @@ struct pasemi_mac { | |||
56 | struct pci_dev *dma_pdev; | 56 | struct pci_dev *dma_pdev; |
57 | struct pci_dev *iob_pdev; | 57 | struct pci_dev *iob_pdev; |
58 | struct phy_device *phydev; | 58 | struct phy_device *phydev; |
59 | struct napi_struct napi; | ||
59 | struct net_device_stats stats; | 60 | struct net_device_stats stats; |
60 | 61 | ||
61 | /* Pointer to the cacheable per-channel status registers */ | 62 | /* Pointer to the cacheable per-channel status registers */ |
diff --git a/drivers/net/pcnet32.c b/drivers/net/pcnet32.c index e6a67531de99..a9973490dba9 100644 --- a/drivers/net/pcnet32.c +++ b/drivers/net/pcnet32.c | |||
@@ -280,6 +280,8 @@ struct pcnet32_private { | |||
280 | unsigned int dirty_rx, /* ring entries to be freed. */ | 280 | unsigned int dirty_rx, /* ring entries to be freed. */ |
281 | dirty_tx; | 281 | dirty_tx; |
282 | 282 | ||
283 | struct net_device *dev; | ||
284 | struct napi_struct napi; | ||
283 | struct net_device_stats stats; | 285 | struct net_device_stats stats; |
284 | char tx_full; | 286 | char tx_full; |
285 | char phycount; /* number of phys found */ | 287 | char phycount; /* number of phys found */ |
@@ -440,15 +442,21 @@ static struct pcnet32_access pcnet32_dwio = { | |||
440 | 442 | ||
441 | static void pcnet32_netif_stop(struct net_device *dev) | 443 | static void pcnet32_netif_stop(struct net_device *dev) |
442 | { | 444 | { |
445 | struct pcnet32_private *lp = netdev_priv(dev); | ||
443 | dev->trans_start = jiffies; | 446 | dev->trans_start = jiffies; |
444 | netif_poll_disable(dev); | 447 | #ifdef CONFIG_PCNET32_NAPI |
448 | napi_disable(&lp->napi); | ||
449 | #endif | ||
445 | netif_tx_disable(dev); | 450 | netif_tx_disable(dev); |
446 | } | 451 | } |
447 | 452 | ||
448 | static void pcnet32_netif_start(struct net_device *dev) | 453 | static void pcnet32_netif_start(struct net_device *dev) |
449 | { | 454 | { |
455 | struct pcnet32_private *lp = netdev_priv(dev); | ||
450 | netif_wake_queue(dev); | 456 | netif_wake_queue(dev); |
451 | netif_poll_enable(dev); | 457 | #ifdef CONFIG_PCNET32_NAPI |
458 | napi_enable(&lp->napi); | ||
459 | #endif | ||
452 | } | 460 | } |
453 | 461 | ||
454 | /* | 462 | /* |
@@ -816,7 +824,7 @@ static int pcnet32_set_ringparam(struct net_device *dev, | |||
816 | if ((1 << i) != lp->rx_ring_size) | 824 | if ((1 << i) != lp->rx_ring_size) |
817 | pcnet32_realloc_rx_ring(dev, lp, i); | 825 | pcnet32_realloc_rx_ring(dev, lp, i); |
818 | 826 | ||
819 | dev->weight = lp->rx_ring_size / 2; | 827 | lp->napi.weight = lp->rx_ring_size / 2; |
820 | 828 | ||
821 | if (netif_running(dev)) { | 829 | if (netif_running(dev)) { |
822 | pcnet32_netif_start(dev); | 830 | pcnet32_netif_start(dev); |
@@ -1255,7 +1263,7 @@ static void pcnet32_rx_entry(struct net_device *dev, | |||
1255 | return; | 1263 | return; |
1256 | } | 1264 | } |
1257 | 1265 | ||
1258 | static int pcnet32_rx(struct net_device *dev, int quota) | 1266 | static int pcnet32_rx(struct net_device *dev, int budget) |
1259 | { | 1267 | { |
1260 | struct pcnet32_private *lp = netdev_priv(dev); | 1268 | struct pcnet32_private *lp = netdev_priv(dev); |
1261 | int entry = lp->cur_rx & lp->rx_mod_mask; | 1269 | int entry = lp->cur_rx & lp->rx_mod_mask; |
@@ -1263,7 +1271,7 @@ static int pcnet32_rx(struct net_device *dev, int quota) | |||
1263 | int npackets = 0; | 1271 | int npackets = 0; |
1264 | 1272 | ||
1265 | /* If we own the next entry, it's a new packet. Send it up. */ | 1273 | /* If we own the next entry, it's a new packet. Send it up. */ |
1266 | while (quota > npackets && (short)le16_to_cpu(rxp->status) >= 0) { | 1274 | while (npackets < budget && (short)le16_to_cpu(rxp->status) >= 0) { |
1267 | pcnet32_rx_entry(dev, lp, rxp, entry); | 1275 | pcnet32_rx_entry(dev, lp, rxp, entry); |
1268 | npackets += 1; | 1276 | npackets += 1; |
1269 | /* | 1277 | /* |
@@ -1379,15 +1387,16 @@ static int pcnet32_tx(struct net_device *dev) | |||
1379 | } | 1387 | } |
1380 | 1388 | ||
1381 | #ifdef CONFIG_PCNET32_NAPI | 1389 | #ifdef CONFIG_PCNET32_NAPI |
1382 | static int pcnet32_poll(struct net_device *dev, int *budget) | 1390 | static int pcnet32_poll(struct napi_struct *napi, int budget) |
1383 | { | 1391 | { |
1384 | struct pcnet32_private *lp = netdev_priv(dev); | 1392 | struct pcnet32_private *lp = container_of(napi, struct pcnet32_private, napi); |
1385 | int quota = min(dev->quota, *budget); | 1393 | struct net_device *dev = lp->dev; |
1386 | unsigned long ioaddr = dev->base_addr; | 1394 | unsigned long ioaddr = dev->base_addr; |
1387 | unsigned long flags; | 1395 | unsigned long flags; |
1396 | int work_done; | ||
1388 | u16 val; | 1397 | u16 val; |
1389 | 1398 | ||
1390 | quota = pcnet32_rx(dev, quota); | 1399 | work_done = pcnet32_rx(dev, budget); |
1391 | 1400 | ||
1392 | spin_lock_irqsave(&lp->lock, flags); | 1401 | spin_lock_irqsave(&lp->lock, flags); |
1393 | if (pcnet32_tx(dev)) { | 1402 | if (pcnet32_tx(dev)) { |
@@ -1399,28 +1408,22 @@ static int pcnet32_poll(struct net_device *dev, int *budget) | |||
1399 | } | 1408 | } |
1400 | spin_unlock_irqrestore(&lp->lock, flags); | 1409 | spin_unlock_irqrestore(&lp->lock, flags); |
1401 | 1410 | ||
1402 | *budget -= quota; | 1411 | if (work_done < budget) { |
1403 | dev->quota -= quota; | 1412 | spin_lock_irqsave(&lp->lock, flags); |
1404 | |||
1405 | if (dev->quota == 0) { | ||
1406 | return 1; | ||
1407 | } | ||
1408 | |||
1409 | netif_rx_complete(dev); | ||
1410 | |||
1411 | spin_lock_irqsave(&lp->lock, flags); | ||
1412 | 1413 | ||
1413 | /* clear interrupt masks */ | 1414 | __netif_rx_complete(dev, napi); |
1414 | val = lp->a.read_csr(ioaddr, CSR3); | ||
1415 | val &= 0x00ff; | ||
1416 | lp->a.write_csr(ioaddr, CSR3, val); | ||
1417 | 1415 | ||
1418 | /* Set interrupt enable. */ | 1416 | /* clear interrupt masks */ |
1419 | lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN); | 1417 | val = lp->a.read_csr(ioaddr, CSR3); |
1420 | mmiowb(); | 1418 | val &= 0x00ff; |
1421 | spin_unlock_irqrestore(&lp->lock, flags); | 1419 | lp->a.write_csr(ioaddr, CSR3, val); |
1422 | 1420 | ||
1423 | return 0; | 1421 | /* Set interrupt enable. */ |
1422 | lp->a.write_csr(ioaddr, CSR0, CSR0_INTEN); | ||
1423 | mmiowb(); | ||
1424 | spin_unlock_irqrestore(&lp->lock, flags); | ||
1425 | } | ||
1426 | return work_done; | ||
1424 | } | 1427 | } |
1425 | #endif | 1428 | #endif |
1426 | 1429 | ||
@@ -1815,6 +1818,8 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) | |||
1815 | } | 1818 | } |
1816 | lp->pci_dev = pdev; | 1819 | lp->pci_dev = pdev; |
1817 | 1820 | ||
1821 | lp->dev = dev; | ||
1822 | |||
1818 | spin_lock_init(&lp->lock); | 1823 | spin_lock_init(&lp->lock); |
1819 | 1824 | ||
1820 | SET_MODULE_OWNER(dev); | 1825 | SET_MODULE_OWNER(dev); |
@@ -1843,6 +1848,10 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) | |||
1843 | lp->mii_if.mdio_read = mdio_read; | 1848 | lp->mii_if.mdio_read = mdio_read; |
1844 | lp->mii_if.mdio_write = mdio_write; | 1849 | lp->mii_if.mdio_write = mdio_write; |
1845 | 1850 | ||
1851 | #ifdef CONFIG_PCNET32_NAPI | ||
1852 | netif_napi_add(dev, &lp->napi, pcnet32_poll, lp->rx_ring_size / 2); | ||
1853 | #endif | ||
1854 | |||
1846 | if (fdx && !(lp->options & PCNET32_PORT_ASEL) && | 1855 | if (fdx && !(lp->options & PCNET32_PORT_ASEL) && |
1847 | ((cards_found >= MAX_UNITS) || full_duplex[cards_found])) | 1856 | ((cards_found >= MAX_UNITS) || full_duplex[cards_found])) |
1848 | lp->options |= PCNET32_PORT_FD; | 1857 | lp->options |= PCNET32_PORT_FD; |
@@ -1953,10 +1962,6 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) | |||
1953 | dev->ethtool_ops = &pcnet32_ethtool_ops; | 1962 | dev->ethtool_ops = &pcnet32_ethtool_ops; |
1954 | dev->tx_timeout = pcnet32_tx_timeout; | 1963 | dev->tx_timeout = pcnet32_tx_timeout; |
1955 | dev->watchdog_timeo = (5 * HZ); | 1964 | dev->watchdog_timeo = (5 * HZ); |
1956 | dev->weight = lp->rx_ring_size / 2; | ||
1957 | #ifdef CONFIG_PCNET32_NAPI | ||
1958 | dev->poll = pcnet32_poll; | ||
1959 | #endif | ||
1960 | 1965 | ||
1961 | #ifdef CONFIG_NET_POLL_CONTROLLER | 1966 | #ifdef CONFIG_NET_POLL_CONTROLLER |
1962 | dev->poll_controller = pcnet32_poll_controller; | 1967 | dev->poll_controller = pcnet32_poll_controller; |
@@ -2276,6 +2281,10 @@ static int pcnet32_open(struct net_device *dev) | |||
2276 | goto err_free_ring; | 2281 | goto err_free_ring; |
2277 | } | 2282 | } |
2278 | 2283 | ||
2284 | #ifdef CONFIG_PCNET32_NAPI | ||
2285 | napi_enable(&lp->napi); | ||
2286 | #endif | ||
2287 | |||
2279 | /* Re-initialize the PCNET32, and start it when done. */ | 2288 | /* Re-initialize the PCNET32, and start it when done. */ |
2280 | lp->a.write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff)); | 2289 | lp->a.write_csr(ioaddr, 1, (lp->init_dma_addr & 0xffff)); |
2281 | lp->a.write_csr(ioaddr, 2, (lp->init_dma_addr >> 16)); | 2290 | lp->a.write_csr(ioaddr, 2, (lp->init_dma_addr >> 16)); |
@@ -2599,18 +2608,18 @@ pcnet32_interrupt(int irq, void *dev_id) | |||
2599 | /* unlike for the lance, there is no restart needed */ | 2608 | /* unlike for the lance, there is no restart needed */ |
2600 | } | 2609 | } |
2601 | #ifdef CONFIG_PCNET32_NAPI | 2610 | #ifdef CONFIG_PCNET32_NAPI |
2602 | if (netif_rx_schedule_prep(dev)) { | 2611 | if (netif_rx_schedule_prep(dev, &lp->napi)) { |
2603 | u16 val; | 2612 | u16 val; |
2604 | /* set interrupt masks */ | 2613 | /* set interrupt masks */ |
2605 | val = lp->a.read_csr(ioaddr, CSR3); | 2614 | val = lp->a.read_csr(ioaddr, CSR3); |
2606 | val |= 0x5f00; | 2615 | val |= 0x5f00; |
2607 | lp->a.write_csr(ioaddr, CSR3, val); | 2616 | lp->a.write_csr(ioaddr, CSR3, val); |
2608 | mmiowb(); | 2617 | mmiowb(); |
2609 | __netif_rx_schedule(dev); | 2618 | __netif_rx_schedule(dev, &lp->napi); |
2610 | break; | 2619 | break; |
2611 | } | 2620 | } |
2612 | #else | 2621 | #else |
2613 | pcnet32_rx(dev, dev->weight); | 2622 | pcnet32_rx(dev, lp->napi.weight); |
2614 | if (pcnet32_tx(dev)) { | 2623 | if (pcnet32_tx(dev)) { |
2615 | /* reset the chip to clear the error condition, then restart */ | 2624 | /* reset the chip to clear the error condition, then restart */ |
2616 | lp->a.reset(ioaddr); | 2625 | lp->a.reset(ioaddr); |
@@ -2645,6 +2654,9 @@ static int pcnet32_close(struct net_device *dev) | |||
2645 | del_timer_sync(&lp->watchdog_timer); | 2654 | del_timer_sync(&lp->watchdog_timer); |
2646 | 2655 | ||
2647 | netif_stop_queue(dev); | 2656 | netif_stop_queue(dev); |
2657 | #ifdef CONFIG_PCNET32_NAPI | ||
2658 | napi_disable(&lp->napi); | ||
2659 | #endif | ||
2648 | 2660 | ||
2649 | spin_lock_irqsave(&lp->lock, flags); | 2661 | spin_lock_irqsave(&lp->lock, flags); |
2650 | 2662 | ||
diff --git a/drivers/net/ps3_gelic_net.c b/drivers/net/ps3_gelic_net.c index e56503918436..92561c0450bc 100644 --- a/drivers/net/ps3_gelic_net.c +++ b/drivers/net/ps3_gelic_net.c | |||
@@ -556,6 +556,7 @@ static int gelic_net_stop(struct net_device *netdev) | |||
556 | { | 556 | { |
557 | struct gelic_net_card *card = netdev_priv(netdev); | 557 | struct gelic_net_card *card = netdev_priv(netdev); |
558 | 558 | ||
559 | napi_disable(&card->napi); | ||
559 | netif_stop_queue(netdev); | 560 | netif_stop_queue(netdev); |
560 | 561 | ||
561 | /* turn off DMA, force end */ | 562 | /* turn off DMA, force end */ |
@@ -987,32 +988,24 @@ refill: | |||
987 | * if the quota is exceeded, but the driver has still packets. | 988 | * if the quota is exceeded, but the driver has still packets. |
988 | * | 989 | * |
989 | */ | 990 | */ |
990 | static int gelic_net_poll(struct net_device *netdev, int *budget) | 991 | static int gelic_net_poll(struct napi_struct *napi, int budget) |
991 | { | 992 | { |
992 | struct gelic_net_card *card = netdev_priv(netdev); | 993 | struct gelic_net_card *card = container_of(napi, struct gelic_net_card, napi); |
993 | int packets_to_do, packets_done = 0; | 994 | struct net_device *netdev = card->netdev; |
994 | int no_more_packets = 0; | 995 | int packets_done = 0; |
995 | |||
996 | packets_to_do = min(*budget, netdev->quota); | ||
997 | 996 | ||
998 | while (packets_to_do) { | 997 | while (packets_done < budget) { |
999 | if (gelic_net_decode_one_descr(card)) { | 998 | if (!gelic_net_decode_one_descr(card)) |
1000 | packets_done++; | ||
1001 | packets_to_do--; | ||
1002 | } else { | ||
1003 | /* no more packets for the stack */ | ||
1004 | no_more_packets = 1; | ||
1005 | break; | 999 | break; |
1006 | } | 1000 | |
1001 | packets_done++; | ||
1007 | } | 1002 | } |
1008 | netdev->quota -= packets_done; | 1003 | |
1009 | *budget -= packets_done; | 1004 | if (packets_done < budget) { |
1010 | if (no_more_packets) { | 1005 | netif_rx_complete(netdev, napi); |
1011 | netif_rx_complete(netdev); | ||
1012 | gelic_net_rx_irq_on(card); | 1006 | gelic_net_rx_irq_on(card); |
1013 | return 0; | 1007 | } |
1014 | } else | 1008 | return packets_done; |
1015 | return 1; | ||
1016 | } | 1009 | } |
1017 | /** | 1010 | /** |
1018 | * gelic_net_change_mtu - changes the MTU of an interface | 1011 | * gelic_net_change_mtu - changes the MTU of an interface |
@@ -1055,7 +1048,7 @@ static irqreturn_t gelic_net_interrupt(int irq, void *ptr) | |||
1055 | 1048 | ||
1056 | if (status & GELIC_NET_RXINT) { | 1049 | if (status & GELIC_NET_RXINT) { |
1057 | gelic_net_rx_irq_off(card); | 1050 | gelic_net_rx_irq_off(card); |
1058 | netif_rx_schedule(netdev); | 1051 | netif_rx_schedule(netdev, &card->napi); |
1059 | } | 1052 | } |
1060 | 1053 | ||
1061 | if (status & GELIC_NET_TXINT) { | 1054 | if (status & GELIC_NET_TXINT) { |
@@ -1159,6 +1152,8 @@ static int gelic_net_open(struct net_device *netdev) | |||
1159 | if (gelic_net_alloc_rx_skbs(card)) | 1152 | if (gelic_net_alloc_rx_skbs(card)) |
1160 | goto alloc_skbs_failed; | 1153 | goto alloc_skbs_failed; |
1161 | 1154 | ||
1155 | napi_enable(&card->napi); | ||
1156 | |||
1162 | card->tx_dma_progress = 0; | 1157 | card->tx_dma_progress = 0; |
1163 | card->ghiintmask = GELIC_NET_RXINT | GELIC_NET_TXINT; | 1158 | card->ghiintmask = GELIC_NET_RXINT | GELIC_NET_TXINT; |
1164 | 1159 | ||
@@ -1360,9 +1355,6 @@ static void gelic_net_setup_netdev_ops(struct net_device *netdev) | |||
1360 | /* tx watchdog */ | 1355 | /* tx watchdog */ |
1361 | netdev->tx_timeout = &gelic_net_tx_timeout; | 1356 | netdev->tx_timeout = &gelic_net_tx_timeout; |
1362 | netdev->watchdog_timeo = GELIC_NET_WATCHDOG_TIMEOUT; | 1357 | netdev->watchdog_timeo = GELIC_NET_WATCHDOG_TIMEOUT; |
1363 | /* NAPI */ | ||
1364 | netdev->poll = &gelic_net_poll; | ||
1365 | netdev->weight = GELIC_NET_NAPI_WEIGHT; | ||
1366 | netdev->ethtool_ops = &gelic_net_ethtool_ops; | 1358 | netdev->ethtool_ops = &gelic_net_ethtool_ops; |
1367 | } | 1359 | } |
1368 | 1360 | ||
@@ -1390,6 +1382,9 @@ static int gelic_net_setup_netdev(struct gelic_net_card *card) | |||
1390 | 1382 | ||
1391 | gelic_net_setup_netdev_ops(netdev); | 1383 | gelic_net_setup_netdev_ops(netdev); |
1392 | 1384 | ||
1385 | netif_napi_add(netdev, &card->napi, | ||
1386 | gelic_net_poll, GELIC_NET_NAPI_WEIGHT); | ||
1387 | |||
1393 | netdev->features = NETIF_F_IP_CSUM; | 1388 | netdev->features = NETIF_F_IP_CSUM; |
1394 | 1389 | ||
1395 | status = lv1_net_control(bus_id(card), dev_id(card), | 1390 | status = lv1_net_control(bus_id(card), dev_id(card), |
diff --git a/drivers/net/ps3_gelic_net.h b/drivers/net/ps3_gelic_net.h index a9c4c4fc2547..968560269a3b 100644 --- a/drivers/net/ps3_gelic_net.h +++ b/drivers/net/ps3_gelic_net.h | |||
@@ -194,6 +194,7 @@ struct gelic_net_descr_chain { | |||
194 | 194 | ||
195 | struct gelic_net_card { | 195 | struct gelic_net_card { |
196 | struct net_device *netdev; | 196 | struct net_device *netdev; |
197 | struct napi_struct napi; | ||
197 | /* | 198 | /* |
198 | * hypervisor requires irq_status should be | 199 | * hypervisor requires irq_status should be |
199 | * 8 bytes aligned, but u64 member is | 200 | * 8 bytes aligned, but u64 member is |
diff --git a/drivers/net/qla3xxx.c b/drivers/net/qla3xxx.c index ea151315050c..bf9f8f64ba67 100755 --- a/drivers/net/qla3xxx.c +++ b/drivers/net/qla3xxx.c | |||
@@ -2310,10 +2310,10 @@ static int ql_tx_rx_clean(struct ql3_adapter *qdev, | |||
2310 | return work_done; | 2310 | return work_done; |
2311 | } | 2311 | } |
2312 | 2312 | ||
2313 | static int ql_poll(struct net_device *ndev, int *budget) | 2313 | static int ql_poll(struct napi_struct *napi, int budget) |
2314 | { | 2314 | { |
2315 | struct ql3_adapter *qdev = netdev_priv(ndev); | 2315 | struct ql3_adapter *qdev = container_of(napi, struct ql3_adapter, napi); |
2316 | int work_to_do = min(*budget, ndev->quota); | 2316 | struct net_device *ndev = qdev->ndev; |
2317 | int rx_cleaned = 0, tx_cleaned = 0; | 2317 | int rx_cleaned = 0, tx_cleaned = 0; |
2318 | unsigned long hw_flags; | 2318 | unsigned long hw_flags; |
2319 | struct ql3xxx_port_registers __iomem *port_regs = qdev->mem_map_registers; | 2319 | struct ql3xxx_port_registers __iomem *port_regs = qdev->mem_map_registers; |
@@ -2321,16 +2321,13 @@ static int ql_poll(struct net_device *ndev, int *budget) | |||
2321 | if (!netif_carrier_ok(ndev)) | 2321 | if (!netif_carrier_ok(ndev)) |
2322 | goto quit_polling; | 2322 | goto quit_polling; |
2323 | 2323 | ||
2324 | ql_tx_rx_clean(qdev, &tx_cleaned, &rx_cleaned, work_to_do); | 2324 | ql_tx_rx_clean(qdev, &tx_cleaned, &rx_cleaned, budget); |
2325 | *budget -= rx_cleaned; | ||
2326 | ndev->quota -= rx_cleaned; | ||
2327 | 2325 | ||
2328 | if( tx_cleaned + rx_cleaned != work_to_do || | 2326 | if (tx_cleaned + rx_cleaned != budget || |
2329 | !netif_running(ndev)) { | 2327 | !netif_running(ndev)) { |
2330 | quit_polling: | 2328 | quit_polling: |
2331 | netif_rx_complete(ndev); | ||
2332 | |||
2333 | spin_lock_irqsave(&qdev->hw_lock, hw_flags); | 2329 | spin_lock_irqsave(&qdev->hw_lock, hw_flags); |
2330 | __netif_rx_complete(ndev, napi); | ||
2334 | ql_update_small_bufq_prod_index(qdev); | 2331 | ql_update_small_bufq_prod_index(qdev); |
2335 | ql_update_lrg_bufq_prod_index(qdev); | 2332 | ql_update_lrg_bufq_prod_index(qdev); |
2336 | writel(qdev->rsp_consumer_index, | 2333 | writel(qdev->rsp_consumer_index, |
@@ -2338,9 +2335,8 @@ quit_polling: | |||
2338 | spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); | 2335 | spin_unlock_irqrestore(&qdev->hw_lock, hw_flags); |
2339 | 2336 | ||
2340 | ql_enable_interrupts(qdev); | 2337 | ql_enable_interrupts(qdev); |
2341 | return 0; | ||
2342 | } | 2338 | } |
2343 | return 1; | 2339 | return tx_cleaned + rx_cleaned; |
2344 | } | 2340 | } |
2345 | 2341 | ||
2346 | static irqreturn_t ql3xxx_isr(int irq, void *dev_id) | 2342 | static irqreturn_t ql3xxx_isr(int irq, void *dev_id) |
@@ -2390,8 +2386,8 @@ static irqreturn_t ql3xxx_isr(int irq, void *dev_id) | |||
2390 | spin_unlock(&qdev->adapter_lock); | 2386 | spin_unlock(&qdev->adapter_lock); |
2391 | } else if (value & ISP_IMR_DISABLE_CMPL_INT) { | 2387 | } else if (value & ISP_IMR_DISABLE_CMPL_INT) { |
2392 | ql_disable_interrupts(qdev); | 2388 | ql_disable_interrupts(qdev); |
2393 | if (likely(netif_rx_schedule_prep(ndev))) { | 2389 | if (likely(netif_rx_schedule_prep(ndev, &qdev->napi))) { |
2394 | __netif_rx_schedule(ndev); | 2390 | __netif_rx_schedule(ndev, &qdev->napi); |
2395 | } | 2391 | } |
2396 | } else { | 2392 | } else { |
2397 | return IRQ_NONE; | 2393 | return IRQ_NONE; |
@@ -3617,7 +3613,7 @@ static int ql_adapter_down(struct ql3_adapter *qdev, int do_reset) | |||
3617 | 3613 | ||
3618 | del_timer_sync(&qdev->adapter_timer); | 3614 | del_timer_sync(&qdev->adapter_timer); |
3619 | 3615 | ||
3620 | netif_poll_disable(ndev); | 3616 | napi_disable(&qdev->napi); |
3621 | 3617 | ||
3622 | if (do_reset) { | 3618 | if (do_reset) { |
3623 | int soft_reset; | 3619 | int soft_reset; |
@@ -3705,7 +3701,7 @@ static int ql_adapter_up(struct ql3_adapter *qdev) | |||
3705 | 3701 | ||
3706 | mod_timer(&qdev->adapter_timer, jiffies + HZ * 1); | 3702 | mod_timer(&qdev->adapter_timer, jiffies + HZ * 1); |
3707 | 3703 | ||
3708 | netif_poll_enable(ndev); | 3704 | napi_enable(&qdev->napi); |
3709 | ql_enable_interrupts(qdev); | 3705 | ql_enable_interrupts(qdev); |
3710 | return 0; | 3706 | return 0; |
3711 | 3707 | ||
@@ -4061,8 +4057,7 @@ static int __devinit ql3xxx_probe(struct pci_dev *pdev, | |||
4061 | ndev->tx_timeout = ql3xxx_tx_timeout; | 4057 | ndev->tx_timeout = ql3xxx_tx_timeout; |
4062 | ndev->watchdog_timeo = 5 * HZ; | 4058 | ndev->watchdog_timeo = 5 * HZ; |
4063 | 4059 | ||
4064 | ndev->poll = &ql_poll; | 4060 | netif_napi_add(ndev, &qdev->napi, ql_poll, 64); |
4065 | ndev->weight = 64; | ||
4066 | 4061 | ||
4067 | ndev->irq = pdev->irq; | 4062 | ndev->irq = pdev->irq; |
4068 | 4063 | ||
diff --git a/drivers/net/qla3xxx.h b/drivers/net/qla3xxx.h index 4a832c46c274..aa2216f0d7b8 100755 --- a/drivers/net/qla3xxx.h +++ b/drivers/net/qla3xxx.h | |||
@@ -1175,6 +1175,8 @@ struct ql3_adapter { | |||
1175 | struct pci_dev *pdev; | 1175 | struct pci_dev *pdev; |
1176 | struct net_device *ndev; /* Parent NET device */ | 1176 | struct net_device *ndev; /* Parent NET device */ |
1177 | 1177 | ||
1178 | struct napi_struct napi; | ||
1179 | |||
1178 | /* Hardware information */ | 1180 | /* Hardware information */ |
1179 | u8 chip_rev_id; | 1181 | u8 chip_rev_id; |
1180 | u8 pci_slot; | 1182 | u8 pci_slot; |
diff --git a/drivers/net/r8169.c b/drivers/net/r8169.c index c76dd29c8e9a..3f2306e3f517 100644 --- a/drivers/net/r8169.c +++ b/drivers/net/r8169.c | |||
@@ -384,6 +384,7 @@ struct rtl8169_private { | |||
384 | void __iomem *mmio_addr; /* memory map physical address */ | 384 | void __iomem *mmio_addr; /* memory map physical address */ |
385 | struct pci_dev *pci_dev; /* Index of PCI device */ | 385 | struct pci_dev *pci_dev; /* Index of PCI device */ |
386 | struct net_device *dev; | 386 | struct net_device *dev; |
387 | struct napi_struct napi; | ||
387 | struct net_device_stats stats; /* statistics of net device */ | 388 | struct net_device_stats stats; /* statistics of net device */ |
388 | spinlock_t lock; /* spin lock flag */ | 389 | spinlock_t lock; /* spin lock flag */ |
389 | u32 msg_enable; | 390 | u32 msg_enable; |
@@ -443,13 +444,13 @@ static void rtl_set_rx_mode(struct net_device *dev); | |||
443 | static void rtl8169_tx_timeout(struct net_device *dev); | 444 | static void rtl8169_tx_timeout(struct net_device *dev); |
444 | static struct net_device_stats *rtl8169_get_stats(struct net_device *dev); | 445 | static struct net_device_stats *rtl8169_get_stats(struct net_device *dev); |
445 | static int rtl8169_rx_interrupt(struct net_device *, struct rtl8169_private *, | 446 | static int rtl8169_rx_interrupt(struct net_device *, struct rtl8169_private *, |
446 | void __iomem *); | 447 | void __iomem *, u32 budget); |
447 | static int rtl8169_change_mtu(struct net_device *dev, int new_mtu); | 448 | static int rtl8169_change_mtu(struct net_device *dev, int new_mtu); |
448 | static void rtl8169_down(struct net_device *dev); | 449 | static void rtl8169_down(struct net_device *dev); |
449 | static void rtl8169_rx_clear(struct rtl8169_private *tp); | 450 | static void rtl8169_rx_clear(struct rtl8169_private *tp); |
450 | 451 | ||
451 | #ifdef CONFIG_R8169_NAPI | 452 | #ifdef CONFIG_R8169_NAPI |
452 | static int rtl8169_poll(struct net_device *dev, int *budget); | 453 | static int rtl8169_poll(struct napi_struct *napi, int budget); |
453 | #endif | 454 | #endif |
454 | 455 | ||
455 | static const unsigned int rtl8169_rx_config = | 456 | static const unsigned int rtl8169_rx_config = |
@@ -1656,8 +1657,7 @@ rtl8169_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
1656 | dev->set_mac_address = rtl_set_mac_address; | 1657 | dev->set_mac_address = rtl_set_mac_address; |
1657 | 1658 | ||
1658 | #ifdef CONFIG_R8169_NAPI | 1659 | #ifdef CONFIG_R8169_NAPI |
1659 | dev->poll = rtl8169_poll; | 1660 | netif_napi_add(dev, &tp->napi, rtl8169_poll, R8169_NAPI_WEIGHT); |
1660 | dev->weight = R8169_NAPI_WEIGHT; | ||
1661 | #endif | 1661 | #endif |
1662 | 1662 | ||
1663 | #ifdef CONFIG_R8169_VLAN | 1663 | #ifdef CONFIG_R8169_VLAN |
@@ -1777,6 +1777,10 @@ static int rtl8169_open(struct net_device *dev) | |||
1777 | if (retval < 0) | 1777 | if (retval < 0) |
1778 | goto err_release_ring_2; | 1778 | goto err_release_ring_2; |
1779 | 1779 | ||
1780 | #ifdef CONFIG_R8169_NAPI | ||
1781 | napi_enable(&tp->napi); | ||
1782 | #endif | ||
1783 | |||
1780 | rtl_hw_start(dev); | 1784 | rtl_hw_start(dev); |
1781 | 1785 | ||
1782 | rtl8169_request_timer(dev); | 1786 | rtl8169_request_timer(dev); |
@@ -2082,7 +2086,9 @@ static int rtl8169_change_mtu(struct net_device *dev, int new_mtu) | |||
2082 | if (ret < 0) | 2086 | if (ret < 0) |
2083 | goto out; | 2087 | goto out; |
2084 | 2088 | ||
2085 | netif_poll_enable(dev); | 2089 | #ifdef CONFIG_R8169_NAPI |
2090 | napi_enable(&tp->napi); | ||
2091 | #endif | ||
2086 | 2092 | ||
2087 | rtl_hw_start(dev); | 2093 | rtl_hw_start(dev); |
2088 | 2094 | ||
@@ -2274,11 +2280,15 @@ static void rtl8169_wait_for_quiescence(struct net_device *dev) | |||
2274 | synchronize_irq(dev->irq); | 2280 | synchronize_irq(dev->irq); |
2275 | 2281 | ||
2276 | /* Wait for any pending NAPI task to complete */ | 2282 | /* Wait for any pending NAPI task to complete */ |
2277 | netif_poll_disable(dev); | 2283 | #ifdef CONFIG_R8169_NAPI |
2284 | napi_disable(&tp->napi); | ||
2285 | #endif | ||
2278 | 2286 | ||
2279 | rtl8169_irq_mask_and_ack(ioaddr); | 2287 | rtl8169_irq_mask_and_ack(ioaddr); |
2280 | 2288 | ||
2281 | netif_poll_enable(dev); | 2289 | #ifdef CONFIG_R8169_NAPI |
2290 | napi_enable(&tp->napi); | ||
2291 | #endif | ||
2282 | } | 2292 | } |
2283 | 2293 | ||
2284 | static void rtl8169_reinit_task(struct work_struct *work) | 2294 | static void rtl8169_reinit_task(struct work_struct *work) |
@@ -2322,7 +2332,7 @@ static void rtl8169_reset_task(struct work_struct *work) | |||
2322 | 2332 | ||
2323 | rtl8169_wait_for_quiescence(dev); | 2333 | rtl8169_wait_for_quiescence(dev); |
2324 | 2334 | ||
2325 | rtl8169_rx_interrupt(dev, tp, tp->mmio_addr); | 2335 | rtl8169_rx_interrupt(dev, tp, tp->mmio_addr, ~(u32)0); |
2326 | rtl8169_tx_clear(tp); | 2336 | rtl8169_tx_clear(tp); |
2327 | 2337 | ||
2328 | if (tp->dirty_rx == tp->cur_rx) { | 2338 | if (tp->dirty_rx == tp->cur_rx) { |
@@ -2636,14 +2646,14 @@ out: | |||
2636 | 2646 | ||
2637 | static int rtl8169_rx_interrupt(struct net_device *dev, | 2647 | static int rtl8169_rx_interrupt(struct net_device *dev, |
2638 | struct rtl8169_private *tp, | 2648 | struct rtl8169_private *tp, |
2639 | void __iomem *ioaddr) | 2649 | void __iomem *ioaddr, u32 budget) |
2640 | { | 2650 | { |
2641 | unsigned int cur_rx, rx_left; | 2651 | unsigned int cur_rx, rx_left; |
2642 | unsigned int delta, count; | 2652 | unsigned int delta, count; |
2643 | 2653 | ||
2644 | cur_rx = tp->cur_rx; | 2654 | cur_rx = tp->cur_rx; |
2645 | rx_left = NUM_RX_DESC + tp->dirty_rx - cur_rx; | 2655 | rx_left = NUM_RX_DESC + tp->dirty_rx - cur_rx; |
2646 | rx_left = rtl8169_rx_quota(rx_left, (u32) dev->quota); | 2656 | rx_left = rtl8169_rx_quota(rx_left, budget); |
2647 | 2657 | ||
2648 | for (; rx_left > 0; rx_left--, cur_rx++) { | 2658 | for (; rx_left > 0; rx_left--, cur_rx++) { |
2649 | unsigned int entry = cur_rx % NUM_RX_DESC; | 2659 | unsigned int entry = cur_rx % NUM_RX_DESC; |
@@ -2792,8 +2802,8 @@ static irqreturn_t rtl8169_interrupt(int irq, void *dev_instance) | |||
2792 | RTL_W16(IntrMask, tp->intr_event & ~tp->napi_event); | 2802 | RTL_W16(IntrMask, tp->intr_event & ~tp->napi_event); |
2793 | tp->intr_mask = ~tp->napi_event; | 2803 | tp->intr_mask = ~tp->napi_event; |
2794 | 2804 | ||
2795 | if (likely(netif_rx_schedule_prep(dev))) | 2805 | if (likely(netif_rx_schedule_prep(dev, &tp->napi))) |
2796 | __netif_rx_schedule(dev); | 2806 | __netif_rx_schedule(dev, &tp->napi); |
2797 | else if (netif_msg_intr(tp)) { | 2807 | else if (netif_msg_intr(tp)) { |
2798 | printk(KERN_INFO "%s: interrupt %04x in poll\n", | 2808 | printk(KERN_INFO "%s: interrupt %04x in poll\n", |
2799 | dev->name, status); | 2809 | dev->name, status); |
@@ -2803,7 +2813,7 @@ static irqreturn_t rtl8169_interrupt(int irq, void *dev_instance) | |||
2803 | #else | 2813 | #else |
2804 | /* Rx interrupt */ | 2814 | /* Rx interrupt */ |
2805 | if (status & (RxOK | RxOverflow | RxFIFOOver)) | 2815 | if (status & (RxOK | RxOverflow | RxFIFOOver)) |
2806 | rtl8169_rx_interrupt(dev, tp, ioaddr); | 2816 | rtl8169_rx_interrupt(dev, tp, ioaddr, ~(u32)0); |
2807 | 2817 | ||
2808 | /* Tx interrupt */ | 2818 | /* Tx interrupt */ |
2809 | if (status & (TxOK | TxErr)) | 2819 | if (status & (TxOK | TxErr)) |
@@ -2826,20 +2836,18 @@ out: | |||
2826 | } | 2836 | } |
2827 | 2837 | ||
2828 | #ifdef CONFIG_R8169_NAPI | 2838 | #ifdef CONFIG_R8169_NAPI |
2829 | static int rtl8169_poll(struct net_device *dev, int *budget) | 2839 | static int rtl8169_poll(struct napi_struct *napi, int budget) |
2830 | { | 2840 | { |
2831 | unsigned int work_done, work_to_do = min(*budget, dev->quota); | 2841 | struct rtl8169_private *tp = container_of(napi, struct rtl8169_private, napi); |
2832 | struct rtl8169_private *tp = netdev_priv(dev); | 2842 | struct net_device *dev = tp->dev; |
2833 | void __iomem *ioaddr = tp->mmio_addr; | 2843 | void __iomem *ioaddr = tp->mmio_addr; |
2844 | int work_done; | ||
2834 | 2845 | ||
2835 | work_done = rtl8169_rx_interrupt(dev, tp, ioaddr); | 2846 | work_done = rtl8169_rx_interrupt(dev, tp, ioaddr, (u32) budget); |
2836 | rtl8169_tx_interrupt(dev, tp, ioaddr); | 2847 | rtl8169_tx_interrupt(dev, tp, ioaddr); |
2837 | 2848 | ||
2838 | *budget -= work_done; | 2849 | if (work_done < budget) { |
2839 | dev->quota -= work_done; | 2850 | netif_rx_complete(dev, napi); |
2840 | |||
2841 | if (work_done < work_to_do) { | ||
2842 | netif_rx_complete(dev); | ||
2843 | tp->intr_mask = 0xffff; | 2851 | tp->intr_mask = 0xffff; |
2844 | /* | 2852 | /* |
2845 | * 20040426: the barrier is not strictly required but the | 2853 | * 20040426: the barrier is not strictly required but the |
@@ -2851,7 +2859,7 @@ static int rtl8169_poll(struct net_device *dev, int *budget) | |||
2851 | RTL_W16(IntrMask, tp->intr_event); | 2859 | RTL_W16(IntrMask, tp->intr_event); |
2852 | } | 2860 | } |
2853 | 2861 | ||
2854 | return (work_done >= work_to_do); | 2862 | return work_done; |
2855 | } | 2863 | } |
2856 | #endif | 2864 | #endif |
2857 | 2865 | ||
@@ -2880,7 +2888,7 @@ core_down: | |||
2880 | synchronize_irq(dev->irq); | 2888 | synchronize_irq(dev->irq); |
2881 | 2889 | ||
2882 | if (!poll_locked) { | 2890 | if (!poll_locked) { |
2883 | netif_poll_disable(dev); | 2891 | napi_disable(&tp->napi); |
2884 | poll_locked++; | 2892 | poll_locked++; |
2885 | } | 2893 | } |
2886 | 2894 | ||
@@ -2918,8 +2926,6 @@ static int rtl8169_close(struct net_device *dev) | |||
2918 | 2926 | ||
2919 | free_irq(dev->irq, dev); | 2927 | free_irq(dev->irq, dev); |
2920 | 2928 | ||
2921 | netif_poll_enable(dev); | ||
2922 | |||
2923 | pci_free_consistent(pdev, R8169_RX_RING_BYTES, tp->RxDescArray, | 2929 | pci_free_consistent(pdev, R8169_RX_RING_BYTES, tp->RxDescArray, |
2924 | tp->RxPhyAddr); | 2930 | tp->RxPhyAddr); |
2925 | pci_free_consistent(pdev, R8169_TX_RING_BYTES, tp->TxDescArray, | 2931 | pci_free_consistent(pdev, R8169_TX_RING_BYTES, tp->TxDescArray, |
diff --git a/drivers/net/s2io.c b/drivers/net/s2io.c index 24feb00600ee..dd012322cdbe 100644 --- a/drivers/net/s2io.c +++ b/drivers/net/s2io.c | |||
@@ -2568,7 +2568,7 @@ static void free_rx_buffers(struct s2io_nic *sp) | |||
2568 | 2568 | ||
2569 | /** | 2569 | /** |
2570 | * s2io_poll - Rx interrupt handler for NAPI support | 2570 | * s2io_poll - Rx interrupt handler for NAPI support |
2571 | * @dev : pointer to the device structure. | 2571 | * @napi : pointer to the napi structure. |
2572 | * @budget : The number of packets that were budgeted to be processed | 2572 | * @budget : The number of packets that were budgeted to be processed |
2573 | * during one pass through the 'Poll" function. | 2573 | * during one pass through the 'Poll" function. |
2574 | * Description: | 2574 | * Description: |
@@ -2579,9 +2579,10 @@ static void free_rx_buffers(struct s2io_nic *sp) | |||
2579 | * 0 on success and 1 if there are No Rx packets to be processed. | 2579 | * 0 on success and 1 if there are No Rx packets to be processed. |
2580 | */ | 2580 | */ |
2581 | 2581 | ||
2582 | static int s2io_poll(struct net_device *dev, int *budget) | 2582 | static int s2io_poll(struct napi_struct *napi, int budget) |
2583 | { | 2583 | { |
2584 | struct s2io_nic *nic = dev->priv; | 2584 | struct s2io_nic *nic = container_of(napi, struct s2io_nic, napi); |
2585 | struct net_device *dev = nic->dev; | ||
2585 | int pkt_cnt = 0, org_pkts_to_process; | 2586 | int pkt_cnt = 0, org_pkts_to_process; |
2586 | struct mac_info *mac_control; | 2587 | struct mac_info *mac_control; |
2587 | struct config_param *config; | 2588 | struct config_param *config; |
@@ -2592,9 +2593,7 @@ static int s2io_poll(struct net_device *dev, int *budget) | |||
2592 | mac_control = &nic->mac_control; | 2593 | mac_control = &nic->mac_control; |
2593 | config = &nic->config; | 2594 | config = &nic->config; |
2594 | 2595 | ||
2595 | nic->pkts_to_process = *budget; | 2596 | nic->pkts_to_process = budget; |
2596 | if (nic->pkts_to_process > dev->quota) | ||
2597 | nic->pkts_to_process = dev->quota; | ||
2598 | org_pkts_to_process = nic->pkts_to_process; | 2597 | org_pkts_to_process = nic->pkts_to_process; |
2599 | 2598 | ||
2600 | writeq(S2IO_MINUS_ONE, &bar0->rx_traffic_int); | 2599 | writeq(S2IO_MINUS_ONE, &bar0->rx_traffic_int); |
@@ -2608,12 +2607,8 @@ static int s2io_poll(struct net_device *dev, int *budget) | |||
2608 | goto no_rx; | 2607 | goto no_rx; |
2609 | } | 2608 | } |
2610 | } | 2609 | } |
2611 | if (!pkt_cnt) | ||
2612 | pkt_cnt = 1; | ||
2613 | 2610 | ||
2614 | dev->quota -= pkt_cnt; | 2611 | netif_rx_complete(dev, napi); |
2615 | *budget -= pkt_cnt; | ||
2616 | netif_rx_complete(dev); | ||
2617 | 2612 | ||
2618 | for (i = 0; i < config->rx_ring_num; i++) { | 2613 | for (i = 0; i < config->rx_ring_num; i++) { |
2619 | if (fill_rx_buffers(nic, i) == -ENOMEM) { | 2614 | if (fill_rx_buffers(nic, i) == -ENOMEM) { |
@@ -2626,12 +2621,9 @@ static int s2io_poll(struct net_device *dev, int *budget) | |||
2626 | writeq(0x0, &bar0->rx_traffic_mask); | 2621 | writeq(0x0, &bar0->rx_traffic_mask); |
2627 | readl(&bar0->rx_traffic_mask); | 2622 | readl(&bar0->rx_traffic_mask); |
2628 | atomic_dec(&nic->isr_cnt); | 2623 | atomic_dec(&nic->isr_cnt); |
2629 | return 0; | 2624 | return pkt_cnt; |
2630 | 2625 | ||
2631 | no_rx: | 2626 | no_rx: |
2632 | dev->quota -= pkt_cnt; | ||
2633 | *budget -= pkt_cnt; | ||
2634 | |||
2635 | for (i = 0; i < config->rx_ring_num; i++) { | 2627 | for (i = 0; i < config->rx_ring_num; i++) { |
2636 | if (fill_rx_buffers(nic, i) == -ENOMEM) { | 2628 | if (fill_rx_buffers(nic, i) == -ENOMEM) { |
2637 | DBG_PRINT(INFO_DBG, "%s:Out of memory", dev->name); | 2629 | DBG_PRINT(INFO_DBG, "%s:Out of memory", dev->name); |
@@ -2640,7 +2632,7 @@ no_rx: | |||
2640 | } | 2632 | } |
2641 | } | 2633 | } |
2642 | atomic_dec(&nic->isr_cnt); | 2634 | atomic_dec(&nic->isr_cnt); |
2643 | return 1; | 2635 | return pkt_cnt; |
2644 | } | 2636 | } |
2645 | 2637 | ||
2646 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2638 | #ifdef CONFIG_NET_POLL_CONTROLLER |
@@ -3809,6 +3801,8 @@ static int s2io_open(struct net_device *dev) | |||
3809 | netif_carrier_off(dev); | 3801 | netif_carrier_off(dev); |
3810 | sp->last_link_state = 0; | 3802 | sp->last_link_state = 0; |
3811 | 3803 | ||
3804 | napi_enable(&sp->napi); | ||
3805 | |||
3812 | /* Initialize H/W and enable interrupts */ | 3806 | /* Initialize H/W and enable interrupts */ |
3813 | err = s2io_card_up(sp); | 3807 | err = s2io_card_up(sp); |
3814 | if (err) { | 3808 | if (err) { |
@@ -3828,6 +3822,7 @@ static int s2io_open(struct net_device *dev) | |||
3828 | return 0; | 3822 | return 0; |
3829 | 3823 | ||
3830 | hw_init_failed: | 3824 | hw_init_failed: |
3825 | napi_disable(&sp->napi); | ||
3831 | if (sp->intr_type == MSI_X) { | 3826 | if (sp->intr_type == MSI_X) { |
3832 | if (sp->entries) { | 3827 | if (sp->entries) { |
3833 | kfree(sp->entries); | 3828 | kfree(sp->entries); |
@@ -3861,6 +3856,7 @@ static int s2io_close(struct net_device *dev) | |||
3861 | struct s2io_nic *sp = dev->priv; | 3856 | struct s2io_nic *sp = dev->priv; |
3862 | 3857 | ||
3863 | netif_stop_queue(dev); | 3858 | netif_stop_queue(dev); |
3859 | napi_disable(&sp->napi); | ||
3864 | /* Reset card, kill tasklet and free Tx and Rx buffers. */ | 3860 | /* Reset card, kill tasklet and free Tx and Rx buffers. */ |
3865 | s2io_card_down(sp); | 3861 | s2io_card_down(sp); |
3866 | 3862 | ||
@@ -4232,8 +4228,8 @@ static irqreturn_t s2io_isr(int irq, void *dev_id) | |||
4232 | 4228 | ||
4233 | if (napi) { | 4229 | if (napi) { |
4234 | if (reason & GEN_INTR_RXTRAFFIC) { | 4230 | if (reason & GEN_INTR_RXTRAFFIC) { |
4235 | if ( likely ( netif_rx_schedule_prep(dev)) ) { | 4231 | if (likely (netif_rx_schedule_prep(dev, &sp->napi))) { |
4236 | __netif_rx_schedule(dev); | 4232 | __netif_rx_schedule(dev, &sp->napi); |
4237 | writeq(S2IO_MINUS_ONE, &bar0->rx_traffic_mask); | 4233 | writeq(S2IO_MINUS_ONE, &bar0->rx_traffic_mask); |
4238 | } | 4234 | } |
4239 | else | 4235 | else |
@@ -7215,8 +7211,7 @@ s2io_init_nic(struct pci_dev *pdev, const struct pci_device_id *pre) | |||
7215 | * will use eth_mac_addr() for dev->set_mac_address | 7211 | * will use eth_mac_addr() for dev->set_mac_address |
7216 | * mac address will be set every time dev->open() is called | 7212 | * mac address will be set every time dev->open() is called |
7217 | */ | 7213 | */ |
7218 | dev->poll = s2io_poll; | 7214 | netif_napi_add(dev, &sp->napi, s2io_poll, 32); |
7219 | dev->weight = 32; | ||
7220 | 7215 | ||
7221 | #ifdef CONFIG_NET_POLL_CONTROLLER | 7216 | #ifdef CONFIG_NET_POLL_CONTROLLER |
7222 | dev->poll_controller = s2io_netpoll; | 7217 | dev->poll_controller = s2io_netpoll; |
diff --git a/drivers/net/s2io.h b/drivers/net/s2io.h index 92983ee7df8c..420fefb99188 100644 --- a/drivers/net/s2io.h +++ b/drivers/net/s2io.h | |||
@@ -786,6 +786,7 @@ struct s2io_nic { | |||
786 | */ | 786 | */ |
787 | int pkts_to_process; | 787 | int pkts_to_process; |
788 | struct net_device *dev; | 788 | struct net_device *dev; |
789 | struct napi_struct napi; | ||
789 | struct mac_info mac_control; | 790 | struct mac_info mac_control; |
790 | struct config_param config; | 791 | struct config_param config; |
791 | struct pci_dev *pdev; | 792 | struct pci_dev *pdev; |
@@ -1019,7 +1020,7 @@ static void s2io_set_multicast(struct net_device *dev); | |||
1019 | static int rx_osm_handler(struct ring_info *ring_data, struct RxD_t * rxdp); | 1020 | static int rx_osm_handler(struct ring_info *ring_data, struct RxD_t * rxdp); |
1020 | static void s2io_link(struct s2io_nic * sp, int link); | 1021 | static void s2io_link(struct s2io_nic * sp, int link); |
1021 | static void s2io_reset(struct s2io_nic * sp); | 1022 | static void s2io_reset(struct s2io_nic * sp); |
1022 | static int s2io_poll(struct net_device *dev, int *budget); | 1023 | static int s2io_poll(struct napi_struct *napi, int budget); |
1023 | static void s2io_init_pci(struct s2io_nic * sp); | 1024 | static void s2io_init_pci(struct s2io_nic * sp); |
1024 | static int s2io_set_mac_addr(struct net_device *dev, u8 * addr); | 1025 | static int s2io_set_mac_addr(struct net_device *dev, u8 * addr); |
1025 | static void s2io_alarm_handle(unsigned long data); | 1026 | static void s2io_alarm_handle(unsigned long data); |
diff --git a/drivers/net/sb1250-mac.c b/drivers/net/sb1250-mac.c index e7fdcf15b5a7..53845ebb649f 100644 --- a/drivers/net/sb1250-mac.c +++ b/drivers/net/sb1250-mac.c | |||
@@ -238,6 +238,7 @@ struct sbmac_softc { | |||
238 | */ | 238 | */ |
239 | 239 | ||
240 | struct net_device *sbm_dev; /* pointer to linux device */ | 240 | struct net_device *sbm_dev; /* pointer to linux device */ |
241 | struct napi_struct napi; | ||
241 | spinlock_t sbm_lock; /* spin lock */ | 242 | spinlock_t sbm_lock; /* spin lock */ |
242 | struct timer_list sbm_timer; /* for monitoring MII */ | 243 | struct timer_list sbm_timer; /* for monitoring MII */ |
243 | struct net_device_stats sbm_stats; | 244 | struct net_device_stats sbm_stats; |
@@ -320,7 +321,7 @@ static struct net_device_stats *sbmac_get_stats(struct net_device *dev); | |||
320 | static void sbmac_set_rx_mode(struct net_device *dev); | 321 | static void sbmac_set_rx_mode(struct net_device *dev); |
321 | static int sbmac_mii_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); | 322 | static int sbmac_mii_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); |
322 | static int sbmac_close(struct net_device *dev); | 323 | static int sbmac_close(struct net_device *dev); |
323 | static int sbmac_poll(struct net_device *poll_dev, int *budget); | 324 | static int sbmac_poll(struct napi_struct *napi, int budget); |
324 | 325 | ||
325 | static int sbmac_mii_poll(struct sbmac_softc *s,int noisy); | 326 | static int sbmac_mii_poll(struct sbmac_softc *s,int noisy); |
326 | static int sbmac_mii_probe(struct net_device *dev); | 327 | static int sbmac_mii_probe(struct net_device *dev); |
@@ -2154,20 +2155,13 @@ static irqreturn_t sbmac_intr(int irq,void *dev_instance) | |||
2154 | * Transmits on channel 0 | 2155 | * Transmits on channel 0 |
2155 | */ | 2156 | */ |
2156 | 2157 | ||
2157 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_TX_CH0)) { | 2158 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_TX_CH0)) |
2158 | sbdma_tx_process(sc,&(sc->sbm_txdma), 0); | 2159 | sbdma_tx_process(sc,&(sc->sbm_txdma), 0); |
2159 | #ifdef CONFIG_NETPOLL_TRAP | ||
2160 | if (netpoll_trap()) { | ||
2161 | if (test_and_clear_bit(__LINK_STATE_XOFF, &dev->state)) | ||
2162 | __netif_schedule(dev); | ||
2163 | } | ||
2164 | #endif | ||
2165 | } | ||
2166 | 2160 | ||
2167 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_RX_CH0)) { | 2161 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_RX_CH0)) { |
2168 | if (netif_rx_schedule_prep(dev)) { | 2162 | if (netif_rx_schedule_prep(dev, &sc->napi)) { |
2169 | __raw_writeq(0, sc->sbm_imr); | 2163 | __raw_writeq(0, sc->sbm_imr); |
2170 | __netif_rx_schedule(dev); | 2164 | __netif_rx_schedule(dev, &sc->napi); |
2171 | /* Depend on the exit from poll to reenable intr */ | 2165 | /* Depend on the exit from poll to reenable intr */ |
2172 | } | 2166 | } |
2173 | else { | 2167 | else { |
@@ -2470,8 +2464,8 @@ static int sbmac_init(struct net_device *dev, int idx) | |||
2470 | dev->do_ioctl = sbmac_mii_ioctl; | 2464 | dev->do_ioctl = sbmac_mii_ioctl; |
2471 | dev->tx_timeout = sbmac_tx_timeout; | 2465 | dev->tx_timeout = sbmac_tx_timeout; |
2472 | dev->watchdog_timeo = TX_TIMEOUT; | 2466 | dev->watchdog_timeo = TX_TIMEOUT; |
2473 | dev->poll = sbmac_poll; | 2467 | |
2474 | dev->weight = 16; | 2468 | netif_napi_add(dev, &sc->napi, sbmac_poll, 16); |
2475 | 2469 | ||
2476 | dev->change_mtu = sb1250_change_mtu; | 2470 | dev->change_mtu = sb1250_change_mtu; |
2477 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2471 | #ifdef CONFIG_NET_POLL_CONTROLLER |
@@ -2537,6 +2531,8 @@ static int sbmac_open(struct net_device *dev) | |||
2537 | return -EINVAL; | 2531 | return -EINVAL; |
2538 | } | 2532 | } |
2539 | 2533 | ||
2534 | napi_enable(&sc->napi); | ||
2535 | |||
2540 | /* | 2536 | /* |
2541 | * Configure default speed | 2537 | * Configure default speed |
2542 | */ | 2538 | */ |
@@ -2850,6 +2846,8 @@ static int sbmac_close(struct net_device *dev) | |||
2850 | unsigned long flags; | 2846 | unsigned long flags; |
2851 | int irq; | 2847 | int irq; |
2852 | 2848 | ||
2849 | napi_disable(&sc->napi); | ||
2850 | |||
2853 | sbmac_set_channel_state(sc,sbmac_state_off); | 2851 | sbmac_set_channel_state(sc,sbmac_state_off); |
2854 | 2852 | ||
2855 | del_timer_sync(&sc->sbm_timer); | 2853 | del_timer_sync(&sc->sbm_timer); |
@@ -2874,26 +2872,17 @@ static int sbmac_close(struct net_device *dev) | |||
2874 | return 0; | 2872 | return 0; |
2875 | } | 2873 | } |
2876 | 2874 | ||
2877 | static int sbmac_poll(struct net_device *dev, int *budget) | 2875 | static int sbmac_poll(struct napi_struct *napi, int budget) |
2878 | { | 2876 | { |
2879 | int work_to_do; | 2877 | struct sbmac_softc *sc = container_of(napi, struct sbmac_softc, napi); |
2878 | struct net_device *dev = sc->sbm_dev; | ||
2880 | int work_done; | 2879 | int work_done; |
2881 | struct sbmac_softc *sc = netdev_priv(dev); | ||
2882 | |||
2883 | work_to_do = min(*budget, dev->quota); | ||
2884 | work_done = sbdma_rx_process(sc, &(sc->sbm_rxdma), work_to_do, 1); | ||
2885 | |||
2886 | if (work_done > work_to_do) | ||
2887 | printk(KERN_ERR "%s exceeded work_to_do budget=%d quota=%d work-done=%d\n", | ||
2888 | sc->sbm_dev->name, *budget, dev->quota, work_done); | ||
2889 | 2880 | ||
2881 | work_done = sbdma_rx_process(sc, &(sc->sbm_rxdma), budget, 1); | ||
2890 | sbdma_tx_process(sc, &(sc->sbm_txdma), 1); | 2882 | sbdma_tx_process(sc, &(sc->sbm_txdma), 1); |
2891 | 2883 | ||
2892 | *budget -= work_done; | 2884 | if (work_done < budget) { |
2893 | dev->quota -= work_done; | 2885 | netif_rx_complete(dev, napi); |
2894 | |||
2895 | if (work_done < work_to_do) { | ||
2896 | netif_rx_complete(dev); | ||
2897 | 2886 | ||
2898 | #ifdef CONFIG_SBMAC_COALESCE | 2887 | #ifdef CONFIG_SBMAC_COALESCE |
2899 | __raw_writeq(((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_TX_CH0) | | 2888 | __raw_writeq(((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_TX_CH0) | |
@@ -2905,7 +2894,7 @@ static int sbmac_poll(struct net_device *dev, int *budget) | |||
2905 | #endif | 2894 | #endif |
2906 | } | 2895 | } |
2907 | 2896 | ||
2908 | return (work_done >= work_to_do); | 2897 | return work_done; |
2909 | } | 2898 | } |
2910 | 2899 | ||
2911 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) | 2900 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) |
diff --git a/drivers/net/sis190.c b/drivers/net/sis190.c index d470b19c0810..038ccfbafdd1 100644 --- a/drivers/net/sis190.c +++ b/drivers/net/sis190.c | |||
@@ -47,24 +47,13 @@ | |||
47 | #define PHY_ID_ANY 0x1f | 47 | #define PHY_ID_ANY 0x1f |
48 | #define MII_REG_ANY 0x1f | 48 | #define MII_REG_ANY 0x1f |
49 | 49 | ||
50 | #ifdef CONFIG_SIS190_NAPI | 50 | #define DRV_VERSION "1.2" |
51 | #define NAPI_SUFFIX "-NAPI" | ||
52 | #else | ||
53 | #define NAPI_SUFFIX "" | ||
54 | #endif | ||
55 | |||
56 | #define DRV_VERSION "1.2" NAPI_SUFFIX | ||
57 | #define DRV_NAME "sis190" | 51 | #define DRV_NAME "sis190" |
58 | #define SIS190_DRIVER_NAME DRV_NAME " Gigabit Ethernet driver " DRV_VERSION | 52 | #define SIS190_DRIVER_NAME DRV_NAME " Gigabit Ethernet driver " DRV_VERSION |
59 | #define PFX DRV_NAME ": " | 53 | #define PFX DRV_NAME ": " |
60 | 54 | ||
61 | #ifdef CONFIG_SIS190_NAPI | ||
62 | #define sis190_rx_skb netif_receive_skb | ||
63 | #define sis190_rx_quota(count, quota) min(count, quota) | ||
64 | #else | ||
65 | #define sis190_rx_skb netif_rx | 55 | #define sis190_rx_skb netif_rx |
66 | #define sis190_rx_quota(count, quota) count | 56 | #define sis190_rx_quota(count, quota) count |
67 | #endif | ||
68 | 57 | ||
69 | #define MAC_ADDR_LEN 6 | 58 | #define MAC_ADDR_LEN 6 |
70 | 59 | ||
@@ -1115,10 +1104,8 @@ static void sis190_down(struct net_device *dev) | |||
1115 | 1104 | ||
1116 | synchronize_irq(dev->irq); | 1105 | synchronize_irq(dev->irq); |
1117 | 1106 | ||
1118 | if (!poll_locked) { | 1107 | if (!poll_locked) |
1119 | netif_poll_disable(dev); | ||
1120 | poll_locked++; | 1108 | poll_locked++; |
1121 | } | ||
1122 | 1109 | ||
1123 | synchronize_sched(); | 1110 | synchronize_sched(); |
1124 | 1111 | ||
@@ -1137,8 +1124,6 @@ static int sis190_close(struct net_device *dev) | |||
1137 | 1124 | ||
1138 | free_irq(dev->irq, dev); | 1125 | free_irq(dev->irq, dev); |
1139 | 1126 | ||
1140 | netif_poll_enable(dev); | ||
1141 | |||
1142 | pci_free_consistent(pdev, TX_RING_BYTES, tp->TxDescRing, tp->tx_dma); | 1127 | pci_free_consistent(pdev, TX_RING_BYTES, tp->TxDescRing, tp->tx_dma); |
1143 | pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); | 1128 | pci_free_consistent(pdev, RX_RING_BYTES, tp->RxDescRing, tp->rx_dma); |
1144 | 1129 | ||
diff --git a/drivers/net/skge.c b/drivers/net/skge.c index e3d8520209b8..0bf46ed4e684 100644 --- a/drivers/net/skge.c +++ b/drivers/net/skge.c | |||
@@ -2528,7 +2528,7 @@ static int skge_up(struct net_device *dev) | |||
2528 | skge_write32(hw, B0_IMSK, hw->intr_mask); | 2528 | skge_write32(hw, B0_IMSK, hw->intr_mask); |
2529 | spin_unlock_irq(&hw->hw_lock); | 2529 | spin_unlock_irq(&hw->hw_lock); |
2530 | 2530 | ||
2531 | netif_poll_enable(dev); | 2531 | napi_enable(&skge->napi); |
2532 | return 0; | 2532 | return 0; |
2533 | 2533 | ||
2534 | free_rx_ring: | 2534 | free_rx_ring: |
@@ -2558,7 +2558,7 @@ static int skge_down(struct net_device *dev) | |||
2558 | if (hw->chip_id == CHIP_ID_GENESIS && hw->phy_type == SK_PHY_XMAC) | 2558 | if (hw->chip_id == CHIP_ID_GENESIS && hw->phy_type == SK_PHY_XMAC) |
2559 | del_timer_sync(&skge->link_timer); | 2559 | del_timer_sync(&skge->link_timer); |
2560 | 2560 | ||
2561 | netif_poll_disable(dev); | 2561 | napi_disable(&skge->napi); |
2562 | netif_carrier_off(dev); | 2562 | netif_carrier_off(dev); |
2563 | 2563 | ||
2564 | spin_lock_irq(&hw->hw_lock); | 2564 | spin_lock_irq(&hw->hw_lock); |
@@ -3044,14 +3044,13 @@ static void skge_tx_done(struct net_device *dev) | |||
3044 | } | 3044 | } |
3045 | } | 3045 | } |
3046 | 3046 | ||
3047 | static int skge_poll(struct net_device *dev, int *budget) | 3047 | static int skge_poll(struct napi_struct *napi, int to_do) |
3048 | { | 3048 | { |
3049 | struct skge_port *skge = netdev_priv(dev); | 3049 | struct skge_port *skge = container_of(napi, struct skge_port, napi); |
3050 | struct net_device *dev = skge->netdev; | ||
3050 | struct skge_hw *hw = skge->hw; | 3051 | struct skge_hw *hw = skge->hw; |
3051 | struct skge_ring *ring = &skge->rx_ring; | 3052 | struct skge_ring *ring = &skge->rx_ring; |
3052 | struct skge_element *e; | 3053 | struct skge_element *e; |
3053 | unsigned long flags; | ||
3054 | int to_do = min(dev->quota, *budget); | ||
3055 | int work_done = 0; | 3054 | int work_done = 0; |
3056 | 3055 | ||
3057 | skge_tx_done(dev); | 3056 | skge_tx_done(dev); |
@@ -3082,20 +3081,16 @@ static int skge_poll(struct net_device *dev, int *budget) | |||
3082 | wmb(); | 3081 | wmb(); |
3083 | skge_write8(hw, Q_ADDR(rxqaddr[skge->port], Q_CSR), CSR_START); | 3082 | skge_write8(hw, Q_ADDR(rxqaddr[skge->port], Q_CSR), CSR_START); |
3084 | 3083 | ||
3085 | *budget -= work_done; | 3084 | if (work_done < to_do) { |
3086 | dev->quota -= work_done; | 3085 | spin_lock_irq(&hw->hw_lock); |
3087 | 3086 | __netif_rx_complete(dev, napi); | |
3088 | if (work_done >= to_do) | 3087 | hw->intr_mask |= napimask[skge->port]; |
3089 | return 1; /* not done */ | 3088 | skge_write32(hw, B0_IMSK, hw->intr_mask); |
3090 | 3089 | skge_read32(hw, B0_IMSK); | |
3091 | spin_lock_irqsave(&hw->hw_lock, flags); | 3090 | spin_unlock_irq(&hw->hw_lock); |
3092 | __netif_rx_complete(dev); | 3091 | } |
3093 | hw->intr_mask |= napimask[skge->port]; | ||
3094 | skge_write32(hw, B0_IMSK, hw->intr_mask); | ||
3095 | skge_read32(hw, B0_IMSK); | ||
3096 | spin_unlock_irqrestore(&hw->hw_lock, flags); | ||
3097 | 3092 | ||
3098 | return 0; | 3093 | return work_done; |
3099 | } | 3094 | } |
3100 | 3095 | ||
3101 | /* Parity errors seem to happen when Genesis is connected to a switch | 3096 | /* Parity errors seem to happen when Genesis is connected to a switch |
@@ -3252,8 +3247,9 @@ static irqreturn_t skge_intr(int irq, void *dev_id) | |||
3252 | } | 3247 | } |
3253 | 3248 | ||
3254 | if (status & (IS_XA1_F|IS_R1_F)) { | 3249 | if (status & (IS_XA1_F|IS_R1_F)) { |
3250 | struct skge_port *skge = netdev_priv(hw->dev[0]); | ||
3255 | hw->intr_mask &= ~(IS_XA1_F|IS_R1_F); | 3251 | hw->intr_mask &= ~(IS_XA1_F|IS_R1_F); |
3256 | netif_rx_schedule(hw->dev[0]); | 3252 | netif_rx_schedule(hw->dev[0], &skge->napi); |
3257 | } | 3253 | } |
3258 | 3254 | ||
3259 | if (status & IS_PA_TO_TX1) | 3255 | if (status & IS_PA_TO_TX1) |
@@ -3271,13 +3267,14 @@ static irqreturn_t skge_intr(int irq, void *dev_id) | |||
3271 | skge_mac_intr(hw, 0); | 3267 | skge_mac_intr(hw, 0); |
3272 | 3268 | ||
3273 | if (hw->dev[1]) { | 3269 | if (hw->dev[1]) { |
3270 | struct skge_port *skge = netdev_priv(hw->dev[1]); | ||
3271 | |||
3274 | if (status & (IS_XA2_F|IS_R2_F)) { | 3272 | if (status & (IS_XA2_F|IS_R2_F)) { |
3275 | hw->intr_mask &= ~(IS_XA2_F|IS_R2_F); | 3273 | hw->intr_mask &= ~(IS_XA2_F|IS_R2_F); |
3276 | netif_rx_schedule(hw->dev[1]); | 3274 | netif_rx_schedule(hw->dev[1], &skge->napi); |
3277 | } | 3275 | } |
3278 | 3276 | ||
3279 | if (status & IS_PA_TO_RX2) { | 3277 | if (status & IS_PA_TO_RX2) { |
3280 | struct skge_port *skge = netdev_priv(hw->dev[1]); | ||
3281 | ++skge->net_stats.rx_over_errors; | 3278 | ++skge->net_stats.rx_over_errors; |
3282 | skge_write16(hw, B3_PA_CTRL, PA_CLR_TO_RX2); | 3279 | skge_write16(hw, B3_PA_CTRL, PA_CLR_TO_RX2); |
3283 | } | 3280 | } |
@@ -3569,8 +3566,6 @@ static struct net_device *skge_devinit(struct skge_hw *hw, int port, | |||
3569 | SET_ETHTOOL_OPS(dev, &skge_ethtool_ops); | 3566 | SET_ETHTOOL_OPS(dev, &skge_ethtool_ops); |
3570 | dev->tx_timeout = skge_tx_timeout; | 3567 | dev->tx_timeout = skge_tx_timeout; |
3571 | dev->watchdog_timeo = TX_WATCHDOG; | 3568 | dev->watchdog_timeo = TX_WATCHDOG; |
3572 | dev->poll = skge_poll; | ||
3573 | dev->weight = NAPI_WEIGHT; | ||
3574 | #ifdef CONFIG_NET_POLL_CONTROLLER | 3569 | #ifdef CONFIG_NET_POLL_CONTROLLER |
3575 | dev->poll_controller = skge_netpoll; | 3570 | dev->poll_controller = skge_netpoll; |
3576 | #endif | 3571 | #endif |
@@ -3580,6 +3575,7 @@ static struct net_device *skge_devinit(struct skge_hw *hw, int port, | |||
3580 | dev->features |= NETIF_F_HIGHDMA; | 3575 | dev->features |= NETIF_F_HIGHDMA; |
3581 | 3576 | ||
3582 | skge = netdev_priv(dev); | 3577 | skge = netdev_priv(dev); |
3578 | netif_napi_add(dev, &skge->napi, skge_poll, NAPI_WEIGHT); | ||
3583 | skge->netdev = dev; | 3579 | skge->netdev = dev; |
3584 | skge->hw = hw; | 3580 | skge->hw = hw; |
3585 | skge->msg_enable = netif_msg_init(debug, default_msg); | 3581 | skge->msg_enable = netif_msg_init(debug, default_msg); |
diff --git a/drivers/net/skge.h b/drivers/net/skge.h index edd71468220c..dd0fd45c7155 100644 --- a/drivers/net/skge.h +++ b/drivers/net/skge.h | |||
@@ -2448,6 +2448,7 @@ enum pause_status { | |||
2448 | struct skge_port { | 2448 | struct skge_port { |
2449 | struct skge_hw *hw; | 2449 | struct skge_hw *hw; |
2450 | struct net_device *netdev; | 2450 | struct net_device *netdev; |
2451 | struct napi_struct napi; | ||
2451 | int port; | 2452 | int port; |
2452 | u32 msg_enable; | 2453 | u32 msg_enable; |
2453 | 2454 | ||
diff --git a/drivers/net/sky2.c b/drivers/net/sky2.c index ea117fc3d5e3..a0d75b0f3798 100644 --- a/drivers/net/sky2.c +++ b/drivers/net/sky2.c | |||
@@ -1130,7 +1130,7 @@ static void sky2_vlan_rx_register(struct net_device *dev, struct vlan_group *grp | |||
1130 | u16 port = sky2->port; | 1130 | u16 port = sky2->port; |
1131 | 1131 | ||
1132 | netif_tx_lock_bh(dev); | 1132 | netif_tx_lock_bh(dev); |
1133 | netif_poll_disable(sky2->hw->dev[0]); | 1133 | napi_disable(&hw->napi); |
1134 | 1134 | ||
1135 | sky2->vlgrp = grp; | 1135 | sky2->vlgrp = grp; |
1136 | if (grp) { | 1136 | if (grp) { |
@@ -1145,7 +1145,7 @@ static void sky2_vlan_rx_register(struct net_device *dev, struct vlan_group *grp | |||
1145 | TX_VLAN_TAG_OFF); | 1145 | TX_VLAN_TAG_OFF); |
1146 | } | 1146 | } |
1147 | 1147 | ||
1148 | netif_poll_enable(sky2->hw->dev[0]); | 1148 | napi_enable(&hw->napi); |
1149 | netif_tx_unlock_bh(dev); | 1149 | netif_tx_unlock_bh(dev); |
1150 | } | 1150 | } |
1151 | #endif | 1151 | #endif |
@@ -1385,9 +1385,13 @@ static int sky2_up(struct net_device *dev) | |||
1385 | sky2_prefetch_init(hw, txqaddr[port], sky2->tx_le_map, | 1385 | sky2_prefetch_init(hw, txqaddr[port], sky2->tx_le_map, |
1386 | TX_RING_SIZE - 1); | 1386 | TX_RING_SIZE - 1); |
1387 | 1387 | ||
1388 | napi_enable(&hw->napi); | ||
1389 | |||
1388 | err = sky2_rx_start(sky2); | 1390 | err = sky2_rx_start(sky2); |
1389 | if (err) | 1391 | if (err) { |
1392 | napi_disable(&hw->napi); | ||
1390 | goto err_out; | 1393 | goto err_out; |
1394 | } | ||
1391 | 1395 | ||
1392 | /* Enable interrupts from phy/mac for port */ | 1396 | /* Enable interrupts from phy/mac for port */ |
1393 | imask = sky2_read32(hw, B0_IMSK); | 1397 | imask = sky2_read32(hw, B0_IMSK); |
@@ -1676,6 +1680,8 @@ static int sky2_down(struct net_device *dev) | |||
1676 | /* Stop more packets from being queued */ | 1680 | /* Stop more packets from being queued */ |
1677 | netif_stop_queue(dev); | 1681 | netif_stop_queue(dev); |
1678 | 1682 | ||
1683 | napi_disable(&hw->napi); | ||
1684 | |||
1679 | /* Disable port IRQ */ | 1685 | /* Disable port IRQ */ |
1680 | imask = sky2_read32(hw, B0_IMSK); | 1686 | imask = sky2_read32(hw, B0_IMSK); |
1681 | imask &= ~portirq_msk[port]; | 1687 | imask &= ~portirq_msk[port]; |
@@ -2016,7 +2022,7 @@ static int sky2_change_mtu(struct net_device *dev, int new_mtu) | |||
2016 | 2022 | ||
2017 | dev->trans_start = jiffies; /* prevent tx timeout */ | 2023 | dev->trans_start = jiffies; /* prevent tx timeout */ |
2018 | netif_stop_queue(dev); | 2024 | netif_stop_queue(dev); |
2019 | netif_poll_disable(hw->dev[0]); | 2025 | napi_disable(&hw->napi); |
2020 | 2026 | ||
2021 | synchronize_irq(hw->pdev->irq); | 2027 | synchronize_irq(hw->pdev->irq); |
2022 | 2028 | ||
@@ -2043,12 +2049,16 @@ static int sky2_change_mtu(struct net_device *dev, int new_mtu) | |||
2043 | err = sky2_rx_start(sky2); | 2049 | err = sky2_rx_start(sky2); |
2044 | sky2_write32(hw, B0_IMSK, imask); | 2050 | sky2_write32(hw, B0_IMSK, imask); |
2045 | 2051 | ||
2052 | /* Unconditionally re-enable NAPI because even if we | ||
2053 | * call dev_close() that will do a napi_disable(). | ||
2054 | */ | ||
2055 | napi_enable(&hw->napi); | ||
2056 | |||
2046 | if (err) | 2057 | if (err) |
2047 | dev_close(dev); | 2058 | dev_close(dev); |
2048 | else { | 2059 | else { |
2049 | gma_write16(hw, port, GM_GP_CTRL, ctl); | 2060 | gma_write16(hw, port, GM_GP_CTRL, ctl); |
2050 | 2061 | ||
2051 | netif_poll_enable(hw->dev[0]); | ||
2052 | netif_wake_queue(dev); | 2062 | netif_wake_queue(dev); |
2053 | } | 2063 | } |
2054 | 2064 | ||
@@ -2544,18 +2554,15 @@ static int sky2_rx_hung(struct net_device *dev) | |||
2544 | static void sky2_watchdog(unsigned long arg) | 2554 | static void sky2_watchdog(unsigned long arg) |
2545 | { | 2555 | { |
2546 | struct sky2_hw *hw = (struct sky2_hw *) arg; | 2556 | struct sky2_hw *hw = (struct sky2_hw *) arg; |
2547 | struct net_device *dev; | ||
2548 | 2557 | ||
2549 | /* Check for lost IRQ once a second */ | 2558 | /* Check for lost IRQ once a second */ |
2550 | if (sky2_read32(hw, B0_ISRC)) { | 2559 | if (sky2_read32(hw, B0_ISRC)) { |
2551 | dev = hw->dev[0]; | 2560 | napi_schedule(&hw->napi); |
2552 | if (__netif_rx_schedule_prep(dev)) | ||
2553 | __netif_rx_schedule(dev); | ||
2554 | } else { | 2561 | } else { |
2555 | int i, active = 0; | 2562 | int i, active = 0; |
2556 | 2563 | ||
2557 | for (i = 0; i < hw->ports; i++) { | 2564 | for (i = 0; i < hw->ports; i++) { |
2558 | dev = hw->dev[i]; | 2565 | struct net_device *dev = hw->dev[i]; |
2559 | if (!netif_running(dev)) | 2566 | if (!netif_running(dev)) |
2560 | continue; | 2567 | continue; |
2561 | ++active; | 2568 | ++active; |
@@ -2605,11 +2612,11 @@ static void sky2_err_intr(struct sky2_hw *hw, u32 status) | |||
2605 | sky2_le_error(hw, 1, Q_XA2, TX_RING_SIZE); | 2612 | sky2_le_error(hw, 1, Q_XA2, TX_RING_SIZE); |
2606 | } | 2613 | } |
2607 | 2614 | ||
2608 | static int sky2_poll(struct net_device *dev0, int *budget) | 2615 | static int sky2_poll(struct napi_struct *napi, int work_limit) |
2609 | { | 2616 | { |
2610 | struct sky2_hw *hw = ((struct sky2_port *) netdev_priv(dev0))->hw; | 2617 | struct sky2_hw *hw = container_of(napi, struct sky2_hw, napi); |
2611 | int work_done; | ||
2612 | u32 status = sky2_read32(hw, B0_Y2_SP_EISR); | 2618 | u32 status = sky2_read32(hw, B0_Y2_SP_EISR); |
2619 | int work_done; | ||
2613 | 2620 | ||
2614 | if (unlikely(status & Y2_IS_ERROR)) | 2621 | if (unlikely(status & Y2_IS_ERROR)) |
2615 | sky2_err_intr(hw, status); | 2622 | sky2_err_intr(hw, status); |
@@ -2620,31 +2627,27 @@ static int sky2_poll(struct net_device *dev0, int *budget) | |||
2620 | if (status & Y2_IS_IRQ_PHY2) | 2627 | if (status & Y2_IS_IRQ_PHY2) |
2621 | sky2_phy_intr(hw, 1); | 2628 | sky2_phy_intr(hw, 1); |
2622 | 2629 | ||
2623 | work_done = sky2_status_intr(hw, min(dev0->quota, *budget)); | 2630 | work_done = sky2_status_intr(hw, work_limit); |
2624 | *budget -= work_done; | ||
2625 | dev0->quota -= work_done; | ||
2626 | 2631 | ||
2627 | /* More work? */ | 2632 | /* More work? */ |
2628 | if (hw->st_idx != sky2_read16(hw, STAT_PUT_IDX)) | 2633 | if (hw->st_idx == sky2_read16(hw, STAT_PUT_IDX)) { |
2629 | return 1; | 2634 | /* Bug/Errata workaround? |
2635 | * Need to kick the TX irq moderation timer. | ||
2636 | */ | ||
2637 | if (sky2_read8(hw, STAT_TX_TIMER_CTRL) == TIM_START) { | ||
2638 | sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_STOP); | ||
2639 | sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_START); | ||
2640 | } | ||
2630 | 2641 | ||
2631 | /* Bug/Errata workaround? | 2642 | napi_complete(napi); |
2632 | * Need to kick the TX irq moderation timer. | 2643 | sky2_read32(hw, B0_Y2_SP_LISR); |
2633 | */ | ||
2634 | if (sky2_read8(hw, STAT_TX_TIMER_CTRL) == TIM_START) { | ||
2635 | sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_STOP); | ||
2636 | sky2_write8(hw, STAT_TX_TIMER_CTRL, TIM_START); | ||
2637 | } | 2644 | } |
2638 | netif_rx_complete(dev0); | 2645 | return work_done; |
2639 | |||
2640 | sky2_read32(hw, B0_Y2_SP_LISR); | ||
2641 | return 0; | ||
2642 | } | 2646 | } |
2643 | 2647 | ||
2644 | static irqreturn_t sky2_intr(int irq, void *dev_id) | 2648 | static irqreturn_t sky2_intr(int irq, void *dev_id) |
2645 | { | 2649 | { |
2646 | struct sky2_hw *hw = dev_id; | 2650 | struct sky2_hw *hw = dev_id; |
2647 | struct net_device *dev0 = hw->dev[0]; | ||
2648 | u32 status; | 2651 | u32 status; |
2649 | 2652 | ||
2650 | /* Reading this mask interrupts as side effect */ | 2653 | /* Reading this mask interrupts as side effect */ |
@@ -2653,8 +2656,8 @@ static irqreturn_t sky2_intr(int irq, void *dev_id) | |||
2653 | return IRQ_NONE; | 2656 | return IRQ_NONE; |
2654 | 2657 | ||
2655 | prefetch(&hw->st_le[hw->st_idx]); | 2658 | prefetch(&hw->st_le[hw->st_idx]); |
2656 | if (likely(__netif_rx_schedule_prep(dev0))) | 2659 | |
2657 | __netif_rx_schedule(dev0); | 2660 | napi_schedule(&hw->napi); |
2658 | 2661 | ||
2659 | return IRQ_HANDLED; | 2662 | return IRQ_HANDLED; |
2660 | } | 2663 | } |
@@ -2663,10 +2666,8 @@ static irqreturn_t sky2_intr(int irq, void *dev_id) | |||
2663 | static void sky2_netpoll(struct net_device *dev) | 2666 | static void sky2_netpoll(struct net_device *dev) |
2664 | { | 2667 | { |
2665 | struct sky2_port *sky2 = netdev_priv(dev); | 2668 | struct sky2_port *sky2 = netdev_priv(dev); |
2666 | struct net_device *dev0 = sky2->hw->dev[0]; | ||
2667 | 2669 | ||
2668 | if (netif_running(dev) && __netif_rx_schedule_prep(dev0)) | 2670 | napi_schedule(&sky2->hw->napi); |
2669 | __netif_rx_schedule(dev0); | ||
2670 | } | 2671 | } |
2671 | #endif | 2672 | #endif |
2672 | 2673 | ||
@@ -2914,8 +2915,6 @@ static void sky2_restart(struct work_struct *work) | |||
2914 | sky2_write32(hw, B0_IMSK, 0); | 2915 | sky2_write32(hw, B0_IMSK, 0); |
2915 | sky2_read32(hw, B0_IMSK); | 2916 | sky2_read32(hw, B0_IMSK); |
2916 | 2917 | ||
2917 | netif_poll_disable(hw->dev[0]); | ||
2918 | |||
2919 | for (i = 0; i < hw->ports; i++) { | 2918 | for (i = 0; i < hw->ports; i++) { |
2920 | dev = hw->dev[i]; | 2919 | dev = hw->dev[i]; |
2921 | if (netif_running(dev)) | 2920 | if (netif_running(dev)) |
@@ -2924,7 +2923,6 @@ static void sky2_restart(struct work_struct *work) | |||
2924 | 2923 | ||
2925 | sky2_reset(hw); | 2924 | sky2_reset(hw); |
2926 | sky2_write32(hw, B0_IMSK, Y2_IS_BASE); | 2925 | sky2_write32(hw, B0_IMSK, Y2_IS_BASE); |
2927 | netif_poll_enable(hw->dev[0]); | ||
2928 | 2926 | ||
2929 | for (i = 0; i < hw->ports; i++) { | 2927 | for (i = 0; i < hw->ports; i++) { |
2930 | dev = hw->dev[i]; | 2928 | dev = hw->dev[i]; |
@@ -3735,7 +3733,7 @@ static int sky2_debug_show(struct seq_file *seq, void *v) | |||
3735 | { | 3733 | { |
3736 | struct net_device *dev = seq->private; | 3734 | struct net_device *dev = seq->private; |
3737 | const struct sky2_port *sky2 = netdev_priv(dev); | 3735 | const struct sky2_port *sky2 = netdev_priv(dev); |
3738 | const struct sky2_hw *hw = sky2->hw; | 3736 | struct sky2_hw *hw = sky2->hw; |
3739 | unsigned port = sky2->port; | 3737 | unsigned port = sky2->port; |
3740 | unsigned idx, last; | 3738 | unsigned idx, last; |
3741 | int sop; | 3739 | int sop; |
@@ -3748,7 +3746,7 @@ static int sky2_debug_show(struct seq_file *seq, void *v) | |||
3748 | sky2_read32(hw, B0_IMSK), | 3746 | sky2_read32(hw, B0_IMSK), |
3749 | sky2_read32(hw, B0_Y2_SP_ICR)); | 3747 | sky2_read32(hw, B0_Y2_SP_ICR)); |
3750 | 3748 | ||
3751 | netif_poll_disable(hw->dev[0]); | 3749 | napi_disable(&hw->napi); |
3752 | last = sky2_read16(hw, STAT_PUT_IDX); | 3750 | last = sky2_read16(hw, STAT_PUT_IDX); |
3753 | 3751 | ||
3754 | if (hw->st_idx == last) | 3752 | if (hw->st_idx == last) |
@@ -3818,7 +3816,7 @@ static int sky2_debug_show(struct seq_file *seq, void *v) | |||
3818 | last = sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_PUT_IDX)), | 3816 | last = sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_PUT_IDX)), |
3819 | sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_LAST_IDX))); | 3817 | sky2_read16(hw, Y2_QADDR(rxqaddr[port], PREF_UNIT_LAST_IDX))); |
3820 | 3818 | ||
3821 | netif_poll_enable(hw->dev[0]); | 3819 | napi_enable(&hw->napi); |
3822 | return 0; | 3820 | return 0; |
3823 | } | 3821 | } |
3824 | 3822 | ||
@@ -3943,15 +3941,8 @@ static __devinit struct net_device *sky2_init_netdev(struct sky2_hw *hw, | |||
3943 | SET_ETHTOOL_OPS(dev, &sky2_ethtool_ops); | 3941 | SET_ETHTOOL_OPS(dev, &sky2_ethtool_ops); |
3944 | dev->tx_timeout = sky2_tx_timeout; | 3942 | dev->tx_timeout = sky2_tx_timeout; |
3945 | dev->watchdog_timeo = TX_WATCHDOG; | 3943 | dev->watchdog_timeo = TX_WATCHDOG; |
3946 | if (port == 0) | ||
3947 | dev->poll = sky2_poll; | ||
3948 | dev->weight = NAPI_WEIGHT; | ||
3949 | #ifdef CONFIG_NET_POLL_CONTROLLER | 3944 | #ifdef CONFIG_NET_POLL_CONTROLLER |
3950 | /* Network console (only works on port 0) | 3945 | dev->poll_controller = sky2_netpoll; |
3951 | * because netpoll makes assumptions about NAPI | ||
3952 | */ | ||
3953 | if (port == 0) | ||
3954 | dev->poll_controller = sky2_netpoll; | ||
3955 | #endif | 3946 | #endif |
3956 | 3947 | ||
3957 | sky2 = netdev_priv(dev); | 3948 | sky2 = netdev_priv(dev); |
@@ -4166,6 +4157,7 @@ static int __devinit sky2_probe(struct pci_dev *pdev, | |||
4166 | err = -ENOMEM; | 4157 | err = -ENOMEM; |
4167 | goto err_out_free_pci; | 4158 | goto err_out_free_pci; |
4168 | } | 4159 | } |
4160 | netif_napi_add(dev, &hw->napi, sky2_poll, NAPI_WEIGHT); | ||
4169 | 4161 | ||
4170 | if (!disable_msi && pci_enable_msi(pdev) == 0) { | 4162 | if (!disable_msi && pci_enable_msi(pdev) == 0) { |
4171 | err = sky2_test_msi(hw); | 4163 | err = sky2_test_msi(hw); |
@@ -4288,8 +4280,6 @@ static int sky2_suspend(struct pci_dev *pdev, pm_message_t state) | |||
4288 | if (!hw) | 4280 | if (!hw) |
4289 | return 0; | 4281 | return 0; |
4290 | 4282 | ||
4291 | netif_poll_disable(hw->dev[0]); | ||
4292 | |||
4293 | for (i = 0; i < hw->ports; i++) { | 4283 | for (i = 0; i < hw->ports; i++) { |
4294 | struct net_device *dev = hw->dev[i]; | 4284 | struct net_device *dev = hw->dev[i]; |
4295 | struct sky2_port *sky2 = netdev_priv(dev); | 4285 | struct sky2_port *sky2 = netdev_priv(dev); |
@@ -4356,8 +4346,6 @@ static int sky2_resume(struct pci_dev *pdev) | |||
4356 | } | 4346 | } |
4357 | } | 4347 | } |
4358 | 4348 | ||
4359 | netif_poll_enable(hw->dev[0]); | ||
4360 | |||
4361 | return 0; | 4349 | return 0; |
4362 | out: | 4350 | out: |
4363 | dev_err(&pdev->dev, "resume failed (%d)\n", err); | 4351 | dev_err(&pdev->dev, "resume failed (%d)\n", err); |
@@ -4374,7 +4362,7 @@ static void sky2_shutdown(struct pci_dev *pdev) | |||
4374 | if (!hw) | 4362 | if (!hw) |
4375 | return; | 4363 | return; |
4376 | 4364 | ||
4377 | netif_poll_disable(hw->dev[0]); | 4365 | napi_disable(&hw->napi); |
4378 | 4366 | ||
4379 | for (i = 0; i < hw->ports; i++) { | 4367 | for (i = 0; i < hw->ports; i++) { |
4380 | struct net_device *dev = hw->dev[i]; | 4368 | struct net_device *dev = hw->dev[i]; |
diff --git a/drivers/net/sky2.h b/drivers/net/sky2.h index 8bc5c54e3efa..f18f8752118e 100644 --- a/drivers/net/sky2.h +++ b/drivers/net/sky2.h | |||
@@ -2057,6 +2057,7 @@ struct sky2_port { | |||
2057 | struct sky2_hw { | 2057 | struct sky2_hw { |
2058 | void __iomem *regs; | 2058 | void __iomem *regs; |
2059 | struct pci_dev *pdev; | 2059 | struct pci_dev *pdev; |
2060 | struct napi_struct napi; | ||
2060 | struct net_device *dev[2]; | 2061 | struct net_device *dev[2]; |
2061 | unsigned long flags; | 2062 | unsigned long flags; |
2062 | #define SKY2_HW_USE_MSI 0x00000001 | 2063 | #define SKY2_HW_USE_MSI 0x00000001 |
diff --git a/drivers/net/spider_net.c b/drivers/net/spider_net.c index 82d837ab4db9..6d8f2bb7e0f9 100644 --- a/drivers/net/spider_net.c +++ b/drivers/net/spider_net.c | |||
@@ -1278,34 +1278,26 @@ bad_desc: | |||
1278 | * (using netif_receive_skb). If all/enough packets are up, the driver | 1278 | * (using netif_receive_skb). If all/enough packets are up, the driver |
1279 | * reenables interrupts and returns 0. If not, 1 is returned. | 1279 | * reenables interrupts and returns 0. If not, 1 is returned. |
1280 | */ | 1280 | */ |
1281 | static int | 1281 | static int spider_net_poll(struct napi_struct *napi, int budget) |
1282 | spider_net_poll(struct net_device *netdev, int *budget) | ||
1283 | { | 1282 | { |
1284 | struct spider_net_card *card = netdev_priv(netdev); | 1283 | struct spider_net_card *card = container_of(napi, struct spider_net_card, napi); |
1285 | int packets_to_do, packets_done = 0; | 1284 | struct net_device *netdev = card->netdev; |
1286 | int no_more_packets = 0; | 1285 | int packets_done = 0; |
1287 | 1286 | ||
1288 | packets_to_do = min(*budget, netdev->quota); | 1287 | while (packets_done < budget) { |
1289 | 1288 | if (!spider_net_decode_one_descr(card)) | |
1290 | while (packets_to_do) { | ||
1291 | if (spider_net_decode_one_descr(card)) { | ||
1292 | packets_done++; | ||
1293 | packets_to_do--; | ||
1294 | } else { | ||
1295 | /* no more packets for the stack */ | ||
1296 | no_more_packets = 1; | ||
1297 | break; | 1289 | break; |
1298 | } | 1290 | |
1291 | packets_done++; | ||
1299 | } | 1292 | } |
1300 | 1293 | ||
1301 | if ((packets_done == 0) && (card->num_rx_ints != 0)) { | 1294 | if ((packets_done == 0) && (card->num_rx_ints != 0)) { |
1302 | no_more_packets = spider_net_resync_tail_ptr(card); | 1295 | if (!spider_net_resync_tail_ptr(card)) |
1296 | packets_done = budget; | ||
1303 | spider_net_resync_head_ptr(card); | 1297 | spider_net_resync_head_ptr(card); |
1304 | } | 1298 | } |
1305 | card->num_rx_ints = 0; | 1299 | card->num_rx_ints = 0; |
1306 | 1300 | ||
1307 | netdev->quota -= packets_done; | ||
1308 | *budget -= packets_done; | ||
1309 | spider_net_refill_rx_chain(card); | 1301 | spider_net_refill_rx_chain(card); |
1310 | spider_net_enable_rxdmac(card); | 1302 | spider_net_enable_rxdmac(card); |
1311 | 1303 | ||
@@ -1313,14 +1305,13 @@ spider_net_poll(struct net_device *netdev, int *budget) | |||
1313 | 1305 | ||
1314 | /* if all packets are in the stack, enable interrupts and return 0 */ | 1306 | /* if all packets are in the stack, enable interrupts and return 0 */ |
1315 | /* if not, return 1 */ | 1307 | /* if not, return 1 */ |
1316 | if (no_more_packets) { | 1308 | if (packets_done < budget) { |
1317 | netif_rx_complete(netdev); | 1309 | netif_rx_complete(netdev, napi); |
1318 | spider_net_rx_irq_on(card); | 1310 | spider_net_rx_irq_on(card); |
1319 | card->ignore_rx_ramfull = 0; | 1311 | card->ignore_rx_ramfull = 0; |
1320 | return 0; | ||
1321 | } | 1312 | } |
1322 | 1313 | ||
1323 | return 1; | 1314 | return packets_done; |
1324 | } | 1315 | } |
1325 | 1316 | ||
1326 | /** | 1317 | /** |
@@ -1560,7 +1551,8 @@ spider_net_handle_error_irq(struct spider_net_card *card, u32 status_reg) | |||
1560 | spider_net_refill_rx_chain(card); | 1551 | spider_net_refill_rx_chain(card); |
1561 | spider_net_enable_rxdmac(card); | 1552 | spider_net_enable_rxdmac(card); |
1562 | card->num_rx_ints ++; | 1553 | card->num_rx_ints ++; |
1563 | netif_rx_schedule(card->netdev); | 1554 | netif_rx_schedule(card->netdev, |
1555 | &card->napi); | ||
1564 | } | 1556 | } |
1565 | show_error = 0; | 1557 | show_error = 0; |
1566 | break; | 1558 | break; |
@@ -1580,7 +1572,8 @@ spider_net_handle_error_irq(struct spider_net_card *card, u32 status_reg) | |||
1580 | spider_net_refill_rx_chain(card); | 1572 | spider_net_refill_rx_chain(card); |
1581 | spider_net_enable_rxdmac(card); | 1573 | spider_net_enable_rxdmac(card); |
1582 | card->num_rx_ints ++; | 1574 | card->num_rx_ints ++; |
1583 | netif_rx_schedule(card->netdev); | 1575 | netif_rx_schedule(card->netdev, |
1576 | &card->napi); | ||
1584 | show_error = 0; | 1577 | show_error = 0; |
1585 | break; | 1578 | break; |
1586 | 1579 | ||
@@ -1594,7 +1587,8 @@ spider_net_handle_error_irq(struct spider_net_card *card, u32 status_reg) | |||
1594 | spider_net_refill_rx_chain(card); | 1587 | spider_net_refill_rx_chain(card); |
1595 | spider_net_enable_rxdmac(card); | 1588 | spider_net_enable_rxdmac(card); |
1596 | card->num_rx_ints ++; | 1589 | card->num_rx_ints ++; |
1597 | netif_rx_schedule(card->netdev); | 1590 | netif_rx_schedule(card->netdev, |
1591 | &card->napi); | ||
1598 | show_error = 0; | 1592 | show_error = 0; |
1599 | break; | 1593 | break; |
1600 | 1594 | ||
@@ -1686,11 +1680,11 @@ spider_net_interrupt(int irq, void *ptr) | |||
1686 | 1680 | ||
1687 | if (status_reg & SPIDER_NET_RXINT ) { | 1681 | if (status_reg & SPIDER_NET_RXINT ) { |
1688 | spider_net_rx_irq_off(card); | 1682 | spider_net_rx_irq_off(card); |
1689 | netif_rx_schedule(netdev); | 1683 | netif_rx_schedule(netdev, &card->napi); |
1690 | card->num_rx_ints ++; | 1684 | card->num_rx_ints ++; |
1691 | } | 1685 | } |
1692 | if (status_reg & SPIDER_NET_TXINT) | 1686 | if (status_reg & SPIDER_NET_TXINT) |
1693 | netif_rx_schedule(netdev); | 1687 | netif_rx_schedule(netdev, &card->napi); |
1694 | 1688 | ||
1695 | if (status_reg & SPIDER_NET_LINKINT) | 1689 | if (status_reg & SPIDER_NET_LINKINT) |
1696 | spider_net_link_reset(netdev); | 1690 | spider_net_link_reset(netdev); |
@@ -2034,7 +2028,7 @@ spider_net_open(struct net_device *netdev) | |||
2034 | 2028 | ||
2035 | netif_start_queue(netdev); | 2029 | netif_start_queue(netdev); |
2036 | netif_carrier_on(netdev); | 2030 | netif_carrier_on(netdev); |
2037 | netif_poll_enable(netdev); | 2031 | napi_enable(&card->napi); |
2038 | 2032 | ||
2039 | spider_net_enable_interrupts(card); | 2033 | spider_net_enable_interrupts(card); |
2040 | 2034 | ||
@@ -2204,7 +2198,7 @@ spider_net_stop(struct net_device *netdev) | |||
2204 | { | 2198 | { |
2205 | struct spider_net_card *card = netdev_priv(netdev); | 2199 | struct spider_net_card *card = netdev_priv(netdev); |
2206 | 2200 | ||
2207 | netif_poll_disable(netdev); | 2201 | napi_disable(&card->napi); |
2208 | netif_carrier_off(netdev); | 2202 | netif_carrier_off(netdev); |
2209 | netif_stop_queue(netdev); | 2203 | netif_stop_queue(netdev); |
2210 | del_timer_sync(&card->tx_timer); | 2204 | del_timer_sync(&card->tx_timer); |
@@ -2304,9 +2298,6 @@ spider_net_setup_netdev_ops(struct net_device *netdev) | |||
2304 | /* tx watchdog */ | 2298 | /* tx watchdog */ |
2305 | netdev->tx_timeout = &spider_net_tx_timeout; | 2299 | netdev->tx_timeout = &spider_net_tx_timeout; |
2306 | netdev->watchdog_timeo = SPIDER_NET_WATCHDOG_TIMEOUT; | 2300 | netdev->watchdog_timeo = SPIDER_NET_WATCHDOG_TIMEOUT; |
2307 | /* NAPI */ | ||
2308 | netdev->poll = &spider_net_poll; | ||
2309 | netdev->weight = SPIDER_NET_NAPI_WEIGHT; | ||
2310 | /* HW VLAN */ | 2301 | /* HW VLAN */ |
2311 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2302 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2312 | /* poll controller */ | 2303 | /* poll controller */ |
@@ -2351,6 +2342,9 @@ spider_net_setup_netdev(struct spider_net_card *card) | |||
2351 | 2342 | ||
2352 | card->options.rx_csum = SPIDER_NET_RX_CSUM_DEFAULT; | 2343 | card->options.rx_csum = SPIDER_NET_RX_CSUM_DEFAULT; |
2353 | 2344 | ||
2345 | netif_napi_add(netdev, &card->napi, | ||
2346 | spider_net_poll, SPIDER_NET_NAPI_WEIGHT); | ||
2347 | |||
2354 | spider_net_setup_netdev_ops(netdev); | 2348 | spider_net_setup_netdev_ops(netdev); |
2355 | 2349 | ||
2356 | netdev->features = NETIF_F_IP_CSUM | NETIF_F_LLTX; | 2350 | netdev->features = NETIF_F_IP_CSUM | NETIF_F_LLTX; |
diff --git a/drivers/net/spider_net.h b/drivers/net/spider_net.h index dbbdb8cee3c6..a2fcdebc3790 100644 --- a/drivers/net/spider_net.h +++ b/drivers/net/spider_net.h | |||
@@ -466,6 +466,8 @@ struct spider_net_card { | |||
466 | struct pci_dev *pdev; | 466 | struct pci_dev *pdev; |
467 | struct mii_phy phy; | 467 | struct mii_phy phy; |
468 | 468 | ||
469 | struct napi_struct napi; | ||
470 | |||
469 | int medium; | 471 | int medium; |
470 | 472 | ||
471 | void __iomem *regs; | 473 | void __iomem *regs; |
diff --git a/drivers/net/starfire.c b/drivers/net/starfire.c index 8b6478663a56..3b9336c34206 100644 --- a/drivers/net/starfire.c +++ b/drivers/net/starfire.c | |||
@@ -178,16 +178,13 @@ static int full_duplex[MAX_UNITS] = {0, }; | |||
178 | #define skb_num_frags(skb) (skb_shinfo(skb)->nr_frags + 1) | 178 | #define skb_num_frags(skb) (skb_shinfo(skb)->nr_frags + 1) |
179 | 179 | ||
180 | #ifdef HAVE_NETDEV_POLL | 180 | #ifdef HAVE_NETDEV_POLL |
181 | #define init_poll(dev) \ | 181 | #define init_poll(dev, np) \ |
182 | do { \ | 182 | netif_napi_add(dev, &np->napi, netdev_poll, max_interrupt_work) |
183 | dev->poll = &netdev_poll; \ | 183 | #define netdev_rx(dev, np, ioaddr) \ |
184 | dev->weight = max_interrupt_work; \ | ||
185 | } while (0) | ||
186 | #define netdev_rx(dev, ioaddr) \ | ||
187 | do { \ | 184 | do { \ |
188 | u32 intr_enable; \ | 185 | u32 intr_enable; \ |
189 | if (netif_rx_schedule_prep(dev)) { \ | 186 | if (netif_rx_schedule_prep(dev, &np->napi)) { \ |
190 | __netif_rx_schedule(dev); \ | 187 | __netif_rx_schedule(dev, &np->napi); \ |
191 | intr_enable = readl(ioaddr + IntrEnable); \ | 188 | intr_enable = readl(ioaddr + IntrEnable); \ |
192 | intr_enable &= ~(IntrRxDone | IntrRxEmpty); \ | 189 | intr_enable &= ~(IntrRxDone | IntrRxEmpty); \ |
193 | writel(intr_enable, ioaddr + IntrEnable); \ | 190 | writel(intr_enable, ioaddr + IntrEnable); \ |
@@ -204,12 +201,12 @@ do { \ | |||
204 | } while (0) | 201 | } while (0) |
205 | #define netdev_receive_skb(skb) netif_receive_skb(skb) | 202 | #define netdev_receive_skb(skb) netif_receive_skb(skb) |
206 | #define vlan_netdev_receive_skb(skb, vlgrp, vlid) vlan_hwaccel_receive_skb(skb, vlgrp, vlid) | 203 | #define vlan_netdev_receive_skb(skb, vlgrp, vlid) vlan_hwaccel_receive_skb(skb, vlgrp, vlid) |
207 | static int netdev_poll(struct net_device *dev, int *budget); | 204 | static int netdev_poll(struct napi_struct *napi, int budget); |
208 | #else /* not HAVE_NETDEV_POLL */ | 205 | #else /* not HAVE_NETDEV_POLL */ |
209 | #define init_poll(dev) | 206 | #define init_poll(dev, np) |
210 | #define netdev_receive_skb(skb) netif_rx(skb) | 207 | #define netdev_receive_skb(skb) netif_rx(skb) |
211 | #define vlan_netdev_receive_skb(skb, vlgrp, vlid) vlan_hwaccel_rx(skb, vlgrp, vlid) | 208 | #define vlan_netdev_receive_skb(skb, vlgrp, vlid) vlan_hwaccel_rx(skb, vlgrp, vlid) |
212 | #define netdev_rx(dev, ioaddr) \ | 209 | #define netdev_rx(dev, np, ioaddr) \ |
213 | do { \ | 210 | do { \ |
214 | int quota = np->dirty_rx + RX_RING_SIZE - np->cur_rx; \ | 211 | int quota = np->dirty_rx + RX_RING_SIZE - np->cur_rx; \ |
215 | __netdev_rx(dev, "a);\ | 212 | __netdev_rx(dev, "a);\ |
@@ -599,6 +596,8 @@ struct netdev_private { | |||
599 | struct tx_done_desc *tx_done_q; | 596 | struct tx_done_desc *tx_done_q; |
600 | dma_addr_t tx_done_q_dma; | 597 | dma_addr_t tx_done_q_dma; |
601 | unsigned int tx_done; | 598 | unsigned int tx_done; |
599 | struct napi_struct napi; | ||
600 | struct net_device *dev; | ||
602 | struct net_device_stats stats; | 601 | struct net_device_stats stats; |
603 | struct pci_dev *pci_dev; | 602 | struct pci_dev *pci_dev; |
604 | #ifdef VLAN_SUPPORT | 603 | #ifdef VLAN_SUPPORT |
@@ -791,6 +790,7 @@ static int __devinit starfire_init_one(struct pci_dev *pdev, | |||
791 | dev->irq = irq; | 790 | dev->irq = irq; |
792 | 791 | ||
793 | np = netdev_priv(dev); | 792 | np = netdev_priv(dev); |
793 | np->dev = dev; | ||
794 | np->base = base; | 794 | np->base = base; |
795 | spin_lock_init(&np->lock); | 795 | spin_lock_init(&np->lock); |
796 | pci_set_drvdata(pdev, dev); | 796 | pci_set_drvdata(pdev, dev); |
@@ -851,7 +851,7 @@ static int __devinit starfire_init_one(struct pci_dev *pdev, | |||
851 | dev->hard_start_xmit = &start_tx; | 851 | dev->hard_start_xmit = &start_tx; |
852 | dev->tx_timeout = tx_timeout; | 852 | dev->tx_timeout = tx_timeout; |
853 | dev->watchdog_timeo = TX_TIMEOUT; | 853 | dev->watchdog_timeo = TX_TIMEOUT; |
854 | init_poll(dev); | 854 | init_poll(dev, np); |
855 | dev->stop = &netdev_close; | 855 | dev->stop = &netdev_close; |
856 | dev->get_stats = &get_stats; | 856 | dev->get_stats = &get_stats; |
857 | dev->set_multicast_list = &set_rx_mode; | 857 | dev->set_multicast_list = &set_rx_mode; |
@@ -1056,6 +1056,9 @@ static int netdev_open(struct net_device *dev) | |||
1056 | 1056 | ||
1057 | writel(np->intr_timer_ctrl, ioaddr + IntrTimerCtrl); | 1057 | writel(np->intr_timer_ctrl, ioaddr + IntrTimerCtrl); |
1058 | 1058 | ||
1059 | #ifdef HAVE_NETDEV_POLL | ||
1060 | napi_enable(&np->napi); | ||
1061 | #endif | ||
1059 | netif_start_queue(dev); | 1062 | netif_start_queue(dev); |
1060 | 1063 | ||
1061 | if (debug > 1) | 1064 | if (debug > 1) |
@@ -1330,7 +1333,7 @@ static irqreturn_t intr_handler(int irq, void *dev_instance) | |||
1330 | handled = 1; | 1333 | handled = 1; |
1331 | 1334 | ||
1332 | if (intr_status & (IntrRxDone | IntrRxEmpty)) | 1335 | if (intr_status & (IntrRxDone | IntrRxEmpty)) |
1333 | netdev_rx(dev, ioaddr); | 1336 | netdev_rx(dev, np, ioaddr); |
1334 | 1337 | ||
1335 | /* Scavenge the skbuff list based on the Tx-done queue. | 1338 | /* Scavenge the skbuff list based on the Tx-done queue. |
1336 | There are redundant checks here that may be cleaned up | 1339 | There are redundant checks here that may be cleaned up |
@@ -1531,36 +1534,35 @@ static int __netdev_rx(struct net_device *dev, int *quota) | |||
1531 | 1534 | ||
1532 | 1535 | ||
1533 | #ifdef HAVE_NETDEV_POLL | 1536 | #ifdef HAVE_NETDEV_POLL |
1534 | static int netdev_poll(struct net_device *dev, int *budget) | 1537 | static int netdev_poll(struct napi_struct *napi, int budget) |
1535 | { | 1538 | { |
1539 | struct netdev_private *np = container_of(napi, struct netdev_private, napi); | ||
1540 | struct net_device *dev = np->dev; | ||
1536 | u32 intr_status; | 1541 | u32 intr_status; |
1537 | struct netdev_private *np = netdev_priv(dev); | ||
1538 | void __iomem *ioaddr = np->base; | 1542 | void __iomem *ioaddr = np->base; |
1539 | int retcode = 0, quota = dev->quota; | 1543 | int quota = budget; |
1540 | 1544 | ||
1541 | do { | 1545 | do { |
1542 | writel(IntrRxDone | IntrRxEmpty, ioaddr + IntrClear); | 1546 | writel(IntrRxDone | IntrRxEmpty, ioaddr + IntrClear); |
1543 | 1547 | ||
1544 | retcode = __netdev_rx(dev, "a); | 1548 | if (__netdev_rx(dev, "a)) |
1545 | *budget -= (dev->quota - quota); | ||
1546 | dev->quota = quota; | ||
1547 | if (retcode) | ||
1548 | goto out; | 1549 | goto out; |
1549 | 1550 | ||
1550 | intr_status = readl(ioaddr + IntrStatus); | 1551 | intr_status = readl(ioaddr + IntrStatus); |
1551 | } while (intr_status & (IntrRxDone | IntrRxEmpty)); | 1552 | } while (intr_status & (IntrRxDone | IntrRxEmpty)); |
1552 | 1553 | ||
1553 | netif_rx_complete(dev); | 1554 | netif_rx_complete(dev, napi); |
1554 | intr_status = readl(ioaddr + IntrEnable); | 1555 | intr_status = readl(ioaddr + IntrEnable); |
1555 | intr_status |= IntrRxDone | IntrRxEmpty; | 1556 | intr_status |= IntrRxDone | IntrRxEmpty; |
1556 | writel(intr_status, ioaddr + IntrEnable); | 1557 | writel(intr_status, ioaddr + IntrEnable); |
1557 | 1558 | ||
1558 | out: | 1559 | out: |
1559 | if (debug > 5) | 1560 | if (debug > 5) |
1560 | printk(KERN_DEBUG " exiting netdev_poll(): %d.\n", retcode); | 1561 | printk(KERN_DEBUG " exiting netdev_poll(): %d.\n", |
1562 | budget - quota); | ||
1561 | 1563 | ||
1562 | /* Restart Rx engine if stopped. */ | 1564 | /* Restart Rx engine if stopped. */ |
1563 | return retcode; | 1565 | return budget - quota; |
1564 | } | 1566 | } |
1565 | #endif /* HAVE_NETDEV_POLL */ | 1567 | #endif /* HAVE_NETDEV_POLL */ |
1566 | 1568 | ||
@@ -1904,6 +1906,9 @@ static int netdev_close(struct net_device *dev) | |||
1904 | int i; | 1906 | int i; |
1905 | 1907 | ||
1906 | netif_stop_queue(dev); | 1908 | netif_stop_queue(dev); |
1909 | #ifdef HAVE_NETDEV_POLL | ||
1910 | napi_disable(&np->napi); | ||
1911 | #endif | ||
1907 | 1912 | ||
1908 | if (debug > 1) { | 1913 | if (debug > 1) { |
1909 | printk(KERN_DEBUG "%s: Shutting down ethercard, Intr status %#8.8x.\n", | 1914 | printk(KERN_DEBUG "%s: Shutting down ethercard, Intr status %#8.8x.\n", |
diff --git a/drivers/net/sungem.c b/drivers/net/sungem.c index 432803855034..bf821e96f7b2 100644 --- a/drivers/net/sungem.c +++ b/drivers/net/sungem.c | |||
@@ -19,7 +19,7 @@ | |||
19 | * | 19 | * |
20 | * gem_change_mtu() and gem_set_multicast() are called with a read_lock() | 20 | * gem_change_mtu() and gem_set_multicast() are called with a read_lock() |
21 | * help by net/core/dev.c, thus they can't schedule. That means they can't | 21 | * help by net/core/dev.c, thus they can't schedule. That means they can't |
22 | * call netif_poll_disable() neither, thus force gem_poll() to keep a spinlock | 22 | * call napi_disable() neither, thus force gem_poll() to keep a spinlock |
23 | * where it could have been dropped. change_mtu especially would love also to | 23 | * where it could have been dropped. change_mtu especially would love also to |
24 | * be able to msleep instead of horrid locked delays when resetting the HW, | 24 | * be able to msleep instead of horrid locked delays when resetting the HW, |
25 | * but that read_lock() makes it impossible, unless I defer it's action to | 25 | * but that read_lock() makes it impossible, unless I defer it's action to |
@@ -878,19 +878,20 @@ static int gem_rx(struct gem *gp, int work_to_do) | |||
878 | return work_done; | 878 | return work_done; |
879 | } | 879 | } |
880 | 880 | ||
881 | static int gem_poll(struct net_device *dev, int *budget) | 881 | static int gem_poll(struct napi_struct *napi, int budget) |
882 | { | 882 | { |
883 | struct gem *gp = dev->priv; | 883 | struct gem *gp = container_of(napi, struct gem, napi); |
884 | struct net_device *dev = gp->dev; | ||
884 | unsigned long flags; | 885 | unsigned long flags; |
886 | int work_done; | ||
885 | 887 | ||
886 | /* | 888 | /* |
887 | * NAPI locking nightmare: See comment at head of driver | 889 | * NAPI locking nightmare: See comment at head of driver |
888 | */ | 890 | */ |
889 | spin_lock_irqsave(&gp->lock, flags); | 891 | spin_lock_irqsave(&gp->lock, flags); |
890 | 892 | ||
893 | work_done = 0; | ||
891 | do { | 894 | do { |
892 | int work_to_do, work_done; | ||
893 | |||
894 | /* Handle anomalies */ | 895 | /* Handle anomalies */ |
895 | if (gp->status & GREG_STAT_ABNORMAL) { | 896 | if (gp->status & GREG_STAT_ABNORMAL) { |
896 | if (gem_abnormal_irq(dev, gp, gp->status)) | 897 | if (gem_abnormal_irq(dev, gp, gp->status)) |
@@ -906,29 +907,25 @@ static int gem_poll(struct net_device *dev, int *budget) | |||
906 | 907 | ||
907 | /* Run RX thread. We don't use any locking here, | 908 | /* Run RX thread. We don't use any locking here, |
908 | * code willing to do bad things - like cleaning the | 909 | * code willing to do bad things - like cleaning the |
909 | * rx ring - must call netif_poll_disable(), which | 910 | * rx ring - must call napi_disable(), which |
910 | * schedule_timeout()'s if polling is already disabled. | 911 | * schedule_timeout()'s if polling is already disabled. |
911 | */ | 912 | */ |
912 | work_to_do = min(*budget, dev->quota); | 913 | work_done += gem_rx(gp, budget); |
913 | |||
914 | work_done = gem_rx(gp, work_to_do); | ||
915 | |||
916 | *budget -= work_done; | ||
917 | dev->quota -= work_done; | ||
918 | 914 | ||
919 | if (work_done >= work_to_do) | 915 | if (work_done >= budget) |
920 | return 1; | 916 | return work_done; |
921 | 917 | ||
922 | spin_lock_irqsave(&gp->lock, flags); | 918 | spin_lock_irqsave(&gp->lock, flags); |
923 | 919 | ||
924 | gp->status = readl(gp->regs + GREG_STAT); | 920 | gp->status = readl(gp->regs + GREG_STAT); |
925 | } while (gp->status & GREG_STAT_NAPI); | 921 | } while (gp->status & GREG_STAT_NAPI); |
926 | 922 | ||
927 | __netif_rx_complete(dev); | 923 | __netif_rx_complete(dev, napi); |
928 | gem_enable_ints(gp); | 924 | gem_enable_ints(gp); |
929 | 925 | ||
930 | spin_unlock_irqrestore(&gp->lock, flags); | 926 | spin_unlock_irqrestore(&gp->lock, flags); |
931 | return 0; | 927 | |
928 | return work_done; | ||
932 | } | 929 | } |
933 | 930 | ||
934 | static irqreturn_t gem_interrupt(int irq, void *dev_id) | 931 | static irqreturn_t gem_interrupt(int irq, void *dev_id) |
@@ -946,17 +943,17 @@ static irqreturn_t gem_interrupt(int irq, void *dev_id) | |||
946 | 943 | ||
947 | spin_lock_irqsave(&gp->lock, flags); | 944 | spin_lock_irqsave(&gp->lock, flags); |
948 | 945 | ||
949 | if (netif_rx_schedule_prep(dev)) { | 946 | if (netif_rx_schedule_prep(dev, &gp->napi)) { |
950 | u32 gem_status = readl(gp->regs + GREG_STAT); | 947 | u32 gem_status = readl(gp->regs + GREG_STAT); |
951 | 948 | ||
952 | if (gem_status == 0) { | 949 | if (gem_status == 0) { |
953 | netif_poll_enable(dev); | 950 | napi_enable(&gp->napi); |
954 | spin_unlock_irqrestore(&gp->lock, flags); | 951 | spin_unlock_irqrestore(&gp->lock, flags); |
955 | return IRQ_NONE; | 952 | return IRQ_NONE; |
956 | } | 953 | } |
957 | gp->status = gem_status; | 954 | gp->status = gem_status; |
958 | gem_disable_ints(gp); | 955 | gem_disable_ints(gp); |
959 | __netif_rx_schedule(dev); | 956 | __netif_rx_schedule(dev, &gp->napi); |
960 | } | 957 | } |
961 | 958 | ||
962 | spin_unlock_irqrestore(&gp->lock, flags); | 959 | spin_unlock_irqrestore(&gp->lock, flags); |
@@ -2284,7 +2281,7 @@ static void gem_reset_task(struct work_struct *work) | |||
2284 | 2281 | ||
2285 | mutex_lock(&gp->pm_mutex); | 2282 | mutex_lock(&gp->pm_mutex); |
2286 | 2283 | ||
2287 | netif_poll_disable(gp->dev); | 2284 | napi_disable(&gp->napi); |
2288 | 2285 | ||
2289 | spin_lock_irq(&gp->lock); | 2286 | spin_lock_irq(&gp->lock); |
2290 | spin_lock(&gp->tx_lock); | 2287 | spin_lock(&gp->tx_lock); |
@@ -2307,7 +2304,7 @@ static void gem_reset_task(struct work_struct *work) | |||
2307 | spin_unlock(&gp->tx_lock); | 2304 | spin_unlock(&gp->tx_lock); |
2308 | spin_unlock_irq(&gp->lock); | 2305 | spin_unlock_irq(&gp->lock); |
2309 | 2306 | ||
2310 | netif_poll_enable(gp->dev); | 2307 | napi_enable(&gp->napi); |
2311 | 2308 | ||
2312 | mutex_unlock(&gp->pm_mutex); | 2309 | mutex_unlock(&gp->pm_mutex); |
2313 | } | 2310 | } |
@@ -2324,6 +2321,8 @@ static int gem_open(struct net_device *dev) | |||
2324 | if (!gp->asleep) | 2321 | if (!gp->asleep) |
2325 | rc = gem_do_start(dev); | 2322 | rc = gem_do_start(dev); |
2326 | gp->opened = (rc == 0); | 2323 | gp->opened = (rc == 0); |
2324 | if (gp->opened) | ||
2325 | napi_enable(&gp->napi); | ||
2327 | 2326 | ||
2328 | mutex_unlock(&gp->pm_mutex); | 2327 | mutex_unlock(&gp->pm_mutex); |
2329 | 2328 | ||
@@ -2334,9 +2333,7 @@ static int gem_close(struct net_device *dev) | |||
2334 | { | 2333 | { |
2335 | struct gem *gp = dev->priv; | 2334 | struct gem *gp = dev->priv; |
2336 | 2335 | ||
2337 | /* Note: we don't need to call netif_poll_disable() here because | 2336 | napi_disable(&gp->napi); |
2338 | * our caller (dev_close) already did it for us | ||
2339 | */ | ||
2340 | 2337 | ||
2341 | mutex_lock(&gp->pm_mutex); | 2338 | mutex_lock(&gp->pm_mutex); |
2342 | 2339 | ||
@@ -2358,7 +2355,7 @@ static int gem_suspend(struct pci_dev *pdev, pm_message_t state) | |||
2358 | 2355 | ||
2359 | mutex_lock(&gp->pm_mutex); | 2356 | mutex_lock(&gp->pm_mutex); |
2360 | 2357 | ||
2361 | netif_poll_disable(dev); | 2358 | napi_disable(&gp->napi); |
2362 | 2359 | ||
2363 | printk(KERN_INFO "%s: suspending, WakeOnLan %s\n", | 2360 | printk(KERN_INFO "%s: suspending, WakeOnLan %s\n", |
2364 | dev->name, | 2361 | dev->name, |
@@ -2482,7 +2479,7 @@ static int gem_resume(struct pci_dev *pdev) | |||
2482 | spin_unlock(&gp->tx_lock); | 2479 | spin_unlock(&gp->tx_lock); |
2483 | spin_unlock_irqrestore(&gp->lock, flags); | 2480 | spin_unlock_irqrestore(&gp->lock, flags); |
2484 | 2481 | ||
2485 | netif_poll_enable(dev); | 2482 | napi_enable(&gp->napi); |
2486 | 2483 | ||
2487 | mutex_unlock(&gp->pm_mutex); | 2484 | mutex_unlock(&gp->pm_mutex); |
2488 | 2485 | ||
@@ -3121,8 +3118,7 @@ static int __devinit gem_init_one(struct pci_dev *pdev, | |||
3121 | dev->get_stats = gem_get_stats; | 3118 | dev->get_stats = gem_get_stats; |
3122 | dev->set_multicast_list = gem_set_multicast; | 3119 | dev->set_multicast_list = gem_set_multicast; |
3123 | dev->do_ioctl = gem_ioctl; | 3120 | dev->do_ioctl = gem_ioctl; |
3124 | dev->poll = gem_poll; | 3121 | netif_napi_add(dev, &gp->napi, gem_poll, 64); |
3125 | dev->weight = 64; | ||
3126 | dev->ethtool_ops = &gem_ethtool_ops; | 3122 | dev->ethtool_ops = &gem_ethtool_ops; |
3127 | dev->tx_timeout = gem_tx_timeout; | 3123 | dev->tx_timeout = gem_tx_timeout; |
3128 | dev->watchdog_timeo = 5 * HZ; | 3124 | dev->watchdog_timeo = 5 * HZ; |
diff --git a/drivers/net/sungem.h b/drivers/net/sungem.h index 58cf87c5751e..76d760acc9e2 100644 --- a/drivers/net/sungem.h +++ b/drivers/net/sungem.h | |||
@@ -993,6 +993,7 @@ struct gem { | |||
993 | u32 msg_enable; | 993 | u32 msg_enable; |
994 | u32 status; | 994 | u32 status; |
995 | 995 | ||
996 | struct napi_struct napi; | ||
996 | struct net_device_stats net_stats; | 997 | struct net_device_stats net_stats; |
997 | 998 | ||
998 | int tx_fifo_sz; | 999 | int tx_fifo_sz; |
diff --git a/drivers/net/tc35815.c b/drivers/net/tc35815.c index ec41469eee82..b5e0dff67230 100644 --- a/drivers/net/tc35815.c +++ b/drivers/net/tc35815.c | |||
@@ -414,6 +414,9 @@ enum tc35815_timer_state { | |||
414 | struct tc35815_local { | 414 | struct tc35815_local { |
415 | struct pci_dev *pci_dev; | 415 | struct pci_dev *pci_dev; |
416 | 416 | ||
417 | struct net_device *dev; | ||
418 | struct napi_struct napi; | ||
419 | |||
417 | /* statistics */ | 420 | /* statistics */ |
418 | struct net_device_stats stats; | 421 | struct net_device_stats stats; |
419 | struct { | 422 | struct { |
@@ -566,7 +569,7 @@ static int tc35815_send_packet(struct sk_buff *skb, struct net_device *dev); | |||
566 | static irqreturn_t tc35815_interrupt(int irq, void *dev_id); | 569 | static irqreturn_t tc35815_interrupt(int irq, void *dev_id); |
567 | #ifdef TC35815_NAPI | 570 | #ifdef TC35815_NAPI |
568 | static int tc35815_rx(struct net_device *dev, int limit); | 571 | static int tc35815_rx(struct net_device *dev, int limit); |
569 | static int tc35815_poll(struct net_device *dev, int *budget); | 572 | static int tc35815_poll(struct napi_struct *napi, int budget); |
570 | #else | 573 | #else |
571 | static void tc35815_rx(struct net_device *dev); | 574 | static void tc35815_rx(struct net_device *dev); |
572 | #endif | 575 | #endif |
@@ -685,6 +688,7 @@ static int __devinit tc35815_init_one (struct pci_dev *pdev, | |||
685 | SET_MODULE_OWNER(dev); | 688 | SET_MODULE_OWNER(dev); |
686 | SET_NETDEV_DEV(dev, &pdev->dev); | 689 | SET_NETDEV_DEV(dev, &pdev->dev); |
687 | lp = dev->priv; | 690 | lp = dev->priv; |
691 | lp->dev = dev; | ||
688 | 692 | ||
689 | /* enable device (incl. PCI PM wakeup), and bus-mastering */ | 693 | /* enable device (incl. PCI PM wakeup), and bus-mastering */ |
690 | rc = pci_enable_device (pdev); | 694 | rc = pci_enable_device (pdev); |
@@ -738,8 +742,7 @@ static int __devinit tc35815_init_one (struct pci_dev *pdev, | |||
738 | dev->tx_timeout = tc35815_tx_timeout; | 742 | dev->tx_timeout = tc35815_tx_timeout; |
739 | dev->watchdog_timeo = TC35815_TX_TIMEOUT; | 743 | dev->watchdog_timeo = TC35815_TX_TIMEOUT; |
740 | #ifdef TC35815_NAPI | 744 | #ifdef TC35815_NAPI |
741 | dev->poll = tc35815_poll; | 745 | netif_napi_add(dev, &lp->napi, tc35815_poll, NAPI_WEIGHT); |
742 | dev->weight = NAPI_WEIGHT; | ||
743 | #endif | 746 | #endif |
744 | #ifdef CONFIG_NET_POLL_CONTROLLER | 747 | #ifdef CONFIG_NET_POLL_CONTROLLER |
745 | dev->poll_controller = tc35815_poll_controller; | 748 | dev->poll_controller = tc35815_poll_controller; |
@@ -748,8 +751,6 @@ static int __devinit tc35815_init_one (struct pci_dev *pdev, | |||
748 | dev->irq = pdev->irq; | 751 | dev->irq = pdev->irq; |
749 | dev->base_addr = (unsigned long) ioaddr; | 752 | dev->base_addr = (unsigned long) ioaddr; |
750 | 753 | ||
751 | /* dev->priv/lp zeroed and aligned in alloc_etherdev */ | ||
752 | lp = dev->priv; | ||
753 | spin_lock_init(&lp->lock); | 754 | spin_lock_init(&lp->lock); |
754 | lp->pci_dev = pdev; | 755 | lp->pci_dev = pdev; |
755 | lp->boardtype = ent->driver_data; | 756 | lp->boardtype = ent->driver_data; |
@@ -1237,6 +1238,10 @@ tc35815_open(struct net_device *dev) | |||
1237 | return -EAGAIN; | 1238 | return -EAGAIN; |
1238 | } | 1239 | } |
1239 | 1240 | ||
1241 | #ifdef TC35815_NAPI | ||
1242 | napi_enable(&lp->napi); | ||
1243 | #endif | ||
1244 | |||
1240 | /* Reset the hardware here. Don't forget to set the station address. */ | 1245 | /* Reset the hardware here. Don't forget to set the station address. */ |
1241 | spin_lock_irq(&lp->lock); | 1246 | spin_lock_irq(&lp->lock); |
1242 | tc35815_chip_init(dev); | 1247 | tc35815_chip_init(dev); |
@@ -1436,6 +1441,7 @@ static int tc35815_do_interrupt(struct net_device *dev, u32 status) | |||
1436 | static irqreturn_t tc35815_interrupt(int irq, void *dev_id) | 1441 | static irqreturn_t tc35815_interrupt(int irq, void *dev_id) |
1437 | { | 1442 | { |
1438 | struct net_device *dev = dev_id; | 1443 | struct net_device *dev = dev_id; |
1444 | struct tc35815_local *lp = netdev_priv(dev); | ||
1439 | struct tc35815_regs __iomem *tr = | 1445 | struct tc35815_regs __iomem *tr = |
1440 | (struct tc35815_regs __iomem *)dev->base_addr; | 1446 | (struct tc35815_regs __iomem *)dev->base_addr; |
1441 | #ifdef TC35815_NAPI | 1447 | #ifdef TC35815_NAPI |
@@ -1444,8 +1450,8 @@ static irqreturn_t tc35815_interrupt(int irq, void *dev_id) | |||
1444 | if (!(dmactl & DMA_IntMask)) { | 1450 | if (!(dmactl & DMA_IntMask)) { |
1445 | /* disable interrupts */ | 1451 | /* disable interrupts */ |
1446 | tc_writel(dmactl | DMA_IntMask, &tr->DMA_Ctl); | 1452 | tc_writel(dmactl | DMA_IntMask, &tr->DMA_Ctl); |
1447 | if (netif_rx_schedule_prep(dev)) | 1453 | if (netif_rx_schedule_prep(dev, &lp->napi)) |
1448 | __netif_rx_schedule(dev); | 1454 | __netif_rx_schedule(dev, &lp->napi); |
1449 | else { | 1455 | else { |
1450 | printk(KERN_ERR "%s: interrupt taken in poll\n", | 1456 | printk(KERN_ERR "%s: interrupt taken in poll\n", |
1451 | dev->name); | 1457 | dev->name); |
@@ -1726,13 +1732,12 @@ tc35815_rx(struct net_device *dev) | |||
1726 | } | 1732 | } |
1727 | 1733 | ||
1728 | #ifdef TC35815_NAPI | 1734 | #ifdef TC35815_NAPI |
1729 | static int | 1735 | static int tc35815_poll(struct napi_struct *napi, int budget) |
1730 | tc35815_poll(struct net_device *dev, int *budget) | ||
1731 | { | 1736 | { |
1732 | struct tc35815_local *lp = dev->priv; | 1737 | struct tc35815_local *lp = container_of(napi, struct tc35815_local, napi); |
1738 | struct net_device *dev = lp->dev; | ||
1733 | struct tc35815_regs __iomem *tr = | 1739 | struct tc35815_regs __iomem *tr = |
1734 | (struct tc35815_regs __iomem *)dev->base_addr; | 1740 | (struct tc35815_regs __iomem *)dev->base_addr; |
1735 | int limit = min(*budget, dev->quota); | ||
1736 | int received = 0, handled; | 1741 | int received = 0, handled; |
1737 | u32 status; | 1742 | u32 status; |
1738 | 1743 | ||
@@ -1744,23 +1749,19 @@ tc35815_poll(struct net_device *dev, int *budget) | |||
1744 | handled = tc35815_do_interrupt(dev, status, limit); | 1749 | handled = tc35815_do_interrupt(dev, status, limit); |
1745 | if (handled >= 0) { | 1750 | if (handled >= 0) { |
1746 | received += handled; | 1751 | received += handled; |
1747 | limit -= handled; | 1752 | if (received >= budget) |
1748 | if (limit <= 0) | ||
1749 | break; | 1753 | break; |
1750 | } | 1754 | } |
1751 | status = tc_readl(&tr->Int_Src); | 1755 | status = tc_readl(&tr->Int_Src); |
1752 | } while (status); | 1756 | } while (status); |
1753 | spin_unlock(&lp->lock); | 1757 | spin_unlock(&lp->lock); |
1754 | 1758 | ||
1755 | dev->quota -= received; | 1759 | if (received < budget) { |
1756 | *budget -= received; | 1760 | netif_rx_complete(dev, napi); |
1757 | if (limit <= 0) | 1761 | /* enable interrupts */ |
1758 | return 1; | 1762 | tc_writel(tc_readl(&tr->DMA_Ctl) & ~DMA_IntMask, &tr->DMA_Ctl); |
1759 | 1763 | } | |
1760 | netif_rx_complete(dev); | 1764 | return received; |
1761 | /* enable interrupts */ | ||
1762 | tc_writel(tc_readl(&tr->DMA_Ctl) & ~DMA_IntMask, &tr->DMA_Ctl); | ||
1763 | return 0; | ||
1764 | } | 1765 | } |
1765 | #endif | 1766 | #endif |
1766 | 1767 | ||
@@ -1949,7 +1950,11 @@ static int | |||
1949 | tc35815_close(struct net_device *dev) | 1950 | tc35815_close(struct net_device *dev) |
1950 | { | 1951 | { |
1951 | struct tc35815_local *lp = dev->priv; | 1952 | struct tc35815_local *lp = dev->priv; |
1953 | |||
1952 | netif_stop_queue(dev); | 1954 | netif_stop_queue(dev); |
1955 | #ifdef TC35815_NAPI | ||
1956 | napi_disable(&lp->napi); | ||
1957 | #endif | ||
1953 | 1958 | ||
1954 | /* Flush the Tx and disable Rx here. */ | 1959 | /* Flush the Tx and disable Rx here. */ |
1955 | 1960 | ||
diff --git a/drivers/net/tg3.c b/drivers/net/tg3.c index 9034a05734ef..ef1e3d1173c4 100644 --- a/drivers/net/tg3.c +++ b/drivers/net/tg3.c | |||
@@ -574,7 +574,7 @@ static void tg3_restart_ints(struct tg3 *tp) | |||
574 | static inline void tg3_netif_stop(struct tg3 *tp) | 574 | static inline void tg3_netif_stop(struct tg3 *tp) |
575 | { | 575 | { |
576 | tp->dev->trans_start = jiffies; /* prevent tx timeout */ | 576 | tp->dev->trans_start = jiffies; /* prevent tx timeout */ |
577 | netif_poll_disable(tp->dev); | 577 | napi_disable(&tp->napi); |
578 | netif_tx_disable(tp->dev); | 578 | netif_tx_disable(tp->dev); |
579 | } | 579 | } |
580 | 580 | ||
@@ -585,7 +585,7 @@ static inline void tg3_netif_start(struct tg3 *tp) | |||
585 | * so long as all callers are assured to have free tx slots | 585 | * so long as all callers are assured to have free tx slots |
586 | * (such as after tg3_init_hw) | 586 | * (such as after tg3_init_hw) |
587 | */ | 587 | */ |
588 | netif_poll_enable(tp->dev); | 588 | napi_enable(&tp->napi); |
589 | tp->hw_status->status |= SD_STATUS_UPDATED; | 589 | tp->hw_status->status |= SD_STATUS_UPDATED; |
590 | tg3_enable_ints(tp); | 590 | tg3_enable_ints(tp); |
591 | } | 591 | } |
@@ -3471,11 +3471,12 @@ next_pkt_nopost: | |||
3471 | return received; | 3471 | return received; |
3472 | } | 3472 | } |
3473 | 3473 | ||
3474 | static int tg3_poll(struct net_device *netdev, int *budget) | 3474 | static int tg3_poll(struct napi_struct *napi, int budget) |
3475 | { | 3475 | { |
3476 | struct tg3 *tp = netdev_priv(netdev); | 3476 | struct tg3 *tp = container_of(napi, struct tg3, napi); |
3477 | struct net_device *netdev = tp->dev; | ||
3477 | struct tg3_hw_status *sblk = tp->hw_status; | 3478 | struct tg3_hw_status *sblk = tp->hw_status; |
3478 | int done; | 3479 | int work_done = 0; |
3479 | 3480 | ||
3480 | /* handle link change and other phy events */ | 3481 | /* handle link change and other phy events */ |
3481 | if (!(tp->tg3_flags & | 3482 | if (!(tp->tg3_flags & |
@@ -3494,7 +3495,7 @@ static int tg3_poll(struct net_device *netdev, int *budget) | |||
3494 | if (sblk->idx[0].tx_consumer != tp->tx_cons) { | 3495 | if (sblk->idx[0].tx_consumer != tp->tx_cons) { |
3495 | tg3_tx(tp); | 3496 | tg3_tx(tp); |
3496 | if (unlikely(tp->tg3_flags & TG3_FLAG_TX_RECOVERY_PENDING)) { | 3497 | if (unlikely(tp->tg3_flags & TG3_FLAG_TX_RECOVERY_PENDING)) { |
3497 | netif_rx_complete(netdev); | 3498 | netif_rx_complete(netdev, napi); |
3498 | schedule_work(&tp->reset_task); | 3499 | schedule_work(&tp->reset_task); |
3499 | return 0; | 3500 | return 0; |
3500 | } | 3501 | } |
@@ -3502,20 +3503,10 @@ static int tg3_poll(struct net_device *netdev, int *budget) | |||
3502 | 3503 | ||
3503 | /* run RX thread, within the bounds set by NAPI. | 3504 | /* run RX thread, within the bounds set by NAPI. |
3504 | * All RX "locking" is done by ensuring outside | 3505 | * All RX "locking" is done by ensuring outside |
3505 | * code synchronizes with dev->poll() | 3506 | * code synchronizes with tg3->napi.poll() |
3506 | */ | 3507 | */ |
3507 | if (sblk->idx[0].rx_producer != tp->rx_rcb_ptr) { | 3508 | if (sblk->idx[0].rx_producer != tp->rx_rcb_ptr) |
3508 | int orig_budget = *budget; | 3509 | work_done = tg3_rx(tp, budget); |
3509 | int work_done; | ||
3510 | |||
3511 | if (orig_budget > netdev->quota) | ||
3512 | orig_budget = netdev->quota; | ||
3513 | |||
3514 | work_done = tg3_rx(tp, orig_budget); | ||
3515 | |||
3516 | *budget -= work_done; | ||
3517 | netdev->quota -= work_done; | ||
3518 | } | ||
3519 | 3510 | ||
3520 | if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) { | 3511 | if (tp->tg3_flags & TG3_FLAG_TAGGED_STATUS) { |
3521 | tp->last_tag = sblk->status_tag; | 3512 | tp->last_tag = sblk->status_tag; |
@@ -3524,13 +3515,12 @@ static int tg3_poll(struct net_device *netdev, int *budget) | |||
3524 | sblk->status &= ~SD_STATUS_UPDATED; | 3515 | sblk->status &= ~SD_STATUS_UPDATED; |
3525 | 3516 | ||
3526 | /* if no more work, tell net stack and NIC we're done */ | 3517 | /* if no more work, tell net stack and NIC we're done */ |
3527 | done = !tg3_has_work(tp); | 3518 | if (!tg3_has_work(tp)) { |
3528 | if (done) { | 3519 | netif_rx_complete(netdev, napi); |
3529 | netif_rx_complete(netdev); | ||
3530 | tg3_restart_ints(tp); | 3520 | tg3_restart_ints(tp); |
3531 | } | 3521 | } |
3532 | 3522 | ||
3533 | return (done ? 0 : 1); | 3523 | return work_done; |
3534 | } | 3524 | } |
3535 | 3525 | ||
3536 | static void tg3_irq_quiesce(struct tg3 *tp) | 3526 | static void tg3_irq_quiesce(struct tg3 *tp) |
@@ -3577,7 +3567,7 @@ static irqreturn_t tg3_msi_1shot(int irq, void *dev_id) | |||
3577 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); | 3567 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); |
3578 | 3568 | ||
3579 | if (likely(!tg3_irq_sync(tp))) | 3569 | if (likely(!tg3_irq_sync(tp))) |
3580 | netif_rx_schedule(dev); /* schedule NAPI poll */ | 3570 | netif_rx_schedule(dev, &tp->napi); |
3581 | 3571 | ||
3582 | return IRQ_HANDLED; | 3572 | return IRQ_HANDLED; |
3583 | } | 3573 | } |
@@ -3602,7 +3592,7 @@ static irqreturn_t tg3_msi(int irq, void *dev_id) | |||
3602 | */ | 3592 | */ |
3603 | tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001); | 3593 | tw32_mailbox(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001); |
3604 | if (likely(!tg3_irq_sync(tp))) | 3594 | if (likely(!tg3_irq_sync(tp))) |
3605 | netif_rx_schedule(dev); /* schedule NAPI poll */ | 3595 | netif_rx_schedule(dev, &tp->napi); |
3606 | 3596 | ||
3607 | return IRQ_RETVAL(1); | 3597 | return IRQ_RETVAL(1); |
3608 | } | 3598 | } |
@@ -3644,7 +3634,7 @@ static irqreturn_t tg3_interrupt(int irq, void *dev_id) | |||
3644 | sblk->status &= ~SD_STATUS_UPDATED; | 3634 | sblk->status &= ~SD_STATUS_UPDATED; |
3645 | if (likely(tg3_has_work(tp))) { | 3635 | if (likely(tg3_has_work(tp))) { |
3646 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); | 3636 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); |
3647 | netif_rx_schedule(dev); /* schedule NAPI poll */ | 3637 | netif_rx_schedule(dev, &tp->napi); |
3648 | } else { | 3638 | } else { |
3649 | /* No work, shared interrupt perhaps? re-enable | 3639 | /* No work, shared interrupt perhaps? re-enable |
3650 | * interrupts, and flush that PCI write | 3640 | * interrupts, and flush that PCI write |
@@ -3690,7 +3680,7 @@ static irqreturn_t tg3_interrupt_tagged(int irq, void *dev_id) | |||
3690 | tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001); | 3680 | tw32_mailbox_f(MAILBOX_INTERRUPT_0 + TG3_64BIT_REG_LOW, 0x00000001); |
3691 | if (tg3_irq_sync(tp)) | 3681 | if (tg3_irq_sync(tp)) |
3692 | goto out; | 3682 | goto out; |
3693 | if (netif_rx_schedule_prep(dev)) { | 3683 | if (netif_rx_schedule_prep(dev, &tp->napi)) { |
3694 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); | 3684 | prefetch(&tp->rx_rcb[tp->rx_rcb_ptr]); |
3695 | /* Update last_tag to mark that this status has been | 3685 | /* Update last_tag to mark that this status has been |
3696 | * seen. Because interrupt may be shared, we may be | 3686 | * seen. Because interrupt may be shared, we may be |
@@ -3698,7 +3688,7 @@ static irqreturn_t tg3_interrupt_tagged(int irq, void *dev_id) | |||
3698 | * if tg3_poll() is not scheduled. | 3688 | * if tg3_poll() is not scheduled. |
3699 | */ | 3689 | */ |
3700 | tp->last_tag = sblk->status_tag; | 3690 | tp->last_tag = sblk->status_tag; |
3701 | __netif_rx_schedule(dev); | 3691 | __netif_rx_schedule(dev, &tp->napi); |
3702 | } | 3692 | } |
3703 | out: | 3693 | out: |
3704 | return IRQ_RETVAL(handled); | 3694 | return IRQ_RETVAL(handled); |
@@ -3737,7 +3727,7 @@ static int tg3_restart_hw(struct tg3 *tp, int reset_phy) | |||
3737 | tg3_full_unlock(tp); | 3727 | tg3_full_unlock(tp); |
3738 | del_timer_sync(&tp->timer); | 3728 | del_timer_sync(&tp->timer); |
3739 | tp->irq_sync = 0; | 3729 | tp->irq_sync = 0; |
3740 | netif_poll_enable(tp->dev); | 3730 | napi_enable(&tp->napi); |
3741 | dev_close(tp->dev); | 3731 | dev_close(tp->dev); |
3742 | tg3_full_lock(tp, 0); | 3732 | tg3_full_lock(tp, 0); |
3743 | } | 3733 | } |
@@ -3932,7 +3922,7 @@ static int tg3_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
3932 | len = skb_headlen(skb); | 3922 | len = skb_headlen(skb); |
3933 | 3923 | ||
3934 | /* We are running in BH disabled context with netif_tx_lock | 3924 | /* We are running in BH disabled context with netif_tx_lock |
3935 | * and TX reclaim runs via tp->poll inside of a software | 3925 | * and TX reclaim runs via tp->napi.poll inside of a software |
3936 | * interrupt. Furthermore, IRQ processing runs lockless so we have | 3926 | * interrupt. Furthermore, IRQ processing runs lockless so we have |
3937 | * no IRQ context deadlocks to worry about either. Rejoice! | 3927 | * no IRQ context deadlocks to worry about either. Rejoice! |
3938 | */ | 3928 | */ |
@@ -4087,7 +4077,7 @@ static int tg3_start_xmit_dma_bug(struct sk_buff *skb, struct net_device *dev) | |||
4087 | len = skb_headlen(skb); | 4077 | len = skb_headlen(skb); |
4088 | 4078 | ||
4089 | /* We are running in BH disabled context with netif_tx_lock | 4079 | /* We are running in BH disabled context with netif_tx_lock |
4090 | * and TX reclaim runs via tp->poll inside of a software | 4080 | * and TX reclaim runs via tp->napi.poll inside of a software |
4091 | * interrupt. Furthermore, IRQ processing runs lockless so we have | 4081 | * interrupt. Furthermore, IRQ processing runs lockless so we have |
4092 | * no IRQ context deadlocks to worry about either. Rejoice! | 4082 | * no IRQ context deadlocks to worry about either. Rejoice! |
4093 | */ | 4083 | */ |
@@ -7147,6 +7137,8 @@ static int tg3_open(struct net_device *dev) | |||
7147 | return err; | 7137 | return err; |
7148 | } | 7138 | } |
7149 | 7139 | ||
7140 | napi_enable(&tp->napi); | ||
7141 | |||
7150 | tg3_full_lock(tp, 0); | 7142 | tg3_full_lock(tp, 0); |
7151 | 7143 | ||
7152 | err = tg3_init_hw(tp, 1); | 7144 | err = tg3_init_hw(tp, 1); |
@@ -7174,6 +7166,7 @@ static int tg3_open(struct net_device *dev) | |||
7174 | tg3_full_unlock(tp); | 7166 | tg3_full_unlock(tp); |
7175 | 7167 | ||
7176 | if (err) { | 7168 | if (err) { |
7169 | napi_disable(&tp->napi); | ||
7177 | free_irq(tp->pdev->irq, dev); | 7170 | free_irq(tp->pdev->irq, dev); |
7178 | if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) { | 7171 | if (tp->tg3_flags2 & TG3_FLG2_USING_MSI) { |
7179 | pci_disable_msi(tp->pdev); | 7172 | pci_disable_msi(tp->pdev); |
@@ -7199,6 +7192,8 @@ static int tg3_open(struct net_device *dev) | |||
7199 | 7192 | ||
7200 | tg3_full_unlock(tp); | 7193 | tg3_full_unlock(tp); |
7201 | 7194 | ||
7195 | napi_disable(&tp->napi); | ||
7196 | |||
7202 | return err; | 7197 | return err; |
7203 | } | 7198 | } |
7204 | 7199 | ||
@@ -7460,6 +7455,7 @@ static int tg3_close(struct net_device *dev) | |||
7460 | { | 7455 | { |
7461 | struct tg3 *tp = netdev_priv(dev); | 7456 | struct tg3 *tp = netdev_priv(dev); |
7462 | 7457 | ||
7458 | napi_disable(&tp->napi); | ||
7463 | cancel_work_sync(&tp->reset_task); | 7459 | cancel_work_sync(&tp->reset_task); |
7464 | 7460 | ||
7465 | netif_stop_queue(dev); | 7461 | netif_stop_queue(dev); |
@@ -11900,9 +11896,8 @@ static int __devinit tg3_init_one(struct pci_dev *pdev, | |||
11900 | dev->set_mac_address = tg3_set_mac_addr; | 11896 | dev->set_mac_address = tg3_set_mac_addr; |
11901 | dev->do_ioctl = tg3_ioctl; | 11897 | dev->do_ioctl = tg3_ioctl; |
11902 | dev->tx_timeout = tg3_tx_timeout; | 11898 | dev->tx_timeout = tg3_tx_timeout; |
11903 | dev->poll = tg3_poll; | 11899 | netif_napi_add(dev, &tp->napi, tg3_poll, 64); |
11904 | dev->ethtool_ops = &tg3_ethtool_ops; | 11900 | dev->ethtool_ops = &tg3_ethtool_ops; |
11905 | dev->weight = 64; | ||
11906 | dev->watchdog_timeo = TG3_TX_TIMEOUT; | 11901 | dev->watchdog_timeo = TG3_TX_TIMEOUT; |
11907 | dev->change_mtu = tg3_change_mtu; | 11902 | dev->change_mtu = tg3_change_mtu; |
11908 | dev->irq = pdev->irq; | 11903 | dev->irq = pdev->irq; |
diff --git a/drivers/net/tg3.h b/drivers/net/tg3.h index 5c21f49026c9..a6a23bbcdfee 100644 --- a/drivers/net/tg3.h +++ b/drivers/net/tg3.h | |||
@@ -2176,6 +2176,7 @@ struct tg3 { | |||
2176 | dma_addr_t tx_desc_mapping; | 2176 | dma_addr_t tx_desc_mapping; |
2177 | 2177 | ||
2178 | /* begin "rx thread" cacheline section */ | 2178 | /* begin "rx thread" cacheline section */ |
2179 | struct napi_struct napi; | ||
2179 | void (*write32_rx_mbox) (struct tg3 *, u32, | 2180 | void (*write32_rx_mbox) (struct tg3 *, u32, |
2180 | u32); | 2181 | u32); |
2181 | u32 rx_rcb_ptr; | 2182 | u32 rx_rcb_ptr; |
diff --git a/drivers/net/tsi108_eth.c b/drivers/net/tsi108_eth.c index 1aabc91f6458..b3069ee34bd2 100644 --- a/drivers/net/tsi108_eth.c +++ b/drivers/net/tsi108_eth.c | |||
@@ -79,6 +79,9 @@ struct tsi108_prv_data { | |||
79 | void __iomem *regs; /* Base of normal regs */ | 79 | void __iomem *regs; /* Base of normal regs */ |
80 | void __iomem *phyregs; /* Base of register bank used for PHY access */ | 80 | void __iomem *phyregs; /* Base of register bank used for PHY access */ |
81 | 81 | ||
82 | struct net_device *dev; | ||
83 | struct napi_struct napi; | ||
84 | |||
82 | unsigned int phy; /* Index of PHY for this interface */ | 85 | unsigned int phy; /* Index of PHY for this interface */ |
83 | unsigned int irq_num; | 86 | unsigned int irq_num; |
84 | unsigned int id; | 87 | unsigned int id; |
@@ -837,13 +840,13 @@ static int tsi108_refill_rx(struct net_device *dev, int budget) | |||
837 | return done; | 840 | return done; |
838 | } | 841 | } |
839 | 842 | ||
840 | static int tsi108_poll(struct net_device *dev, int *budget) | 843 | static int tsi108_poll(struct napi_struct *napi, int budget) |
841 | { | 844 | { |
842 | struct tsi108_prv_data *data = netdev_priv(dev); | 845 | struct tsi108_prv_data *data = container_of(napi, struct tsi108_prv_data, napi); |
846 | struct net_device *dev = data->dev; | ||
843 | u32 estat = TSI_READ(TSI108_EC_RXESTAT); | 847 | u32 estat = TSI_READ(TSI108_EC_RXESTAT); |
844 | u32 intstat = TSI_READ(TSI108_EC_INTSTAT); | 848 | u32 intstat = TSI_READ(TSI108_EC_INTSTAT); |
845 | int total_budget = min(*budget, dev->quota); | 849 | int num_received = 0, num_filled = 0; |
846 | int num_received = 0, num_filled = 0, budget_used; | ||
847 | 850 | ||
848 | intstat &= TSI108_INT_RXQUEUE0 | TSI108_INT_RXTHRESH | | 851 | intstat &= TSI108_INT_RXQUEUE0 | TSI108_INT_RXTHRESH | |
849 | TSI108_INT_RXOVERRUN | TSI108_INT_RXERROR | TSI108_INT_RXWAIT; | 852 | TSI108_INT_RXOVERRUN | TSI108_INT_RXERROR | TSI108_INT_RXWAIT; |
@@ -852,7 +855,7 @@ static int tsi108_poll(struct net_device *dev, int *budget) | |||
852 | TSI_WRITE(TSI108_EC_INTSTAT, intstat); | 855 | TSI_WRITE(TSI108_EC_INTSTAT, intstat); |
853 | 856 | ||
854 | if (data->rxpending || (estat & TSI108_EC_RXESTAT_Q0_DESCINT)) | 857 | if (data->rxpending || (estat & TSI108_EC_RXESTAT_Q0_DESCINT)) |
855 | num_received = tsi108_complete_rx(dev, total_budget); | 858 | num_received = tsi108_complete_rx(dev, budget); |
856 | 859 | ||
857 | /* This should normally fill no more slots than the number of | 860 | /* This should normally fill no more slots than the number of |
858 | * packets received in tsi108_complete_rx(). The exception | 861 | * packets received in tsi108_complete_rx(). The exception |
@@ -867,7 +870,7 @@ static int tsi108_poll(struct net_device *dev, int *budget) | |||
867 | */ | 870 | */ |
868 | 871 | ||
869 | if (data->rxfree < TSI108_RXRING_LEN) | 872 | if (data->rxfree < TSI108_RXRING_LEN) |
870 | num_filled = tsi108_refill_rx(dev, total_budget * 2); | 873 | num_filled = tsi108_refill_rx(dev, budget * 2); |
871 | 874 | ||
872 | if (intstat & TSI108_INT_RXERROR) { | 875 | if (intstat & TSI108_INT_RXERROR) { |
873 | u32 err = TSI_READ(TSI108_EC_RXERR); | 876 | u32 err = TSI_READ(TSI108_EC_RXERR); |
@@ -890,14 +893,9 @@ static int tsi108_poll(struct net_device *dev, int *budget) | |||
890 | spin_unlock_irq(&data->misclock); | 893 | spin_unlock_irq(&data->misclock); |
891 | } | 894 | } |
892 | 895 | ||
893 | budget_used = max(num_received, num_filled / 2); | 896 | if (num_received < budget) { |
894 | |||
895 | *budget -= budget_used; | ||
896 | dev->quota -= budget_used; | ||
897 | |||
898 | if (budget_used != total_budget) { | ||
899 | data->rxpending = 0; | 897 | data->rxpending = 0; |
900 | netif_rx_complete(dev); | 898 | netif_rx_complete(dev, napi); |
901 | 899 | ||
902 | TSI_WRITE(TSI108_EC_INTMASK, | 900 | TSI_WRITE(TSI108_EC_INTMASK, |
903 | TSI_READ(TSI108_EC_INTMASK) | 901 | TSI_READ(TSI108_EC_INTMASK) |
@@ -906,14 +904,11 @@ static int tsi108_poll(struct net_device *dev, int *budget) | |||
906 | TSI108_INT_RXOVERRUN | | 904 | TSI108_INT_RXOVERRUN | |
907 | TSI108_INT_RXERROR | | 905 | TSI108_INT_RXERROR | |
908 | TSI108_INT_RXWAIT)); | 906 | TSI108_INT_RXWAIT)); |
909 | |||
910 | /* IRQs are level-triggered, so no need to re-check */ | ||
911 | return 0; | ||
912 | } else { | 907 | } else { |
913 | data->rxpending = 1; | 908 | data->rxpending = 1; |
914 | } | 909 | } |
915 | 910 | ||
916 | return 1; | 911 | return num_received; |
917 | } | 912 | } |
918 | 913 | ||
919 | static void tsi108_rx_int(struct net_device *dev) | 914 | static void tsi108_rx_int(struct net_device *dev) |
@@ -931,7 +926,7 @@ static void tsi108_rx_int(struct net_device *dev) | |||
931 | * from tsi108_check_rxring(). | 926 | * from tsi108_check_rxring(). |
932 | */ | 927 | */ |
933 | 928 | ||
934 | if (netif_rx_schedule_prep(dev)) { | 929 | if (netif_rx_schedule_prep(dev, &data->napi)) { |
935 | /* Mask, rather than ack, the receive interrupts. The ack | 930 | /* Mask, rather than ack, the receive interrupts. The ack |
936 | * will happen in tsi108_poll(). | 931 | * will happen in tsi108_poll(). |
937 | */ | 932 | */ |
@@ -942,7 +937,7 @@ static void tsi108_rx_int(struct net_device *dev) | |||
942 | | TSI108_INT_RXTHRESH | | 937 | | TSI108_INT_RXTHRESH | |
943 | TSI108_INT_RXOVERRUN | TSI108_INT_RXERROR | | 938 | TSI108_INT_RXOVERRUN | TSI108_INT_RXERROR | |
944 | TSI108_INT_RXWAIT); | 939 | TSI108_INT_RXWAIT); |
945 | __netif_rx_schedule(dev); | 940 | __netif_rx_schedule(dev, &data->napi); |
946 | } else { | 941 | } else { |
947 | if (!netif_running(dev)) { | 942 | if (!netif_running(dev)) { |
948 | /* This can happen if an interrupt occurs while the | 943 | /* This can happen if an interrupt occurs while the |
@@ -1401,6 +1396,8 @@ static int tsi108_open(struct net_device *dev) | |||
1401 | TSI_WRITE(TSI108_EC_TXQ_PTRLOW, data->txdma); | 1396 | TSI_WRITE(TSI108_EC_TXQ_PTRLOW, data->txdma); |
1402 | tsi108_init_phy(dev); | 1397 | tsi108_init_phy(dev); |
1403 | 1398 | ||
1399 | napi_enable(&data->napi); | ||
1400 | |||
1404 | setup_timer(&data->timer, tsi108_timed_checker, (unsigned long)dev); | 1401 | setup_timer(&data->timer, tsi108_timed_checker, (unsigned long)dev); |
1405 | mod_timer(&data->timer, jiffies + 1); | 1402 | mod_timer(&data->timer, jiffies + 1); |
1406 | 1403 | ||
@@ -1425,6 +1422,7 @@ static int tsi108_close(struct net_device *dev) | |||
1425 | struct tsi108_prv_data *data = netdev_priv(dev); | 1422 | struct tsi108_prv_data *data = netdev_priv(dev); |
1426 | 1423 | ||
1427 | netif_stop_queue(dev); | 1424 | netif_stop_queue(dev); |
1425 | napi_disable(&data->napi); | ||
1428 | 1426 | ||
1429 | del_timer_sync(&data->timer); | 1427 | del_timer_sync(&data->timer); |
1430 | 1428 | ||
@@ -1562,6 +1560,7 @@ tsi108_init_one(struct platform_device *pdev) | |||
1562 | 1560 | ||
1563 | printk("tsi108_eth%d: probe...\n", pdev->id); | 1561 | printk("tsi108_eth%d: probe...\n", pdev->id); |
1564 | data = netdev_priv(dev); | 1562 | data = netdev_priv(dev); |
1563 | data->dev = dev; | ||
1565 | 1564 | ||
1566 | pr_debug("tsi108_eth%d:regs:phyresgs:phy:irq_num=0x%x:0x%x:0x%x:0x%x\n", | 1565 | pr_debug("tsi108_eth%d:regs:phyresgs:phy:irq_num=0x%x:0x%x:0x%x:0x%x\n", |
1567 | pdev->id, einfo->regs, einfo->phyregs, | 1566 | pdev->id, einfo->regs, einfo->phyregs, |
@@ -1597,9 +1596,8 @@ tsi108_init_one(struct platform_device *pdev) | |||
1597 | dev->set_mac_address = tsi108_set_mac; | 1596 | dev->set_mac_address = tsi108_set_mac; |
1598 | dev->set_multicast_list = tsi108_set_rx_mode; | 1597 | dev->set_multicast_list = tsi108_set_rx_mode; |
1599 | dev->get_stats = tsi108_get_stats; | 1598 | dev->get_stats = tsi108_get_stats; |
1600 | dev->poll = tsi108_poll; | 1599 | netif_napi_add(dev, &data->napi, tsi108_poll, 64); |
1601 | dev->do_ioctl = tsi108_do_ioctl; | 1600 | dev->do_ioctl = tsi108_do_ioctl; |
1602 | dev->weight = 64; /* 64 is more suitable for GigE interface - klai */ | ||
1603 | 1601 | ||
1604 | /* Apparently, the Linux networking code won't use scatter-gather | 1602 | /* Apparently, the Linux networking code won't use scatter-gather |
1605 | * if the hardware doesn't do checksums. However, it's faster | 1603 | * if the hardware doesn't do checksums. However, it's faster |
diff --git a/drivers/net/tulip/interrupt.c b/drivers/net/tulip/interrupt.c index 53efd6694e75..365331446387 100644 --- a/drivers/net/tulip/interrupt.c +++ b/drivers/net/tulip/interrupt.c | |||
@@ -103,28 +103,29 @@ int tulip_refill_rx(struct net_device *dev) | |||
103 | void oom_timer(unsigned long data) | 103 | void oom_timer(unsigned long data) |
104 | { | 104 | { |
105 | struct net_device *dev = (struct net_device *)data; | 105 | struct net_device *dev = (struct net_device *)data; |
106 | netif_rx_schedule(dev); | 106 | struct tulip_private *tp = netdev_priv(dev); |
107 | netif_rx_schedule(dev, &tp->napi); | ||
107 | } | 108 | } |
108 | 109 | ||
109 | int tulip_poll(struct net_device *dev, int *budget) | 110 | int tulip_poll(struct napi_struct *napi, int budget) |
110 | { | 111 | { |
111 | struct tulip_private *tp = netdev_priv(dev); | 112 | struct tulip_private *tp = container_of(napi, struct tulip_private, napi); |
113 | struct net_device *dev = tp->dev; | ||
112 | int entry = tp->cur_rx % RX_RING_SIZE; | 114 | int entry = tp->cur_rx % RX_RING_SIZE; |
113 | int rx_work_limit = *budget; | 115 | int work_done = 0; |
116 | #ifdef CONFIG_TULIP_NAPI_HW_MITIGATION | ||
114 | int received = 0; | 117 | int received = 0; |
118 | #endif | ||
115 | 119 | ||
116 | if (!netif_running(dev)) | 120 | if (!netif_running(dev)) |
117 | goto done; | 121 | goto done; |
118 | 122 | ||
119 | if (rx_work_limit > dev->quota) | ||
120 | rx_work_limit = dev->quota; | ||
121 | |||
122 | #ifdef CONFIG_TULIP_NAPI_HW_MITIGATION | 123 | #ifdef CONFIG_TULIP_NAPI_HW_MITIGATION |
123 | 124 | ||
124 | /* that one buffer is needed for mit activation; or might be a | 125 | /* that one buffer is needed for mit activation; or might be a |
125 | bug in the ring buffer code; check later -- JHS*/ | 126 | bug in the ring buffer code; check later -- JHS*/ |
126 | 127 | ||
127 | if (rx_work_limit >=RX_RING_SIZE) rx_work_limit--; | 128 | if (budget >=RX_RING_SIZE) budget--; |
128 | #endif | 129 | #endif |
129 | 130 | ||
130 | if (tulip_debug > 4) | 131 | if (tulip_debug > 4) |
@@ -144,14 +145,13 @@ int tulip_poll(struct net_device *dev, int *budget) | |||
144 | while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) { | 145 | while ( ! (tp->rx_ring[entry].status & cpu_to_le32(DescOwned))) { |
145 | s32 status = le32_to_cpu(tp->rx_ring[entry].status); | 146 | s32 status = le32_to_cpu(tp->rx_ring[entry].status); |
146 | 147 | ||
147 | |||
148 | if (tp->dirty_rx + RX_RING_SIZE == tp->cur_rx) | 148 | if (tp->dirty_rx + RX_RING_SIZE == tp->cur_rx) |
149 | break; | 149 | break; |
150 | 150 | ||
151 | if (tulip_debug > 5) | 151 | if (tulip_debug > 5) |
152 | printk(KERN_DEBUG "%s: In tulip_rx(), entry %d %8.8x.\n", | 152 | printk(KERN_DEBUG "%s: In tulip_rx(), entry %d %8.8x.\n", |
153 | dev->name, entry, status); | 153 | dev->name, entry, status); |
154 | if (--rx_work_limit < 0) | 154 | if (work_done++ >= budget) |
155 | goto not_done; | 155 | goto not_done; |
156 | 156 | ||
157 | if ((status & 0x38008300) != 0x0300) { | 157 | if ((status & 0x38008300) != 0x0300) { |
@@ -238,7 +238,9 @@ int tulip_poll(struct net_device *dev, int *budget) | |||
238 | tp->stats.rx_packets++; | 238 | tp->stats.rx_packets++; |
239 | tp->stats.rx_bytes += pkt_len; | 239 | tp->stats.rx_bytes += pkt_len; |
240 | } | 240 | } |
241 | received++; | 241 | #ifdef CONFIG_TULIP_NAPI_HW_MITIGATION |
242 | received++; | ||
243 | #endif | ||
242 | 244 | ||
243 | entry = (++tp->cur_rx) % RX_RING_SIZE; | 245 | entry = (++tp->cur_rx) % RX_RING_SIZE; |
244 | if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/4) | 246 | if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/4) |
@@ -296,17 +298,15 @@ done: | |||
296 | 298 | ||
297 | #endif /* CONFIG_TULIP_NAPI_HW_MITIGATION */ | 299 | #endif /* CONFIG_TULIP_NAPI_HW_MITIGATION */ |
298 | 300 | ||
299 | dev->quota -= received; | ||
300 | *budget -= received; | ||
301 | |||
302 | tulip_refill_rx(dev); | 301 | tulip_refill_rx(dev); |
303 | 302 | ||
304 | /* If RX ring is not full we are out of memory. */ | 303 | /* If RX ring is not full we are out of memory. */ |
305 | if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) goto oom; | 304 | if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) |
305 | goto oom; | ||
306 | 306 | ||
307 | /* Remove us from polling list and enable RX intr. */ | 307 | /* Remove us from polling list and enable RX intr. */ |
308 | 308 | ||
309 | netif_rx_complete(dev); | 309 | netif_rx_complete(dev, napi); |
310 | iowrite32(tulip_tbl[tp->chip_id].valid_intrs, tp->base_addr+CSR7); | 310 | iowrite32(tulip_tbl[tp->chip_id].valid_intrs, tp->base_addr+CSR7); |
311 | 311 | ||
312 | /* The last op happens after poll completion. Which means the following: | 312 | /* The last op happens after poll completion. Which means the following: |
@@ -320,28 +320,20 @@ done: | |||
320 | * processed irqs. But it must not result in losing events. | 320 | * processed irqs. But it must not result in losing events. |
321 | */ | 321 | */ |
322 | 322 | ||
323 | return 0; | 323 | return work_done; |
324 | 324 | ||
325 | not_done: | 325 | not_done: |
326 | if (!received) { | ||
327 | |||
328 | received = dev->quota; /* Not to happen */ | ||
329 | } | ||
330 | dev->quota -= received; | ||
331 | *budget -= received; | ||
332 | |||
333 | if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/2 || | 326 | if (tp->cur_rx - tp->dirty_rx > RX_RING_SIZE/2 || |
334 | tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) | 327 | tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) |
335 | tulip_refill_rx(dev); | 328 | tulip_refill_rx(dev); |
336 | 329 | ||
337 | if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) goto oom; | 330 | if (tp->rx_buffers[tp->dirty_rx % RX_RING_SIZE].skb == NULL) |
338 | 331 | goto oom; | |
339 | return 1; | ||
340 | 332 | ||
333 | return work_done; | ||
341 | 334 | ||
342 | oom: /* Executed with RX ints disabled */ | 335 | oom: /* Executed with RX ints disabled */ |
343 | 336 | ||
344 | |||
345 | /* Start timer, stop polling, but do not enable rx interrupts. */ | 337 | /* Start timer, stop polling, but do not enable rx interrupts. */ |
346 | mod_timer(&tp->oom_timer, jiffies+1); | 338 | mod_timer(&tp->oom_timer, jiffies+1); |
347 | 339 | ||
@@ -350,9 +342,9 @@ done: | |||
350 | * before we did netif_rx_complete(). See? We would lose it. */ | 342 | * before we did netif_rx_complete(). See? We would lose it. */ |
351 | 343 | ||
352 | /* remove ourselves from the polling list */ | 344 | /* remove ourselves from the polling list */ |
353 | netif_rx_complete(dev); | 345 | netif_rx_complete(dev, napi); |
354 | 346 | ||
355 | return 0; | 347 | return work_done; |
356 | } | 348 | } |
357 | 349 | ||
358 | #else /* CONFIG_TULIP_NAPI */ | 350 | #else /* CONFIG_TULIP_NAPI */ |
@@ -534,7 +526,7 @@ irqreturn_t tulip_interrupt(int irq, void *dev_instance) | |||
534 | rxd++; | 526 | rxd++; |
535 | /* Mask RX intrs and add the device to poll list. */ | 527 | /* Mask RX intrs and add the device to poll list. */ |
536 | iowrite32(tulip_tbl[tp->chip_id].valid_intrs&~RxPollInt, ioaddr + CSR7); | 528 | iowrite32(tulip_tbl[tp->chip_id].valid_intrs&~RxPollInt, ioaddr + CSR7); |
537 | netif_rx_schedule(dev); | 529 | netif_rx_schedule(dev, &tp->napi); |
538 | 530 | ||
539 | if (!(csr5&~(AbnormalIntr|NormalIntr|RxPollInt|TPLnkPass))) | 531 | if (!(csr5&~(AbnormalIntr|NormalIntr|RxPollInt|TPLnkPass))) |
540 | break; | 532 | break; |
diff --git a/drivers/net/tulip/tulip.h b/drivers/net/tulip/tulip.h index 16f26a8364f0..5a4d7270973e 100644 --- a/drivers/net/tulip/tulip.h +++ b/drivers/net/tulip/tulip.h | |||
@@ -353,6 +353,7 @@ struct tulip_private { | |||
353 | int chip_id; | 353 | int chip_id; |
354 | int revision; | 354 | int revision; |
355 | int flags; | 355 | int flags; |
356 | struct napi_struct napi; | ||
356 | struct net_device_stats stats; | 357 | struct net_device_stats stats; |
357 | struct timer_list timer; /* Media selection timer. */ | 358 | struct timer_list timer; /* Media selection timer. */ |
358 | struct timer_list oom_timer; /* Out of memory timer. */ | 359 | struct timer_list oom_timer; /* Out of memory timer. */ |
@@ -429,7 +430,7 @@ extern int tulip_rx_copybreak; | |||
429 | irqreturn_t tulip_interrupt(int irq, void *dev_instance); | 430 | irqreturn_t tulip_interrupt(int irq, void *dev_instance); |
430 | int tulip_refill_rx(struct net_device *dev); | 431 | int tulip_refill_rx(struct net_device *dev); |
431 | #ifdef CONFIG_TULIP_NAPI | 432 | #ifdef CONFIG_TULIP_NAPI |
432 | int tulip_poll(struct net_device *dev, int *budget); | 433 | int tulip_poll(struct napi_struct *napi, int budget); |
433 | #endif | 434 | #endif |
434 | 435 | ||
435 | 436 | ||
diff --git a/drivers/net/tulip/tulip_core.c b/drivers/net/tulip/tulip_core.c index eca984f89bbf..7040a59fa3c9 100644 --- a/drivers/net/tulip/tulip_core.c +++ b/drivers/net/tulip/tulip_core.c | |||
@@ -294,6 +294,10 @@ static void tulip_up(struct net_device *dev) | |||
294 | int next_tick = 3*HZ; | 294 | int next_tick = 3*HZ; |
295 | int i; | 295 | int i; |
296 | 296 | ||
297 | #ifdef CONFIG_TULIP_NAPI | ||
298 | napi_enable(&tp->napi); | ||
299 | #endif | ||
300 | |||
297 | /* Wake the chip from sleep/snooze mode. */ | 301 | /* Wake the chip from sleep/snooze mode. */ |
298 | tulip_set_power_state (tp, 0, 0); | 302 | tulip_set_power_state (tp, 0, 0); |
299 | 303 | ||
@@ -728,6 +732,10 @@ static void tulip_down (struct net_device *dev) | |||
728 | 732 | ||
729 | flush_scheduled_work(); | 733 | flush_scheduled_work(); |
730 | 734 | ||
735 | #ifdef CONFIG_TULIP_NAPI | ||
736 | napi_disable(&tp->napi); | ||
737 | #endif | ||
738 | |||
731 | del_timer_sync (&tp->timer); | 739 | del_timer_sync (&tp->timer); |
732 | #ifdef CONFIG_TULIP_NAPI | 740 | #ifdef CONFIG_TULIP_NAPI |
733 | del_timer_sync (&tp->oom_timer); | 741 | del_timer_sync (&tp->oom_timer); |
@@ -1606,8 +1614,7 @@ static int __devinit tulip_init_one (struct pci_dev *pdev, | |||
1606 | dev->tx_timeout = tulip_tx_timeout; | 1614 | dev->tx_timeout = tulip_tx_timeout; |
1607 | dev->watchdog_timeo = TX_TIMEOUT; | 1615 | dev->watchdog_timeo = TX_TIMEOUT; |
1608 | #ifdef CONFIG_TULIP_NAPI | 1616 | #ifdef CONFIG_TULIP_NAPI |
1609 | dev->poll = tulip_poll; | 1617 | netif_napi_add(dev, &tp->napi, tulip_poll, 16); |
1610 | dev->weight = 16; | ||
1611 | #endif | 1618 | #endif |
1612 | dev->stop = tulip_close; | 1619 | dev->stop = tulip_close; |
1613 | dev->get_stats = tulip_get_stats; | 1620 | dev->get_stats = tulip_get_stats; |
diff --git a/drivers/net/typhoon.c b/drivers/net/typhoon.c index 03587205546e..0377b8b64c78 100644 --- a/drivers/net/typhoon.c +++ b/drivers/net/typhoon.c | |||
@@ -284,6 +284,7 @@ struct typhoon { | |||
284 | struct basic_ring rxLoRing; | 284 | struct basic_ring rxLoRing; |
285 | struct pci_dev * pdev; | 285 | struct pci_dev * pdev; |
286 | struct net_device * dev; | 286 | struct net_device * dev; |
287 | struct napi_struct napi; | ||
287 | spinlock_t state_lock; | 288 | spinlock_t state_lock; |
288 | struct vlan_group * vlgrp; | 289 | struct vlan_group * vlgrp; |
289 | struct basic_ring rxHiRing; | 290 | struct basic_ring rxHiRing; |
@@ -1759,12 +1760,12 @@ typhoon_fill_free_ring(struct typhoon *tp) | |||
1759 | } | 1760 | } |
1760 | 1761 | ||
1761 | static int | 1762 | static int |
1762 | typhoon_poll(struct net_device *dev, int *total_budget) | 1763 | typhoon_poll(struct napi_struct *napi, int budget) |
1763 | { | 1764 | { |
1764 | struct typhoon *tp = netdev_priv(dev); | 1765 | struct typhoon *tp = container_of(napi, struct typhoon, napi); |
1766 | struct net_device *dev = tp->dev; | ||
1765 | struct typhoon_indexes *indexes = tp->indexes; | 1767 | struct typhoon_indexes *indexes = tp->indexes; |
1766 | int orig_budget = *total_budget; | 1768 | int work_done; |
1767 | int budget, work_done, done; | ||
1768 | 1769 | ||
1769 | rmb(); | 1770 | rmb(); |
1770 | if(!tp->awaiting_resp && indexes->respReady != indexes->respCleared) | 1771 | if(!tp->awaiting_resp && indexes->respReady != indexes->respCleared) |
@@ -1773,30 +1774,16 @@ typhoon_poll(struct net_device *dev, int *total_budget) | |||
1773 | if(le32_to_cpu(indexes->txLoCleared) != tp->txLoRing.lastRead) | 1774 | if(le32_to_cpu(indexes->txLoCleared) != tp->txLoRing.lastRead) |
1774 | typhoon_tx_complete(tp, &tp->txLoRing, &indexes->txLoCleared); | 1775 | typhoon_tx_complete(tp, &tp->txLoRing, &indexes->txLoCleared); |
1775 | 1776 | ||
1776 | if(orig_budget > dev->quota) | ||
1777 | orig_budget = dev->quota; | ||
1778 | |||
1779 | budget = orig_budget; | ||
1780 | work_done = 0; | 1777 | work_done = 0; |
1781 | done = 1; | ||
1782 | 1778 | ||
1783 | if(indexes->rxHiCleared != indexes->rxHiReady) { | 1779 | if(indexes->rxHiCleared != indexes->rxHiReady) { |
1784 | work_done = typhoon_rx(tp, &tp->rxHiRing, &indexes->rxHiReady, | 1780 | work_done += typhoon_rx(tp, &tp->rxHiRing, &indexes->rxHiReady, |
1785 | &indexes->rxHiCleared, budget); | 1781 | &indexes->rxHiCleared, budget); |
1786 | budget -= work_done; | ||
1787 | } | 1782 | } |
1788 | 1783 | ||
1789 | if(indexes->rxLoCleared != indexes->rxLoReady) { | 1784 | if(indexes->rxLoCleared != indexes->rxLoReady) { |
1790 | work_done += typhoon_rx(tp, &tp->rxLoRing, &indexes->rxLoReady, | 1785 | work_done += typhoon_rx(tp, &tp->rxLoRing, &indexes->rxLoReady, |
1791 | &indexes->rxLoCleared, budget); | 1786 | &indexes->rxLoCleared, budget - work_done); |
1792 | } | ||
1793 | |||
1794 | if(work_done) { | ||
1795 | *total_budget -= work_done; | ||
1796 | dev->quota -= work_done; | ||
1797 | |||
1798 | if(work_done >= orig_budget) | ||
1799 | done = 0; | ||
1800 | } | 1787 | } |
1801 | 1788 | ||
1802 | if(le32_to_cpu(indexes->rxBuffCleared) == tp->rxBuffRing.lastWrite) { | 1789 | if(le32_to_cpu(indexes->rxBuffCleared) == tp->rxBuffRing.lastWrite) { |
@@ -1804,14 +1791,14 @@ typhoon_poll(struct net_device *dev, int *total_budget) | |||
1804 | typhoon_fill_free_ring(tp); | 1791 | typhoon_fill_free_ring(tp); |
1805 | } | 1792 | } |
1806 | 1793 | ||
1807 | if(done) { | 1794 | if (work_done < budget) { |
1808 | netif_rx_complete(dev); | 1795 | netif_rx_complete(dev, napi); |
1809 | iowrite32(TYPHOON_INTR_NONE, | 1796 | iowrite32(TYPHOON_INTR_NONE, |
1810 | tp->ioaddr + TYPHOON_REG_INTR_MASK); | 1797 | tp->ioaddr + TYPHOON_REG_INTR_MASK); |
1811 | typhoon_post_pci_writes(tp->ioaddr); | 1798 | typhoon_post_pci_writes(tp->ioaddr); |
1812 | } | 1799 | } |
1813 | 1800 | ||
1814 | return (done ? 0 : 1); | 1801 | return work_done; |
1815 | } | 1802 | } |
1816 | 1803 | ||
1817 | static irqreturn_t | 1804 | static irqreturn_t |
@@ -1828,10 +1815,10 @@ typhoon_interrupt(int irq, void *dev_instance) | |||
1828 | 1815 | ||
1829 | iowrite32(intr_status, ioaddr + TYPHOON_REG_INTR_STATUS); | 1816 | iowrite32(intr_status, ioaddr + TYPHOON_REG_INTR_STATUS); |
1830 | 1817 | ||
1831 | if(netif_rx_schedule_prep(dev)) { | 1818 | if (netif_rx_schedule_prep(dev, &tp->napi)) { |
1832 | iowrite32(TYPHOON_INTR_ALL, ioaddr + TYPHOON_REG_INTR_MASK); | 1819 | iowrite32(TYPHOON_INTR_ALL, ioaddr + TYPHOON_REG_INTR_MASK); |
1833 | typhoon_post_pci_writes(ioaddr); | 1820 | typhoon_post_pci_writes(ioaddr); |
1834 | __netif_rx_schedule(dev); | 1821 | __netif_rx_schedule(dev, &tp->napi); |
1835 | } else { | 1822 | } else { |
1836 | printk(KERN_ERR "%s: Error, poll already scheduled\n", | 1823 | printk(KERN_ERR "%s: Error, poll already scheduled\n", |
1837 | dev->name); | 1824 | dev->name); |
@@ -2119,9 +2106,13 @@ typhoon_open(struct net_device *dev) | |||
2119 | if(err < 0) | 2106 | if(err < 0) |
2120 | goto out_sleep; | 2107 | goto out_sleep; |
2121 | 2108 | ||
2109 | napi_enable(&tp->napi); | ||
2110 | |||
2122 | err = typhoon_start_runtime(tp); | 2111 | err = typhoon_start_runtime(tp); |
2123 | if(err < 0) | 2112 | if(err < 0) { |
2113 | napi_disable(&tp->napi); | ||
2124 | goto out_irq; | 2114 | goto out_irq; |
2115 | } | ||
2125 | 2116 | ||
2126 | netif_start_queue(dev); | 2117 | netif_start_queue(dev); |
2127 | return 0; | 2118 | return 0; |
@@ -2150,6 +2141,7 @@ typhoon_close(struct net_device *dev) | |||
2150 | struct typhoon *tp = netdev_priv(dev); | 2141 | struct typhoon *tp = netdev_priv(dev); |
2151 | 2142 | ||
2152 | netif_stop_queue(dev); | 2143 | netif_stop_queue(dev); |
2144 | napi_disable(&tp->napi); | ||
2153 | 2145 | ||
2154 | if(typhoon_stop_runtime(tp, WaitSleep) < 0) | 2146 | if(typhoon_stop_runtime(tp, WaitSleep) < 0) |
2155 | printk(KERN_ERR "%s: unable to stop runtime\n", dev->name); | 2147 | printk(KERN_ERR "%s: unable to stop runtime\n", dev->name); |
@@ -2521,8 +2513,7 @@ typhoon_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
2521 | dev->stop = typhoon_close; | 2513 | dev->stop = typhoon_close; |
2522 | dev->set_multicast_list = typhoon_set_rx_mode; | 2514 | dev->set_multicast_list = typhoon_set_rx_mode; |
2523 | dev->tx_timeout = typhoon_tx_timeout; | 2515 | dev->tx_timeout = typhoon_tx_timeout; |
2524 | dev->poll = typhoon_poll; | 2516 | netif_napi_add(dev, &tp->napi, typhoon_poll, 16); |
2525 | dev->weight = 16; | ||
2526 | dev->watchdog_timeo = TX_TIMEOUT; | 2517 | dev->watchdog_timeo = TX_TIMEOUT; |
2527 | dev->get_stats = typhoon_get_stats; | 2518 | dev->get_stats = typhoon_get_stats; |
2528 | dev->set_mac_address = typhoon_set_mac_address; | 2519 | dev->set_mac_address = typhoon_set_mac_address; |
diff --git a/drivers/net/ucc_geth.c b/drivers/net/ucc_geth.c index 9a38dfe45f8f..72f617bf2520 100644 --- a/drivers/net/ucc_geth.c +++ b/drivers/net/ucc_geth.c | |||
@@ -3582,41 +3582,31 @@ static int ucc_geth_tx(struct net_device *dev, u8 txQ) | |||
3582 | } | 3582 | } |
3583 | 3583 | ||
3584 | #ifdef CONFIG_UGETH_NAPI | 3584 | #ifdef CONFIG_UGETH_NAPI |
3585 | static int ucc_geth_poll(struct net_device *dev, int *budget) | 3585 | static int ucc_geth_poll(struct napi_struct *napi, int budget) |
3586 | { | 3586 | { |
3587 | struct ucc_geth_private *ugeth = netdev_priv(dev); | 3587 | struct ucc_geth_private *ugeth = container_of(napi, struct ucc_geth_private, napi); |
3588 | struct net_device *dev = ugeth->dev; | ||
3588 | struct ucc_geth_info *ug_info; | 3589 | struct ucc_geth_info *ug_info; |
3589 | struct ucc_fast_private *uccf; | 3590 | int howmany, i; |
3590 | int howmany; | ||
3591 | u8 i; | ||
3592 | int rx_work_limit; | ||
3593 | register u32 uccm; | ||
3594 | 3591 | ||
3595 | ug_info = ugeth->ug_info; | 3592 | ug_info = ugeth->ug_info; |
3596 | 3593 | ||
3597 | rx_work_limit = *budget; | ||
3598 | if (rx_work_limit > dev->quota) | ||
3599 | rx_work_limit = dev->quota; | ||
3600 | |||
3601 | howmany = 0; | 3594 | howmany = 0; |
3595 | for (i = 0; i < ug_info->numQueuesRx; i++) | ||
3596 | howmany += ucc_geth_rx(ugeth, i, budget - howmany); | ||
3602 | 3597 | ||
3603 | for (i = 0; i < ug_info->numQueuesRx; i++) { | 3598 | if (howmany < budget) { |
3604 | howmany += ucc_geth_rx(ugeth, i, rx_work_limit); | 3599 | struct ucc_fast_private *uccf; |
3605 | } | 3600 | u32 uccm; |
3606 | |||
3607 | dev->quota -= howmany; | ||
3608 | rx_work_limit -= howmany; | ||
3609 | *budget -= howmany; | ||
3610 | 3601 | ||
3611 | if (rx_work_limit > 0) { | 3602 | netif_rx_complete(dev, napi); |
3612 | netif_rx_complete(dev); | ||
3613 | uccf = ugeth->uccf; | 3603 | uccf = ugeth->uccf; |
3614 | uccm = in_be32(uccf->p_uccm); | 3604 | uccm = in_be32(uccf->p_uccm); |
3615 | uccm |= UCCE_RX_EVENTS; | 3605 | uccm |= UCCE_RX_EVENTS; |
3616 | out_be32(uccf->p_uccm, uccm); | 3606 | out_be32(uccf->p_uccm, uccm); |
3617 | } | 3607 | } |
3618 | 3608 | ||
3619 | return (rx_work_limit > 0) ? 0 : 1; | 3609 | return howmany; |
3620 | } | 3610 | } |
3621 | #endif /* CONFIG_UGETH_NAPI */ | 3611 | #endif /* CONFIG_UGETH_NAPI */ |
3622 | 3612 | ||
@@ -3651,10 +3641,10 @@ static irqreturn_t ucc_geth_irq_handler(int irq, void *info) | |||
3651 | /* check for receive events that require processing */ | 3641 | /* check for receive events that require processing */ |
3652 | if (ucce & UCCE_RX_EVENTS) { | 3642 | if (ucce & UCCE_RX_EVENTS) { |
3653 | #ifdef CONFIG_UGETH_NAPI | 3643 | #ifdef CONFIG_UGETH_NAPI |
3654 | if (netif_rx_schedule_prep(dev)) { | 3644 | if (netif_rx_schedule_prep(dev, &ugeth->napi)) { |
3655 | uccm &= ~UCCE_RX_EVENTS; | 3645 | uccm &= ~UCCE_RX_EVENTS; |
3656 | out_be32(uccf->p_uccm, uccm); | 3646 | out_be32(uccf->p_uccm, uccm); |
3657 | __netif_rx_schedule(dev); | 3647 | __netif_rx_schedule(dev, &ugeth->napi); |
3658 | } | 3648 | } |
3659 | #else | 3649 | #else |
3660 | rx_mask = UCCE_RXBF_SINGLE_MASK; | 3650 | rx_mask = UCCE_RXBF_SINGLE_MASK; |
@@ -3717,12 +3707,15 @@ static int ucc_geth_open(struct net_device *dev) | |||
3717 | return err; | 3707 | return err; |
3718 | } | 3708 | } |
3719 | 3709 | ||
3710 | #ifdef CONFIG_UGETH_NAPI | ||
3711 | napi_enable(&ugeth->napi); | ||
3712 | #endif | ||
3720 | err = ucc_geth_startup(ugeth); | 3713 | err = ucc_geth_startup(ugeth); |
3721 | if (err) { | 3714 | if (err) { |
3722 | if (netif_msg_ifup(ugeth)) | 3715 | if (netif_msg_ifup(ugeth)) |
3723 | ugeth_err("%s: Cannot configure net device, aborting.", | 3716 | ugeth_err("%s: Cannot configure net device, aborting.", |
3724 | dev->name); | 3717 | dev->name); |
3725 | return err; | 3718 | goto out_err; |
3726 | } | 3719 | } |
3727 | 3720 | ||
3728 | err = adjust_enet_interface(ugeth); | 3721 | err = adjust_enet_interface(ugeth); |
@@ -3730,7 +3723,7 @@ static int ucc_geth_open(struct net_device *dev) | |||
3730 | if (netif_msg_ifup(ugeth)) | 3723 | if (netif_msg_ifup(ugeth)) |
3731 | ugeth_err("%s: Cannot configure net device, aborting.", | 3724 | ugeth_err("%s: Cannot configure net device, aborting.", |
3732 | dev->name); | 3725 | dev->name); |
3733 | return err; | 3726 | goto out_err; |
3734 | } | 3727 | } |
3735 | 3728 | ||
3736 | /* Set MACSTNADDR1, MACSTNADDR2 */ | 3729 | /* Set MACSTNADDR1, MACSTNADDR2 */ |
@@ -3748,7 +3741,7 @@ static int ucc_geth_open(struct net_device *dev) | |||
3748 | if (err) { | 3741 | if (err) { |
3749 | if (netif_msg_ifup(ugeth)) | 3742 | if (netif_msg_ifup(ugeth)) |
3750 | ugeth_err("%s: Cannot initialize PHY, aborting.", dev->name); | 3743 | ugeth_err("%s: Cannot initialize PHY, aborting.", dev->name); |
3751 | return err; | 3744 | goto out_err; |
3752 | } | 3745 | } |
3753 | 3746 | ||
3754 | phy_start(ugeth->phydev); | 3747 | phy_start(ugeth->phydev); |
@@ -3761,7 +3754,7 @@ static int ucc_geth_open(struct net_device *dev) | |||
3761 | ugeth_err("%s: Cannot get IRQ for net device, aborting.", | 3754 | ugeth_err("%s: Cannot get IRQ for net device, aborting.", |
3762 | dev->name); | 3755 | dev->name); |
3763 | ucc_geth_stop(ugeth); | 3756 | ucc_geth_stop(ugeth); |
3764 | return err; | 3757 | goto out_err; |
3765 | } | 3758 | } |
3766 | 3759 | ||
3767 | err = ugeth_enable(ugeth, COMM_DIR_RX_AND_TX); | 3760 | err = ugeth_enable(ugeth, COMM_DIR_RX_AND_TX); |
@@ -3769,12 +3762,18 @@ static int ucc_geth_open(struct net_device *dev) | |||
3769 | if (netif_msg_ifup(ugeth)) | 3762 | if (netif_msg_ifup(ugeth)) |
3770 | ugeth_err("%s: Cannot enable net device, aborting.", dev->name); | 3763 | ugeth_err("%s: Cannot enable net device, aborting.", dev->name); |
3771 | ucc_geth_stop(ugeth); | 3764 | ucc_geth_stop(ugeth); |
3772 | return err; | 3765 | goto out_err; |
3773 | } | 3766 | } |
3774 | 3767 | ||
3775 | netif_start_queue(dev); | 3768 | netif_start_queue(dev); |
3776 | 3769 | ||
3777 | return err; | 3770 | return err; |
3771 | |||
3772 | out_err: | ||
3773 | #ifdef CONFIG_UGETH_NAPI | ||
3774 | napi_disable(&ugeth->napi); | ||
3775 | #endif | ||
3776 | return err; | ||
3778 | } | 3777 | } |
3779 | 3778 | ||
3780 | /* Stops the kernel queue, and halts the controller */ | 3779 | /* Stops the kernel queue, and halts the controller */ |
@@ -3784,6 +3783,10 @@ static int ucc_geth_close(struct net_device *dev) | |||
3784 | 3783 | ||
3785 | ugeth_vdbg("%s: IN", __FUNCTION__); | 3784 | ugeth_vdbg("%s: IN", __FUNCTION__); |
3786 | 3785 | ||
3786 | #ifdef CONFIG_UGETH_NAPI | ||
3787 | napi_disable(&ugeth->napi); | ||
3788 | #endif | ||
3789 | |||
3787 | ucc_geth_stop(ugeth); | 3790 | ucc_geth_stop(ugeth); |
3788 | 3791 | ||
3789 | phy_disconnect(ugeth->phydev); | 3792 | phy_disconnect(ugeth->phydev); |
@@ -3964,8 +3967,7 @@ static int ucc_geth_probe(struct of_device* ofdev, const struct of_device_id *ma | |||
3964 | dev->tx_timeout = ucc_geth_timeout; | 3967 | dev->tx_timeout = ucc_geth_timeout; |
3965 | dev->watchdog_timeo = TX_TIMEOUT; | 3968 | dev->watchdog_timeo = TX_TIMEOUT; |
3966 | #ifdef CONFIG_UGETH_NAPI | 3969 | #ifdef CONFIG_UGETH_NAPI |
3967 | dev->poll = ucc_geth_poll; | 3970 | netif_napi_add(dev, &ugeth->napi, ucc_geth_poll, UCC_GETH_DEV_WEIGHT); |
3968 | dev->weight = UCC_GETH_DEV_WEIGHT; | ||
3969 | #endif /* CONFIG_UGETH_NAPI */ | 3971 | #endif /* CONFIG_UGETH_NAPI */ |
3970 | dev->stop = ucc_geth_close; | 3972 | dev->stop = ucc_geth_close; |
3971 | dev->get_stats = ucc_geth_get_stats; | 3973 | dev->get_stats = ucc_geth_get_stats; |
diff --git a/drivers/net/ucc_geth.h b/drivers/net/ucc_geth.h index bb4dac8c0c65..0579ba081aa5 100644 --- a/drivers/net/ucc_geth.h +++ b/drivers/net/ucc_geth.h | |||
@@ -1184,6 +1184,7 @@ struct ucc_geth_private { | |||
1184 | struct ucc_geth_info *ug_info; | 1184 | struct ucc_geth_info *ug_info; |
1185 | struct ucc_fast_private *uccf; | 1185 | struct ucc_fast_private *uccf; |
1186 | struct net_device *dev; | 1186 | struct net_device *dev; |
1187 | struct napi_struct napi; | ||
1187 | struct net_device_stats stats; /* linux network statistics */ | 1188 | struct net_device_stats stats; /* linux network statistics */ |
1188 | struct ucc_geth *ug_regs; | 1189 | struct ucc_geth *ug_regs; |
1189 | struct ucc_geth_init_pram *p_init_enet_param_shadow; | 1190 | struct ucc_geth_init_pram *p_init_enet_param_shadow; |
diff --git a/drivers/net/via-rhine.c b/drivers/net/via-rhine.c index b56dff26772d..7a5899059c44 100644 --- a/drivers/net/via-rhine.c +++ b/drivers/net/via-rhine.c | |||
@@ -389,6 +389,8 @@ struct rhine_private { | |||
389 | 389 | ||
390 | struct pci_dev *pdev; | 390 | struct pci_dev *pdev; |
391 | long pioaddr; | 391 | long pioaddr; |
392 | struct net_device *dev; | ||
393 | struct napi_struct napi; | ||
392 | struct net_device_stats stats; | 394 | struct net_device_stats stats; |
393 | spinlock_t lock; | 395 | spinlock_t lock; |
394 | 396 | ||
@@ -582,28 +584,25 @@ static void rhine_poll(struct net_device *dev) | |||
582 | #endif | 584 | #endif |
583 | 585 | ||
584 | #ifdef CONFIG_VIA_RHINE_NAPI | 586 | #ifdef CONFIG_VIA_RHINE_NAPI |
585 | static int rhine_napipoll(struct net_device *dev, int *budget) | 587 | static int rhine_napipoll(struct napi_struct *napi, int budget) |
586 | { | 588 | { |
587 | struct rhine_private *rp = netdev_priv(dev); | 589 | struct rhine_private *rp = container_of(napi, struct rhine_private, napi); |
590 | struct net_device *dev = rp->dev; | ||
588 | void __iomem *ioaddr = rp->base; | 591 | void __iomem *ioaddr = rp->base; |
589 | int done, limit = min(dev->quota, *budget); | 592 | int work_done; |
590 | 593 | ||
591 | done = rhine_rx(dev, limit); | 594 | work_done = rhine_rx(dev, budget); |
592 | *budget -= done; | ||
593 | dev->quota -= done; | ||
594 | 595 | ||
595 | if (done < limit) { | 596 | if (work_done < budget) { |
596 | netif_rx_complete(dev); | 597 | netif_rx_complete(dev, napi); |
597 | 598 | ||
598 | iowrite16(IntrRxDone | IntrRxErr | IntrRxEmpty| IntrRxOverflow | | 599 | iowrite16(IntrRxDone | IntrRxErr | IntrRxEmpty| IntrRxOverflow | |
599 | IntrRxDropped | IntrRxNoBuf | IntrTxAborted | | 600 | IntrRxDropped | IntrRxNoBuf | IntrTxAborted | |
600 | IntrTxDone | IntrTxError | IntrTxUnderrun | | 601 | IntrTxDone | IntrTxError | IntrTxUnderrun | |
601 | IntrPCIErr | IntrStatsMax | IntrLinkChange, | 602 | IntrPCIErr | IntrStatsMax | IntrLinkChange, |
602 | ioaddr + IntrEnable); | 603 | ioaddr + IntrEnable); |
603 | return 0; | ||
604 | } | 604 | } |
605 | else | 605 | return work_done; |
606 | return 1; | ||
607 | } | 606 | } |
608 | #endif | 607 | #endif |
609 | 608 | ||
@@ -707,6 +706,7 @@ static int __devinit rhine_init_one(struct pci_dev *pdev, | |||
707 | SET_NETDEV_DEV(dev, &pdev->dev); | 706 | SET_NETDEV_DEV(dev, &pdev->dev); |
708 | 707 | ||
709 | rp = netdev_priv(dev); | 708 | rp = netdev_priv(dev); |
709 | rp->dev = dev; | ||
710 | rp->quirks = quirks; | 710 | rp->quirks = quirks; |
711 | rp->pioaddr = pioaddr; | 711 | rp->pioaddr = pioaddr; |
712 | rp->pdev = pdev; | 712 | rp->pdev = pdev; |
@@ -785,8 +785,7 @@ static int __devinit rhine_init_one(struct pci_dev *pdev, | |||
785 | dev->poll_controller = rhine_poll; | 785 | dev->poll_controller = rhine_poll; |
786 | #endif | 786 | #endif |
787 | #ifdef CONFIG_VIA_RHINE_NAPI | 787 | #ifdef CONFIG_VIA_RHINE_NAPI |
788 | dev->poll = rhine_napipoll; | 788 | netif_napi_add(dev, &rp->napi, rhine_napipoll, 64); |
789 | dev->weight = 64; | ||
790 | #endif | 789 | #endif |
791 | if (rp->quirks & rqRhineI) | 790 | if (rp->quirks & rqRhineI) |
792 | dev->features |= NETIF_F_SG|NETIF_F_HW_CSUM; | 791 | dev->features |= NETIF_F_SG|NETIF_F_HW_CSUM; |
@@ -1061,7 +1060,9 @@ static void init_registers(struct net_device *dev) | |||
1061 | 1060 | ||
1062 | rhine_set_rx_mode(dev); | 1061 | rhine_set_rx_mode(dev); |
1063 | 1062 | ||
1064 | netif_poll_enable(dev); | 1063 | #ifdef CONFIG_VIA_RHINE_NAPI |
1064 | napi_enable(&rp->napi); | ||
1065 | #endif | ||
1065 | 1066 | ||
1066 | /* Enable interrupts by setting the interrupt mask. */ | 1067 | /* Enable interrupts by setting the interrupt mask. */ |
1067 | iowrite16(IntrRxDone | IntrRxErr | IntrRxEmpty| IntrRxOverflow | | 1068 | iowrite16(IntrRxDone | IntrRxErr | IntrRxEmpty| IntrRxOverflow | |
@@ -1196,6 +1197,10 @@ static void rhine_tx_timeout(struct net_device *dev) | |||
1196 | /* protect against concurrent rx interrupts */ | 1197 | /* protect against concurrent rx interrupts */ |
1197 | disable_irq(rp->pdev->irq); | 1198 | disable_irq(rp->pdev->irq); |
1198 | 1199 | ||
1200 | #ifdef CONFIG_VIA_RHINE_NAPI | ||
1201 | napi_disable(&rp->napi); | ||
1202 | #endif | ||
1203 | |||
1199 | spin_lock(&rp->lock); | 1204 | spin_lock(&rp->lock); |
1200 | 1205 | ||
1201 | /* clear all descriptors */ | 1206 | /* clear all descriptors */ |
@@ -1324,7 +1329,7 @@ static irqreturn_t rhine_interrupt(int irq, void *dev_instance) | |||
1324 | IntrPCIErr | IntrStatsMax | IntrLinkChange, | 1329 | IntrPCIErr | IntrStatsMax | IntrLinkChange, |
1325 | ioaddr + IntrEnable); | 1330 | ioaddr + IntrEnable); |
1326 | 1331 | ||
1327 | netif_rx_schedule(dev); | 1332 | netif_rx_schedule(dev, &rp->napi); |
1328 | #else | 1333 | #else |
1329 | rhine_rx(dev, RX_RING_SIZE); | 1334 | rhine_rx(dev, RX_RING_SIZE); |
1330 | #endif | 1335 | #endif |
@@ -1837,7 +1842,9 @@ static int rhine_close(struct net_device *dev) | |||
1837 | spin_lock_irq(&rp->lock); | 1842 | spin_lock_irq(&rp->lock); |
1838 | 1843 | ||
1839 | netif_stop_queue(dev); | 1844 | netif_stop_queue(dev); |
1840 | netif_poll_disable(dev); | 1845 | #ifdef CONFIG_VIA_RHINE_NAPI |
1846 | napi_disable(&rp->napi); | ||
1847 | #endif | ||
1841 | 1848 | ||
1842 | if (debug > 1) | 1849 | if (debug > 1) |
1843 | printk(KERN_DEBUG "%s: Shutting down ethercard, " | 1850 | printk(KERN_DEBUG "%s: Shutting down ethercard, " |
@@ -1936,6 +1943,9 @@ static int rhine_suspend(struct pci_dev *pdev, pm_message_t state) | |||
1936 | if (!netif_running(dev)) | 1943 | if (!netif_running(dev)) |
1937 | return 0; | 1944 | return 0; |
1938 | 1945 | ||
1946 | #ifdef CONFIG_VIA_RHINE_NAPI | ||
1947 | napi_disable(&rp->napi); | ||
1948 | #endif | ||
1939 | netif_device_detach(dev); | 1949 | netif_device_detach(dev); |
1940 | pci_save_state(pdev); | 1950 | pci_save_state(pdev); |
1941 | 1951 | ||
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index 4445810335a8..70e551c19e3a 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c | |||
@@ -72,6 +72,7 @@ struct netfront_info { | |||
72 | struct list_head list; | 72 | struct list_head list; |
73 | struct net_device *netdev; | 73 | struct net_device *netdev; |
74 | 74 | ||
75 | struct napi_struct napi; | ||
75 | struct net_device_stats stats; | 76 | struct net_device_stats stats; |
76 | 77 | ||
77 | struct xen_netif_tx_front_ring tx; | 78 | struct xen_netif_tx_front_ring tx; |
@@ -185,7 +186,8 @@ static int xennet_can_sg(struct net_device *dev) | |||
185 | static void rx_refill_timeout(unsigned long data) | 186 | static void rx_refill_timeout(unsigned long data) |
186 | { | 187 | { |
187 | struct net_device *dev = (struct net_device *)data; | 188 | struct net_device *dev = (struct net_device *)data; |
188 | netif_rx_schedule(dev); | 189 | struct netfront_info *np = netdev_priv(dev); |
190 | netif_rx_schedule(dev, &np->napi); | ||
189 | } | 191 | } |
190 | 192 | ||
191 | static int netfront_tx_slot_available(struct netfront_info *np) | 193 | static int netfront_tx_slot_available(struct netfront_info *np) |
@@ -342,12 +344,14 @@ static int xennet_open(struct net_device *dev) | |||
342 | 344 | ||
343 | memset(&np->stats, 0, sizeof(np->stats)); | 345 | memset(&np->stats, 0, sizeof(np->stats)); |
344 | 346 | ||
347 | napi_enable(&np->napi); | ||
348 | |||
345 | spin_lock_bh(&np->rx_lock); | 349 | spin_lock_bh(&np->rx_lock); |
346 | if (netif_carrier_ok(dev)) { | 350 | if (netif_carrier_ok(dev)) { |
347 | xennet_alloc_rx_buffers(dev); | 351 | xennet_alloc_rx_buffers(dev); |
348 | np->rx.sring->rsp_event = np->rx.rsp_cons + 1; | 352 | np->rx.sring->rsp_event = np->rx.rsp_cons + 1; |
349 | if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) | 353 | if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) |
350 | netif_rx_schedule(dev); | 354 | netif_rx_schedule(dev, &np->napi); |
351 | } | 355 | } |
352 | spin_unlock_bh(&np->rx_lock); | 356 | spin_unlock_bh(&np->rx_lock); |
353 | 357 | ||
@@ -589,6 +593,7 @@ static int xennet_close(struct net_device *dev) | |||
589 | { | 593 | { |
590 | struct netfront_info *np = netdev_priv(dev); | 594 | struct netfront_info *np = netdev_priv(dev); |
591 | netif_stop_queue(np->netdev); | 595 | netif_stop_queue(np->netdev); |
596 | napi_disable(&np->napi); | ||
592 | return 0; | 597 | return 0; |
593 | } | 598 | } |
594 | 599 | ||
@@ -872,15 +877,16 @@ static int handle_incoming_queue(struct net_device *dev, | |||
872 | return packets_dropped; | 877 | return packets_dropped; |
873 | } | 878 | } |
874 | 879 | ||
875 | static int xennet_poll(struct net_device *dev, int *pbudget) | 880 | static int xennet_poll(struct napi_struct *napi, int budget) |
876 | { | 881 | { |
877 | struct netfront_info *np = netdev_priv(dev); | 882 | struct netfront_info *np = container_of(napi, struct netfront_info, napi); |
883 | struct net_device *dev = np->netdev; | ||
878 | struct sk_buff *skb; | 884 | struct sk_buff *skb; |
879 | struct netfront_rx_info rinfo; | 885 | struct netfront_rx_info rinfo; |
880 | struct xen_netif_rx_response *rx = &rinfo.rx; | 886 | struct xen_netif_rx_response *rx = &rinfo.rx; |
881 | struct xen_netif_extra_info *extras = rinfo.extras; | 887 | struct xen_netif_extra_info *extras = rinfo.extras; |
882 | RING_IDX i, rp; | 888 | RING_IDX i, rp; |
883 | int work_done, budget, more_to_do = 1; | 889 | int work_done; |
884 | struct sk_buff_head rxq; | 890 | struct sk_buff_head rxq; |
885 | struct sk_buff_head errq; | 891 | struct sk_buff_head errq; |
886 | struct sk_buff_head tmpq; | 892 | struct sk_buff_head tmpq; |
@@ -899,9 +905,6 @@ static int xennet_poll(struct net_device *dev, int *pbudget) | |||
899 | skb_queue_head_init(&errq); | 905 | skb_queue_head_init(&errq); |
900 | skb_queue_head_init(&tmpq); | 906 | skb_queue_head_init(&tmpq); |
901 | 907 | ||
902 | budget = *pbudget; | ||
903 | if (budget > dev->quota) | ||
904 | budget = dev->quota; | ||
905 | rp = np->rx.sring->rsp_prod; | 908 | rp = np->rx.sring->rsp_prod; |
906 | rmb(); /* Ensure we see queued responses up to 'rp'. */ | 909 | rmb(); /* Ensure we see queued responses up to 'rp'. */ |
907 | 910 | ||
@@ -1006,22 +1009,21 @@ err: | |||
1006 | 1009 | ||
1007 | xennet_alloc_rx_buffers(dev); | 1010 | xennet_alloc_rx_buffers(dev); |
1008 | 1011 | ||
1009 | *pbudget -= work_done; | ||
1010 | dev->quota -= work_done; | ||
1011 | |||
1012 | if (work_done < budget) { | 1012 | if (work_done < budget) { |
1013 | int more_to_do = 0; | ||
1014 | |||
1013 | local_irq_save(flags); | 1015 | local_irq_save(flags); |
1014 | 1016 | ||
1015 | RING_FINAL_CHECK_FOR_RESPONSES(&np->rx, more_to_do); | 1017 | RING_FINAL_CHECK_FOR_RESPONSES(&np->rx, more_to_do); |
1016 | if (!more_to_do) | 1018 | if (!more_to_do) |
1017 | __netif_rx_complete(dev); | 1019 | __netif_rx_complete(dev, napi); |
1018 | 1020 | ||
1019 | local_irq_restore(flags); | 1021 | local_irq_restore(flags); |
1020 | } | 1022 | } |
1021 | 1023 | ||
1022 | spin_unlock(&np->rx_lock); | 1024 | spin_unlock(&np->rx_lock); |
1023 | 1025 | ||
1024 | return more_to_do; | 1026 | return work_done; |
1025 | } | 1027 | } |
1026 | 1028 | ||
1027 | static int xennet_change_mtu(struct net_device *dev, int mtu) | 1029 | static int xennet_change_mtu(struct net_device *dev, int mtu) |
@@ -1201,10 +1203,9 @@ static struct net_device * __devinit xennet_create_dev(struct xenbus_device *dev | |||
1201 | netdev->hard_start_xmit = xennet_start_xmit; | 1203 | netdev->hard_start_xmit = xennet_start_xmit; |
1202 | netdev->stop = xennet_close; | 1204 | netdev->stop = xennet_close; |
1203 | netdev->get_stats = xennet_get_stats; | 1205 | netdev->get_stats = xennet_get_stats; |
1204 | netdev->poll = xennet_poll; | 1206 | netif_napi_add(netdev, &np->napi, xennet_poll, 64); |
1205 | netdev->uninit = xennet_uninit; | 1207 | netdev->uninit = xennet_uninit; |
1206 | netdev->change_mtu = xennet_change_mtu; | 1208 | netdev->change_mtu = xennet_change_mtu; |
1207 | netdev->weight = 64; | ||
1208 | netdev->features = NETIF_F_IP_CSUM; | 1209 | netdev->features = NETIF_F_IP_CSUM; |
1209 | 1210 | ||
1210 | SET_ETHTOOL_OPS(netdev, &xennet_ethtool_ops); | 1211 | SET_ETHTOOL_OPS(netdev, &xennet_ethtool_ops); |
@@ -1349,7 +1350,7 @@ static irqreturn_t xennet_interrupt(int irq, void *dev_id) | |||
1349 | xennet_tx_buf_gc(dev); | 1350 | xennet_tx_buf_gc(dev); |
1350 | /* Under tx_lock: protects access to rx shared-ring indexes. */ | 1351 | /* Under tx_lock: protects access to rx shared-ring indexes. */ |
1351 | if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) | 1352 | if (RING_HAS_UNCONSUMED_RESPONSES(&np->rx)) |
1352 | netif_rx_schedule(dev); | 1353 | netif_rx_schedule(dev, &np->napi); |
1353 | } | 1354 | } |
1354 | 1355 | ||
1355 | spin_unlock_irqrestore(&np->tx_lock, flags); | 1356 | spin_unlock_irqrestore(&np->tx_lock, flags); |