diff options
author | Mark Mason <mmason@upwardaccess.com> | 2007-04-26 03:23:22 -0400 |
---|---|---|
committer | Jeff Garzik <jeff@garzik.org> | 2007-04-28 11:01:07 -0400 |
commit | 693aa9470d8273a0ded8b211a8f5f7c0835adf30 (patch) | |
tree | efad4a82527e9285ed6dae51e775d58360cc622f /drivers/net | |
parent | 55e924cf5772cbcf00549e448be35b392ff3084c (diff) |
add NAPI support to sb1250-mac.c
Patch to add NAPI support to sb1250-mac.c (rev 2). This patch differs from
the last in that the NAPI support isn't marked as experimental, nor is it
configurable (ie. once applied - NAPI is enabled all the time). This was
based on feedback from Ralf and others.
Signed-off-by: Mark Mason <mason@broadcom.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Jeff Garzik <jeff@garzik.org>
Diffstat (limited to 'drivers/net')
-rw-r--r-- | drivers/net/sb1250-mac.c | 271 |
1 files changed, 179 insertions, 92 deletions
diff --git a/drivers/net/sb1250-mac.c b/drivers/net/sb1250-mac.c index e3f1a0707267..132e2148b21c 100644 --- a/drivers/net/sb1250-mac.c +++ b/drivers/net/sb1250-mac.c | |||
@@ -95,19 +95,28 @@ MODULE_PARM_DESC(full_duplex, "1-" __MODULE_STRING(MAX_UNITS)); | |||
95 | #endif | 95 | #endif |
96 | 96 | ||
97 | #ifdef CONFIG_SBMAC_COALESCE | 97 | #ifdef CONFIG_SBMAC_COALESCE |
98 | static int int_pktcnt = 0; | 98 | static int int_pktcnt_tx = 255; |
99 | module_param(int_pktcnt, int, S_IRUGO); | 99 | module_param(int_pktcnt_tx, int, S_IRUGO); |
100 | MODULE_PARM_DESC(int_pktcnt, "Packet count"); | 100 | MODULE_PARM_DESC(int_pktcnt_tx, "TX packet count"); |
101 | 101 | ||
102 | static int int_timeout = 0; | 102 | static int int_timeout_tx = 255; |
103 | module_param(int_timeout, int, S_IRUGO); | 103 | module_param(int_timeout_tx, int, S_IRUGO); |
104 | MODULE_PARM_DESC(int_timeout, "Timeout value"); | 104 | MODULE_PARM_DESC(int_timeout_tx, "TX timeout value"); |
105 | |||
106 | static int int_pktcnt_rx = 64; | ||
107 | module_param(int_pktcnt_rx, int, S_IRUGO); | ||
108 | MODULE_PARM_DESC(int_pktcnt_rx, "RX packet count"); | ||
109 | |||
110 | static int int_timeout_rx = 64; | ||
111 | module_param(int_timeout_rx, int, S_IRUGO); | ||
112 | MODULE_PARM_DESC(int_timeout_rx, "RX timeout value"); | ||
105 | #endif | 113 | #endif |
106 | 114 | ||
107 | #include <asm/sibyte/sb1250.h> | 115 | #include <asm/sibyte/sb1250.h> |
108 | #if defined(CONFIG_SIBYTE_BCM1x55) || defined(CONFIG_SIBYTE_BCM1x80) | 116 | #if defined(CONFIG_SIBYTE_BCM1x55) || defined(CONFIG_SIBYTE_BCM1x80) |
109 | #include <asm/sibyte/bcm1480_regs.h> | 117 | #include <asm/sibyte/bcm1480_regs.h> |
110 | #include <asm/sibyte/bcm1480_int.h> | 118 | #include <asm/sibyte/bcm1480_int.h> |
119 | #define R_MAC_DMA_OODPKTLOST_RX R_MAC_DMA_OODPKTLOST | ||
111 | #elif defined(CONFIG_SIBYTE_SB1250) || defined(CONFIG_SIBYTE_BCM112X) | 120 | #elif defined(CONFIG_SIBYTE_SB1250) || defined(CONFIG_SIBYTE_BCM112X) |
112 | #include <asm/sibyte/sb1250_regs.h> | 121 | #include <asm/sibyte/sb1250_regs.h> |
113 | #include <asm/sibyte/sb1250_int.h> | 122 | #include <asm/sibyte/sb1250_int.h> |
@@ -155,8 +164,8 @@ typedef enum { sbmac_state_uninit, sbmac_state_off, sbmac_state_on, | |||
155 | 164 | ||
156 | #define NUMCACHEBLKS(x) (((x)+SMP_CACHE_BYTES-1)/SMP_CACHE_BYTES) | 165 | #define NUMCACHEBLKS(x) (((x)+SMP_CACHE_BYTES-1)/SMP_CACHE_BYTES) |
157 | 166 | ||
158 | #define SBMAC_MAX_TXDESCR 32 | 167 | #define SBMAC_MAX_TXDESCR 256 |
159 | #define SBMAC_MAX_RXDESCR 32 | 168 | #define SBMAC_MAX_RXDESCR 256 |
160 | 169 | ||
161 | #define ETHER_ALIGN 2 | 170 | #define ETHER_ALIGN 2 |
162 | #define ETHER_ADDR_LEN 6 | 171 | #define ETHER_ADDR_LEN 6 |
@@ -185,10 +194,10 @@ typedef struct sbmacdma_s { | |||
185 | * associated with it. | 194 | * associated with it. |
186 | */ | 195 | */ |
187 | 196 | ||
188 | struct sbmac_softc *sbdma_eth; /* back pointer to associated MAC */ | 197 | struct sbmac_softc *sbdma_eth; /* back pointer to associated MAC */ |
189 | int sbdma_channel; /* channel number */ | 198 | int sbdma_channel; /* channel number */ |
190 | int sbdma_txdir; /* direction (1=transmit) */ | 199 | int sbdma_txdir; /* direction (1=transmit) */ |
191 | int sbdma_maxdescr; /* total # of descriptors in ring */ | 200 | int sbdma_maxdescr; /* total # of descriptors in ring */ |
192 | #ifdef CONFIG_SBMAC_COALESCE | 201 | #ifdef CONFIG_SBMAC_COALESCE |
193 | int sbdma_int_pktcnt; /* # descriptors rx/tx before interrupt*/ | 202 | int sbdma_int_pktcnt; /* # descriptors rx/tx before interrupt*/ |
194 | int sbdma_int_timeout; /* # usec rx/tx interrupt */ | 203 | int sbdma_int_timeout; /* # usec rx/tx interrupt */ |
@@ -197,13 +206,16 @@ typedef struct sbmacdma_s { | |||
197 | volatile void __iomem *sbdma_config0; /* DMA config register 0 */ | 206 | volatile void __iomem *sbdma_config0; /* DMA config register 0 */ |
198 | volatile void __iomem *sbdma_config1; /* DMA config register 1 */ | 207 | volatile void __iomem *sbdma_config1; /* DMA config register 1 */ |
199 | volatile void __iomem *sbdma_dscrbase; /* Descriptor base address */ | 208 | volatile void __iomem *sbdma_dscrbase; /* Descriptor base address */ |
200 | volatile void __iomem *sbdma_dscrcnt; /* Descriptor count register */ | 209 | volatile void __iomem *sbdma_dscrcnt; /* Descriptor count register */ |
201 | volatile void __iomem *sbdma_curdscr; /* current descriptor address */ | 210 | volatile void __iomem *sbdma_curdscr; /* current descriptor address */ |
211 | volatile void __iomem *sbdma_oodpktlost;/* pkt drop (rx only) */ | ||
212 | |||
202 | 213 | ||
203 | /* | 214 | /* |
204 | * This stuff is for maintenance of the ring | 215 | * This stuff is for maintenance of the ring |
205 | */ | 216 | */ |
206 | 217 | ||
218 | sbdmadscr_t *sbdma_dscrtable_unaligned; | ||
207 | sbdmadscr_t *sbdma_dscrtable; /* base of descriptor table */ | 219 | sbdmadscr_t *sbdma_dscrtable; /* base of descriptor table */ |
208 | sbdmadscr_t *sbdma_dscrtable_end; /* end of descriptor table */ | 220 | sbdmadscr_t *sbdma_dscrtable_end; /* end of descriptor table */ |
209 | 221 | ||
@@ -286,8 +298,8 @@ static int sbdma_add_rcvbuffer(sbmacdma_t *d,struct sk_buff *m); | |||
286 | static int sbdma_add_txbuffer(sbmacdma_t *d,struct sk_buff *m); | 298 | static int sbdma_add_txbuffer(sbmacdma_t *d,struct sk_buff *m); |
287 | static void sbdma_emptyring(sbmacdma_t *d); | 299 | static void sbdma_emptyring(sbmacdma_t *d); |
288 | static void sbdma_fillring(sbmacdma_t *d); | 300 | static void sbdma_fillring(sbmacdma_t *d); |
289 | static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d); | 301 | static int sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d, int work_to_do, int poll); |
290 | static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d); | 302 | static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d, int poll); |
291 | static int sbmac_initctx(struct sbmac_softc *s); | 303 | static int sbmac_initctx(struct sbmac_softc *s); |
292 | static void sbmac_channel_start(struct sbmac_softc *s); | 304 | static void sbmac_channel_start(struct sbmac_softc *s); |
293 | static void sbmac_channel_stop(struct sbmac_softc *s); | 305 | static void sbmac_channel_stop(struct sbmac_softc *s); |
@@ -308,6 +320,8 @@ static struct net_device_stats *sbmac_get_stats(struct net_device *dev); | |||
308 | static void sbmac_set_rx_mode(struct net_device *dev); | 320 | static void sbmac_set_rx_mode(struct net_device *dev); |
309 | static int sbmac_mii_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); | 321 | static int sbmac_mii_ioctl(struct net_device *dev, struct ifreq *rq, int cmd); |
310 | static int sbmac_close(struct net_device *dev); | 322 | static int sbmac_close(struct net_device *dev); |
323 | static int sbmac_poll(struct net_device *poll_dev, int *budget); | ||
324 | |||
311 | static int sbmac_mii_poll(struct sbmac_softc *s,int noisy); | 325 | static int sbmac_mii_poll(struct sbmac_softc *s,int noisy); |
312 | static int sbmac_mii_probe(struct net_device *dev); | 326 | static int sbmac_mii_probe(struct net_device *dev); |
313 | 327 | ||
@@ -679,6 +693,10 @@ static void sbdma_initctx(sbmacdma_t *d, | |||
679 | int txrx, | 693 | int txrx, |
680 | int maxdescr) | 694 | int maxdescr) |
681 | { | 695 | { |
696 | #ifdef CONFIG_SBMAC_COALESCE | ||
697 | int int_pktcnt, int_timeout; | ||
698 | #endif | ||
699 | |||
682 | /* | 700 | /* |
683 | * Save away interesting stuff in the structure | 701 | * Save away interesting stuff in the structure |
684 | */ | 702 | */ |
@@ -728,6 +746,11 @@ static void sbdma_initctx(sbmacdma_t *d, | |||
728 | s->sbm_base + R_MAC_DMA_REGISTER(txrx,chan,R_MAC_DMA_DSCR_CNT); | 746 | s->sbm_base + R_MAC_DMA_REGISTER(txrx,chan,R_MAC_DMA_DSCR_CNT); |
729 | d->sbdma_curdscr = | 747 | d->sbdma_curdscr = |
730 | s->sbm_base + R_MAC_DMA_REGISTER(txrx,chan,R_MAC_DMA_CUR_DSCRADDR); | 748 | s->sbm_base + R_MAC_DMA_REGISTER(txrx,chan,R_MAC_DMA_CUR_DSCRADDR); |
749 | if (d->sbdma_txdir) | ||
750 | d->sbdma_oodpktlost = NULL; | ||
751 | else | ||
752 | d->sbdma_oodpktlost = | ||
753 | s->sbm_base + R_MAC_DMA_REGISTER(txrx,chan,R_MAC_DMA_OODPKTLOST_RX); | ||
731 | 754 | ||
732 | /* | 755 | /* |
733 | * Allocate memory for the ring | 756 | * Allocate memory for the ring |
@@ -735,6 +758,7 @@ static void sbdma_initctx(sbmacdma_t *d, | |||
735 | 758 | ||
736 | d->sbdma_maxdescr = maxdescr; | 759 | d->sbdma_maxdescr = maxdescr; |
737 | 760 | ||
761 | d->sbdma_dscrtable_unaligned = | ||
738 | d->sbdma_dscrtable = (sbdmadscr_t *) | 762 | d->sbdma_dscrtable = (sbdmadscr_t *) |
739 | kmalloc((d->sbdma_maxdescr+1)*sizeof(sbdmadscr_t), GFP_KERNEL); | 763 | kmalloc((d->sbdma_maxdescr+1)*sizeof(sbdmadscr_t), GFP_KERNEL); |
740 | 764 | ||
@@ -765,12 +789,14 @@ static void sbdma_initctx(sbmacdma_t *d, | |||
765 | * Setup Rx/Tx DMA coalescing defaults | 789 | * Setup Rx/Tx DMA coalescing defaults |
766 | */ | 790 | */ |
767 | 791 | ||
792 | int_pktcnt = (txrx == DMA_TX) ? int_pktcnt_tx : int_pktcnt_rx; | ||
768 | if ( int_pktcnt ) { | 793 | if ( int_pktcnt ) { |
769 | d->sbdma_int_pktcnt = int_pktcnt; | 794 | d->sbdma_int_pktcnt = int_pktcnt; |
770 | } else { | 795 | } else { |
771 | d->sbdma_int_pktcnt = 1; | 796 | d->sbdma_int_pktcnt = 1; |
772 | } | 797 | } |
773 | 798 | ||
799 | int_timeout = (txrx == DMA_TX) ? int_timeout_tx : int_timeout_rx; | ||
774 | if ( int_timeout ) { | 800 | if ( int_timeout ) { |
775 | d->sbdma_int_timeout = int_timeout; | 801 | d->sbdma_int_timeout = int_timeout; |
776 | } else { | 802 | } else { |
@@ -1147,30 +1173,41 @@ static void sbmac_netpoll(struct net_device *netdev) | |||
1147 | #endif | 1173 | #endif |
1148 | 1174 | ||
1149 | /********************************************************************** | 1175 | /********************************************************************** |
1150 | * SBDMA_RX_PROCESS(sc,d) | 1176 | * SBDMA_RX_PROCESS(sc,d,work_to_do,poll) |
1151 | * | 1177 | * |
1152 | * Process "completed" receive buffers on the specified DMA channel. | 1178 | * Process "completed" receive buffers on the specified DMA channel. |
1153 | * Note that this isn't really ideal for priority channels, since | ||
1154 | * it processes all of the packets on a given channel before | ||
1155 | * returning. | ||
1156 | * | 1179 | * |
1157 | * Input parameters: | 1180 | * Input parameters: |
1158 | * sc - softc structure | 1181 | * sc - softc structure |
1159 | * d - DMA channel context | 1182 | * d - DMA channel context |
1183 | * work_to_do - no. of packets to process before enabling interrupt | ||
1184 | * again (for NAPI) | ||
1185 | * poll - 1: using polling (for NAPI) | ||
1160 | * | 1186 | * |
1161 | * Return value: | 1187 | * Return value: |
1162 | * nothing | 1188 | * nothing |
1163 | ********************************************************************* */ | 1189 | ********************************************************************* */ |
1164 | 1190 | ||
1165 | static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | 1191 | static int sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d, |
1192 | int work_to_do, int poll) | ||
1166 | { | 1193 | { |
1167 | int curidx; | 1194 | int curidx; |
1168 | int hwidx; | 1195 | int hwidx; |
1169 | sbdmadscr_t *dsc; | 1196 | sbdmadscr_t *dsc; |
1170 | struct sk_buff *sb; | 1197 | struct sk_buff *sb; |
1171 | int len; | 1198 | int len; |
1199 | int work_done = 0; | ||
1200 | int dropped = 0; | ||
1172 | 1201 | ||
1173 | for (;;) { | 1202 | prefetch(d); |
1203 | |||
1204 | again: | ||
1205 | /* Check if the HW dropped any frames */ | ||
1206 | sc->sbm_stats.rx_fifo_errors | ||
1207 | += __raw_readq(sc->sbm_rxdma.sbdma_oodpktlost) & 0xffff; | ||
1208 | __raw_writeq(0, sc->sbm_rxdma.sbdma_oodpktlost); | ||
1209 | |||
1210 | while (work_to_do-- > 0) { | ||
1174 | /* | 1211 | /* |
1175 | * figure out where we are (as an index) and where | 1212 | * figure out where we are (as an index) and where |
1176 | * the hardware is (also as an index) | 1213 | * the hardware is (also as an index) |
@@ -1182,7 +1219,12 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1182 | * (sbdma_remptr) and the physical address (sbdma_curdscr CSR) | 1219 | * (sbdma_remptr) and the physical address (sbdma_curdscr CSR) |
1183 | */ | 1220 | */ |
1184 | 1221 | ||
1185 | curidx = d->sbdma_remptr - d->sbdma_dscrtable; | 1222 | dsc = d->sbdma_remptr; |
1223 | curidx = dsc - d->sbdma_dscrtable; | ||
1224 | |||
1225 | prefetch(dsc); | ||
1226 | prefetch(&d->sbdma_ctxtable[curidx]); | ||
1227 | |||
1186 | hwidx = (int) (((__raw_readq(d->sbdma_curdscr) & M_DMA_CURDSCR_ADDR) - | 1228 | hwidx = (int) (((__raw_readq(d->sbdma_curdscr) & M_DMA_CURDSCR_ADDR) - |
1187 | d->sbdma_dscrtable_phys) / sizeof(sbdmadscr_t)); | 1229 | d->sbdma_dscrtable_phys) / sizeof(sbdmadscr_t)); |
1188 | 1230 | ||
@@ -1193,13 +1235,12 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1193 | */ | 1235 | */ |
1194 | 1236 | ||
1195 | if (curidx == hwidx) | 1237 | if (curidx == hwidx) |
1196 | break; | 1238 | goto done; |
1197 | 1239 | ||
1198 | /* | 1240 | /* |
1199 | * Otherwise, get the packet's sk_buff ptr back | 1241 | * Otherwise, get the packet's sk_buff ptr back |
1200 | */ | 1242 | */ |
1201 | 1243 | ||
1202 | dsc = &(d->sbdma_dscrtable[curidx]); | ||
1203 | sb = d->sbdma_ctxtable[curidx]; | 1244 | sb = d->sbdma_ctxtable[curidx]; |
1204 | d->sbdma_ctxtable[curidx] = NULL; | 1245 | d->sbdma_ctxtable[curidx] = NULL; |
1205 | 1246 | ||
@@ -1211,7 +1252,7 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1211 | * receive ring. | 1252 | * receive ring. |
1212 | */ | 1253 | */ |
1213 | 1254 | ||
1214 | if (!(dsc->dscr_a & M_DMA_ETHRX_BAD)) { | 1255 | if (likely (!(dsc->dscr_a & M_DMA_ETHRX_BAD))) { |
1215 | 1256 | ||
1216 | /* | 1257 | /* |
1217 | * Add a new buffer to replace the old one. If we fail | 1258 | * Add a new buffer to replace the old one. If we fail |
@@ -1219,9 +1260,14 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1219 | * packet and put it right back on the receive ring. | 1260 | * packet and put it right back on the receive ring. |
1220 | */ | 1261 | */ |
1221 | 1262 | ||
1222 | if (sbdma_add_rcvbuffer(d,NULL) == -ENOBUFS) { | 1263 | if (unlikely (sbdma_add_rcvbuffer(d,NULL) == |
1223 | sc->sbm_stats.rx_dropped++; | 1264 | -ENOBUFS)) { |
1265 | sc->sbm_stats.rx_dropped++; | ||
1224 | sbdma_add_rcvbuffer(d,sb); /* re-add old buffer */ | 1266 | sbdma_add_rcvbuffer(d,sb); /* re-add old buffer */ |
1267 | /* No point in continuing at the moment */ | ||
1268 | printk(KERN_ERR "dropped packet (1)\n"); | ||
1269 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); | ||
1270 | goto done; | ||
1225 | } else { | 1271 | } else { |
1226 | /* | 1272 | /* |
1227 | * Set length into the packet | 1273 | * Set length into the packet |
@@ -1233,8 +1279,6 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1233 | * receive ring. Pass the buffer to | 1279 | * receive ring. Pass the buffer to |
1234 | * the kernel | 1280 | * the kernel |
1235 | */ | 1281 | */ |
1236 | sc->sbm_stats.rx_bytes += len; | ||
1237 | sc->sbm_stats.rx_packets++; | ||
1238 | sb->protocol = eth_type_trans(sb,d->sbdma_eth->sbm_dev); | 1282 | sb->protocol = eth_type_trans(sb,d->sbdma_eth->sbm_dev); |
1239 | /* Check hw IPv4/TCP checksum if supported */ | 1283 | /* Check hw IPv4/TCP checksum if supported */ |
1240 | if (sc->rx_hw_checksum == ENABLE) { | 1284 | if (sc->rx_hw_checksum == ENABLE) { |
@@ -1246,8 +1290,22 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1246 | sb->ip_summed = CHECKSUM_NONE; | 1290 | sb->ip_summed = CHECKSUM_NONE; |
1247 | } | 1291 | } |
1248 | } | 1292 | } |
1249 | 1293 | prefetch(sb->data); | |
1250 | netif_rx(sb); | 1294 | prefetch((const void *)(((char *)sb->data)+32)); |
1295 | if (poll) | ||
1296 | dropped = netif_receive_skb(sb); | ||
1297 | else | ||
1298 | dropped = netif_rx(sb); | ||
1299 | |||
1300 | if (dropped == NET_RX_DROP) { | ||
1301 | sc->sbm_stats.rx_dropped++; | ||
1302 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); | ||
1303 | goto done; | ||
1304 | } | ||
1305 | else { | ||
1306 | sc->sbm_stats.rx_bytes += len; | ||
1307 | sc->sbm_stats.rx_packets++; | ||
1308 | } | ||
1251 | } | 1309 | } |
1252 | } else { | 1310 | } else { |
1253 | /* | 1311 | /* |
@@ -1264,12 +1322,16 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1264 | */ | 1322 | */ |
1265 | 1323 | ||
1266 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); | 1324 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); |
1267 | 1325 | work_done++; | |
1326 | } | ||
1327 | if (!poll) { | ||
1328 | work_to_do = 32; | ||
1329 | goto again; /* collect fifo drop statistics again */ | ||
1268 | } | 1330 | } |
1331 | done: | ||
1332 | return work_done; | ||
1269 | } | 1333 | } |
1270 | 1334 | ||
1271 | |||
1272 | |||
1273 | /********************************************************************** | 1335 | /********************************************************************** |
1274 | * SBDMA_TX_PROCESS(sc,d) | 1336 | * SBDMA_TX_PROCESS(sc,d) |
1275 | * | 1337 | * |
@@ -1281,22 +1343,30 @@ static void sbdma_rx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1281 | * | 1343 | * |
1282 | * Input parameters: | 1344 | * Input parameters: |
1283 | * sc - softc structure | 1345 | * sc - softc structure |
1284 | * d - DMA channel context | 1346 | * d - DMA channel context |
1347 | * poll - 1: using polling (for NAPI) | ||
1285 | * | 1348 | * |
1286 | * Return value: | 1349 | * Return value: |
1287 | * nothing | 1350 | * nothing |
1288 | ********************************************************************* */ | 1351 | ********************************************************************* */ |
1289 | 1352 | ||
1290 | static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d) | 1353 | static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d, int poll) |
1291 | { | 1354 | { |
1292 | int curidx; | 1355 | int curidx; |
1293 | int hwidx; | 1356 | int hwidx; |
1294 | sbdmadscr_t *dsc; | 1357 | sbdmadscr_t *dsc; |
1295 | struct sk_buff *sb; | 1358 | struct sk_buff *sb; |
1296 | unsigned long flags; | 1359 | unsigned long flags; |
1360 | int packets_handled = 0; | ||
1297 | 1361 | ||
1298 | spin_lock_irqsave(&(sc->sbm_lock), flags); | 1362 | spin_lock_irqsave(&(sc->sbm_lock), flags); |
1299 | 1363 | ||
1364 | if (d->sbdma_remptr == d->sbdma_addptr) | ||
1365 | goto end_unlock; | ||
1366 | |||
1367 | hwidx = (int) (((__raw_readq(d->sbdma_curdscr) & M_DMA_CURDSCR_ADDR) - | ||
1368 | d->sbdma_dscrtable_phys) / sizeof(sbdmadscr_t)); | ||
1369 | |||
1300 | for (;;) { | 1370 | for (;;) { |
1301 | /* | 1371 | /* |
1302 | * figure out where we are (as an index) and where | 1372 | * figure out where we are (as an index) and where |
@@ -1310,8 +1380,6 @@ static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1310 | */ | 1380 | */ |
1311 | 1381 | ||
1312 | curidx = d->sbdma_remptr - d->sbdma_dscrtable; | 1382 | curidx = d->sbdma_remptr - d->sbdma_dscrtable; |
1313 | hwidx = (int) (((__raw_readq(d->sbdma_curdscr) & M_DMA_CURDSCR_ADDR) - | ||
1314 | d->sbdma_dscrtable_phys) / sizeof(sbdmadscr_t)); | ||
1315 | 1383 | ||
1316 | /* | 1384 | /* |
1317 | * If they're the same, that means we've processed all | 1385 | * If they're the same, that means we've processed all |
@@ -1349,6 +1417,8 @@ static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1349 | 1417 | ||
1350 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); | 1418 | d->sbdma_remptr = SBDMA_NEXTBUF(d,sbdma_remptr); |
1351 | 1419 | ||
1420 | packets_handled++; | ||
1421 | |||
1352 | } | 1422 | } |
1353 | 1423 | ||
1354 | /* | 1424 | /* |
@@ -1357,8 +1427,10 @@ static void sbdma_tx_process(struct sbmac_softc *sc,sbmacdma_t *d) | |||
1357 | * watermark on the transmit queue. | 1427 | * watermark on the transmit queue. |
1358 | */ | 1428 | */ |
1359 | 1429 | ||
1360 | netif_wake_queue(d->sbdma_eth->sbm_dev); | 1430 | if (packets_handled) |
1431 | netif_wake_queue(d->sbdma_eth->sbm_dev); | ||
1361 | 1432 | ||
1433 | end_unlock: | ||
1362 | spin_unlock_irqrestore(&(sc->sbm_lock), flags); | 1434 | spin_unlock_irqrestore(&(sc->sbm_lock), flags); |
1363 | 1435 | ||
1364 | } | 1436 | } |
@@ -1432,9 +1504,9 @@ static int sbmac_initctx(struct sbmac_softc *s) | |||
1432 | 1504 | ||
1433 | static void sbdma_uninitctx(struct sbmacdma_s *d) | 1505 | static void sbdma_uninitctx(struct sbmacdma_s *d) |
1434 | { | 1506 | { |
1435 | if (d->sbdma_dscrtable) { | 1507 | if (d->sbdma_dscrtable_unaligned) { |
1436 | kfree(d->sbdma_dscrtable); | 1508 | kfree(d->sbdma_dscrtable_unaligned); |
1437 | d->sbdma_dscrtable = NULL; | 1509 | d->sbdma_dscrtable_unaligned = d->sbdma_dscrtable = NULL; |
1438 | } | 1510 | } |
1439 | 1511 | ||
1440 | if (d->sbdma_ctxtable) { | 1512 | if (d->sbdma_ctxtable) { |
@@ -1632,15 +1704,9 @@ static void sbmac_channel_start(struct sbmac_softc *s) | |||
1632 | #endif | 1704 | #endif |
1633 | 1705 | ||
1634 | #ifdef CONFIG_SBMAC_COALESCE | 1706 | #ifdef CONFIG_SBMAC_COALESCE |
1635 | /* | ||
1636 | * Accept any TX interrupt and EOP count/timer RX interrupts on ch 0 | ||
1637 | */ | ||
1638 | __raw_writeq(((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_TX_CH0) | | 1707 | __raw_writeq(((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_TX_CH0) | |
1639 | ((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_RX_CH0), s->sbm_imr); | 1708 | ((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_RX_CH0), s->sbm_imr); |
1640 | #else | 1709 | #else |
1641 | /* | ||
1642 | * Accept any kind of interrupt on TX and RX DMA channel 0 | ||
1643 | */ | ||
1644 | __raw_writeq((M_MAC_INT_CHANNEL << S_MAC_TX_CH0) | | 1710 | __raw_writeq((M_MAC_INT_CHANNEL << S_MAC_TX_CH0) | |
1645 | (M_MAC_INT_CHANNEL << S_MAC_RX_CH0), s->sbm_imr); | 1711 | (M_MAC_INT_CHANNEL << S_MAC_RX_CH0), s->sbm_imr); |
1646 | #endif | 1712 | #endif |
@@ -2073,57 +2139,46 @@ static irqreturn_t sbmac_intr(int irq,void *dev_instance) | |||
2073 | uint64_t isr; | 2139 | uint64_t isr; |
2074 | int handled = 0; | 2140 | int handled = 0; |
2075 | 2141 | ||
2076 | for (;;) { | 2142 | /* |
2077 | 2143 | * Read the ISR (this clears the bits in the real | |
2078 | /* | 2144 | * register, except for counter addr) |
2079 | * Read the ISR (this clears the bits in the real | 2145 | */ |
2080 | * register, except for counter addr) | ||
2081 | */ | ||
2082 | |||
2083 | isr = __raw_readq(sc->sbm_isr) & ~M_MAC_COUNTER_ADDR; | ||
2084 | 2146 | ||
2085 | if (isr == 0) | 2147 | isr = __raw_readq(sc->sbm_isr) & ~M_MAC_COUNTER_ADDR; |
2086 | break; | ||
2087 | 2148 | ||
2088 | handled = 1; | 2149 | if (isr == 0) |
2150 | return IRQ_RETVAL(0); | ||
2151 | handled = 1; | ||
2089 | 2152 | ||
2090 | /* | 2153 | /* |
2091 | * Transmits on channel 0 | 2154 | * Transmits on channel 0 |
2092 | */ | 2155 | */ |
2093 | 2156 | ||
2094 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_TX_CH0)) { | 2157 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_TX_CH0)) { |
2095 | sbdma_tx_process(sc,&(sc->sbm_txdma)); | 2158 | sbdma_tx_process(sc,&(sc->sbm_txdma), 0); |
2159 | #ifdef CONFIG_NETPOLL_TRAP | ||
2160 | if (netpoll_trap()) { | ||
2161 | if (test_and_clear_bit(__LINK_STATE_XOFF, &dev->state)) | ||
2162 | __netif_schedule(dev); | ||
2096 | } | 2163 | } |
2164 | #endif | ||
2165 | } | ||
2097 | 2166 | ||
2098 | /* | 2167 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_RX_CH0)) { |
2099 | * Receives on channel 0 | 2168 | if (netif_rx_schedule_prep(dev)) { |
2100 | */ | 2169 | __raw_writeq(0, sc->sbm_imr); |
2101 | 2170 | __netif_rx_schedule(dev); | |
2102 | /* | 2171 | /* Depend on the exit from poll to reenable intr */ |
2103 | * It's important to test all the bits (or at least the | 2172 | } |
2104 | * EOP_SEEN bit) when deciding to do the RX process | 2173 | else { |
2105 | * particularly when coalescing, to make sure we | 2174 | /* may leave some packets behind */ |
2106 | * take care of the following: | 2175 | sbdma_rx_process(sc,&(sc->sbm_rxdma), |
2107 | * | 2176 | SBMAC_MAX_RXDESCR * 2, 0); |
2108 | * If you have some packets waiting (have been received | ||
2109 | * but no interrupt) and get a TX interrupt before | ||
2110 | * the RX timer or counter expires, reading the ISR | ||
2111 | * above will clear the timer and counter, and you | ||
2112 | * won't get another interrupt until a packet shows | ||
2113 | * up to start the timer again. Testing | ||
2114 | * EOP_SEEN here takes care of this case. | ||
2115 | * (EOP_SEEN is part of M_MAC_INT_CHANNEL << S_MAC_RX_CH0) | ||
2116 | */ | ||
2117 | |||
2118 | |||
2119 | if (isr & (M_MAC_INT_CHANNEL << S_MAC_RX_CH0)) { | ||
2120 | sbdma_rx_process(sc,&(sc->sbm_rxdma)); | ||
2121 | } | 2177 | } |
2122 | } | 2178 | } |
2123 | return IRQ_RETVAL(handled); | 2179 | return IRQ_RETVAL(handled); |
2124 | } | 2180 | } |
2125 | 2181 | ||
2126 | |||
2127 | /********************************************************************** | 2182 | /********************************************************************** |
2128 | * SBMAC_START_TX(skb,dev) | 2183 | * SBMAC_START_TX(skb,dev) |
2129 | * | 2184 | * |
@@ -2253,8 +2308,6 @@ static void sbmac_setmulti(struct sbmac_softc *sc) | |||
2253 | } | 2308 | } |
2254 | } | 2309 | } |
2255 | 2310 | ||
2256 | |||
2257 | |||
2258 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) | 2311 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) |
2259 | /********************************************************************** | 2312 | /********************************************************************** |
2260 | * SBMAC_PARSE_XDIGIT(str) | 2313 | * SBMAC_PARSE_XDIGIT(str) |
@@ -2417,6 +2470,8 @@ static int sbmac_init(struct net_device *dev, int idx) | |||
2417 | dev->do_ioctl = sbmac_mii_ioctl; | 2470 | dev->do_ioctl = sbmac_mii_ioctl; |
2418 | dev->tx_timeout = sbmac_tx_timeout; | 2471 | dev->tx_timeout = sbmac_tx_timeout; |
2419 | dev->watchdog_timeo = TX_TIMEOUT; | 2472 | dev->watchdog_timeo = TX_TIMEOUT; |
2473 | dev->poll = sbmac_poll; | ||
2474 | dev->weight = 16; | ||
2420 | 2475 | ||
2421 | dev->change_mtu = sb1250_change_mtu; | 2476 | dev->change_mtu = sb1250_change_mtu; |
2422 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2477 | #ifdef CONFIG_NET_POLL_CONTROLLER |
@@ -2819,7 +2874,39 @@ static int sbmac_close(struct net_device *dev) | |||
2819 | return 0; | 2874 | return 0; |
2820 | } | 2875 | } |
2821 | 2876 | ||
2877 | static int sbmac_poll(struct net_device *dev, int *budget) | ||
2878 | { | ||
2879 | int work_to_do; | ||
2880 | int work_done; | ||
2881 | struct sbmac_softc *sc = netdev_priv(dev); | ||
2882 | |||
2883 | work_to_do = min(*budget, dev->quota); | ||
2884 | work_done = sbdma_rx_process(sc, &(sc->sbm_rxdma), work_to_do, 1); | ||
2885 | |||
2886 | if (work_done > work_to_do) | ||
2887 | printk(KERN_ERR "%s exceeded work_to_do budget=%d quota=%d work-done=%d\n", | ||
2888 | sc->sbm_dev->name, *budget, dev->quota, work_done); | ||
2822 | 2889 | ||
2890 | sbdma_tx_process(sc, &(sc->sbm_txdma), 1); | ||
2891 | |||
2892 | *budget -= work_done; | ||
2893 | dev->quota -= work_done; | ||
2894 | |||
2895 | if (work_done < work_to_do) { | ||
2896 | netif_rx_complete(dev); | ||
2897 | |||
2898 | #ifdef CONFIG_SBMAC_COALESCE | ||
2899 | __raw_writeq(((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_TX_CH0) | | ||
2900 | ((M_MAC_INT_EOP_COUNT | M_MAC_INT_EOP_TIMER) << S_MAC_RX_CH0), | ||
2901 | sc->sbm_imr); | ||
2902 | #else | ||
2903 | __raw_writeq((M_MAC_INT_CHANNEL << S_MAC_TX_CH0) | | ||
2904 | (M_MAC_INT_CHANNEL << S_MAC_RX_CH0), sc->sbm_imr); | ||
2905 | #endif | ||
2906 | } | ||
2907 | |||
2908 | return (work_done >= work_to_do); | ||
2909 | } | ||
2823 | 2910 | ||
2824 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) | 2911 | #if defined(SBMAC_ETH0_HWADDR) || defined(SBMAC_ETH1_HWADDR) || defined(SBMAC_ETH2_HWADDR) || defined(SBMAC_ETH3_HWADDR) |
2825 | static void | 2912 | static void |
@@ -2906,7 +2993,7 @@ sbmac_init_module(void) | |||
2906 | 2993 | ||
2907 | /* | 2994 | /* |
2908 | * The R_MAC_ETHERNET_ADDR register will be set to some nonzero | 2995 | * The R_MAC_ETHERNET_ADDR register will be set to some nonzero |
2909 | * value for us by the firmware if we're going to use this MAC. | 2996 | * value for us by the firmware if we are going to use this MAC. |
2910 | * If we find a zero, skip this MAC. | 2997 | * If we find a zero, skip this MAC. |
2911 | */ | 2998 | */ |
2912 | 2999 | ||