aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/net/sky2.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/net/sky2.c')
-rw-r--r--drivers/net/sky2.c95
1 files changed, 47 insertions, 48 deletions
diff --git a/drivers/net/sky2.c b/drivers/net/sky2.c
index b908596bd89a..64af764d0087 100644
--- a/drivers/net/sky2.c
+++ b/drivers/net/sky2.c
@@ -769,9 +769,16 @@ static inline struct sky2_tx_le *get_tx_le(struct sky2_port *sky2)
769 struct sky2_tx_le *le = sky2->tx_le + sky2->tx_prod; 769 struct sky2_tx_le *le = sky2->tx_le + sky2->tx_prod;
770 770
771 sky2->tx_prod = RING_NEXT(sky2->tx_prod, TX_RING_SIZE); 771 sky2->tx_prod = RING_NEXT(sky2->tx_prod, TX_RING_SIZE);
772 le->ctrl = 0;
772 return le; 773 return le;
773} 774}
774 775
776static inline struct tx_ring_info *tx_le_re(struct sky2_port *sky2,
777 struct sky2_tx_le *le)
778{
779 return sky2->tx_ring + (le - sky2->tx_le);
780}
781
775/* Update chip's next pointer */ 782/* Update chip's next pointer */
776static inline void sky2_put_idx(struct sky2_hw *hw, unsigned q, u16 idx) 783static inline void sky2_put_idx(struct sky2_hw *hw, unsigned q, u16 idx)
777{ 784{
@@ -786,6 +793,7 @@ static inline struct sky2_rx_le *sky2_next_rx(struct sky2_port *sky2)
786{ 793{
787 struct sky2_rx_le *le = sky2->rx_le + sky2->rx_put; 794 struct sky2_rx_le *le = sky2->rx_le + sky2->rx_put;
788 sky2->rx_put = RING_NEXT(sky2->rx_put, RX_LE_SIZE); 795 sky2->rx_put = RING_NEXT(sky2->rx_put, RX_LE_SIZE);
796 le->ctrl = 0;
789 return le; 797 return le;
790} 798}
791 799
@@ -805,7 +813,6 @@ static void sky2_rx_add(struct sky2_port *sky2, dma_addr_t map)
805 if (sky2->rx_addr64 != hi) { 813 if (sky2->rx_addr64 != hi) {
806 le = sky2_next_rx(sky2); 814 le = sky2_next_rx(sky2);
807 le->addr = cpu_to_le32(hi); 815 le->addr = cpu_to_le32(hi);
808 le->ctrl = 0;
809 le->opcode = OP_ADDR64 | HW_OWNER; 816 le->opcode = OP_ADDR64 | HW_OWNER;
810 sky2->rx_addr64 = high32(map + len); 817 sky2->rx_addr64 = high32(map + len);
811 } 818 }
@@ -813,7 +820,6 @@ static void sky2_rx_add(struct sky2_port *sky2, dma_addr_t map)
813 le = sky2_next_rx(sky2); 820 le = sky2_next_rx(sky2);
814 le->addr = cpu_to_le32((u32) map); 821 le->addr = cpu_to_le32((u32) map);
815 le->length = cpu_to_le16(len); 822 le->length = cpu_to_le16(len);
816 le->ctrl = 0;
817 le->opcode = OP_PACKET | HW_OWNER; 823 le->opcode = OP_PACKET | HW_OWNER;
818} 824}
819 825
@@ -877,7 +883,7 @@ static void sky2_rx_clean(struct sky2_port *sky2)
877 883
878 memset(sky2->rx_le, 0, RX_LE_BYTES); 884 memset(sky2->rx_le, 0, RX_LE_BYTES);
879 for (i = 0; i < sky2->rx_pending; i++) { 885 for (i = 0; i < sky2->rx_pending; i++) {
880 struct ring_info *re = sky2->rx_ring + i; 886 struct rx_ring_info *re = sky2->rx_ring + i;
881 887
882 if (re->skb) { 888 if (re->skb) {
883 pci_unmap_single(sky2->hw->pdev, 889 pci_unmap_single(sky2->hw->pdev,
@@ -1008,7 +1014,7 @@ static int sky2_rx_start(struct sky2_port *sky2)
1008 1014
1009 rx_set_checksum(sky2); 1015 rx_set_checksum(sky2);
1010 for (i = 0; i < sky2->rx_pending; i++) { 1016 for (i = 0; i < sky2->rx_pending; i++) {
1011 struct ring_info *re = sky2->rx_ring + i; 1017 struct rx_ring_info *re = sky2->rx_ring + i;
1012 1018
1013 re->skb = sky2_alloc_skb(sky2->netdev, sky2->rx_bufsize, 1019 re->skb = sky2_alloc_skb(sky2->netdev, sky2->rx_bufsize,
1014 GFP_KERNEL); 1020 GFP_KERNEL);
@@ -1094,7 +1100,7 @@ static int sky2_up(struct net_device *dev)
1094 goto err_out; 1100 goto err_out;
1095 memset(sky2->rx_le, 0, RX_LE_BYTES); 1101 memset(sky2->rx_le, 0, RX_LE_BYTES);
1096 1102
1097 sky2->rx_ring = kcalloc(sky2->rx_pending, sizeof(struct ring_info), 1103 sky2->rx_ring = kcalloc(sky2->rx_pending, sizeof(struct rx_ring_info),
1098 GFP_KERNEL); 1104 GFP_KERNEL);
1099 if (!sky2->rx_ring) 1105 if (!sky2->rx_ring)
1100 goto err_out; 1106 goto err_out;
@@ -1241,13 +1247,10 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1241 mapping = pci_map_single(hw->pdev, skb->data, len, PCI_DMA_TODEVICE); 1247 mapping = pci_map_single(hw->pdev, skb->data, len, PCI_DMA_TODEVICE);
1242 addr64 = high32(mapping); 1248 addr64 = high32(mapping);
1243 1249
1244 re = sky2->tx_ring + sky2->tx_prod;
1245
1246 /* Send high bits if changed or crosses boundary */ 1250 /* Send high bits if changed or crosses boundary */
1247 if (addr64 != sky2->tx_addr64 || high32(mapping + len) != sky2->tx_addr64) { 1251 if (addr64 != sky2->tx_addr64 || high32(mapping + len) != sky2->tx_addr64) {
1248 le = get_tx_le(sky2); 1252 le = get_tx_le(sky2);
1249 le->addr = cpu_to_le32(addr64); 1253 le->addr = cpu_to_le32(addr64);
1250 le->ctrl = 0;
1251 le->opcode = OP_ADDR64 | HW_OWNER; 1254 le->opcode = OP_ADDR64 | HW_OWNER;
1252 sky2->tx_addr64 = high32(mapping + len); 1255 sky2->tx_addr64 = high32(mapping + len);
1253 } 1256 }
@@ -1263,7 +1266,6 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1263 le = get_tx_le(sky2); 1266 le = get_tx_le(sky2);
1264 le->addr = cpu_to_le32(mss); 1267 le->addr = cpu_to_le32(mss);
1265 le->opcode = OP_LRGLEN | HW_OWNER; 1268 le->opcode = OP_LRGLEN | HW_OWNER;
1266 le->ctrl = 0;
1267 sky2->tx_last_mss = mss; 1269 sky2->tx_last_mss = mss;
1268 } 1270 }
1269 } 1271 }
@@ -1276,7 +1278,6 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1276 le = get_tx_le(sky2); 1278 le = get_tx_le(sky2);
1277 le->addr = 0; 1279 le->addr = 0;
1278 le->opcode = OP_VLAN|HW_OWNER; 1280 le->opcode = OP_VLAN|HW_OWNER;
1279 le->ctrl = 0;
1280 } else 1281 } else
1281 le->opcode |= OP_VLAN; 1282 le->opcode |= OP_VLAN;
1282 le->length = cpu_to_be16(vlan_tx_tag_get(skb)); 1283 le->length = cpu_to_be16(vlan_tx_tag_get(skb));
@@ -1313,13 +1314,13 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1313 le->ctrl = ctrl; 1314 le->ctrl = ctrl;
1314 le->opcode = mss ? (OP_LARGESEND | HW_OWNER) : (OP_PACKET | HW_OWNER); 1315 le->opcode = mss ? (OP_LARGESEND | HW_OWNER) : (OP_PACKET | HW_OWNER);
1315 1316
1316 /* Record the transmit mapping info */ 1317 re = tx_le_re(sky2, le);
1317 re->skb = skb; 1318 re->skb = skb;
1318 pci_unmap_addr_set(re, mapaddr, mapping); 1319 pci_unmap_addr_set(re, mapaddr, mapping);
1320 pci_unmap_len_set(re, maplen, len);
1319 1321
1320 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { 1322 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
1321 skb_frag_t *frag = &skb_shinfo(skb)->frags[i]; 1323 const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
1322 struct tx_ring_info *fre;
1323 1324
1324 mapping = pci_map_page(hw->pdev, frag->page, frag->page_offset, 1325 mapping = pci_map_page(hw->pdev, frag->page, frag->page_offset,
1325 frag->size, PCI_DMA_TODEVICE); 1326 frag->size, PCI_DMA_TODEVICE);
@@ -1338,12 +1339,12 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1338 le->ctrl = ctrl; 1339 le->ctrl = ctrl;
1339 le->opcode = OP_BUFFER | HW_OWNER; 1340 le->opcode = OP_BUFFER | HW_OWNER;
1340 1341
1341 fre = sky2->tx_ring 1342 re = tx_le_re(sky2, le);
1342 + RING_NEXT((re - sky2->tx_ring) + i, TX_RING_SIZE); 1343 re->skb = skb;
1343 pci_unmap_addr_set(fre, mapaddr, mapping); 1344 pci_unmap_addr_set(re, mapaddr, mapping);
1345 pci_unmap_len_set(re, maplen, frag->size);
1344 } 1346 }
1345 1347
1346 re->idx = sky2->tx_prod;
1347 le->ctrl |= EOP; 1348 le->ctrl |= EOP;
1348 1349
1349 if (tx_avail(sky2) <= MAX_SKB_TX_LE) 1350 if (tx_avail(sky2) <= MAX_SKB_TX_LE)
@@ -1361,49 +1362,47 @@ static int sky2_xmit_frame(struct sk_buff *skb, struct net_device *dev)
1361 * Free ring elements from starting at tx_cons until "done" 1362 * Free ring elements from starting at tx_cons until "done"
1362 * 1363 *
1363 * NB: the hardware will tell us about partial completion of multi-part 1364 * NB: the hardware will tell us about partial completion of multi-part
1364 * buffers; these are deferred until completion. 1365 * buffers so make sure not to free skb to early.
1365 */ 1366 */
1366static void sky2_tx_complete(struct sky2_port *sky2, u16 done) 1367static void sky2_tx_complete(struct sky2_port *sky2, u16 done)
1367{ 1368{
1368 struct net_device *dev = sky2->netdev; 1369 struct net_device *dev = sky2->netdev;
1369 struct pci_dev *pdev = sky2->hw->pdev; 1370 struct pci_dev *pdev = sky2->hw->pdev;
1370 u16 nxt, put; 1371 unsigned idx;
1371 unsigned i;
1372 1372
1373 BUG_ON(done >= TX_RING_SIZE); 1373 BUG_ON(done >= TX_RING_SIZE);
1374 1374
1375 if (unlikely(netif_msg_tx_done(sky2))) 1375 for (idx = sky2->tx_cons; idx != done;
1376 printk(KERN_DEBUG "%s: tx done, up to %u\n", 1376 idx = RING_NEXT(idx, TX_RING_SIZE)) {
1377 dev->name, done); 1377 struct sky2_tx_le *le = sky2->tx_le + idx;
1378 1378 struct tx_ring_info *re = sky2->tx_ring + idx;
1379 for (put = sky2->tx_cons; put != done; put = nxt) { 1379
1380 struct tx_ring_info *re = sky2->tx_ring + put; 1380 switch(le->opcode & ~HW_OWNER) {
1381 struct sk_buff *skb = re->skb; 1381 case OP_LARGESEND:
1382 1382 case OP_PACKET:
1383 nxt = re->idx; 1383 pci_unmap_single(pdev,
1384 BUG_ON(nxt >= TX_RING_SIZE); 1384 pci_unmap_addr(re, mapaddr),
1385 prefetch(sky2->tx_ring + nxt); 1385 pci_unmap_len(re, maplen),
1386 1386 PCI_DMA_TODEVICE);
1387 /* Check for partial status */
1388 if (tx_dist(put, done) < tx_dist(put, nxt))
1389 break; 1387 break;
1390 1388 case OP_BUFFER:
1391 skb = re->skb; 1389 pci_unmap_page(pdev, pci_unmap_addr(re, mapaddr),
1392 pci_unmap_single(pdev, pci_unmap_addr(re, mapaddr), 1390 pci_unmap_len(re, maplen),
1393 skb_headlen(skb), PCI_DMA_TODEVICE);
1394
1395 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
1396 struct tx_ring_info *fre;
1397 fre = sky2->tx_ring + RING_NEXT(put + i, TX_RING_SIZE);
1398 pci_unmap_page(pdev, pci_unmap_addr(fre, mapaddr),
1399 skb_shinfo(skb)->frags[i].size,
1400 PCI_DMA_TODEVICE); 1391 PCI_DMA_TODEVICE);
1392 break;
1393 }
1394
1395 if (le->ctrl & EOP) {
1396 if (unlikely(netif_msg_tx_done(sky2)))
1397 printk(KERN_DEBUG "%s: tx done %u\n",
1398 dev->name, idx);
1399 dev_kfree_skb(re->skb);
1401 } 1400 }
1402 1401
1403 dev_kfree_skb(skb); 1402 le->opcode = 0; /* paranoia */
1404 } 1403 }
1405 1404
1406 sky2->tx_cons = put; 1405 sky2->tx_cons = idx;
1407 if (tx_avail(sky2) > MAX_SKB_TX_LE + 4) 1406 if (tx_avail(sky2) > MAX_SKB_TX_LE + 4)
1408 netif_wake_queue(dev); 1407 netif_wake_queue(dev);
1409} 1408}
@@ -1843,7 +1842,7 @@ static struct sk_buff *sky2_receive(struct net_device *dev,
1843 u16 length, u32 status) 1842 u16 length, u32 status)
1844{ 1843{
1845 struct sky2_port *sky2 = netdev_priv(dev); 1844 struct sky2_port *sky2 = netdev_priv(dev);
1846 struct ring_info *re = sky2->rx_ring + sky2->rx_next; 1845 struct rx_ring_info *re = sky2->rx_ring + sky2->rx_next;
1847 struct sk_buff *skb = NULL; 1846 struct sk_buff *skb = NULL;
1848 1847
1849 if (unlikely(netif_msg_rx_status(sky2))) 1848 if (unlikely(netif_msg_rx_status(sky2)))
@@ -1889,7 +1888,7 @@ static struct sk_buff *sky2_receive(struct net_device *dev,
1889 prefetch(skb->data); 1888 prefetch(skb->data);
1890 1889
1891 re->mapaddr = pci_map_single(sky2->hw->pdev, nskb->data, 1890 re->mapaddr = pci_map_single(sky2->hw->pdev, nskb->data,
1892 sky2->rx_bufsize, PCI_DMA_FROMDEVICE); 1891 sky2->rx_bufsize, PCI_DMA_FROMDEVICE);
1893 } 1892 }
1894 1893
1895 skb_put(skb, length); 1894 skb_put(skb, length);