diff options
author | Jeff Garzik <jeff@garzik.org> | 2006-09-13 13:24:59 -0400 |
---|---|---|
committer | Jeff Garzik <jeff@garzik.org> | 2006-09-13 13:24:59 -0400 |
commit | 6aa20a2235535605db6d6d2bd850298b2fe7f31e (patch) | |
tree | df0b855043407b831d57f2f2c271f8aab48444f4 /drivers/net/cassini.c | |
parent | 7a291083225af6e22ffaa46b3d91cfc1a1ccaab4 (diff) |
drivers/net: Trim trailing whitespace
Signed-off-by: Jeff Garzik <jeff@garzik.org>
Diffstat (limited to 'drivers/net/cassini.c')
-rw-r--r-- | drivers/net/cassini.c | 536 |
1 files changed, 268 insertions, 268 deletions
diff --git a/drivers/net/cassini.c b/drivers/net/cassini.c index 26040abfef62..7ed376448a68 100644 --- a/drivers/net/cassini.c +++ b/drivers/net/cassini.c | |||
@@ -43,7 +43,7 @@ | |||
43 | * -- on page reclamation, the driver swaps the page with a spare page. | 43 | * -- on page reclamation, the driver swaps the page with a spare page. |
44 | * if that page is still in use, it frees its reference to that page, | 44 | * if that page is still in use, it frees its reference to that page, |
45 | * and allocates a new page for use. otherwise, it just recycles the | 45 | * and allocates a new page for use. otherwise, it just recycles the |
46 | * the page. | 46 | * the page. |
47 | * | 47 | * |
48 | * NOTE: cassini can parse the header. however, it's not worth it | 48 | * NOTE: cassini can parse the header. however, it's not worth it |
49 | * as long as the network stack requires a header copy. | 49 | * as long as the network stack requires a header copy. |
@@ -60,10 +60,10 @@ | |||
60 | * interrupts, but the INT# assignment needs to be set up properly by | 60 | * interrupts, but the INT# assignment needs to be set up properly by |
61 | * the BIOS and conveyed to the driver. PCI BIOSes don't know how to do | 61 | * the BIOS and conveyed to the driver. PCI BIOSes don't know how to do |
62 | * that. also, the two descriptor rings are designed to distinguish between | 62 | * that. also, the two descriptor rings are designed to distinguish between |
63 | * encrypted and non-encrypted packets, but we use them for buffering | 63 | * encrypted and non-encrypted packets, but we use them for buffering |
64 | * instead. | 64 | * instead. |
65 | * | 65 | * |
66 | * by default, the selective clear mask is set up to process rx packets. | 66 | * by default, the selective clear mask is set up to process rx packets. |
67 | */ | 67 | */ |
68 | 68 | ||
69 | 69 | ||
@@ -112,7 +112,7 @@ | |||
112 | #endif | 112 | #endif |
113 | 113 | ||
114 | /* select which firmware to use */ | 114 | /* select which firmware to use */ |
115 | #define USE_HP_WORKAROUND | 115 | #define USE_HP_WORKAROUND |
116 | #define HP_WORKAROUND_DEFAULT /* select which firmware to use as default */ | 116 | #define HP_WORKAROUND_DEFAULT /* select which firmware to use as default */ |
117 | #define CAS_HP_ALT_FIRMWARE cas_prog_null /* alternate firmware */ | 117 | #define CAS_HP_ALT_FIRMWARE cas_prog_null /* alternate firmware */ |
118 | 118 | ||
@@ -168,7 +168,7 @@ | |||
168 | #define STOP_TRIES_PHY 1000 | 168 | #define STOP_TRIES_PHY 1000 |
169 | #define STOP_TRIES 5000 | 169 | #define STOP_TRIES 5000 |
170 | 170 | ||
171 | /* specify a minimum frame size to deal with some fifo issues | 171 | /* specify a minimum frame size to deal with some fifo issues |
172 | * max mtu == 2 * page size - ethernet header - 64 - swivel = | 172 | * max mtu == 2 * page size - ethernet header - 64 - swivel = |
173 | * 2 * page_size - 0x50 | 173 | * 2 * page_size - 0x50 |
174 | */ | 174 | */ |
@@ -207,7 +207,7 @@ MODULE_PARM_DESC(link_mode, "default link mode"); | |||
207 | * being confused and never showing a link status of "up." | 207 | * being confused and never showing a link status of "up." |
208 | */ | 208 | */ |
209 | #define DEFAULT_LINKDOWN_TIMEOUT 5 | 209 | #define DEFAULT_LINKDOWN_TIMEOUT 5 |
210 | /* | 210 | /* |
211 | * Value in seconds, for user input. | 211 | * Value in seconds, for user input. |
212 | */ | 212 | */ |
213 | static int linkdown_timeout = DEFAULT_LINKDOWN_TIMEOUT; | 213 | static int linkdown_timeout = DEFAULT_LINKDOWN_TIMEOUT; |
@@ -249,7 +249,7 @@ static inline void cas_lock_tx(struct cas *cp) | |||
249 | { | 249 | { |
250 | int i; | 250 | int i; |
251 | 251 | ||
252 | for (i = 0; i < N_TX_RINGS; i++) | 252 | for (i = 0; i < N_TX_RINGS; i++) |
253 | spin_lock(&cp->tx_lock[i]); | 253 | spin_lock(&cp->tx_lock[i]); |
254 | } | 254 | } |
255 | 255 | ||
@@ -278,8 +278,8 @@ static inline void cas_unlock_tx(struct cas *cp) | |||
278 | { | 278 | { |
279 | int i; | 279 | int i; |
280 | 280 | ||
281 | for (i = N_TX_RINGS; i > 0; i--) | 281 | for (i = N_TX_RINGS; i > 0; i--) |
282 | spin_unlock(&cp->tx_lock[i - 1]); | 282 | spin_unlock(&cp->tx_lock[i - 1]); |
283 | } | 283 | } |
284 | 284 | ||
285 | static inline void cas_unlock_all(struct cas *cp) | 285 | static inline void cas_unlock_all(struct cas *cp) |
@@ -316,7 +316,7 @@ static void cas_disable_irq(struct cas *cp, const int ring) | |||
316 | #ifdef USE_PCI_INTD | 316 | #ifdef USE_PCI_INTD |
317 | case 3: | 317 | case 3: |
318 | #endif | 318 | #endif |
319 | writel(INTRN_MASK_CLEAR_ALL | INTRN_MASK_RX_EN, | 319 | writel(INTRN_MASK_CLEAR_ALL | INTRN_MASK_RX_EN, |
320 | cp->regs + REG_PLUS_INTRN_MASK(ring)); | 320 | cp->regs + REG_PLUS_INTRN_MASK(ring)); |
321 | break; | 321 | break; |
322 | #endif | 322 | #endif |
@@ -415,7 +415,7 @@ static inline void cas_entropy_reset(struct cas *cp) | |||
415 | if ((cp->cas_flags & CAS_FLAG_ENTROPY_DEV) == 0) | 415 | if ((cp->cas_flags & CAS_FLAG_ENTROPY_DEV) == 0) |
416 | return; | 416 | return; |
417 | 417 | ||
418 | writel(BIM_LOCAL_DEV_PAD | BIM_LOCAL_DEV_PROM | BIM_LOCAL_DEV_EXT, | 418 | writel(BIM_LOCAL_DEV_PAD | BIM_LOCAL_DEV_PROM | BIM_LOCAL_DEV_EXT, |
419 | cp->regs + REG_BIM_LOCAL_DEV_EN); | 419 | cp->regs + REG_BIM_LOCAL_DEV_EN); |
420 | writeb(ENTROPY_RESET_STC_MODE, cp->regs + REG_ENTROPY_RESET); | 420 | writeb(ENTROPY_RESET_STC_MODE, cp->regs + REG_ENTROPY_RESET); |
421 | writeb(0x55, cp->regs + REG_ENTROPY_RAND_REG); | 421 | writeb(0x55, cp->regs + REG_ENTROPY_RAND_REG); |
@@ -426,7 +426,7 @@ static inline void cas_entropy_reset(struct cas *cp) | |||
426 | #endif | 426 | #endif |
427 | } | 427 | } |
428 | 428 | ||
429 | /* access to the phy. the following assumes that we've initialized the MIF to | 429 | /* access to the phy. the following assumes that we've initialized the MIF to |
430 | * be in frame rather than bit-bang mode | 430 | * be in frame rather than bit-bang mode |
431 | */ | 431 | */ |
432 | static u16 cas_phy_read(struct cas *cp, int reg) | 432 | static u16 cas_phy_read(struct cas *cp, int reg) |
@@ -439,7 +439,7 @@ static u16 cas_phy_read(struct cas *cp, int reg) | |||
439 | cmd |= CAS_BASE(MIF_FRAME_REG_ADDR, reg); | 439 | cmd |= CAS_BASE(MIF_FRAME_REG_ADDR, reg); |
440 | cmd |= MIF_FRAME_TURN_AROUND_MSB; | 440 | cmd |= MIF_FRAME_TURN_AROUND_MSB; |
441 | writel(cmd, cp->regs + REG_MIF_FRAME); | 441 | writel(cmd, cp->regs + REG_MIF_FRAME); |
442 | 442 | ||
443 | /* poll for completion */ | 443 | /* poll for completion */ |
444 | while (limit-- > 0) { | 444 | while (limit-- > 0) { |
445 | udelay(10); | 445 | udelay(10); |
@@ -461,7 +461,7 @@ static int cas_phy_write(struct cas *cp, int reg, u16 val) | |||
461 | cmd |= MIF_FRAME_TURN_AROUND_MSB; | 461 | cmd |= MIF_FRAME_TURN_AROUND_MSB; |
462 | cmd |= val & MIF_FRAME_DATA_MASK; | 462 | cmd |= val & MIF_FRAME_DATA_MASK; |
463 | writel(cmd, cp->regs + REG_MIF_FRAME); | 463 | writel(cmd, cp->regs + REG_MIF_FRAME); |
464 | 464 | ||
465 | /* poll for completion */ | 465 | /* poll for completion */ |
466 | while (limit-- > 0) { | 466 | while (limit-- > 0) { |
467 | udelay(10); | 467 | udelay(10); |
@@ -474,7 +474,7 @@ static int cas_phy_write(struct cas *cp, int reg, u16 val) | |||
474 | 474 | ||
475 | static void cas_phy_powerup(struct cas *cp) | 475 | static void cas_phy_powerup(struct cas *cp) |
476 | { | 476 | { |
477 | u16 ctl = cas_phy_read(cp, MII_BMCR); | 477 | u16 ctl = cas_phy_read(cp, MII_BMCR); |
478 | 478 | ||
479 | if ((ctl & BMCR_PDOWN) == 0) | 479 | if ((ctl & BMCR_PDOWN) == 0) |
480 | return; | 480 | return; |
@@ -484,7 +484,7 @@ static void cas_phy_powerup(struct cas *cp) | |||
484 | 484 | ||
485 | static void cas_phy_powerdown(struct cas *cp) | 485 | static void cas_phy_powerdown(struct cas *cp) |
486 | { | 486 | { |
487 | u16 ctl = cas_phy_read(cp, MII_BMCR); | 487 | u16 ctl = cas_phy_read(cp, MII_BMCR); |
488 | 488 | ||
489 | if (ctl & BMCR_PDOWN) | 489 | if (ctl & BMCR_PDOWN) |
490 | return; | 490 | return; |
@@ -495,7 +495,7 @@ static void cas_phy_powerdown(struct cas *cp) | |||
495 | /* cp->lock held. note: the last put_page will free the buffer */ | 495 | /* cp->lock held. note: the last put_page will free the buffer */ |
496 | static int cas_page_free(struct cas *cp, cas_page_t *page) | 496 | static int cas_page_free(struct cas *cp, cas_page_t *page) |
497 | { | 497 | { |
498 | pci_unmap_page(cp->pdev, page->dma_addr, cp->page_size, | 498 | pci_unmap_page(cp->pdev, page->dma_addr, cp->page_size, |
499 | PCI_DMA_FROMDEVICE); | 499 | PCI_DMA_FROMDEVICE); |
500 | cas_buffer_dec(page); | 500 | cas_buffer_dec(page); |
501 | __free_pages(page->buffer, cp->page_order); | 501 | __free_pages(page->buffer, cp->page_order); |
@@ -507,7 +507,7 @@ static int cas_page_free(struct cas *cp, cas_page_t *page) | |||
507 | #define RX_USED_ADD(x, y) ((x)->used += (y)) | 507 | #define RX_USED_ADD(x, y) ((x)->used += (y)) |
508 | #define RX_USED_SET(x, y) ((x)->used = (y)) | 508 | #define RX_USED_SET(x, y) ((x)->used = (y)) |
509 | #else | 509 | #else |
510 | #define RX_USED_ADD(x, y) | 510 | #define RX_USED_ADD(x, y) |
511 | #define RX_USED_SET(x, y) | 511 | #define RX_USED_SET(x, y) |
512 | #endif | 512 | #endif |
513 | 513 | ||
@@ -602,7 +602,7 @@ static void cas_spare_recover(struct cas *cp, const gfp_t flags) | |||
602 | list_splice(&cp->rx_inuse_list, &list); | 602 | list_splice(&cp->rx_inuse_list, &list); |
603 | INIT_LIST_HEAD(&cp->rx_inuse_list); | 603 | INIT_LIST_HEAD(&cp->rx_inuse_list); |
604 | spin_unlock(&cp->rx_inuse_lock); | 604 | spin_unlock(&cp->rx_inuse_lock); |
605 | 605 | ||
606 | list_for_each_safe(elem, tmp, &list) { | 606 | list_for_each_safe(elem, tmp, &list) { |
607 | cas_page_t *page = list_entry(elem, cas_page_t, list); | 607 | cas_page_t *page = list_entry(elem, cas_page_t, list); |
608 | 608 | ||
@@ -627,7 +627,7 @@ static void cas_spare_recover(struct cas *cp, const gfp_t flags) | |||
627 | list_splice(&list, &cp->rx_inuse_list); | 627 | list_splice(&list, &cp->rx_inuse_list); |
628 | spin_unlock(&cp->rx_inuse_lock); | 628 | spin_unlock(&cp->rx_inuse_lock); |
629 | } | 629 | } |
630 | 630 | ||
631 | spin_lock(&cp->rx_spare_lock); | 631 | spin_lock(&cp->rx_spare_lock); |
632 | needed = cp->rx_spares_needed; | 632 | needed = cp->rx_spares_needed; |
633 | spin_unlock(&cp->rx_spare_lock); | 633 | spin_unlock(&cp->rx_spare_lock); |
@@ -639,7 +639,7 @@ static void cas_spare_recover(struct cas *cp, const gfp_t flags) | |||
639 | i = 0; | 639 | i = 0; |
640 | while (i < needed) { | 640 | while (i < needed) { |
641 | cas_page_t *spare = cas_page_alloc(cp, flags); | 641 | cas_page_t *spare = cas_page_alloc(cp, flags); |
642 | if (!spare) | 642 | if (!spare) |
643 | break; | 643 | break; |
644 | list_add(&spare->list, &list); | 644 | list_add(&spare->list, &list); |
645 | i++; | 645 | i++; |
@@ -695,12 +695,12 @@ static cas_page_t *cas_page_dequeue(struct cas *cp) | |||
695 | static void cas_mif_poll(struct cas *cp, const int enable) | 695 | static void cas_mif_poll(struct cas *cp, const int enable) |
696 | { | 696 | { |
697 | u32 cfg; | 697 | u32 cfg; |
698 | 698 | ||
699 | cfg = readl(cp->regs + REG_MIF_CFG); | 699 | cfg = readl(cp->regs + REG_MIF_CFG); |
700 | cfg &= (MIF_CFG_MDIO_0 | MIF_CFG_MDIO_1); | 700 | cfg &= (MIF_CFG_MDIO_0 | MIF_CFG_MDIO_1); |
701 | 701 | ||
702 | if (cp->phy_type & CAS_PHY_MII_MDIO1) | 702 | if (cp->phy_type & CAS_PHY_MII_MDIO1) |
703 | cfg |= MIF_CFG_PHY_SELECT; | 703 | cfg |= MIF_CFG_PHY_SELECT; |
704 | 704 | ||
705 | /* poll and interrupt on link status change. */ | 705 | /* poll and interrupt on link status change. */ |
706 | if (enable) { | 706 | if (enable) { |
@@ -708,8 +708,8 @@ static void cas_mif_poll(struct cas *cp, const int enable) | |||
708 | cfg |= CAS_BASE(MIF_CFG_POLL_REG, MII_BMSR); | 708 | cfg |= CAS_BASE(MIF_CFG_POLL_REG, MII_BMSR); |
709 | cfg |= CAS_BASE(MIF_CFG_POLL_PHY, cp->phy_addr); | 709 | cfg |= CAS_BASE(MIF_CFG_POLL_PHY, cp->phy_addr); |
710 | } | 710 | } |
711 | writel((enable) ? ~(BMSR_LSTATUS | BMSR_ANEGCOMPLETE) : 0xFFFF, | 711 | writel((enable) ? ~(BMSR_LSTATUS | BMSR_ANEGCOMPLETE) : 0xFFFF, |
712 | cp->regs + REG_MIF_MASK); | 712 | cp->regs + REG_MIF_MASK); |
713 | writel(cfg, cp->regs + REG_MIF_CFG); | 713 | writel(cfg, cp->regs + REG_MIF_CFG); |
714 | } | 714 | } |
715 | 715 | ||
@@ -759,7 +759,7 @@ start_aneg: | |||
759 | /* | 759 | /* |
760 | * WTZ: If the old state was link_up, we turn off the carrier | 760 | * WTZ: If the old state was link_up, we turn off the carrier |
761 | * to replicate everything we do elsewhere on a link-down | 761 | * to replicate everything we do elsewhere on a link-down |
762 | * event when we were already in a link-up state.. | 762 | * event when we were already in a link-up state.. |
763 | */ | 763 | */ |
764 | if (oldstate == link_up) | 764 | if (oldstate == link_up) |
765 | netif_carrier_off(cp->dev); | 765 | netif_carrier_off(cp->dev); |
@@ -767,7 +767,7 @@ start_aneg: | |||
767 | /* | 767 | /* |
768 | * WTZ: This branch will simply schedule a full reset after | 768 | * WTZ: This branch will simply schedule a full reset after |
769 | * we explicitly changed link modes in an ioctl. See if this | 769 | * we explicitly changed link modes in an ioctl. See if this |
770 | * fixes the link-problems we were having for forced mode. | 770 | * fixes the link-problems we were having for forced mode. |
771 | */ | 771 | */ |
772 | atomic_inc(&cp->reset_task_pending); | 772 | atomic_inc(&cp->reset_task_pending); |
773 | atomic_inc(&cp->reset_task_pending_all); | 773 | atomic_inc(&cp->reset_task_pending_all); |
@@ -795,7 +795,7 @@ start_aneg: | |||
795 | } else { | 795 | } else { |
796 | cas_mif_poll(cp, 0); | 796 | cas_mif_poll(cp, 0); |
797 | ctl = cas_phy_read(cp, MII_BMCR); | 797 | ctl = cas_phy_read(cp, MII_BMCR); |
798 | ctl &= ~(BMCR_FULLDPLX | BMCR_SPEED100 | | 798 | ctl &= ~(BMCR_FULLDPLX | BMCR_SPEED100 | |
799 | CAS_BMCR_SPEED1000 | BMCR_ANENABLE); | 799 | CAS_BMCR_SPEED1000 | BMCR_ANENABLE); |
800 | ctl |= cp->link_cntl; | 800 | ctl |= cp->link_cntl; |
801 | if (ctl & BMCR_ANENABLE) { | 801 | if (ctl & BMCR_ANENABLE) { |
@@ -818,7 +818,7 @@ static int cas_reset_mii_phy(struct cas *cp) | |||
818 | { | 818 | { |
819 | int limit = STOP_TRIES_PHY; | 819 | int limit = STOP_TRIES_PHY; |
820 | u16 val; | 820 | u16 val; |
821 | 821 | ||
822 | cas_phy_write(cp, MII_BMCR, BMCR_RESET); | 822 | cas_phy_write(cp, MII_BMCR, BMCR_RESET); |
823 | udelay(100); | 823 | udelay(100); |
824 | while (limit--) { | 824 | while (limit--) { |
@@ -901,17 +901,17 @@ static void cas_phy_init(struct cas *cp) | |||
901 | val = cas_phy_read(cp, BROADCOM_MII_REG4); | 901 | val = cas_phy_read(cp, BROADCOM_MII_REG4); |
902 | if (val & 0x0080) { | 902 | if (val & 0x0080) { |
903 | /* link workaround */ | 903 | /* link workaround */ |
904 | cas_phy_write(cp, BROADCOM_MII_REG4, | 904 | cas_phy_write(cp, BROADCOM_MII_REG4, |
905 | val & ~0x0080); | 905 | val & ~0x0080); |
906 | } | 906 | } |
907 | 907 | ||
908 | } else if (cp->cas_flags & CAS_FLAG_SATURN) { | 908 | } else if (cp->cas_flags & CAS_FLAG_SATURN) { |
909 | writel((cp->phy_type & CAS_PHY_MII_MDIO0) ? | 909 | writel((cp->phy_type & CAS_PHY_MII_MDIO0) ? |
910 | SATURN_PCFG_FSI : 0x0, | 910 | SATURN_PCFG_FSI : 0x0, |
911 | cp->regs + REG_SATURN_PCFG); | 911 | cp->regs + REG_SATURN_PCFG); |
912 | 912 | ||
913 | /* load firmware to address 10Mbps auto-negotiation | 913 | /* load firmware to address 10Mbps auto-negotiation |
914 | * issue. NOTE: this will need to be changed if the | 914 | * issue. NOTE: this will need to be changed if the |
915 | * default firmware gets fixed. | 915 | * default firmware gets fixed. |
916 | */ | 916 | */ |
917 | if (PHY_NS_DP83065 == cp->phy_id) { | 917 | if (PHY_NS_DP83065 == cp->phy_id) { |
@@ -930,9 +930,9 @@ static void cas_phy_init(struct cas *cp) | |||
930 | cas_phy_read(cp, MII_ADVERTISE) | | 930 | cas_phy_read(cp, MII_ADVERTISE) | |
931 | (ADVERTISE_10HALF | ADVERTISE_10FULL | | 931 | (ADVERTISE_10HALF | ADVERTISE_10FULL | |
932 | ADVERTISE_100HALF | ADVERTISE_100FULL | | 932 | ADVERTISE_100HALF | ADVERTISE_100FULL | |
933 | CAS_ADVERTISE_PAUSE | | 933 | CAS_ADVERTISE_PAUSE | |
934 | CAS_ADVERTISE_ASYM_PAUSE)); | 934 | CAS_ADVERTISE_ASYM_PAUSE)); |
935 | 935 | ||
936 | if (cp->cas_flags & CAS_FLAG_1000MB_CAP) { | 936 | if (cp->cas_flags & CAS_FLAG_1000MB_CAP) { |
937 | /* make sure that we don't advertise half | 937 | /* make sure that we don't advertise half |
938 | * duplex to avoid a chip issue | 938 | * duplex to avoid a chip issue |
@@ -963,7 +963,7 @@ static void cas_phy_init(struct cas *cp) | |||
963 | limit = STOP_TRIES; | 963 | limit = STOP_TRIES; |
964 | while (limit-- > 0) { | 964 | while (limit-- > 0) { |
965 | udelay(10); | 965 | udelay(10); |
966 | if ((readl(cp->regs + REG_PCS_MII_CTRL) & | 966 | if ((readl(cp->regs + REG_PCS_MII_CTRL) & |
967 | PCS_MII_RESET) == 0) | 967 | PCS_MII_RESET) == 0) |
968 | break; | 968 | break; |
969 | } | 969 | } |
@@ -980,7 +980,7 @@ static void cas_phy_init(struct cas *cp) | |||
980 | /* Advertise all capabilities except half-duplex. */ | 980 | /* Advertise all capabilities except half-duplex. */ |
981 | val = readl(cp->regs + REG_PCS_MII_ADVERT); | 981 | val = readl(cp->regs + REG_PCS_MII_ADVERT); |
982 | val &= ~PCS_MII_ADVERT_HD; | 982 | val &= ~PCS_MII_ADVERT_HD; |
983 | val |= (PCS_MII_ADVERT_FD | PCS_MII_ADVERT_SYM_PAUSE | | 983 | val |= (PCS_MII_ADVERT_FD | PCS_MII_ADVERT_SYM_PAUSE | |
984 | PCS_MII_ADVERT_ASYM_PAUSE); | 984 | PCS_MII_ADVERT_ASYM_PAUSE); |
985 | writel(val, cp->regs + REG_PCS_MII_ADVERT); | 985 | writel(val, cp->regs + REG_PCS_MII_ADVERT); |
986 | 986 | ||
@@ -1014,7 +1014,7 @@ static int cas_pcs_link_check(struct cas *cp) | |||
1014 | PCS_MII_STATUS_REMOTE_FAULT)) == | 1014 | PCS_MII_STATUS_REMOTE_FAULT)) == |
1015 | (PCS_MII_STATUS_AUTONEG_COMP | PCS_MII_STATUS_REMOTE_FAULT)) { | 1015 | (PCS_MII_STATUS_AUTONEG_COMP | PCS_MII_STATUS_REMOTE_FAULT)) { |
1016 | if (netif_msg_link(cp)) | 1016 | if (netif_msg_link(cp)) |
1017 | printk(KERN_INFO "%s: PCS RemoteFault\n", | 1017 | printk(KERN_INFO "%s: PCS RemoteFault\n", |
1018 | cp->dev->name); | 1018 | cp->dev->name); |
1019 | } | 1019 | } |
1020 | 1020 | ||
@@ -1033,7 +1033,7 @@ static int cas_pcs_link_check(struct cas *cp) | |||
1033 | if (cp->opened) { | 1033 | if (cp->opened) { |
1034 | cp->lstate = link_up; | 1034 | cp->lstate = link_up; |
1035 | cp->link_transition = LINK_TRANSITION_LINK_UP; | 1035 | cp->link_transition = LINK_TRANSITION_LINK_UP; |
1036 | 1036 | ||
1037 | cas_set_link_modes(cp); | 1037 | cas_set_link_modes(cp); |
1038 | netif_carrier_on(cp->dev); | 1038 | netif_carrier_on(cp->dev); |
1039 | } | 1039 | } |
@@ -1044,8 +1044,8 @@ static int cas_pcs_link_check(struct cas *cp) | |||
1044 | cp->link_transition != LINK_TRANSITION_REQUESTED_RESET && | 1044 | cp->link_transition != LINK_TRANSITION_REQUESTED_RESET && |
1045 | !cp->link_transition_jiffies_valid) { | 1045 | !cp->link_transition_jiffies_valid) { |
1046 | /* | 1046 | /* |
1047 | * force a reset, as a workaround for the | 1047 | * force a reset, as a workaround for the |
1048 | * link-failure problem. May want to move this to a | 1048 | * link-failure problem. May want to move this to a |
1049 | * point a bit earlier in the sequence. If we had | 1049 | * point a bit earlier in the sequence. If we had |
1050 | * generated a reset a short time ago, we'll wait for | 1050 | * generated a reset a short time ago, we'll wait for |
1051 | * the link timer to check the status until a | 1051 | * the link timer to check the status until a |
@@ -1103,17 +1103,17 @@ static int cas_pcs_link_check(struct cas *cp) | |||
1103 | return retval; | 1103 | return retval; |
1104 | } | 1104 | } |
1105 | 1105 | ||
1106 | static int cas_pcs_interrupt(struct net_device *dev, | 1106 | static int cas_pcs_interrupt(struct net_device *dev, |
1107 | struct cas *cp, u32 status) | 1107 | struct cas *cp, u32 status) |
1108 | { | 1108 | { |
1109 | u32 stat = readl(cp->regs + REG_PCS_INTR_STATUS); | 1109 | u32 stat = readl(cp->regs + REG_PCS_INTR_STATUS); |
1110 | 1110 | ||
1111 | if ((stat & PCS_INTR_STATUS_LINK_CHANGE) == 0) | 1111 | if ((stat & PCS_INTR_STATUS_LINK_CHANGE) == 0) |
1112 | return 0; | 1112 | return 0; |
1113 | return cas_pcs_link_check(cp); | 1113 | return cas_pcs_link_check(cp); |
1114 | } | 1114 | } |
1115 | 1115 | ||
1116 | static int cas_txmac_interrupt(struct net_device *dev, | 1116 | static int cas_txmac_interrupt(struct net_device *dev, |
1117 | struct cas *cp, u32 status) | 1117 | struct cas *cp, u32 status) |
1118 | { | 1118 | { |
1119 | u32 txmac_stat = readl(cp->regs + REG_MAC_TX_STATUS); | 1119 | u32 txmac_stat = readl(cp->regs + REG_MAC_TX_STATUS); |
@@ -1168,7 +1168,7 @@ static int cas_txmac_interrupt(struct net_device *dev, | |||
1168 | return 0; | 1168 | return 0; |
1169 | } | 1169 | } |
1170 | 1170 | ||
1171 | static void cas_load_firmware(struct cas *cp, cas_hp_inst_t *firmware) | 1171 | static void cas_load_firmware(struct cas *cp, cas_hp_inst_t *firmware) |
1172 | { | 1172 | { |
1173 | cas_hp_inst_t *inst; | 1173 | cas_hp_inst_t *inst; |
1174 | u32 val; | 1174 | u32 val; |
@@ -1203,12 +1203,12 @@ static void cas_load_firmware(struct cas *cp, cas_hp_inst_t *firmware) | |||
1203 | 1203 | ||
1204 | static void cas_init_rx_dma(struct cas *cp) | 1204 | static void cas_init_rx_dma(struct cas *cp) |
1205 | { | 1205 | { |
1206 | u64 desc_dma = cp->block_dvma; | 1206 | u64 desc_dma = cp->block_dvma; |
1207 | u32 val; | 1207 | u32 val; |
1208 | int i, size; | 1208 | int i, size; |
1209 | 1209 | ||
1210 | /* rx free descriptors */ | 1210 | /* rx free descriptors */ |
1211 | val = CAS_BASE(RX_CFG_SWIVEL, RX_SWIVEL_OFF_VAL); | 1211 | val = CAS_BASE(RX_CFG_SWIVEL, RX_SWIVEL_OFF_VAL); |
1212 | val |= CAS_BASE(RX_CFG_DESC_RING, RX_DESC_RINGN_INDEX(0)); | 1212 | val |= CAS_BASE(RX_CFG_DESC_RING, RX_DESC_RINGN_INDEX(0)); |
1213 | val |= CAS_BASE(RX_CFG_COMP_RING, RX_COMP_RINGN_INDEX(0)); | 1213 | val |= CAS_BASE(RX_CFG_COMP_RING, RX_COMP_RINGN_INDEX(0)); |
1214 | if ((N_RX_DESC_RINGS > 1) && | 1214 | if ((N_RX_DESC_RINGS > 1) && |
@@ -1216,27 +1216,27 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1216 | val |= CAS_BASE(RX_CFG_DESC_RING1, RX_DESC_RINGN_INDEX(1)); | 1216 | val |= CAS_BASE(RX_CFG_DESC_RING1, RX_DESC_RINGN_INDEX(1)); |
1217 | writel(val, cp->regs + REG_RX_CFG); | 1217 | writel(val, cp->regs + REG_RX_CFG); |
1218 | 1218 | ||
1219 | val = (unsigned long) cp->init_rxds[0] - | 1219 | val = (unsigned long) cp->init_rxds[0] - |
1220 | (unsigned long) cp->init_block; | 1220 | (unsigned long) cp->init_block; |
1221 | writel((desc_dma + val) >> 32, cp->regs + REG_RX_DB_HI); | 1221 | writel((desc_dma + val) >> 32, cp->regs + REG_RX_DB_HI); |
1222 | writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_DB_LOW); | 1222 | writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_DB_LOW); |
1223 | writel(RX_DESC_RINGN_SIZE(0) - 4, cp->regs + REG_RX_KICK); | 1223 | writel(RX_DESC_RINGN_SIZE(0) - 4, cp->regs + REG_RX_KICK); |
1224 | 1224 | ||
1225 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { | 1225 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { |
1226 | /* rx desc 2 is for IPSEC packets. however, | 1226 | /* rx desc 2 is for IPSEC packets. however, |
1227 | * we don't it that for that purpose. | 1227 | * we don't it that for that purpose. |
1228 | */ | 1228 | */ |
1229 | val = (unsigned long) cp->init_rxds[1] - | 1229 | val = (unsigned long) cp->init_rxds[1] - |
1230 | (unsigned long) cp->init_block; | 1230 | (unsigned long) cp->init_block; |
1231 | writel((desc_dma + val) >> 32, cp->regs + REG_PLUS_RX_DB1_HI); | 1231 | writel((desc_dma + val) >> 32, cp->regs + REG_PLUS_RX_DB1_HI); |
1232 | writel((desc_dma + val) & 0xffffffff, cp->regs + | 1232 | writel((desc_dma + val) & 0xffffffff, cp->regs + |
1233 | REG_PLUS_RX_DB1_LOW); | 1233 | REG_PLUS_RX_DB1_LOW); |
1234 | writel(RX_DESC_RINGN_SIZE(1) - 4, cp->regs + | 1234 | writel(RX_DESC_RINGN_SIZE(1) - 4, cp->regs + |
1235 | REG_PLUS_RX_KICK1); | 1235 | REG_PLUS_RX_KICK1); |
1236 | } | 1236 | } |
1237 | 1237 | ||
1238 | /* rx completion registers */ | 1238 | /* rx completion registers */ |
1239 | val = (unsigned long) cp->init_rxcs[0] - | 1239 | val = (unsigned long) cp->init_rxcs[0] - |
1240 | (unsigned long) cp->init_block; | 1240 | (unsigned long) cp->init_block; |
1241 | writel((desc_dma + val) >> 32, cp->regs + REG_RX_CB_HI); | 1241 | writel((desc_dma + val) >> 32, cp->regs + REG_RX_CB_HI); |
1242 | writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_CB_LOW); | 1242 | writel((desc_dma + val) & 0xffffffff, cp->regs + REG_RX_CB_LOW); |
@@ -1244,11 +1244,11 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1244 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { | 1244 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { |
1245 | /* rx comp 2-4 */ | 1245 | /* rx comp 2-4 */ |
1246 | for (i = 1; i < MAX_RX_COMP_RINGS; i++) { | 1246 | for (i = 1; i < MAX_RX_COMP_RINGS; i++) { |
1247 | val = (unsigned long) cp->init_rxcs[i] - | 1247 | val = (unsigned long) cp->init_rxcs[i] - |
1248 | (unsigned long) cp->init_block; | 1248 | (unsigned long) cp->init_block; |
1249 | writel((desc_dma + val) >> 32, cp->regs + | 1249 | writel((desc_dma + val) >> 32, cp->regs + |
1250 | REG_PLUS_RX_CBN_HI(i)); | 1250 | REG_PLUS_RX_CBN_HI(i)); |
1251 | writel((desc_dma + val) & 0xffffffff, cp->regs + | 1251 | writel((desc_dma + val) & 0xffffffff, cp->regs + |
1252 | REG_PLUS_RX_CBN_LOW(i)); | 1252 | REG_PLUS_RX_CBN_LOW(i)); |
1253 | } | 1253 | } |
1254 | } | 1254 | } |
@@ -1265,21 +1265,21 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1265 | 1265 | ||
1266 | /* 2 is different from 3 and 4 */ | 1266 | /* 2 is different from 3 and 4 */ |
1267 | if (N_RX_COMP_RINGS > 1) | 1267 | if (N_RX_COMP_RINGS > 1) |
1268 | writel(INTR_RX_DONE_ALT | INTR_RX_BUF_UNAVAIL_1, | 1268 | writel(INTR_RX_DONE_ALT | INTR_RX_BUF_UNAVAIL_1, |
1269 | cp->regs + REG_PLUS_ALIASN_CLEAR(1)); | 1269 | cp->regs + REG_PLUS_ALIASN_CLEAR(1)); |
1270 | 1270 | ||
1271 | for (i = 2; i < N_RX_COMP_RINGS; i++) | 1271 | for (i = 2; i < N_RX_COMP_RINGS; i++) |
1272 | writel(INTR_RX_DONE_ALT, | 1272 | writel(INTR_RX_DONE_ALT, |
1273 | cp->regs + REG_PLUS_ALIASN_CLEAR(i)); | 1273 | cp->regs + REG_PLUS_ALIASN_CLEAR(i)); |
1274 | } | 1274 | } |
1275 | 1275 | ||
1276 | /* set up pause thresholds */ | 1276 | /* set up pause thresholds */ |
1277 | val = CAS_BASE(RX_PAUSE_THRESH_OFF, | 1277 | val = CAS_BASE(RX_PAUSE_THRESH_OFF, |
1278 | cp->rx_pause_off / RX_PAUSE_THRESH_QUANTUM); | 1278 | cp->rx_pause_off / RX_PAUSE_THRESH_QUANTUM); |
1279 | val |= CAS_BASE(RX_PAUSE_THRESH_ON, | 1279 | val |= CAS_BASE(RX_PAUSE_THRESH_ON, |
1280 | cp->rx_pause_on / RX_PAUSE_THRESH_QUANTUM); | 1280 | cp->rx_pause_on / RX_PAUSE_THRESH_QUANTUM); |
1281 | writel(val, cp->regs + REG_RX_PAUSE_THRESH); | 1281 | writel(val, cp->regs + REG_RX_PAUSE_THRESH); |
1282 | 1282 | ||
1283 | /* zero out dma reassembly buffers */ | 1283 | /* zero out dma reassembly buffers */ |
1284 | for (i = 0; i < 64; i++) { | 1284 | for (i = 0; i < 64; i++) { |
1285 | writel(i, cp->regs + REG_RX_TABLE_ADDR); | 1285 | writel(i, cp->regs + REG_RX_TABLE_ADDR); |
@@ -1318,7 +1318,7 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1318 | * this should be tunable. | 1318 | * this should be tunable. |
1319 | */ | 1319 | */ |
1320 | writel(0x0, cp->regs + REG_RX_RED); | 1320 | writel(0x0, cp->regs + REG_RX_RED); |
1321 | 1321 | ||
1322 | /* receive page sizes. default == 2K (0x800) */ | 1322 | /* receive page sizes. default == 2K (0x800) */ |
1323 | val = 0; | 1323 | val = 0; |
1324 | if (cp->page_size == 0x1000) | 1324 | if (cp->page_size == 0x1000) |
@@ -1327,7 +1327,7 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1327 | val = 0x2; | 1327 | val = 0x2; |
1328 | else if (cp->page_size == 0x4000) | 1328 | else if (cp->page_size == 0x4000) |
1329 | val = 0x3; | 1329 | val = 0x3; |
1330 | 1330 | ||
1331 | /* round mtu + offset. constrain to page size. */ | 1331 | /* round mtu + offset. constrain to page size. */ |
1332 | size = cp->dev->mtu + 64; | 1332 | size = cp->dev->mtu + 64; |
1333 | if (size > cp->page_size) | 1333 | if (size > cp->page_size) |
@@ -1344,11 +1344,11 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1344 | 1344 | ||
1345 | cp->mtu_stride = 1 << (i + 10); | 1345 | cp->mtu_stride = 1 << (i + 10); |
1346 | val = CAS_BASE(RX_PAGE_SIZE, val); | 1346 | val = CAS_BASE(RX_PAGE_SIZE, val); |
1347 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_STRIDE, i); | 1347 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_STRIDE, i); |
1348 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_COUNT, cp->page_size >> (i + 10)); | 1348 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_COUNT, cp->page_size >> (i + 10)); |
1349 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_OFF, 0x1); | 1349 | val |= CAS_BASE(RX_PAGE_SIZE_MTU_OFF, 0x1); |
1350 | writel(val, cp->regs + REG_RX_PAGE_SIZE); | 1350 | writel(val, cp->regs + REG_RX_PAGE_SIZE); |
1351 | 1351 | ||
1352 | /* enable the header parser if desired */ | 1352 | /* enable the header parser if desired */ |
1353 | if (CAS_HP_FIRMWARE == cas_prog_null) | 1353 | if (CAS_HP_FIRMWARE == cas_prog_null) |
1354 | return; | 1354 | return; |
@@ -1362,7 +1362,7 @@ static void cas_init_rx_dma(struct cas *cp) | |||
1362 | static inline void cas_rxc_init(struct cas_rx_comp *rxc) | 1362 | static inline void cas_rxc_init(struct cas_rx_comp *rxc) |
1363 | { | 1363 | { |
1364 | memset(rxc, 0, sizeof(*rxc)); | 1364 | memset(rxc, 0, sizeof(*rxc)); |
1365 | rxc->word4 = cpu_to_le64(RX_COMP4_ZERO); | 1365 | rxc->word4 = cpu_to_le64(RX_COMP4_ZERO); |
1366 | } | 1366 | } |
1367 | 1367 | ||
1368 | /* NOTE: we use the ENC RX DESC ring for spares. the rx_page[0,1] | 1368 | /* NOTE: we use the ENC RX DESC ring for spares. the rx_page[0,1] |
@@ -1385,9 +1385,9 @@ static inline cas_page_t *cas_page_spare(struct cas *cp, const int index) | |||
1385 | } | 1385 | } |
1386 | return new; | 1386 | return new; |
1387 | } | 1387 | } |
1388 | 1388 | ||
1389 | /* this needs to be changed if we actually use the ENC RX DESC ring */ | 1389 | /* this needs to be changed if we actually use the ENC RX DESC ring */ |
1390 | static cas_page_t *cas_page_swap(struct cas *cp, const int ring, | 1390 | static cas_page_t *cas_page_swap(struct cas *cp, const int ring, |
1391 | const int index) | 1391 | const int index) |
1392 | { | 1392 | { |
1393 | cas_page_t **page0 = cp->rx_pages[0]; | 1393 | cas_page_t **page0 = cp->rx_pages[0]; |
@@ -1400,7 +1400,7 @@ static cas_page_t *cas_page_swap(struct cas *cp, const int ring, | |||
1400 | page1[index] = page0[index]; | 1400 | page1[index] = page0[index]; |
1401 | page0[index] = new; | 1401 | page0[index] = new; |
1402 | } | 1402 | } |
1403 | } | 1403 | } |
1404 | RX_USED_SET(page0[index], 0); | 1404 | RX_USED_SET(page0[index], 0); |
1405 | return page0[index]; | 1405 | return page0[index]; |
1406 | } | 1406 | } |
@@ -1424,11 +1424,11 @@ static void cas_clean_rxds(struct cas *cp) | |||
1424 | for (i = 0; i < size; i++) { | 1424 | for (i = 0; i < size; i++) { |
1425 | cas_page_t *page = cas_page_swap(cp, 0, i); | 1425 | cas_page_t *page = cas_page_swap(cp, 0, i); |
1426 | rxd[i].buffer = cpu_to_le64(page->dma_addr); | 1426 | rxd[i].buffer = cpu_to_le64(page->dma_addr); |
1427 | rxd[i].index = cpu_to_le64(CAS_BASE(RX_INDEX_NUM, i) | | 1427 | rxd[i].index = cpu_to_le64(CAS_BASE(RX_INDEX_NUM, i) | |
1428 | CAS_BASE(RX_INDEX_RING, 0)); | 1428 | CAS_BASE(RX_INDEX_RING, 0)); |
1429 | } | 1429 | } |
1430 | 1430 | ||
1431 | cp->rx_old[0] = RX_DESC_RINGN_SIZE(0) - 4; | 1431 | cp->rx_old[0] = RX_DESC_RINGN_SIZE(0) - 4; |
1432 | cp->rx_last[0] = 0; | 1432 | cp->rx_last[0] = 0; |
1433 | cp->cas_flags &= ~CAS_FLAG_RXD_POST(0); | 1433 | cp->cas_flags &= ~CAS_FLAG_RXD_POST(0); |
1434 | } | 1434 | } |
@@ -1533,7 +1533,7 @@ static int cas_rxmac_interrupt(struct net_device *dev, struct cas *cp, | |||
1533 | 1533 | ||
1534 | /* these are all rollovers */ | 1534 | /* these are all rollovers */ |
1535 | spin_lock(&cp->stat_lock[0]); | 1535 | spin_lock(&cp->stat_lock[0]); |
1536 | if (stat & MAC_RX_ALIGN_ERR) | 1536 | if (stat & MAC_RX_ALIGN_ERR) |
1537 | cp->net_stats[0].rx_frame_errors += 0x10000; | 1537 | cp->net_stats[0].rx_frame_errors += 0x10000; |
1538 | 1538 | ||
1539 | if (stat & MAC_RX_CRC_ERR) | 1539 | if (stat & MAC_RX_CRC_ERR) |
@@ -1579,12 +1579,12 @@ static int cas_mac_interrupt(struct net_device *dev, struct cas *cp, | |||
1579 | return 0; | 1579 | return 0; |
1580 | } | 1580 | } |
1581 | 1581 | ||
1582 | 1582 | ||
1583 | /* Must be invoked under cp->lock. */ | 1583 | /* Must be invoked under cp->lock. */ |
1584 | static inline int cas_mdio_link_not_up(struct cas *cp) | 1584 | static inline int cas_mdio_link_not_up(struct cas *cp) |
1585 | { | 1585 | { |
1586 | u16 val; | 1586 | u16 val; |
1587 | 1587 | ||
1588 | switch (cp->lstate) { | 1588 | switch (cp->lstate) { |
1589 | case link_force_ret: | 1589 | case link_force_ret: |
1590 | if (netif_msg_link(cp)) | 1590 | if (netif_msg_link(cp)) |
@@ -1595,7 +1595,7 @@ static inline int cas_mdio_link_not_up(struct cas *cp) | |||
1595 | cp->lstate = link_force_ok; | 1595 | cp->lstate = link_force_ok; |
1596 | cp->link_transition = LINK_TRANSITION_LINK_CONFIG; | 1596 | cp->link_transition = LINK_TRANSITION_LINK_CONFIG; |
1597 | break; | 1597 | break; |
1598 | 1598 | ||
1599 | case link_aneg: | 1599 | case link_aneg: |
1600 | val = cas_phy_read(cp, MII_BMCR); | 1600 | val = cas_phy_read(cp, MII_BMCR); |
1601 | 1601 | ||
@@ -1604,7 +1604,7 @@ static inline int cas_mdio_link_not_up(struct cas *cp) | |||
1604 | */ | 1604 | */ |
1605 | val &= ~(BMCR_ANRESTART | BMCR_ANENABLE); | 1605 | val &= ~(BMCR_ANRESTART | BMCR_ANENABLE); |
1606 | val |= BMCR_FULLDPLX; | 1606 | val |= BMCR_FULLDPLX; |
1607 | val |= (cp->cas_flags & CAS_FLAG_1000MB_CAP) ? | 1607 | val |= (cp->cas_flags & CAS_FLAG_1000MB_CAP) ? |
1608 | CAS_BMCR_SPEED1000 : BMCR_SPEED100; | 1608 | CAS_BMCR_SPEED1000 : BMCR_SPEED100; |
1609 | cas_phy_write(cp, MII_BMCR, val); | 1609 | cas_phy_write(cp, MII_BMCR, val); |
1610 | cp->timer_ticks = 5; | 1610 | cp->timer_ticks = 5; |
@@ -1646,11 +1646,11 @@ static int cas_mii_link_check(struct cas *cp, const u16 bmsr) | |||
1646 | 1646 | ||
1647 | if (bmsr & BMSR_LSTATUS) { | 1647 | if (bmsr & BMSR_LSTATUS) { |
1648 | /* Ok, here we got a link. If we had it due to a forced | 1648 | /* Ok, here we got a link. If we had it due to a forced |
1649 | * fallback, and we were configured for autoneg, we | 1649 | * fallback, and we were configured for autoneg, we |
1650 | * retry a short autoneg pass. If you know your hub is | 1650 | * retry a short autoneg pass. If you know your hub is |
1651 | * broken, use ethtool ;) | 1651 | * broken, use ethtool ;) |
1652 | */ | 1652 | */ |
1653 | if ((cp->lstate == link_force_try) && | 1653 | if ((cp->lstate == link_force_try) && |
1654 | (cp->link_cntl & BMCR_ANENABLE)) { | 1654 | (cp->link_cntl & BMCR_ANENABLE)) { |
1655 | cp->lstate = link_force_ret; | 1655 | cp->lstate = link_force_ret; |
1656 | cp->link_transition = LINK_TRANSITION_LINK_CONFIG; | 1656 | cp->link_transition = LINK_TRANSITION_LINK_CONFIG; |
@@ -1690,10 +1690,10 @@ static int cas_mii_link_check(struct cas *cp, const u16 bmsr) | |||
1690 | printk(KERN_INFO "%s: Link down\n", | 1690 | printk(KERN_INFO "%s: Link down\n", |
1691 | cp->dev->name); | 1691 | cp->dev->name); |
1692 | restart = 1; | 1692 | restart = 1; |
1693 | 1693 | ||
1694 | } else if (++cp->timer_ticks > 10) | 1694 | } else if (++cp->timer_ticks > 10) |
1695 | cas_mdio_link_not_up(cp); | 1695 | cas_mdio_link_not_up(cp); |
1696 | 1696 | ||
1697 | return restart; | 1697 | return restart; |
1698 | } | 1698 | } |
1699 | 1699 | ||
@@ -1908,7 +1908,7 @@ static inline void cas_tx_ringN(struct cas *cp, int ring, int limit) | |||
1908 | 1908 | ||
1909 | skbs[entry] = NULL; | 1909 | skbs[entry] = NULL; |
1910 | cp->tx_tiny_use[ring][entry].nbufs = 0; | 1910 | cp->tx_tiny_use[ring][entry].nbufs = 0; |
1911 | 1911 | ||
1912 | for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { | 1912 | for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) { |
1913 | struct cas_tx_desc *txd = txds + entry; | 1913 | struct cas_tx_desc *txd = txds + entry; |
1914 | 1914 | ||
@@ -1923,7 +1923,7 @@ static inline void cas_tx_ringN(struct cas *cp, int ring, int limit) | |||
1923 | if (cp->tx_tiny_use[ring][entry].used) { | 1923 | if (cp->tx_tiny_use[ring][entry].used) { |
1924 | cp->tx_tiny_use[ring][entry].used = 0; | 1924 | cp->tx_tiny_use[ring][entry].used = 0; |
1925 | entry = TX_DESC_NEXT(ring, entry); | 1925 | entry = TX_DESC_NEXT(ring, entry); |
1926 | } | 1926 | } |
1927 | } | 1927 | } |
1928 | 1928 | ||
1929 | spin_lock(&cp->stat_lock[ring]); | 1929 | spin_lock(&cp->stat_lock[ring]); |
@@ -1964,14 +1964,14 @@ static void cas_tx(struct net_device *dev, struct cas *cp, | |||
1964 | #else | 1964 | #else |
1965 | limit = readl(cp->regs + REG_TX_COMPN(ring)); | 1965 | limit = readl(cp->regs + REG_TX_COMPN(ring)); |
1966 | #endif | 1966 | #endif |
1967 | if (cp->tx_old[ring] != limit) | 1967 | if (cp->tx_old[ring] != limit) |
1968 | cas_tx_ringN(cp, ring, limit); | 1968 | cas_tx_ringN(cp, ring, limit); |
1969 | } | 1969 | } |
1970 | } | 1970 | } |
1971 | 1971 | ||
1972 | 1972 | ||
1973 | static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | 1973 | static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, |
1974 | int entry, const u64 *words, | 1974 | int entry, const u64 *words, |
1975 | struct sk_buff **skbref) | 1975 | struct sk_buff **skbref) |
1976 | { | 1976 | { |
1977 | int dlen, hlen, len, i, alloclen; | 1977 | int dlen, hlen, len, i, alloclen; |
@@ -1979,19 +1979,19 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
1979 | struct cas_page *page; | 1979 | struct cas_page *page; |
1980 | struct sk_buff *skb; | 1980 | struct sk_buff *skb; |
1981 | void *addr, *crcaddr; | 1981 | void *addr, *crcaddr; |
1982 | char *p; | 1982 | char *p; |
1983 | 1983 | ||
1984 | hlen = CAS_VAL(RX_COMP2_HDR_SIZE, words[1]); | 1984 | hlen = CAS_VAL(RX_COMP2_HDR_SIZE, words[1]); |
1985 | dlen = CAS_VAL(RX_COMP1_DATA_SIZE, words[0]); | 1985 | dlen = CAS_VAL(RX_COMP1_DATA_SIZE, words[0]); |
1986 | len = hlen + dlen; | 1986 | len = hlen + dlen; |
1987 | 1987 | ||
1988 | if (RX_COPY_ALWAYS || (words[2] & RX_COMP3_SMALL_PKT)) | 1988 | if (RX_COPY_ALWAYS || (words[2] & RX_COMP3_SMALL_PKT)) |
1989 | alloclen = len; | 1989 | alloclen = len; |
1990 | else | 1990 | else |
1991 | alloclen = max(hlen, RX_COPY_MIN); | 1991 | alloclen = max(hlen, RX_COPY_MIN); |
1992 | 1992 | ||
1993 | skb = dev_alloc_skb(alloclen + swivel + cp->crc_size); | 1993 | skb = dev_alloc_skb(alloclen + swivel + cp->crc_size); |
1994 | if (skb == NULL) | 1994 | if (skb == NULL) |
1995 | return -1; | 1995 | return -1; |
1996 | 1996 | ||
1997 | *skbref = skb; | 1997 | *skbref = skb; |
@@ -2003,7 +2003,7 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2003 | if (hlen) { /* always copy header pages */ | 2003 | if (hlen) { /* always copy header pages */ |
2004 | i = CAS_VAL(RX_COMP2_HDR_INDEX, words[1]); | 2004 | i = CAS_VAL(RX_COMP2_HDR_INDEX, words[1]); |
2005 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; | 2005 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; |
2006 | off = CAS_VAL(RX_COMP2_HDR_OFF, words[1]) * 0x100 + | 2006 | off = CAS_VAL(RX_COMP2_HDR_OFF, words[1]) * 0x100 + |
2007 | swivel; | 2007 | swivel; |
2008 | 2008 | ||
2009 | i = hlen; | 2009 | i = hlen; |
@@ -2019,7 +2019,7 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2019 | RX_USED_ADD(page, 0x100); | 2019 | RX_USED_ADD(page, 0x100); |
2020 | p += hlen; | 2020 | p += hlen; |
2021 | swivel = 0; | 2021 | swivel = 0; |
2022 | } | 2022 | } |
2023 | 2023 | ||
2024 | 2024 | ||
2025 | if (alloclen < (hlen + dlen)) { | 2025 | if (alloclen < (hlen + dlen)) { |
@@ -2070,7 +2070,7 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2070 | frag->page = page->buffer; | 2070 | frag->page = page->buffer; |
2071 | frag->page_offset = off; | 2071 | frag->page_offset = off; |
2072 | frag->size = hlen - swivel; | 2072 | frag->size = hlen - swivel; |
2073 | 2073 | ||
2074 | /* any more data? */ | 2074 | /* any more data? */ |
2075 | if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { | 2075 | if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { |
2076 | hlen = dlen; | 2076 | hlen = dlen; |
@@ -2078,8 +2078,8 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2078 | 2078 | ||
2079 | i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); | 2079 | i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); |
2080 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; | 2080 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; |
2081 | pci_dma_sync_single_for_cpu(cp->pdev, page->dma_addr, | 2081 | pci_dma_sync_single_for_cpu(cp->pdev, page->dma_addr, |
2082 | hlen + cp->crc_size, | 2082 | hlen + cp->crc_size, |
2083 | PCI_DMA_FROMDEVICE); | 2083 | PCI_DMA_FROMDEVICE); |
2084 | pci_dma_sync_single_for_device(cp->pdev, page->dma_addr, | 2084 | pci_dma_sync_single_for_device(cp->pdev, page->dma_addr, |
2085 | hlen + cp->crc_size, | 2085 | hlen + cp->crc_size, |
@@ -2087,7 +2087,7 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2087 | 2087 | ||
2088 | skb_shinfo(skb)->nr_frags++; | 2088 | skb_shinfo(skb)->nr_frags++; |
2089 | skb->data_len += hlen; | 2089 | skb->data_len += hlen; |
2090 | skb->len += hlen; | 2090 | skb->len += hlen; |
2091 | frag++; | 2091 | frag++; |
2092 | 2092 | ||
2093 | get_page(page->buffer); | 2093 | get_page(page->buffer); |
@@ -2134,14 +2134,14 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2134 | RX_USED_ADD(page, cp->mtu_stride); | 2134 | RX_USED_ADD(page, cp->mtu_stride); |
2135 | else | 2135 | else |
2136 | RX_USED_ADD(page, i); | 2136 | RX_USED_ADD(page, i); |
2137 | 2137 | ||
2138 | /* any more data? */ | 2138 | /* any more data? */ |
2139 | if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { | 2139 | if ((words[0] & RX_COMP1_SPLIT_PKT) && ((dlen -= hlen) > 0)) { |
2140 | p += hlen; | 2140 | p += hlen; |
2141 | i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); | 2141 | i = CAS_VAL(RX_COMP2_NEXT_INDEX, words[1]); |
2142 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; | 2142 | page = cp->rx_pages[CAS_VAL(RX_INDEX_RING, i)][CAS_VAL(RX_INDEX_NUM, i)]; |
2143 | pci_dma_sync_single_for_cpu(cp->pdev, page->dma_addr, | 2143 | pci_dma_sync_single_for_cpu(cp->pdev, page->dma_addr, |
2144 | dlen + cp->crc_size, | 2144 | dlen + cp->crc_size, |
2145 | PCI_DMA_FROMDEVICE); | 2145 | PCI_DMA_FROMDEVICE); |
2146 | addr = cas_page_map(page->buffer); | 2146 | addr = cas_page_map(page->buffer); |
2147 | memcpy(p, addr, dlen + cp->crc_size); | 2147 | memcpy(p, addr, dlen + cp->crc_size); |
@@ -2149,7 +2149,7 @@ static int cas_rx_process_pkt(struct cas *cp, struct cas_rx_comp *rxc, | |||
2149 | dlen + cp->crc_size, | 2149 | dlen + cp->crc_size, |
2150 | PCI_DMA_FROMDEVICE); | 2150 | PCI_DMA_FROMDEVICE); |
2151 | cas_page_unmap(addr); | 2151 | cas_page_unmap(addr); |
2152 | RX_USED_ADD(page, dlen + cp->crc_size); | 2152 | RX_USED_ADD(page, dlen + cp->crc_size); |
2153 | } | 2153 | } |
2154 | end_copy_pkt: | 2154 | end_copy_pkt: |
2155 | if (cp->crc_size) { | 2155 | if (cp->crc_size) { |
@@ -2174,7 +2174,7 @@ end_copy_pkt: | |||
2174 | 2174 | ||
2175 | 2175 | ||
2176 | /* we can handle up to 64 rx flows at a time. we do the same thing | 2176 | /* we can handle up to 64 rx flows at a time. we do the same thing |
2177 | * as nonreassm except that we batch up the buffers. | 2177 | * as nonreassm except that we batch up the buffers. |
2178 | * NOTE: we currently just treat each flow as a bunch of packets that | 2178 | * NOTE: we currently just treat each flow as a bunch of packets that |
2179 | * we pass up. a better way would be to coalesce the packets | 2179 | * we pass up. a better way would be to coalesce the packets |
2180 | * into a jumbo packet. to do that, we need to do the following: | 2180 | * into a jumbo packet. to do that, we need to do the following: |
@@ -2184,7 +2184,7 @@ end_copy_pkt: | |||
2184 | * data length and merge the checksums. | 2184 | * data length and merge the checksums. |
2185 | * 3) on flow release, fix up the header. | 2185 | * 3) on flow release, fix up the header. |
2186 | * 4) make sure the higher layer doesn't care. | 2186 | * 4) make sure the higher layer doesn't care. |
2187 | * because packets get coalesced, we shouldn't run into fragment count | 2187 | * because packets get coalesced, we shouldn't run into fragment count |
2188 | * issues. | 2188 | * issues. |
2189 | */ | 2189 | */ |
2190 | static inline void cas_rx_flow_pkt(struct cas *cp, const u64 *words, | 2190 | static inline void cas_rx_flow_pkt(struct cas *cp, const u64 *words, |
@@ -2192,8 +2192,8 @@ static inline void cas_rx_flow_pkt(struct cas *cp, const u64 *words, | |||
2192 | { | 2192 | { |
2193 | int flowid = CAS_VAL(RX_COMP3_FLOWID, words[2]) & (N_RX_FLOWS - 1); | 2193 | int flowid = CAS_VAL(RX_COMP3_FLOWID, words[2]) & (N_RX_FLOWS - 1); |
2194 | struct sk_buff_head *flow = &cp->rx_flows[flowid]; | 2194 | struct sk_buff_head *flow = &cp->rx_flows[flowid]; |
2195 | 2195 | ||
2196 | /* this is protected at a higher layer, so no need to | 2196 | /* this is protected at a higher layer, so no need to |
2197 | * do any additional locking here. stick the buffer | 2197 | * do any additional locking here. stick the buffer |
2198 | * at the end. | 2198 | * at the end. |
2199 | */ | 2199 | */ |
@@ -2218,19 +2218,19 @@ static void cas_post_page(struct cas *cp, const int ring, const int index) | |||
2218 | new = cas_page_swap(cp, ring, index); | 2218 | new = cas_page_swap(cp, ring, index); |
2219 | cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr); | 2219 | cp->init_rxds[ring][entry].buffer = cpu_to_le64(new->dma_addr); |
2220 | cp->init_rxds[ring][entry].index = | 2220 | cp->init_rxds[ring][entry].index = |
2221 | cpu_to_le64(CAS_BASE(RX_INDEX_NUM, index) | | 2221 | cpu_to_le64(CAS_BASE(RX_INDEX_NUM, index) | |
2222 | CAS_BASE(RX_INDEX_RING, ring)); | 2222 | CAS_BASE(RX_INDEX_RING, ring)); |
2223 | 2223 | ||
2224 | entry = RX_DESC_ENTRY(ring, entry + 1); | 2224 | entry = RX_DESC_ENTRY(ring, entry + 1); |
2225 | cp->rx_old[ring] = entry; | 2225 | cp->rx_old[ring] = entry; |
2226 | 2226 | ||
2227 | if (entry % 4) | 2227 | if (entry % 4) |
2228 | return; | 2228 | return; |
2229 | 2229 | ||
2230 | if (ring == 0) | 2230 | if (ring == 0) |
2231 | writel(entry, cp->regs + REG_RX_KICK); | 2231 | writel(entry, cp->regs + REG_RX_KICK); |
2232 | else if ((N_RX_DESC_RINGS > 1) && | 2232 | else if ((N_RX_DESC_RINGS > 1) && |
2233 | (cp->cas_flags & CAS_FLAG_REG_PLUS)) | 2233 | (cp->cas_flags & CAS_FLAG_REG_PLUS)) |
2234 | writel(entry, cp->regs + REG_PLUS_RX_KICK1); | 2234 | writel(entry, cp->regs + REG_PLUS_RX_KICK1); |
2235 | } | 2235 | } |
2236 | 2236 | ||
@@ -2249,7 +2249,7 @@ static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) | |||
2249 | cp->dev->name, ring, entry); | 2249 | cp->dev->name, ring, entry); |
2250 | 2250 | ||
2251 | cluster = -1; | 2251 | cluster = -1; |
2252 | count = entry & 0x3; | 2252 | count = entry & 0x3; |
2253 | last = RX_DESC_ENTRY(ring, num ? entry + num - 4: entry - 4); | 2253 | last = RX_DESC_ENTRY(ring, num ? entry + num - 4: entry - 4); |
2254 | released = 0; | 2254 | released = 0; |
2255 | while (entry != last) { | 2255 | while (entry != last) { |
@@ -2257,12 +2257,12 @@ static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) | |||
2257 | if (cas_buffer_count(page[entry]) > 1) { | 2257 | if (cas_buffer_count(page[entry]) > 1) { |
2258 | cas_page_t *new = cas_page_dequeue(cp); | 2258 | cas_page_t *new = cas_page_dequeue(cp); |
2259 | if (!new) { | 2259 | if (!new) { |
2260 | /* let the timer know that we need to | 2260 | /* let the timer know that we need to |
2261 | * do this again | 2261 | * do this again |
2262 | */ | 2262 | */ |
2263 | cp->cas_flags |= CAS_FLAG_RXD_POST(ring); | 2263 | cp->cas_flags |= CAS_FLAG_RXD_POST(ring); |
2264 | if (!timer_pending(&cp->link_timer)) | 2264 | if (!timer_pending(&cp->link_timer)) |
2265 | mod_timer(&cp->link_timer, jiffies + | 2265 | mod_timer(&cp->link_timer, jiffies + |
2266 | CAS_LINK_FAST_TIMEOUT); | 2266 | CAS_LINK_FAST_TIMEOUT); |
2267 | cp->rx_old[ring] = entry; | 2267 | cp->rx_old[ring] = entry; |
2268 | cp->rx_last[ring] = num ? num - released : 0; | 2268 | cp->rx_last[ring] = num ? num - released : 0; |
@@ -2271,10 +2271,10 @@ static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) | |||
2271 | spin_lock(&cp->rx_inuse_lock); | 2271 | spin_lock(&cp->rx_inuse_lock); |
2272 | list_add(&page[entry]->list, &cp->rx_inuse_list); | 2272 | list_add(&page[entry]->list, &cp->rx_inuse_list); |
2273 | spin_unlock(&cp->rx_inuse_lock); | 2273 | spin_unlock(&cp->rx_inuse_lock); |
2274 | cp->init_rxds[ring][entry].buffer = | 2274 | cp->init_rxds[ring][entry].buffer = |
2275 | cpu_to_le64(new->dma_addr); | 2275 | cpu_to_le64(new->dma_addr); |
2276 | page[entry] = new; | 2276 | page[entry] = new; |
2277 | 2277 | ||
2278 | } | 2278 | } |
2279 | 2279 | ||
2280 | if (++count == 4) { | 2280 | if (++count == 4) { |
@@ -2286,13 +2286,13 @@ static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) | |||
2286 | } | 2286 | } |
2287 | cp->rx_old[ring] = entry; | 2287 | cp->rx_old[ring] = entry; |
2288 | 2288 | ||
2289 | if (cluster < 0) | 2289 | if (cluster < 0) |
2290 | return 0; | 2290 | return 0; |
2291 | 2291 | ||
2292 | if (ring == 0) | 2292 | if (ring == 0) |
2293 | writel(cluster, cp->regs + REG_RX_KICK); | 2293 | writel(cluster, cp->regs + REG_RX_KICK); |
2294 | else if ((N_RX_DESC_RINGS > 1) && | 2294 | else if ((N_RX_DESC_RINGS > 1) && |
2295 | (cp->cas_flags & CAS_FLAG_REG_PLUS)) | 2295 | (cp->cas_flags & CAS_FLAG_REG_PLUS)) |
2296 | writel(cluster, cp->regs + REG_PLUS_RX_KICK1); | 2296 | writel(cluster, cp->regs + REG_PLUS_RX_KICK1); |
2297 | return 0; | 2297 | return 0; |
2298 | } | 2298 | } |
@@ -2301,14 +2301,14 @@ static int cas_post_rxds_ringN(struct cas *cp, int ring, int num) | |||
2301 | /* process a completion ring. packets are set up in three basic ways: | 2301 | /* process a completion ring. packets are set up in three basic ways: |
2302 | * small packets: should be copied header + data in single buffer. | 2302 | * small packets: should be copied header + data in single buffer. |
2303 | * large packets: header and data in a single buffer. | 2303 | * large packets: header and data in a single buffer. |
2304 | * split packets: header in a separate buffer from data. | 2304 | * split packets: header in a separate buffer from data. |
2305 | * data may be in multiple pages. data may be > 256 | 2305 | * data may be in multiple pages. data may be > 256 |
2306 | * bytes but in a single page. | 2306 | * bytes but in a single page. |
2307 | * | 2307 | * |
2308 | * NOTE: RX page posting is done in this routine as well. while there's | 2308 | * NOTE: RX page posting is done in this routine as well. while there's |
2309 | * the capability of using multiple RX completion rings, it isn't | 2309 | * the capability of using multiple RX completion rings, it isn't |
2310 | * really worthwhile due to the fact that the page posting will | 2310 | * really worthwhile due to the fact that the page posting will |
2311 | * force serialization on the single descriptor ring. | 2311 | * force serialization on the single descriptor ring. |
2312 | */ | 2312 | */ |
2313 | static int cas_rx_ringN(struct cas *cp, int ring, int budget) | 2313 | static int cas_rx_ringN(struct cas *cp, int ring, int budget) |
2314 | { | 2314 | { |
@@ -2319,7 +2319,7 @@ static int cas_rx_ringN(struct cas *cp, int ring, int budget) | |||
2319 | if (netif_msg_intr(cp)) | 2319 | if (netif_msg_intr(cp)) |
2320 | printk(KERN_DEBUG "%s: rx[%d] interrupt, done: %d/%d\n", | 2320 | printk(KERN_DEBUG "%s: rx[%d] interrupt, done: %d/%d\n", |
2321 | cp->dev->name, ring, | 2321 | cp->dev->name, ring, |
2322 | readl(cp->regs + REG_RX_COMP_HEAD), | 2322 | readl(cp->regs + REG_RX_COMP_HEAD), |
2323 | cp->rx_new[ring]); | 2323 | cp->rx_new[ring]); |
2324 | 2324 | ||
2325 | entry = cp->rx_new[ring]; | 2325 | entry = cp->rx_new[ring]; |
@@ -2375,7 +2375,7 @@ static int cas_rx_ringN(struct cas *cp, int ring, int budget) | |||
2375 | */ | 2375 | */ |
2376 | if (RX_DONT_BATCH || (type == 0x2)) { | 2376 | if (RX_DONT_BATCH || (type == 0x2)) { |
2377 | /* non-reassm: these always get released */ | 2377 | /* non-reassm: these always get released */ |
2378 | cas_skb_release(skb); | 2378 | cas_skb_release(skb); |
2379 | } else { | 2379 | } else { |
2380 | cas_rx_flow_pkt(cp, words, skb); | 2380 | cas_rx_flow_pkt(cp, words, skb); |
2381 | } | 2381 | } |
@@ -2396,7 +2396,7 @@ static int cas_rx_ringN(struct cas *cp, int ring, int budget) | |||
2396 | i = CAS_VAL(RX_INDEX_NUM, i); | 2396 | i = CAS_VAL(RX_INDEX_NUM, i); |
2397 | cas_post_page(cp, dring, i); | 2397 | cas_post_page(cp, dring, i); |
2398 | } | 2398 | } |
2399 | 2399 | ||
2400 | if (words[0] & RX_COMP1_RELEASE_DATA) { | 2400 | if (words[0] & RX_COMP1_RELEASE_DATA) { |
2401 | i = CAS_VAL(RX_COMP1_DATA_INDEX, words[0]); | 2401 | i = CAS_VAL(RX_COMP1_DATA_INDEX, words[0]); |
2402 | dring = CAS_VAL(RX_INDEX_RING, i); | 2402 | dring = CAS_VAL(RX_INDEX_RING, i); |
@@ -2412,7 +2412,7 @@ static int cas_rx_ringN(struct cas *cp, int ring, int budget) | |||
2412 | } | 2412 | } |
2413 | 2413 | ||
2414 | /* skip to the next entry */ | 2414 | /* skip to the next entry */ |
2415 | entry = RX_COMP_ENTRY(ring, entry + 1 + | 2415 | entry = RX_COMP_ENTRY(ring, entry + 1 + |
2416 | CAS_VAL(RX_COMP1_SKIP, words[0])); | 2416 | CAS_VAL(RX_COMP1_SKIP, words[0])); |
2417 | #ifdef USE_NAPI | 2417 | #ifdef USE_NAPI |
2418 | if (budget && (npackets >= budget)) | 2418 | if (budget && (npackets >= budget)) |
@@ -2436,12 +2436,12 @@ static void cas_post_rxcs_ringN(struct net_device *dev, | |||
2436 | int last, entry; | 2436 | int last, entry; |
2437 | 2437 | ||
2438 | last = cp->rx_cur[ring]; | 2438 | last = cp->rx_cur[ring]; |
2439 | entry = cp->rx_new[ring]; | 2439 | entry = cp->rx_new[ring]; |
2440 | if (netif_msg_intr(cp)) | 2440 | if (netif_msg_intr(cp)) |
2441 | printk(KERN_DEBUG "%s: rxc[%d] interrupt, done: %d/%d\n", | 2441 | printk(KERN_DEBUG "%s: rxc[%d] interrupt, done: %d/%d\n", |
2442 | dev->name, ring, readl(cp->regs + REG_RX_COMP_HEAD), | 2442 | dev->name, ring, readl(cp->regs + REG_RX_COMP_HEAD), |
2443 | entry); | 2443 | entry); |
2444 | 2444 | ||
2445 | /* zero and re-mark descriptors */ | 2445 | /* zero and re-mark descriptors */ |
2446 | while (last != entry) { | 2446 | while (last != entry) { |
2447 | cas_rxc_init(rxc + last); | 2447 | cas_rxc_init(rxc + last); |
@@ -2451,21 +2451,21 @@ static void cas_post_rxcs_ringN(struct net_device *dev, | |||
2451 | 2451 | ||
2452 | if (ring == 0) | 2452 | if (ring == 0) |
2453 | writel(last, cp->regs + REG_RX_COMP_TAIL); | 2453 | writel(last, cp->regs + REG_RX_COMP_TAIL); |
2454 | else if (cp->cas_flags & CAS_FLAG_REG_PLUS) | 2454 | else if (cp->cas_flags & CAS_FLAG_REG_PLUS) |
2455 | writel(last, cp->regs + REG_PLUS_RX_COMPN_TAIL(ring)); | 2455 | writel(last, cp->regs + REG_PLUS_RX_COMPN_TAIL(ring)); |
2456 | } | 2456 | } |
2457 | 2457 | ||
2458 | 2458 | ||
2459 | 2459 | ||
2460 | /* cassini can use all four PCI interrupts for the completion ring. | 2460 | /* cassini can use all four PCI interrupts for the completion ring. |
2461 | * rings 3 and 4 are identical | 2461 | * rings 3 and 4 are identical |
2462 | */ | 2462 | */ |
2463 | #if defined(USE_PCI_INTC) || defined(USE_PCI_INTD) | 2463 | #if defined(USE_PCI_INTC) || defined(USE_PCI_INTD) |
2464 | static inline void cas_handle_irqN(struct net_device *dev, | 2464 | static inline void cas_handle_irqN(struct net_device *dev, |
2465 | struct cas *cp, const u32 status, | 2465 | struct cas *cp, const u32 status, |
2466 | const int ring) | 2466 | const int ring) |
2467 | { | 2467 | { |
2468 | if (status & (INTR_RX_COMP_FULL_ALT | INTR_RX_COMP_AF_ALT)) | 2468 | if (status & (INTR_RX_COMP_FULL_ALT | INTR_RX_COMP_AF_ALT)) |
2469 | cas_post_rxcs_ringN(dev, cp, ring); | 2469 | cas_post_rxcs_ringN(dev, cp, ring); |
2470 | } | 2470 | } |
2471 | 2471 | ||
@@ -2505,7 +2505,7 @@ static irqreturn_t cas_interruptN(int irq, void *dev_id, struct pt_regs *regs) | |||
2505 | static inline void cas_handle_irq1(struct cas *cp, const u32 status) | 2505 | static inline void cas_handle_irq1(struct cas *cp, const u32 status) |
2506 | { | 2506 | { |
2507 | if (status & INTR_RX_BUF_UNAVAIL_1) { | 2507 | if (status & INTR_RX_BUF_UNAVAIL_1) { |
2508 | /* Frame arrived, no free RX buffers available. | 2508 | /* Frame arrived, no free RX buffers available. |
2509 | * NOTE: we can get this on a link transition. */ | 2509 | * NOTE: we can get this on a link transition. */ |
2510 | cas_post_rxds_ringN(cp, 1, 0); | 2510 | cas_post_rxds_ringN(cp, 1, 0); |
2511 | spin_lock(&cp->stat_lock[1]); | 2511 | spin_lock(&cp->stat_lock[1]); |
@@ -2513,8 +2513,8 @@ static inline void cas_handle_irq1(struct cas *cp, const u32 status) | |||
2513 | spin_unlock(&cp->stat_lock[1]); | 2513 | spin_unlock(&cp->stat_lock[1]); |
2514 | } | 2514 | } |
2515 | 2515 | ||
2516 | if (status & INTR_RX_BUF_AE_1) | 2516 | if (status & INTR_RX_BUF_AE_1) |
2517 | cas_post_rxds_ringN(cp, 1, RX_DESC_RINGN_SIZE(1) - | 2517 | cas_post_rxds_ringN(cp, 1, RX_DESC_RINGN_SIZE(1) - |
2518 | RX_AE_FREEN_VAL(1)); | 2518 | RX_AE_FREEN_VAL(1)); |
2519 | 2519 | ||
2520 | if (status & (INTR_RX_COMP_AF | INTR_RX_COMP_FULL)) | 2520 | if (status & (INTR_RX_COMP_AF | INTR_RX_COMP_FULL)) |
@@ -2558,7 +2558,7 @@ static inline void cas_handle_irq(struct net_device *dev, | |||
2558 | cas_abnormal_irq(dev, cp, status); | 2558 | cas_abnormal_irq(dev, cp, status); |
2559 | 2559 | ||
2560 | if (status & INTR_RX_BUF_UNAVAIL) { | 2560 | if (status & INTR_RX_BUF_UNAVAIL) { |
2561 | /* Frame arrived, no free RX buffers available. | 2561 | /* Frame arrived, no free RX buffers available. |
2562 | * NOTE: we can get this on a link transition. | 2562 | * NOTE: we can get this on a link transition. |
2563 | */ | 2563 | */ |
2564 | cas_post_rxds_ringN(cp, 0, 0); | 2564 | cas_post_rxds_ringN(cp, 0, 0); |
@@ -2625,7 +2625,7 @@ static int cas_poll(struct net_device *dev, int *budget) | |||
2625 | todo = min(*budget, dev->quota); | 2625 | todo = min(*budget, dev->quota); |
2626 | 2626 | ||
2627 | /* to make sure we're fair with the work we loop through each | 2627 | /* to make sure we're fair with the work we loop through each |
2628 | * ring N_RX_COMP_RING times with a request of | 2628 | * ring N_RX_COMP_RING times with a request of |
2629 | * todo / N_RX_COMP_RINGS | 2629 | * todo / N_RX_COMP_RINGS |
2630 | */ | 2630 | */ |
2631 | enable_intr = 1; | 2631 | enable_intr = 1; |
@@ -2784,13 +2784,13 @@ static void cas_write_txd(struct cas *cp, int ring, int entry, | |||
2784 | txd->buffer = cpu_to_le64(mapping); | 2784 | txd->buffer = cpu_to_le64(mapping); |
2785 | } | 2785 | } |
2786 | 2786 | ||
2787 | static inline void *tx_tiny_buf(struct cas *cp, const int ring, | 2787 | static inline void *tx_tiny_buf(struct cas *cp, const int ring, |
2788 | const int entry) | 2788 | const int entry) |
2789 | { | 2789 | { |
2790 | return cp->tx_tiny_bufs[ring] + TX_TINY_BUF_LEN*entry; | 2790 | return cp->tx_tiny_bufs[ring] + TX_TINY_BUF_LEN*entry; |
2791 | } | 2791 | } |
2792 | 2792 | ||
2793 | static inline dma_addr_t tx_tiny_map(struct cas *cp, const int ring, | 2793 | static inline dma_addr_t tx_tiny_map(struct cas *cp, const int ring, |
2794 | const int entry, const int tentry) | 2794 | const int entry, const int tentry) |
2795 | { | 2795 | { |
2796 | cp->tx_tiny_use[ring][tentry].nbufs++; | 2796 | cp->tx_tiny_use[ring][tentry].nbufs++; |
@@ -2798,7 +2798,7 @@ static inline dma_addr_t tx_tiny_map(struct cas *cp, const int ring, | |||
2798 | return cp->tx_tiny_dvma[ring] + TX_TINY_BUF_LEN*entry; | 2798 | return cp->tx_tiny_dvma[ring] + TX_TINY_BUF_LEN*entry; |
2799 | } | 2799 | } |
2800 | 2800 | ||
2801 | static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | 2801 | static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, |
2802 | struct sk_buff *skb) | 2802 | struct sk_buff *skb) |
2803 | { | 2803 | { |
2804 | struct net_device *dev = cp->dev; | 2804 | struct net_device *dev = cp->dev; |
@@ -2811,7 +2811,7 @@ static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | |||
2811 | spin_lock_irqsave(&cp->tx_lock[ring], flags); | 2811 | spin_lock_irqsave(&cp->tx_lock[ring], flags); |
2812 | 2812 | ||
2813 | /* This is a hard error, log it. */ | 2813 | /* This is a hard error, log it. */ |
2814 | if (TX_BUFFS_AVAIL(cp, ring) <= | 2814 | if (TX_BUFFS_AVAIL(cp, ring) <= |
2815 | CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { | 2815 | CAS_TABORT(cp)*(skb_shinfo(skb)->nr_frags + 1)) { |
2816 | netif_stop_queue(dev); | 2816 | netif_stop_queue(dev); |
2817 | spin_unlock_irqrestore(&cp->tx_lock[ring], flags); | 2817 | spin_unlock_irqrestore(&cp->tx_lock[ring], flags); |
@@ -2827,7 +2827,7 @@ static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | |||
2827 | csum_start_off = (u64) (skb->h.raw - skb->data); | 2827 | csum_start_off = (u64) (skb->h.raw - skb->data); |
2828 | csum_stuff_off = (u64) ((skb->h.raw + skb->csum) - skb->data); | 2828 | csum_stuff_off = (u64) ((skb->h.raw + skb->csum) - skb->data); |
2829 | 2829 | ||
2830 | ctrl = TX_DESC_CSUM_EN | | 2830 | ctrl = TX_DESC_CSUM_EN | |
2831 | CAS_BASE(TX_DESC_CSUM_START, csum_start_off) | | 2831 | CAS_BASE(TX_DESC_CSUM_START, csum_start_off) | |
2832 | CAS_BASE(TX_DESC_CSUM_STUFF, csum_stuff_off); | 2832 | CAS_BASE(TX_DESC_CSUM_STUFF, csum_stuff_off); |
2833 | } | 2833 | } |
@@ -2845,17 +2845,17 @@ static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | |||
2845 | tabort = cas_calc_tabort(cp, (unsigned long) skb->data, len); | 2845 | tabort = cas_calc_tabort(cp, (unsigned long) skb->data, len); |
2846 | if (unlikely(tabort)) { | 2846 | if (unlikely(tabort)) { |
2847 | /* NOTE: len is always > tabort */ | 2847 | /* NOTE: len is always > tabort */ |
2848 | cas_write_txd(cp, ring, entry, mapping, len - tabort, | 2848 | cas_write_txd(cp, ring, entry, mapping, len - tabort, |
2849 | ctrl | TX_DESC_SOF, 0); | 2849 | ctrl | TX_DESC_SOF, 0); |
2850 | entry = TX_DESC_NEXT(ring, entry); | 2850 | entry = TX_DESC_NEXT(ring, entry); |
2851 | 2851 | ||
2852 | memcpy(tx_tiny_buf(cp, ring, entry), skb->data + | 2852 | memcpy(tx_tiny_buf(cp, ring, entry), skb->data + |
2853 | len - tabort, tabort); | 2853 | len - tabort, tabort); |
2854 | mapping = tx_tiny_map(cp, ring, entry, tentry); | 2854 | mapping = tx_tiny_map(cp, ring, entry, tentry); |
2855 | cas_write_txd(cp, ring, entry, mapping, tabort, ctrl, | 2855 | cas_write_txd(cp, ring, entry, mapping, tabort, ctrl, |
2856 | (nr_frags == 0)); | 2856 | (nr_frags == 0)); |
2857 | } else { | 2857 | } else { |
2858 | cas_write_txd(cp, ring, entry, mapping, len, ctrl | | 2858 | cas_write_txd(cp, ring, entry, mapping, len, ctrl | |
2859 | TX_DESC_SOF, (nr_frags == 0)); | 2859 | TX_DESC_SOF, (nr_frags == 0)); |
2860 | } | 2860 | } |
2861 | entry = TX_DESC_NEXT(ring, entry); | 2861 | entry = TX_DESC_NEXT(ring, entry); |
@@ -2876,10 +2876,10 @@ static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | |||
2876 | cas_write_txd(cp, ring, entry, mapping, len - tabort, | 2876 | cas_write_txd(cp, ring, entry, mapping, len - tabort, |
2877 | ctrl, 0); | 2877 | ctrl, 0); |
2878 | entry = TX_DESC_NEXT(ring, entry); | 2878 | entry = TX_DESC_NEXT(ring, entry); |
2879 | 2879 | ||
2880 | addr = cas_page_map(fragp->page); | 2880 | addr = cas_page_map(fragp->page); |
2881 | memcpy(tx_tiny_buf(cp, ring, entry), | 2881 | memcpy(tx_tiny_buf(cp, ring, entry), |
2882 | addr + fragp->page_offset + len - tabort, | 2882 | addr + fragp->page_offset + len - tabort, |
2883 | tabort); | 2883 | tabort); |
2884 | cas_page_unmap(addr); | 2884 | cas_page_unmap(addr); |
2885 | mapping = tx_tiny_map(cp, ring, entry, tentry); | 2885 | mapping = tx_tiny_map(cp, ring, entry, tentry); |
@@ -2898,12 +2898,12 @@ static inline int cas_xmit_tx_ringN(struct cas *cp, int ring, | |||
2898 | if (netif_msg_tx_queued(cp)) | 2898 | if (netif_msg_tx_queued(cp)) |
2899 | printk(KERN_DEBUG "%s: tx[%d] queued, slot %d, skblen %d, " | 2899 | printk(KERN_DEBUG "%s: tx[%d] queued, slot %d, skblen %d, " |
2900 | "avail %d\n", | 2900 | "avail %d\n", |
2901 | dev->name, ring, entry, skb->len, | 2901 | dev->name, ring, entry, skb->len, |
2902 | TX_BUFFS_AVAIL(cp, ring)); | 2902 | TX_BUFFS_AVAIL(cp, ring)); |
2903 | writel(entry, cp->regs + REG_TX_KICKN(ring)); | 2903 | writel(entry, cp->regs + REG_TX_KICKN(ring)); |
2904 | spin_unlock_irqrestore(&cp->tx_lock[ring], flags); | 2904 | spin_unlock_irqrestore(&cp->tx_lock[ring], flags); |
2905 | return 0; | 2905 | return 0; |
2906 | } | 2906 | } |
2907 | 2907 | ||
2908 | static int cas_start_xmit(struct sk_buff *skb, struct net_device *dev) | 2908 | static int cas_start_xmit(struct sk_buff *skb, struct net_device *dev) |
2909 | { | 2909 | { |
@@ -2912,7 +2912,7 @@ static int cas_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
2912 | /* this is only used as a load-balancing hint, so it doesn't | 2912 | /* this is only used as a load-balancing hint, so it doesn't |
2913 | * need to be SMP safe | 2913 | * need to be SMP safe |
2914 | */ | 2914 | */ |
2915 | static int ring; | 2915 | static int ring; |
2916 | 2916 | ||
2917 | if (skb_padto(skb, cp->min_frame_size)) | 2917 | if (skb_padto(skb, cp->min_frame_size)) |
2918 | return 0; | 2918 | return 0; |
@@ -2943,14 +2943,14 @@ static void cas_init_tx_dma(struct cas *cp) | |||
2943 | /* enable completion writebacks, enable paced mode, | 2943 | /* enable completion writebacks, enable paced mode, |
2944 | * disable read pipe, and disable pre-interrupt compwbs | 2944 | * disable read pipe, and disable pre-interrupt compwbs |
2945 | */ | 2945 | */ |
2946 | val = TX_CFG_COMPWB_Q1 | TX_CFG_COMPWB_Q2 | | 2946 | val = TX_CFG_COMPWB_Q1 | TX_CFG_COMPWB_Q2 | |
2947 | TX_CFG_COMPWB_Q3 | TX_CFG_COMPWB_Q4 | | 2947 | TX_CFG_COMPWB_Q3 | TX_CFG_COMPWB_Q4 | |
2948 | TX_CFG_DMA_RDPIPE_DIS | TX_CFG_PACED_MODE | | 2948 | TX_CFG_DMA_RDPIPE_DIS | TX_CFG_PACED_MODE | |
2949 | TX_CFG_INTR_COMPWB_DIS; | 2949 | TX_CFG_INTR_COMPWB_DIS; |
2950 | 2950 | ||
2951 | /* write out tx ring info and tx desc bases */ | 2951 | /* write out tx ring info and tx desc bases */ |
2952 | for (i = 0; i < MAX_TX_RINGS; i++) { | 2952 | for (i = 0; i < MAX_TX_RINGS; i++) { |
2953 | off = (unsigned long) cp->init_txds[i] - | 2953 | off = (unsigned long) cp->init_txds[i] - |
2954 | (unsigned long) cp->init_block; | 2954 | (unsigned long) cp->init_block; |
2955 | 2955 | ||
2956 | val |= CAS_TX_RINGN_BASE(i); | 2956 | val |= CAS_TX_RINGN_BASE(i); |
@@ -2991,7 +2991,7 @@ static u32 cas_setup_multicast(struct cas *cp) | |||
2991 | { | 2991 | { |
2992 | u32 rxcfg = 0; | 2992 | u32 rxcfg = 0; |
2993 | int i; | 2993 | int i; |
2994 | 2994 | ||
2995 | if (cp->dev->flags & IFF_PROMISC) { | 2995 | if (cp->dev->flags & IFF_PROMISC) { |
2996 | rxcfg |= MAC_RX_CFG_PROMISC_EN; | 2996 | rxcfg |= MAC_RX_CFG_PROMISC_EN; |
2997 | 2997 | ||
@@ -3016,16 +3016,16 @@ static u32 cas_setup_multicast(struct cas *cp) | |||
3016 | writel(0x0, cp->regs + REG_MAC_ADDRN(i*3 + 2)); | 3016 | writel(0x0, cp->regs + REG_MAC_ADDRN(i*3 + 2)); |
3017 | continue; | 3017 | continue; |
3018 | } | 3018 | } |
3019 | writel((dmi->dmi_addr[4] << 8) | dmi->dmi_addr[5], | 3019 | writel((dmi->dmi_addr[4] << 8) | dmi->dmi_addr[5], |
3020 | cp->regs + REG_MAC_ADDRN(i*3 + 0)); | 3020 | cp->regs + REG_MAC_ADDRN(i*3 + 0)); |
3021 | writel((dmi->dmi_addr[2] << 8) | dmi->dmi_addr[3], | 3021 | writel((dmi->dmi_addr[2] << 8) | dmi->dmi_addr[3], |
3022 | cp->regs + REG_MAC_ADDRN(i*3 + 1)); | 3022 | cp->regs + REG_MAC_ADDRN(i*3 + 1)); |
3023 | writel((dmi->dmi_addr[0] << 8) | dmi->dmi_addr[1], | 3023 | writel((dmi->dmi_addr[0] << 8) | dmi->dmi_addr[1], |
3024 | cp->regs + REG_MAC_ADDRN(i*3 + 2)); | 3024 | cp->regs + REG_MAC_ADDRN(i*3 + 2)); |
3025 | dmi = dmi->next; | 3025 | dmi = dmi->next; |
3026 | } | 3026 | } |
3027 | 3027 | ||
3028 | /* use hw hash table for the next series of | 3028 | /* use hw hash table for the next series of |
3029 | * multicast addresses | 3029 | * multicast addresses |
3030 | */ | 3030 | */ |
3031 | memset(hash_table, 0, sizeof(hash_table)); | 3031 | memset(hash_table, 0, sizeof(hash_table)); |
@@ -3036,7 +3036,7 @@ static u32 cas_setup_multicast(struct cas *cp) | |||
3036 | dmi = dmi->next; | 3036 | dmi = dmi->next; |
3037 | } | 3037 | } |
3038 | for (i=0; i < 16; i++) | 3038 | for (i=0; i < 16; i++) |
3039 | writel(hash_table[i], cp->regs + | 3039 | writel(hash_table[i], cp->regs + |
3040 | REG_MAC_HASH_TABLEN(i)); | 3040 | REG_MAC_HASH_TABLEN(i)); |
3041 | rxcfg |= MAC_RX_CFG_HASH_FILTER_EN; | 3041 | rxcfg |= MAC_RX_CFG_HASH_FILTER_EN; |
3042 | } | 3042 | } |
@@ -3121,23 +3121,23 @@ static void cas_init_mac(struct cas *cp) | |||
3121 | writel(0x00, cp->regs + REG_MAC_IPG0); | 3121 | writel(0x00, cp->regs + REG_MAC_IPG0); |
3122 | writel(0x08, cp->regs + REG_MAC_IPG1); | 3122 | writel(0x08, cp->regs + REG_MAC_IPG1); |
3123 | writel(0x04, cp->regs + REG_MAC_IPG2); | 3123 | writel(0x04, cp->regs + REG_MAC_IPG2); |
3124 | 3124 | ||
3125 | /* change later for 802.3z */ | 3125 | /* change later for 802.3z */ |
3126 | writel(0x40, cp->regs + REG_MAC_SLOT_TIME); | 3126 | writel(0x40, cp->regs + REG_MAC_SLOT_TIME); |
3127 | 3127 | ||
3128 | /* min frame + FCS */ | 3128 | /* min frame + FCS */ |
3129 | writel(ETH_ZLEN + 4, cp->regs + REG_MAC_FRAMESIZE_MIN); | 3129 | writel(ETH_ZLEN + 4, cp->regs + REG_MAC_FRAMESIZE_MIN); |
3130 | 3130 | ||
3131 | /* Ethernet payload + header + FCS + optional VLAN tag. NOTE: we | 3131 | /* Ethernet payload + header + FCS + optional VLAN tag. NOTE: we |
3132 | * specify the maximum frame size to prevent RX tag errors on | 3132 | * specify the maximum frame size to prevent RX tag errors on |
3133 | * oversized frames. | 3133 | * oversized frames. |
3134 | */ | 3134 | */ |
3135 | writel(CAS_BASE(MAC_FRAMESIZE_MAX_BURST, 0x2000) | | 3135 | writel(CAS_BASE(MAC_FRAMESIZE_MAX_BURST, 0x2000) | |
3136 | CAS_BASE(MAC_FRAMESIZE_MAX_FRAME, | 3136 | CAS_BASE(MAC_FRAMESIZE_MAX_FRAME, |
3137 | (CAS_MAX_MTU + ETH_HLEN + 4 + 4)), | 3137 | (CAS_MAX_MTU + ETH_HLEN + 4 + 4)), |
3138 | cp->regs + REG_MAC_FRAMESIZE_MAX); | 3138 | cp->regs + REG_MAC_FRAMESIZE_MAX); |
3139 | 3139 | ||
3140 | /* NOTE: crc_size is used as a surrogate for half-duplex. | 3140 | /* NOTE: crc_size is used as a surrogate for half-duplex. |
3141 | * workaround saturn half-duplex issue by increasing preamble | 3141 | * workaround saturn half-duplex issue by increasing preamble |
3142 | * size to 65 bytes. | 3142 | * size to 65 bytes. |
3143 | */ | 3143 | */ |
@@ -3180,7 +3180,7 @@ static void cas_init_mac(struct cas *cp) | |||
3180 | * spin_lock_irqsave, but we are called only in cas_init_hw and | 3180 | * spin_lock_irqsave, but we are called only in cas_init_hw and |
3181 | * cas_init_hw is protected by cas_lock_all, which calls | 3181 | * cas_init_hw is protected by cas_lock_all, which calls |
3182 | * spin_lock_irq (so it doesn't need to save the flags, and | 3182 | * spin_lock_irq (so it doesn't need to save the flags, and |
3183 | * we should be OK for the writel, as that is the only | 3183 | * we should be OK for the writel, as that is the only |
3184 | * difference). | 3184 | * difference). |
3185 | */ | 3185 | */ |
3186 | cp->mac_rx_cfg = rxcfg = cas_setup_multicast(cp); | 3186 | cp->mac_rx_cfg = rxcfg = cas_setup_multicast(cp); |
@@ -3229,7 +3229,7 @@ static int cas_vpd_match(const void __iomem *p, const char *str) | |||
3229 | { | 3229 | { |
3230 | int len = strlen(str) + 1; | 3230 | int len = strlen(str) + 1; |
3231 | int i; | 3231 | int i; |
3232 | 3232 | ||
3233 | for (i = 0; i < len; i++) { | 3233 | for (i = 0; i < len; i++) { |
3234 | if (readb(p + i) != str[i]) | 3234 | if (readb(p + i) != str[i]) |
3235 | return 0; | 3235 | return 0; |
@@ -3246,7 +3246,7 @@ static int cas_vpd_match(const void __iomem *p, const char *str) | |||
3246 | * number. | 3246 | * number. |
3247 | * 3) fiber cards don't have bridges, so their slot numbers don't | 3247 | * 3) fiber cards don't have bridges, so their slot numbers don't |
3248 | * mean anything. | 3248 | * mean anything. |
3249 | * 4) we don't actually know we have a fiber card until after | 3249 | * 4) we don't actually know we have a fiber card until after |
3250 | * the mac addresses are parsed. | 3250 | * the mac addresses are parsed. |
3251 | */ | 3251 | */ |
3252 | static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | 3252 | static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, |
@@ -3278,15 +3278,15 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3278 | (readb(p + i + 1) == 0x43) && | 3278 | (readb(p + i + 1) == 0x43) && |
3279 | (readb(p + i + 2) == 0x49) && | 3279 | (readb(p + i + 2) == 0x49) && |
3280 | (readb(p + i + 3) == 0x52)) { | 3280 | (readb(p + i + 3) == 0x52)) { |
3281 | base = p + (readb(p + i + 8) | | 3281 | base = p + (readb(p + i + 8) | |
3282 | (readb(p + i + 9) << 8)); | 3282 | (readb(p + i + 9) << 8)); |
3283 | break; | 3283 | break; |
3284 | } | 3284 | } |
3285 | } | 3285 | } |
3286 | 3286 | ||
3287 | if (!base || (readb(base) != 0x82)) | 3287 | if (!base || (readb(base) != 0x82)) |
3288 | goto use_random_mac_addr; | 3288 | goto use_random_mac_addr; |
3289 | 3289 | ||
3290 | i = (readb(base + 1) | (readb(base + 2) << 8)) + 3; | 3290 | i = (readb(base + 1) | (readb(base + 2) << 8)) + 3; |
3291 | while (i < EXPANSION_ROM_SIZE) { | 3291 | while (i < EXPANSION_ROM_SIZE) { |
3292 | if (readb(base + i) != 0x90) /* no vpd found */ | 3292 | if (readb(base + i) != 0x90) /* no vpd found */ |
@@ -3304,20 +3304,20 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3304 | char type; | 3304 | char type; |
3305 | 3305 | ||
3306 | p += 3; | 3306 | p += 3; |
3307 | 3307 | ||
3308 | /* look for the following things: | 3308 | /* look for the following things: |
3309 | * -- correct length == 29 | 3309 | * -- correct length == 29 |
3310 | * 3 (type) + 2 (size) + | 3310 | * 3 (type) + 2 (size) + |
3311 | * 18 (strlen("local-mac-address") + 1) + | 3311 | * 18 (strlen("local-mac-address") + 1) + |
3312 | * 6 (mac addr) | 3312 | * 6 (mac addr) |
3313 | * -- VPD Instance 'I' | 3313 | * -- VPD Instance 'I' |
3314 | * -- VPD Type Bytes 'B' | 3314 | * -- VPD Type Bytes 'B' |
3315 | * -- VPD data length == 6 | 3315 | * -- VPD data length == 6 |
3316 | * -- property string == local-mac-address | 3316 | * -- property string == local-mac-address |
3317 | * | 3317 | * |
3318 | * -- correct length == 24 | 3318 | * -- correct length == 24 |
3319 | * 3 (type) + 2 (size) + | 3319 | * 3 (type) + 2 (size) + |
3320 | * 12 (strlen("entropy-dev") + 1) + | 3320 | * 12 (strlen("entropy-dev") + 1) + |
3321 | * 7 (strlen("vms110") + 1) | 3321 | * 7 (strlen("vms110") + 1) |
3322 | * -- VPD Instance 'I' | 3322 | * -- VPD Instance 'I' |
3323 | * -- VPD Type String 'B' | 3323 | * -- VPD Type String 'B' |
@@ -3325,17 +3325,17 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3325 | * -- property string == entropy-dev | 3325 | * -- property string == entropy-dev |
3326 | * | 3326 | * |
3327 | * -- correct length == 18 | 3327 | * -- correct length == 18 |
3328 | * 3 (type) + 2 (size) + | 3328 | * 3 (type) + 2 (size) + |
3329 | * 9 (strlen("phy-type") + 1) + | 3329 | * 9 (strlen("phy-type") + 1) + |
3330 | * 4 (strlen("pcs") + 1) | 3330 | * 4 (strlen("pcs") + 1) |
3331 | * -- VPD Instance 'I' | 3331 | * -- VPD Instance 'I' |
3332 | * -- VPD Type String 'S' | 3332 | * -- VPD Type String 'S' |
3333 | * -- VPD data length == 4 | 3333 | * -- VPD data length == 4 |
3334 | * -- property string == phy-type | 3334 | * -- property string == phy-type |
3335 | * | 3335 | * |
3336 | * -- correct length == 23 | 3336 | * -- correct length == 23 |
3337 | * 3 (type) + 2 (size) + | 3337 | * 3 (type) + 2 (size) + |
3338 | * 14 (strlen("phy-interface") + 1) + | 3338 | * 14 (strlen("phy-interface") + 1) + |
3339 | * 4 (strlen("pcs") + 1) | 3339 | * 4 (strlen("pcs") + 1) |
3340 | * -- VPD Instance 'I' | 3340 | * -- VPD Instance 'I' |
3341 | * -- VPD Type String 'S' | 3341 | * -- VPD Type String 'S' |
@@ -3349,14 +3349,14 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3349 | type = readb(p + 3); | 3349 | type = readb(p + 3); |
3350 | if (type == 'B') { | 3350 | if (type == 'B') { |
3351 | if ((klen == 29) && readb(p + 4) == 6 && | 3351 | if ((klen == 29) && readb(p + 4) == 6 && |
3352 | cas_vpd_match(p + 5, | 3352 | cas_vpd_match(p + 5, |
3353 | "local-mac-address")) { | 3353 | "local-mac-address")) { |
3354 | if (mac_off++ > offset) | 3354 | if (mac_off++ > offset) |
3355 | goto next; | 3355 | goto next; |
3356 | 3356 | ||
3357 | /* set mac address */ | 3357 | /* set mac address */ |
3358 | for (j = 0; j < 6; j++) | 3358 | for (j = 0; j < 6; j++) |
3359 | dev_addr[j] = | 3359 | dev_addr[j] = |
3360 | readb(p + 23 + j); | 3360 | readb(p + 23 + j); |
3361 | goto found_mac; | 3361 | goto found_mac; |
3362 | } | 3362 | } |
@@ -3366,7 +3366,7 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3366 | goto next; | 3366 | goto next; |
3367 | 3367 | ||
3368 | #ifdef USE_ENTROPY_DEV | 3368 | #ifdef USE_ENTROPY_DEV |
3369 | if ((klen == 24) && | 3369 | if ((klen == 24) && |
3370 | cas_vpd_match(p + 5, "entropy-dev") && | 3370 | cas_vpd_match(p + 5, "entropy-dev") && |
3371 | cas_vpd_match(p + 17, "vms110")) { | 3371 | cas_vpd_match(p + 17, "vms110")) { |
3372 | cp->cas_flags |= CAS_FLAG_ENTROPY_DEV; | 3372 | cp->cas_flags |= CAS_FLAG_ENTROPY_DEV; |
@@ -3384,7 +3384,7 @@ static int cas_get_vpd_info(struct cas *cp, unsigned char *dev_addr, | |||
3384 | goto found_phy; | 3384 | goto found_phy; |
3385 | } | 3385 | } |
3386 | } | 3386 | } |
3387 | 3387 | ||
3388 | if ((klen == 23) && readb(p + 4) == 4 && | 3388 | if ((klen == 23) && readb(p + 4) == 4 && |
3389 | cas_vpd_match(p + 5, "phy-interface")) { | 3389 | cas_vpd_match(p + 5, "phy-interface")) { |
3390 | if (cas_vpd_match(p + 19, "pcs")) { | 3390 | if (cas_vpd_match(p + 19, "pcs")) { |
@@ -3462,12 +3462,12 @@ static int cas_check_invariants(struct cas *cp) | |||
3462 | int i; | 3462 | int i; |
3463 | 3463 | ||
3464 | /* get page size for rx buffers. */ | 3464 | /* get page size for rx buffers. */ |
3465 | cp->page_order = 0; | 3465 | cp->page_order = 0; |
3466 | #ifdef USE_PAGE_ORDER | 3466 | #ifdef USE_PAGE_ORDER |
3467 | if (PAGE_SHIFT < CAS_JUMBO_PAGE_SHIFT) { | 3467 | if (PAGE_SHIFT < CAS_JUMBO_PAGE_SHIFT) { |
3468 | /* see if we can allocate larger pages */ | 3468 | /* see if we can allocate larger pages */ |
3469 | struct page *page = alloc_pages(GFP_ATOMIC, | 3469 | struct page *page = alloc_pages(GFP_ATOMIC, |
3470 | CAS_JUMBO_PAGE_SHIFT - | 3470 | CAS_JUMBO_PAGE_SHIFT - |
3471 | PAGE_SHIFT); | 3471 | PAGE_SHIFT); |
3472 | if (page) { | 3472 | if (page) { |
3473 | __free_pages(page, CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT); | 3473 | __free_pages(page, CAS_JUMBO_PAGE_SHIFT - PAGE_SHIFT); |
@@ -3483,15 +3483,15 @@ static int cas_check_invariants(struct cas *cp) | |||
3483 | cp->tx_fifo_size = readl(cp->regs + REG_TX_FIFO_SIZE) * 64; | 3483 | cp->tx_fifo_size = readl(cp->regs + REG_TX_FIFO_SIZE) * 64; |
3484 | cp->rx_fifo_size = RX_FIFO_SIZE; | 3484 | cp->rx_fifo_size = RX_FIFO_SIZE; |
3485 | 3485 | ||
3486 | /* finish phy determination. MDIO1 takes precedence over MDIO0 if | 3486 | /* finish phy determination. MDIO1 takes precedence over MDIO0 if |
3487 | * they're both connected. | 3487 | * they're both connected. |
3488 | */ | 3488 | */ |
3489 | cp->phy_type = cas_get_vpd_info(cp, cp->dev->dev_addr, | 3489 | cp->phy_type = cas_get_vpd_info(cp, cp->dev->dev_addr, |
3490 | PCI_SLOT(pdev->devfn)); | 3490 | PCI_SLOT(pdev->devfn)); |
3491 | if (cp->phy_type & CAS_PHY_SERDES) { | 3491 | if (cp->phy_type & CAS_PHY_SERDES) { |
3492 | cp->cas_flags |= CAS_FLAG_1000MB_CAP; | 3492 | cp->cas_flags |= CAS_FLAG_1000MB_CAP; |
3493 | return 0; /* no more checking needed */ | 3493 | return 0; /* no more checking needed */ |
3494 | } | 3494 | } |
3495 | 3495 | ||
3496 | /* MII */ | 3496 | /* MII */ |
3497 | cfg = readl(cp->regs + REG_MIF_CFG); | 3497 | cfg = readl(cp->regs + REG_MIF_CFG); |
@@ -3525,7 +3525,7 @@ static int cas_check_invariants(struct cas *cp) | |||
3525 | done: | 3525 | done: |
3526 | /* see if we can do gigabit */ | 3526 | /* see if we can do gigabit */ |
3527 | cfg = cas_phy_read(cp, MII_BMSR); | 3527 | cfg = cas_phy_read(cp, MII_BMSR); |
3528 | if ((cfg & CAS_BMSR_1000_EXTEND) && | 3528 | if ((cfg & CAS_BMSR_1000_EXTEND) && |
3529 | cas_phy_read(cp, CAS_MII_1000_EXTEND)) | 3529 | cas_phy_read(cp, CAS_MII_1000_EXTEND)) |
3530 | cp->cas_flags |= CAS_FLAG_1000MB_CAP; | 3530 | cp->cas_flags |= CAS_FLAG_1000MB_CAP; |
3531 | return 0; | 3531 | return 0; |
@@ -3537,7 +3537,7 @@ static inline void cas_start_dma(struct cas *cp) | |||
3537 | int i; | 3537 | int i; |
3538 | u32 val; | 3538 | u32 val; |
3539 | int txfailed = 0; | 3539 | int txfailed = 0; |
3540 | 3540 | ||
3541 | /* enable dma */ | 3541 | /* enable dma */ |
3542 | val = readl(cp->regs + REG_TX_CFG) | TX_CFG_DMA_EN; | 3542 | val = readl(cp->regs + REG_TX_CFG) | TX_CFG_DMA_EN; |
3543 | writel(val, cp->regs + REG_TX_CFG); | 3543 | writel(val, cp->regs + REG_TX_CFG); |
@@ -3563,8 +3563,8 @@ static inline void cas_start_dma(struct cas *cp) | |||
3563 | val = readl(cp->regs + REG_MAC_RX_CFG); | 3563 | val = readl(cp->regs + REG_MAC_RX_CFG); |
3564 | if ((val & MAC_RX_CFG_EN)) { | 3564 | if ((val & MAC_RX_CFG_EN)) { |
3565 | if (txfailed) { | 3565 | if (txfailed) { |
3566 | printk(KERN_ERR | 3566 | printk(KERN_ERR |
3567 | "%s: enabling mac failed [tx:%08x:%08x].\n", | 3567 | "%s: enabling mac failed [tx:%08x:%08x].\n", |
3568 | cp->dev->name, | 3568 | cp->dev->name, |
3569 | readl(cp->regs + REG_MIF_STATE_MACHINE), | 3569 | readl(cp->regs + REG_MIF_STATE_MACHINE), |
3570 | readl(cp->regs + REG_MAC_STATE_MACHINE)); | 3570 | readl(cp->regs + REG_MAC_STATE_MACHINE)); |
@@ -3573,7 +3573,7 @@ static inline void cas_start_dma(struct cas *cp) | |||
3573 | } | 3573 | } |
3574 | udelay(10); | 3574 | udelay(10); |
3575 | } | 3575 | } |
3576 | printk(KERN_ERR "%s: enabling mac failed [%s:%08x:%08x].\n", | 3576 | printk(KERN_ERR "%s: enabling mac failed [%s:%08x:%08x].\n", |
3577 | cp->dev->name, | 3577 | cp->dev->name, |
3578 | (txfailed? "tx,rx":"rx"), | 3578 | (txfailed? "tx,rx":"rx"), |
3579 | readl(cp->regs + REG_MIF_STATE_MACHINE), | 3579 | readl(cp->regs + REG_MIF_STATE_MACHINE), |
@@ -3585,11 +3585,11 @@ enable_rx_done: | |||
3585 | writel(0, cp->regs + REG_RX_COMP_TAIL); | 3585 | writel(0, cp->regs + REG_RX_COMP_TAIL); |
3586 | 3586 | ||
3587 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { | 3587 | if (cp->cas_flags & CAS_FLAG_REG_PLUS) { |
3588 | if (N_RX_DESC_RINGS > 1) | 3588 | if (N_RX_DESC_RINGS > 1) |
3589 | writel(RX_DESC_RINGN_SIZE(1) - 4, | 3589 | writel(RX_DESC_RINGN_SIZE(1) - 4, |
3590 | cp->regs + REG_PLUS_RX_KICK1); | 3590 | cp->regs + REG_PLUS_RX_KICK1); |
3591 | 3591 | ||
3592 | for (i = 1; i < N_RX_COMP_RINGS; i++) | 3592 | for (i = 1; i < N_RX_COMP_RINGS; i++) |
3593 | writel(0, cp->regs + REG_PLUS_RX_COMPN_TAIL(i)); | 3593 | writel(0, cp->regs + REG_PLUS_RX_COMPN_TAIL(i)); |
3594 | } | 3594 | } |
3595 | } | 3595 | } |
@@ -3615,7 +3615,7 @@ static void cas_read_mii_link_mode(struct cas *cp, int *fd, int *spd, | |||
3615 | *fd = 0; | 3615 | *fd = 0; |
3616 | *spd = 10; | 3616 | *spd = 10; |
3617 | *pause = 0; | 3617 | *pause = 0; |
3618 | 3618 | ||
3619 | /* use GMII registers */ | 3619 | /* use GMII registers */ |
3620 | val = cas_phy_read(cp, MII_LPA); | 3620 | val = cas_phy_read(cp, MII_LPA); |
3621 | if (val & CAS_LPA_PAUSE) | 3621 | if (val & CAS_LPA_PAUSE) |
@@ -3656,7 +3656,7 @@ static void cas_set_link_modes(struct cas *cp) | |||
3656 | cas_mif_poll(cp, 0); | 3656 | cas_mif_poll(cp, 0); |
3657 | val = cas_phy_read(cp, MII_BMCR); | 3657 | val = cas_phy_read(cp, MII_BMCR); |
3658 | if (val & BMCR_ANENABLE) { | 3658 | if (val & BMCR_ANENABLE) { |
3659 | cas_read_mii_link_mode(cp, &full_duplex, &speed, | 3659 | cas_read_mii_link_mode(cp, &full_duplex, &speed, |
3660 | &pause); | 3660 | &pause); |
3661 | } else { | 3661 | } else { |
3662 | if (val & BMCR_FULLDPLX) | 3662 | if (val & BMCR_FULLDPLX) |
@@ -3689,7 +3689,7 @@ static void cas_set_link_modes(struct cas *cp) | |||
3689 | if (!full_duplex) | 3689 | if (!full_duplex) |
3690 | val |= MAC_XIF_DISABLE_ECHO; | 3690 | val |= MAC_XIF_DISABLE_ECHO; |
3691 | } | 3691 | } |
3692 | if (full_duplex) | 3692 | if (full_duplex) |
3693 | val |= MAC_XIF_FDPLX_LED; | 3693 | val |= MAC_XIF_FDPLX_LED; |
3694 | if (speed == 1000) | 3694 | if (speed == 1000) |
3695 | val |= MAC_XIF_GMII_MODE; | 3695 | val |= MAC_XIF_GMII_MODE; |
@@ -3709,17 +3709,17 @@ static void cas_set_link_modes(struct cas *cp) | |||
3709 | /* val now set up for REG_MAC_TX_CFG */ | 3709 | /* val now set up for REG_MAC_TX_CFG */ |
3710 | 3710 | ||
3711 | /* If gigabit and half-duplex, enable carrier extension | 3711 | /* If gigabit and half-duplex, enable carrier extension |
3712 | * mode. increase slot time to 512 bytes as well. | 3712 | * mode. increase slot time to 512 bytes as well. |
3713 | * else, disable it and make sure slot time is 64 bytes. | 3713 | * else, disable it and make sure slot time is 64 bytes. |
3714 | * also activate checksum bug workaround | 3714 | * also activate checksum bug workaround |
3715 | */ | 3715 | */ |
3716 | if ((speed == 1000) && !full_duplex) { | 3716 | if ((speed == 1000) && !full_duplex) { |
3717 | writel(val | MAC_TX_CFG_CARRIER_EXTEND, | 3717 | writel(val | MAC_TX_CFG_CARRIER_EXTEND, |
3718 | cp->regs + REG_MAC_TX_CFG); | 3718 | cp->regs + REG_MAC_TX_CFG); |
3719 | 3719 | ||
3720 | val = readl(cp->regs + REG_MAC_RX_CFG); | 3720 | val = readl(cp->regs + REG_MAC_RX_CFG); |
3721 | val &= ~MAC_RX_CFG_STRIP_FCS; /* checksum workaround */ | 3721 | val &= ~MAC_RX_CFG_STRIP_FCS; /* checksum workaround */ |
3722 | writel(val | MAC_RX_CFG_CARRIER_EXTEND, | 3722 | writel(val | MAC_RX_CFG_CARRIER_EXTEND, |
3723 | cp->regs + REG_MAC_RX_CFG); | 3723 | cp->regs + REG_MAC_RX_CFG); |
3724 | 3724 | ||
3725 | writel(0x200, cp->regs + REG_MAC_SLOT_TIME); | 3725 | writel(0x200, cp->regs + REG_MAC_SLOT_TIME); |
@@ -3731,7 +3731,7 @@ static void cas_set_link_modes(struct cas *cp) | |||
3731 | } else { | 3731 | } else { |
3732 | writel(val, cp->regs + REG_MAC_TX_CFG); | 3732 | writel(val, cp->regs + REG_MAC_TX_CFG); |
3733 | 3733 | ||
3734 | /* checksum bug workaround. don't strip FCS when in | 3734 | /* checksum bug workaround. don't strip FCS when in |
3735 | * half-duplex mode | 3735 | * half-duplex mode |
3736 | */ | 3736 | */ |
3737 | val = readl(cp->regs + REG_MAC_RX_CFG); | 3737 | val = readl(cp->regs + REG_MAC_RX_CFG); |
@@ -3744,7 +3744,7 @@ static void cas_set_link_modes(struct cas *cp) | |||
3744 | cp->crc_size = 4; | 3744 | cp->crc_size = 4; |
3745 | cp->min_frame_size = CAS_MIN_FRAME; | 3745 | cp->min_frame_size = CAS_MIN_FRAME; |
3746 | } | 3746 | } |
3747 | writel(val & ~MAC_RX_CFG_CARRIER_EXTEND, | 3747 | writel(val & ~MAC_RX_CFG_CARRIER_EXTEND, |
3748 | cp->regs + REG_MAC_RX_CFG); | 3748 | cp->regs + REG_MAC_RX_CFG); |
3749 | writel(0x40, cp->regs + REG_MAC_SLOT_TIME); | 3749 | writel(0x40, cp->regs + REG_MAC_SLOT_TIME); |
3750 | } | 3750 | } |
@@ -3772,7 +3772,7 @@ static void cas_set_link_modes(struct cas *cp) | |||
3772 | val |= MAC_CTRL_CFG_SEND_PAUSE_EN; | 3772 | val |= MAC_CTRL_CFG_SEND_PAUSE_EN; |
3773 | if (pause & 0x01) { /* symmetric pause */ | 3773 | if (pause & 0x01) { /* symmetric pause */ |
3774 | val |= MAC_CTRL_CFG_RECV_PAUSE_EN; | 3774 | val |= MAC_CTRL_CFG_RECV_PAUSE_EN; |
3775 | } | 3775 | } |
3776 | } | 3776 | } |
3777 | writel(val, cp->regs + REG_MAC_CTRL_CFG); | 3777 | writel(val, cp->regs + REG_MAC_CTRL_CFG); |
3778 | cas_start_dma(cp); | 3778 | cas_start_dma(cp); |
@@ -3804,7 +3804,7 @@ static void cas_init_hw(struct cas *cp, int restart_link) | |||
3804 | */ | 3804 | */ |
3805 | static void cas_hard_reset(struct cas *cp) | 3805 | static void cas_hard_reset(struct cas *cp) |
3806 | { | 3806 | { |
3807 | writel(BIM_LOCAL_DEV_SOFT_0, cp->regs + REG_BIM_LOCAL_DEV_EN); | 3807 | writel(BIM_LOCAL_DEV_SOFT_0, cp->regs + REG_BIM_LOCAL_DEV_EN); |
3808 | udelay(20); | 3808 | udelay(20); |
3809 | pci_restore_state(cp->pdev); | 3809 | pci_restore_state(cp->pdev); |
3810 | } | 3810 | } |
@@ -3822,7 +3822,7 @@ static void cas_global_reset(struct cas *cp, int blkflag) | |||
3822 | * need some special handling if the chip is set into a | 3822 | * need some special handling if the chip is set into a |
3823 | * loopback mode. | 3823 | * loopback mode. |
3824 | */ | 3824 | */ |
3825 | writel((SW_RESET_TX | SW_RESET_RX | SW_RESET_BLOCK_PCS_SLINK), | 3825 | writel((SW_RESET_TX | SW_RESET_RX | SW_RESET_BLOCK_PCS_SLINK), |
3826 | cp->regs + REG_SW_RESET); | 3826 | cp->regs + REG_SW_RESET); |
3827 | } else { | 3827 | } else { |
3828 | writel(SW_RESET_TX | SW_RESET_RX, cp->regs + REG_SW_RESET); | 3828 | writel(SW_RESET_TX | SW_RESET_RX, cp->regs + REG_SW_RESET); |
@@ -3842,16 +3842,16 @@ static void cas_global_reset(struct cas *cp, int blkflag) | |||
3842 | 3842 | ||
3843 | done: | 3843 | done: |
3844 | /* enable various BIM interrupts */ | 3844 | /* enable various BIM interrupts */ |
3845 | writel(BIM_CFG_DPAR_INTR_ENABLE | BIM_CFG_RMA_INTR_ENABLE | | 3845 | writel(BIM_CFG_DPAR_INTR_ENABLE | BIM_CFG_RMA_INTR_ENABLE | |
3846 | BIM_CFG_RTA_INTR_ENABLE, cp->regs + REG_BIM_CFG); | 3846 | BIM_CFG_RTA_INTR_ENABLE, cp->regs + REG_BIM_CFG); |
3847 | 3847 | ||
3848 | /* clear out pci error status mask for handled errors. | 3848 | /* clear out pci error status mask for handled errors. |
3849 | * we don't deal with DMA counter overflows as they happen | 3849 | * we don't deal with DMA counter overflows as they happen |
3850 | * all the time. | 3850 | * all the time. |
3851 | */ | 3851 | */ |
3852 | writel(0xFFFFFFFFU & ~(PCI_ERR_BADACK | PCI_ERR_DTRTO | | 3852 | writel(0xFFFFFFFFU & ~(PCI_ERR_BADACK | PCI_ERR_DTRTO | |
3853 | PCI_ERR_OTHER | PCI_ERR_BIM_DMA_WRITE | | 3853 | PCI_ERR_OTHER | PCI_ERR_BIM_DMA_WRITE | |
3854 | PCI_ERR_BIM_DMA_READ), cp->regs + | 3854 | PCI_ERR_BIM_DMA_READ), cp->regs + |
3855 | REG_PCI_ERR_STATUS_MASK); | 3855 | REG_PCI_ERR_STATUS_MASK); |
3856 | 3856 | ||
3857 | /* set up for MII by default to address mac rx reset timeout | 3857 | /* set up for MII by default to address mac rx reset timeout |
@@ -3912,7 +3912,7 @@ static void cas_shutdown(struct cas *cp) | |||
3912 | #else | 3912 | #else |
3913 | while (atomic_read(&cp->reset_task_pending)) | 3913 | while (atomic_read(&cp->reset_task_pending)) |
3914 | schedule(); | 3914 | schedule(); |
3915 | #endif | 3915 | #endif |
3916 | /* Actually stop the chip */ | 3916 | /* Actually stop the chip */ |
3917 | cas_lock_all_save(cp, flags); | 3917 | cas_lock_all_save(cp, flags); |
3918 | cas_reset(cp, 0); | 3918 | cas_reset(cp, 0); |
@@ -3942,7 +3942,7 @@ static int cas_change_mtu(struct net_device *dev, int new_mtu) | |||
3942 | } | 3942 | } |
3943 | schedule_work(&cp->reset_task); | 3943 | schedule_work(&cp->reset_task); |
3944 | #else | 3944 | #else |
3945 | atomic_set(&cp->reset_task_pending, (cp->phy_type & CAS_PHY_SERDES) ? | 3945 | atomic_set(&cp->reset_task_pending, (cp->phy_type & CAS_PHY_SERDES) ? |
3946 | CAS_RESET_ALL : CAS_RESET_MTU); | 3946 | CAS_RESET_ALL : CAS_RESET_MTU); |
3947 | printk(KERN_ERR "reset called in cas_change_mtu\n"); | 3947 | printk(KERN_ERR "reset called in cas_change_mtu\n"); |
3948 | schedule_work(&cp->reset_task); | 3948 | schedule_work(&cp->reset_task); |
@@ -3976,7 +3976,7 @@ static void cas_clean_txd(struct cas *cp, int ring) | |||
3976 | * needs to be unmapped. | 3976 | * needs to be unmapped. |
3977 | */ | 3977 | */ |
3978 | daddr = le64_to_cpu(txd[ent].buffer); | 3978 | daddr = le64_to_cpu(txd[ent].buffer); |
3979 | dlen = CAS_VAL(TX_DESC_BUFLEN, | 3979 | dlen = CAS_VAL(TX_DESC_BUFLEN, |
3980 | le64_to_cpu(txd[ent].control)); | 3980 | le64_to_cpu(txd[ent].control)); |
3981 | pci_unmap_page(cp->pdev, daddr, dlen, | 3981 | pci_unmap_page(cp->pdev, daddr, dlen, |
3982 | PCI_DMA_TODEVICE); | 3982 | PCI_DMA_TODEVICE); |
@@ -4047,7 +4047,7 @@ static inline int cas_alloc_rx_desc(struct cas *cp, int ring) | |||
4047 | 4047 | ||
4048 | size = RX_DESC_RINGN_SIZE(ring); | 4048 | size = RX_DESC_RINGN_SIZE(ring); |
4049 | for (i = 0; i < size; i++) { | 4049 | for (i = 0; i < size; i++) { |
4050 | if ((page[i] = cas_page_alloc(cp, GFP_KERNEL)) == NULL) | 4050 | if ((page[i] = cas_page_alloc(cp, GFP_KERNEL)) == NULL) |
4051 | return -1; | 4051 | return -1; |
4052 | } | 4052 | } |
4053 | return 0; | 4053 | return 0; |
@@ -4114,7 +4114,7 @@ static void cas_reset_task(void *data) | |||
4114 | * call to cas_init_hw will restart auto negotiation. | 4114 | * call to cas_init_hw will restart auto negotiation. |
4115 | * Setting the second argument of cas_reset to | 4115 | * Setting the second argument of cas_reset to |
4116 | * !(pending == CAS_RESET_ALL) will set this argument | 4116 | * !(pending == CAS_RESET_ALL) will set this argument |
4117 | * to 1 (avoiding reinitializing the PHY for the normal | 4117 | * to 1 (avoiding reinitializing the PHY for the normal |
4118 | * PCS case) when auto negotiation is not restarted. | 4118 | * PCS case) when auto negotiation is not restarted. |
4119 | */ | 4119 | */ |
4120 | #if 1 | 4120 | #if 1 |
@@ -4151,9 +4151,9 @@ static void cas_link_timer(unsigned long data) | |||
4151 | 4151 | ||
4152 | if (link_transition_timeout != 0 && | 4152 | if (link_transition_timeout != 0 && |
4153 | cp->link_transition_jiffies_valid && | 4153 | cp->link_transition_jiffies_valid && |
4154 | ((jiffies - cp->link_transition_jiffies) > | 4154 | ((jiffies - cp->link_transition_jiffies) > |
4155 | (link_transition_timeout))) { | 4155 | (link_transition_timeout))) { |
4156 | /* One-second counter so link-down workaround doesn't | 4156 | /* One-second counter so link-down workaround doesn't |
4157 | * cause resets to occur so fast as to fool the switch | 4157 | * cause resets to occur so fast as to fool the switch |
4158 | * into thinking the link is down. | 4158 | * into thinking the link is down. |
4159 | */ | 4159 | */ |
@@ -4173,10 +4173,10 @@ static void cas_link_timer(unsigned long data) | |||
4173 | #if 1 | 4173 | #if 1 |
4174 | if (atomic_read(&cp->reset_task_pending_all) || | 4174 | if (atomic_read(&cp->reset_task_pending_all) || |
4175 | atomic_read(&cp->reset_task_pending_spare) || | 4175 | atomic_read(&cp->reset_task_pending_spare) || |
4176 | atomic_read(&cp->reset_task_pending_mtu)) | 4176 | atomic_read(&cp->reset_task_pending_mtu)) |
4177 | goto done; | 4177 | goto done; |
4178 | #else | 4178 | #else |
4179 | if (atomic_read(&cp->reset_task_pending)) | 4179 | if (atomic_read(&cp->reset_task_pending)) |
4180 | goto done; | 4180 | goto done; |
4181 | #endif | 4181 | #endif |
4182 | 4182 | ||
@@ -4268,7 +4268,7 @@ done: | |||
4268 | spin_unlock_irqrestore(&cp->lock, flags); | 4268 | spin_unlock_irqrestore(&cp->lock, flags); |
4269 | } | 4269 | } |
4270 | 4270 | ||
4271 | /* tiny buffers are used to avoid target abort issues with | 4271 | /* tiny buffers are used to avoid target abort issues with |
4272 | * older cassini's | 4272 | * older cassini's |
4273 | */ | 4273 | */ |
4274 | static void cas_tx_tiny_free(struct cas *cp) | 4274 | static void cas_tx_tiny_free(struct cas *cp) |
@@ -4280,7 +4280,7 @@ static void cas_tx_tiny_free(struct cas *cp) | |||
4280 | if (!cp->tx_tiny_bufs[i]) | 4280 | if (!cp->tx_tiny_bufs[i]) |
4281 | continue; | 4281 | continue; |
4282 | 4282 | ||
4283 | pci_free_consistent(pdev, TX_TINY_BUF_BLOCK, | 4283 | pci_free_consistent(pdev, TX_TINY_BUF_BLOCK, |
4284 | cp->tx_tiny_bufs[i], | 4284 | cp->tx_tiny_bufs[i], |
4285 | cp->tx_tiny_dvma[i]); | 4285 | cp->tx_tiny_dvma[i]); |
4286 | cp->tx_tiny_bufs[i] = NULL; | 4286 | cp->tx_tiny_bufs[i] = NULL; |
@@ -4293,7 +4293,7 @@ static int cas_tx_tiny_alloc(struct cas *cp) | |||
4293 | int i; | 4293 | int i; |
4294 | 4294 | ||
4295 | for (i = 0; i < N_TX_RINGS; i++) { | 4295 | for (i = 0; i < N_TX_RINGS; i++) { |
4296 | cp->tx_tiny_bufs[i] = | 4296 | cp->tx_tiny_bufs[i] = |
4297 | pci_alloc_consistent(pdev, TX_TINY_BUF_BLOCK, | 4297 | pci_alloc_consistent(pdev, TX_TINY_BUF_BLOCK, |
4298 | &cp->tx_tiny_dvma[i]); | 4298 | &cp->tx_tiny_dvma[i]); |
4299 | if (!cp->tx_tiny_bufs[i]) { | 4299 | if (!cp->tx_tiny_bufs[i]) { |
@@ -4322,7 +4322,7 @@ static int cas_open(struct net_device *dev) | |||
4322 | /* Reset the chip */ | 4322 | /* Reset the chip */ |
4323 | cas_lock_all_save(cp, flags); | 4323 | cas_lock_all_save(cp, flags); |
4324 | /* We set the second arg to cas_reset to zero | 4324 | /* We set the second arg to cas_reset to zero |
4325 | * because cas_init_hw below will have its second | 4325 | * because cas_init_hw below will have its second |
4326 | * argument set to non-zero, which will force | 4326 | * argument set to non-zero, which will force |
4327 | * autonegotiation to start. | 4327 | * autonegotiation to start. |
4328 | */ | 4328 | */ |
@@ -4338,19 +4338,19 @@ static int cas_open(struct net_device *dev) | |||
4338 | err = -ENOMEM; | 4338 | err = -ENOMEM; |
4339 | if (cas_alloc_rxds(cp) < 0) | 4339 | if (cas_alloc_rxds(cp) < 0) |
4340 | goto err_tx_tiny; | 4340 | goto err_tx_tiny; |
4341 | 4341 | ||
4342 | /* allocate spares */ | 4342 | /* allocate spares */ |
4343 | cas_spare_init(cp); | 4343 | cas_spare_init(cp); |
4344 | cas_spare_recover(cp, GFP_KERNEL); | 4344 | cas_spare_recover(cp, GFP_KERNEL); |
4345 | 4345 | ||
4346 | /* We can now request the interrupt as we know it's masked | 4346 | /* We can now request the interrupt as we know it's masked |
4347 | * on the controller. cassini+ has up to 4 interrupts | 4347 | * on the controller. cassini+ has up to 4 interrupts |
4348 | * that can be used, but you need to do explicit pci interrupt | 4348 | * that can be used, but you need to do explicit pci interrupt |
4349 | * mapping to expose them | 4349 | * mapping to expose them |
4350 | */ | 4350 | */ |
4351 | if (request_irq(cp->pdev->irq, cas_interrupt, | 4351 | if (request_irq(cp->pdev->irq, cas_interrupt, |
4352 | IRQF_SHARED, dev->name, (void *) dev)) { | 4352 | IRQF_SHARED, dev->name, (void *) dev)) { |
4353 | printk(KERN_ERR "%s: failed to request irq !\n", | 4353 | printk(KERN_ERR "%s: failed to request irq !\n", |
4354 | cp->dev->name); | 4354 | cp->dev->name); |
4355 | err = -EAGAIN; | 4355 | err = -EAGAIN; |
4356 | goto err_spare; | 4356 | goto err_spare; |
@@ -4388,9 +4388,9 @@ static int cas_close(struct net_device *dev) | |||
4388 | 4388 | ||
4389 | /* Stop traffic, mark us closed */ | 4389 | /* Stop traffic, mark us closed */ |
4390 | cas_lock_all_save(cp, flags); | 4390 | cas_lock_all_save(cp, flags); |
4391 | cp->opened = 0; | 4391 | cp->opened = 0; |
4392 | cas_reset(cp, 0); | 4392 | cas_reset(cp, 0); |
4393 | cas_phy_init(cp); | 4393 | cas_phy_init(cp); |
4394 | cas_begin_auto_negotiation(cp, NULL); | 4394 | cas_begin_auto_negotiation(cp, NULL); |
4395 | cas_clean_rings(cp); | 4395 | cas_clean_rings(cp); |
4396 | cas_unlock_all_restore(cp, flags); | 4396 | cas_unlock_all_restore(cp, flags); |
@@ -4483,7 +4483,7 @@ static struct net_device_stats *cas_get_stats(struct net_device *dev) | |||
4483 | /* we collate all of the stats into net_stats[N_TX_RING] */ | 4483 | /* we collate all of the stats into net_stats[N_TX_RING] */ |
4484 | if (!cp->hw_running) | 4484 | if (!cp->hw_running) |
4485 | return stats + N_TX_RINGS; | 4485 | return stats + N_TX_RINGS; |
4486 | 4486 | ||
4487 | /* collect outstanding stats */ | 4487 | /* collect outstanding stats */ |
4488 | /* WTZ: the Cassini spec gives these as 16 bit counters but | 4488 | /* WTZ: the Cassini spec gives these as 16 bit counters but |
4489 | * stored in 32-bit words. Added a mask of 0xffff to be safe, | 4489 | * stored in 32-bit words. Added a mask of 0xffff to be safe, |
@@ -4493,11 +4493,11 @@ static struct net_device_stats *cas_get_stats(struct net_device *dev) | |||
4493 | * that consistent. | 4493 | * that consistent. |
4494 | */ | 4494 | */ |
4495 | spin_lock_irqsave(&cp->stat_lock[N_TX_RINGS], flags); | 4495 | spin_lock_irqsave(&cp->stat_lock[N_TX_RINGS], flags); |
4496 | stats[N_TX_RINGS].rx_crc_errors += | 4496 | stats[N_TX_RINGS].rx_crc_errors += |
4497 | readl(cp->regs + REG_MAC_FCS_ERR) & 0xffff; | 4497 | readl(cp->regs + REG_MAC_FCS_ERR) & 0xffff; |
4498 | stats[N_TX_RINGS].rx_frame_errors += | 4498 | stats[N_TX_RINGS].rx_frame_errors += |
4499 | readl(cp->regs + REG_MAC_ALIGN_ERR) &0xffff; | 4499 | readl(cp->regs + REG_MAC_ALIGN_ERR) &0xffff; |
4500 | stats[N_TX_RINGS].rx_length_errors += | 4500 | stats[N_TX_RINGS].rx_length_errors += |
4501 | readl(cp->regs + REG_MAC_LEN_ERR) & 0xffff; | 4501 | readl(cp->regs + REG_MAC_LEN_ERR) & 0xffff; |
4502 | #if 1 | 4502 | #if 1 |
4503 | tmp = (readl(cp->regs + REG_MAC_COLL_EXCESS) & 0xffff) + | 4503 | tmp = (readl(cp->regs + REG_MAC_COLL_EXCESS) & 0xffff) + |
@@ -4506,7 +4506,7 @@ static struct net_device_stats *cas_get_stats(struct net_device *dev) | |||
4506 | stats[N_TX_RINGS].collisions += | 4506 | stats[N_TX_RINGS].collisions += |
4507 | tmp + (readl(cp->regs + REG_MAC_COLL_NORMAL) & 0xffff); | 4507 | tmp + (readl(cp->regs + REG_MAC_COLL_NORMAL) & 0xffff); |
4508 | #else | 4508 | #else |
4509 | stats[N_TX_RINGS].tx_aborted_errors += | 4509 | stats[N_TX_RINGS].tx_aborted_errors += |
4510 | readl(cp->regs + REG_MAC_COLL_EXCESS); | 4510 | readl(cp->regs + REG_MAC_COLL_EXCESS); |
4511 | stats[N_TX_RINGS].collisions += readl(cp->regs + REG_MAC_COLL_EXCESS) + | 4511 | stats[N_TX_RINGS].collisions += readl(cp->regs + REG_MAC_COLL_EXCESS) + |
4512 | readl(cp->regs + REG_MAC_COLL_LATE); | 4512 | readl(cp->regs + REG_MAC_COLL_LATE); |
@@ -4525,7 +4525,7 @@ static struct net_device_stats *cas_get_stats(struct net_device *dev) | |||
4525 | 4525 | ||
4526 | for (i = 0; i < N_TX_RINGS; i++) { | 4526 | for (i = 0; i < N_TX_RINGS; i++) { |
4527 | spin_lock(&cp->stat_lock[i]); | 4527 | spin_lock(&cp->stat_lock[i]); |
4528 | stats[N_TX_RINGS].rx_length_errors += | 4528 | stats[N_TX_RINGS].rx_length_errors += |
4529 | stats[i].rx_length_errors; | 4529 | stats[i].rx_length_errors; |
4530 | stats[N_TX_RINGS].rx_crc_errors += stats[i].rx_crc_errors; | 4530 | stats[N_TX_RINGS].rx_crc_errors += stats[i].rx_crc_errors; |
4531 | stats[N_TX_RINGS].rx_packets += stats[i].rx_packets; | 4531 | stats[N_TX_RINGS].rx_packets += stats[i].rx_packets; |
@@ -4550,10 +4550,10 @@ static void cas_set_multicast(struct net_device *dev) | |||
4550 | u32 rxcfg, rxcfg_new; | 4550 | u32 rxcfg, rxcfg_new; |
4551 | unsigned long flags; | 4551 | unsigned long flags; |
4552 | int limit = STOP_TRIES; | 4552 | int limit = STOP_TRIES; |
4553 | 4553 | ||
4554 | if (!cp->hw_running) | 4554 | if (!cp->hw_running) |
4555 | return; | 4555 | return; |
4556 | 4556 | ||
4557 | spin_lock_irqsave(&cp->lock, flags); | 4557 | spin_lock_irqsave(&cp->lock, flags); |
4558 | rxcfg = readl(cp->regs + REG_MAC_RX_CFG); | 4558 | rxcfg = readl(cp->regs + REG_MAC_RX_CFG); |
4559 | 4559 | ||
@@ -4619,22 +4619,22 @@ static int cas_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) | |||
4619 | XCVR_INTERNAL : XCVR_EXTERNAL; | 4619 | XCVR_INTERNAL : XCVR_EXTERNAL; |
4620 | cmd->phy_address = cp->phy_addr; | 4620 | cmd->phy_address = cp->phy_addr; |
4621 | cmd->advertising |= ADVERTISED_TP | ADVERTISED_MII | | 4621 | cmd->advertising |= ADVERTISED_TP | ADVERTISED_MII | |
4622 | ADVERTISED_10baseT_Half | | 4622 | ADVERTISED_10baseT_Half | |
4623 | ADVERTISED_10baseT_Full | | 4623 | ADVERTISED_10baseT_Full | |
4624 | ADVERTISED_100baseT_Half | | 4624 | ADVERTISED_100baseT_Half | |
4625 | ADVERTISED_100baseT_Full; | 4625 | ADVERTISED_100baseT_Full; |
4626 | 4626 | ||
4627 | cmd->supported |= | 4627 | cmd->supported |= |
4628 | (SUPPORTED_10baseT_Half | | 4628 | (SUPPORTED_10baseT_Half | |
4629 | SUPPORTED_10baseT_Full | | 4629 | SUPPORTED_10baseT_Full | |
4630 | SUPPORTED_100baseT_Half | | 4630 | SUPPORTED_100baseT_Half | |
4631 | SUPPORTED_100baseT_Full | | 4631 | SUPPORTED_100baseT_Full | |
4632 | SUPPORTED_TP | SUPPORTED_MII); | 4632 | SUPPORTED_TP | SUPPORTED_MII); |
4633 | 4633 | ||
4634 | if (cp->hw_running) { | 4634 | if (cp->hw_running) { |
4635 | cas_mif_poll(cp, 0); | 4635 | cas_mif_poll(cp, 0); |
4636 | bmcr = cas_phy_read(cp, MII_BMCR); | 4636 | bmcr = cas_phy_read(cp, MII_BMCR); |
4637 | cas_read_mii_link_mode(cp, &full_duplex, | 4637 | cas_read_mii_link_mode(cp, &full_duplex, |
4638 | &speed, &pause); | 4638 | &speed, &pause); |
4639 | cas_mif_poll(cp, 1); | 4639 | cas_mif_poll(cp, 1); |
4640 | } | 4640 | } |
@@ -4647,9 +4647,9 @@ static int cas_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) | |||
4647 | cmd->advertising |= ADVERTISED_FIBRE; | 4647 | cmd->advertising |= ADVERTISED_FIBRE; |
4648 | 4648 | ||
4649 | if (cp->hw_running) { | 4649 | if (cp->hw_running) { |
4650 | /* pcs uses the same bits as mii */ | 4650 | /* pcs uses the same bits as mii */ |
4651 | bmcr = readl(cp->regs + REG_PCS_MII_CTRL); | 4651 | bmcr = readl(cp->regs + REG_PCS_MII_CTRL); |
4652 | cas_read_pcs_link_mode(cp, &full_duplex, | 4652 | cas_read_pcs_link_mode(cp, &full_duplex, |
4653 | &speed, &pause); | 4653 | &speed, &pause); |
4654 | } | 4654 | } |
4655 | } | 4655 | } |
@@ -4667,8 +4667,8 @@ static int cas_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) | |||
4667 | cmd->autoneg = AUTONEG_DISABLE; | 4667 | cmd->autoneg = AUTONEG_DISABLE; |
4668 | cmd->speed = | 4668 | cmd->speed = |
4669 | (bmcr & CAS_BMCR_SPEED1000) ? | 4669 | (bmcr & CAS_BMCR_SPEED1000) ? |
4670 | SPEED_1000 : | 4670 | SPEED_1000 : |
4671 | ((bmcr & BMCR_SPEED100) ? SPEED_100: | 4671 | ((bmcr & BMCR_SPEED100) ? SPEED_100: |
4672 | SPEED_10); | 4672 | SPEED_10); |
4673 | cmd->duplex = | 4673 | cmd->duplex = |
4674 | (bmcr & BMCR_FULLDPLX) ? | 4674 | (bmcr & BMCR_FULLDPLX) ? |
@@ -4676,7 +4676,7 @@ static int cas_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) | |||
4676 | } | 4676 | } |
4677 | if (linkstate != link_up) { | 4677 | if (linkstate != link_up) { |
4678 | /* Force these to "unknown" if the link is not up and | 4678 | /* Force these to "unknown" if the link is not up and |
4679 | * autonogotiation in enabled. We can set the link | 4679 | * autonogotiation in enabled. We can set the link |
4680 | * speed to 0, but not cmd->duplex, | 4680 | * speed to 0, but not cmd->duplex, |
4681 | * because its legal values are 0 and 1. Ethtool will | 4681 | * because its legal values are 0 and 1. Ethtool will |
4682 | * print the value reported in parentheses after the | 4682 | * print the value reported in parentheses after the |
@@ -4783,7 +4783,7 @@ static int cas_get_stats_count(struct net_device *dev) | |||
4783 | 4783 | ||
4784 | static void cas_get_strings(struct net_device *dev, u32 stringset, u8 *data) | 4784 | static void cas_get_strings(struct net_device *dev, u32 stringset, u8 *data) |
4785 | { | 4785 | { |
4786 | memcpy(data, ðtool_cassini_statnames, | 4786 | memcpy(data, ðtool_cassini_statnames, |
4787 | CAS_NUM_STAT_KEYS * ETH_GSTRING_LEN); | 4787 | CAS_NUM_STAT_KEYS * ETH_GSTRING_LEN); |
4788 | } | 4788 | } |
4789 | 4789 | ||
@@ -4833,7 +4833,7 @@ static int cas_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) | |||
4833 | struct mii_ioctl_data *data = if_mii(ifr); | 4833 | struct mii_ioctl_data *data = if_mii(ifr); |
4834 | unsigned long flags; | 4834 | unsigned long flags; |
4835 | int rc = -EOPNOTSUPP; | 4835 | int rc = -EOPNOTSUPP; |
4836 | 4836 | ||
4837 | /* Hold the PM mutex while doing ioctl's or we may collide | 4837 | /* Hold the PM mutex while doing ioctl's or we may collide |
4838 | * with open/close and power management and oops. | 4838 | * with open/close and power management and oops. |
4839 | */ | 4839 | */ |
@@ -4933,11 +4933,11 @@ static int __devinit cas_init_one(struct pci_dev *pdev, | |||
4933 | pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, | 4933 | pci_read_config_byte(pdev, PCI_CACHE_LINE_SIZE, |
4934 | &orig_cacheline_size); | 4934 | &orig_cacheline_size); |
4935 | if (orig_cacheline_size < CAS_PREF_CACHELINE_SIZE) { | 4935 | if (orig_cacheline_size < CAS_PREF_CACHELINE_SIZE) { |
4936 | cas_cacheline_size = | 4936 | cas_cacheline_size = |
4937 | (CAS_PREF_CACHELINE_SIZE < SMP_CACHE_BYTES) ? | 4937 | (CAS_PREF_CACHELINE_SIZE < SMP_CACHE_BYTES) ? |
4938 | CAS_PREF_CACHELINE_SIZE : SMP_CACHE_BYTES; | 4938 | CAS_PREF_CACHELINE_SIZE : SMP_CACHE_BYTES; |
4939 | if (pci_write_config_byte(pdev, | 4939 | if (pci_write_config_byte(pdev, |
4940 | PCI_CACHE_LINE_SIZE, | 4940 | PCI_CACHE_LINE_SIZE, |
4941 | cas_cacheline_size)) { | 4941 | cas_cacheline_size)) { |
4942 | dev_err(&pdev->dev, "Could not set PCI cache " | 4942 | dev_err(&pdev->dev, "Could not set PCI cache " |
4943 | "line size\n"); | 4943 | "line size\n"); |
@@ -4977,7 +4977,7 @@ static int __devinit cas_init_one(struct pci_dev *pdev, | |||
4977 | cp->orig_cacheline_size = cas_cacheline_size ? orig_cacheline_size: 0; | 4977 | cp->orig_cacheline_size = cas_cacheline_size ? orig_cacheline_size: 0; |
4978 | #endif | 4978 | #endif |
4979 | cp->dev = dev; | 4979 | cp->dev = dev; |
4980 | cp->msg_enable = (cassini_debug < 0) ? CAS_DEF_MSG_ENABLE : | 4980 | cp->msg_enable = (cassini_debug < 0) ? CAS_DEF_MSG_ENABLE : |
4981 | cassini_debug; | 4981 | cassini_debug; |
4982 | 4982 | ||
4983 | cp->link_transition = LINK_TRANSITION_UNKNOWN; | 4983 | cp->link_transition = LINK_TRANSITION_UNKNOWN; |
@@ -5041,13 +5041,13 @@ static int __devinit cas_init_one(struct pci_dev *pdev, | |||
5041 | goto err_out_iounmap; | 5041 | goto err_out_iounmap; |
5042 | } | 5042 | } |
5043 | 5043 | ||
5044 | for (i = 0; i < N_TX_RINGS; i++) | 5044 | for (i = 0; i < N_TX_RINGS; i++) |
5045 | cp->init_txds[i] = cp->init_block->txds[i]; | 5045 | cp->init_txds[i] = cp->init_block->txds[i]; |
5046 | 5046 | ||
5047 | for (i = 0; i < N_RX_DESC_RINGS; i++) | 5047 | for (i = 0; i < N_RX_DESC_RINGS; i++) |
5048 | cp->init_rxds[i] = cp->init_block->rxds[i]; | 5048 | cp->init_rxds[i] = cp->init_block->rxds[i]; |
5049 | 5049 | ||
5050 | for (i = 0; i < N_RX_COMP_RINGS; i++) | 5050 | for (i = 0; i < N_RX_COMP_RINGS; i++) |
5051 | cp->init_rxcs[i] = cp->init_block->rxcs[i]; | 5051 | cp->init_rxcs[i] = cp->init_block->rxcs[i]; |
5052 | 5052 | ||
5053 | for (i = 0; i < N_RX_FLOWS; i++) | 5053 | for (i = 0; i < N_RX_FLOWS; i++) |
@@ -5087,11 +5087,11 @@ static int __devinit cas_init_one(struct pci_dev *pdev, | |||
5087 | 5087 | ||
5088 | i = readl(cp->regs + REG_BIM_CFG); | 5088 | i = readl(cp->regs + REG_BIM_CFG); |
5089 | printk(KERN_INFO "%s: Sun Cassini%s (%sbit/%sMHz PCI/%s) " | 5089 | printk(KERN_INFO "%s: Sun Cassini%s (%sbit/%sMHz PCI/%s) " |
5090 | "Ethernet[%d] ", dev->name, | 5090 | "Ethernet[%d] ", dev->name, |
5091 | (cp->cas_flags & CAS_FLAG_REG_PLUS) ? "+" : "", | 5091 | (cp->cas_flags & CAS_FLAG_REG_PLUS) ? "+" : "", |
5092 | (i & BIM_CFG_32BIT) ? "32" : "64", | 5092 | (i & BIM_CFG_32BIT) ? "32" : "64", |
5093 | (i & BIM_CFG_66MHZ) ? "66" : "33", | 5093 | (i & BIM_CFG_66MHZ) ? "66" : "33", |
5094 | (cp->phy_type == CAS_PHY_SERDES) ? "Fi" : "Cu", pdev->irq); | 5094 | (cp->phy_type == CAS_PHY_SERDES) ? "Fi" : "Cu", pdev->irq); |
5095 | 5095 | ||
5096 | for (i = 0; i < 6; i++) | 5096 | for (i = 0; i < 6; i++) |
5097 | printk("%2.2x%c", dev->dev_addr[i], | 5097 | printk("%2.2x%c", dev->dev_addr[i], |
@@ -5123,7 +5123,7 @@ err_out_free_res: | |||
5123 | 5123 | ||
5124 | err_write_cacheline: | 5124 | err_write_cacheline: |
5125 | /* Try to restore it in case the error occured after we | 5125 | /* Try to restore it in case the error occured after we |
5126 | * set it. | 5126 | * set it. |
5127 | */ | 5127 | */ |
5128 | pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, orig_cacheline_size); | 5128 | pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, orig_cacheline_size); |
5129 | 5129 | ||
@@ -5157,7 +5157,7 @@ static void __devexit cas_remove_one(struct pci_dev *pdev) | |||
5157 | /* Restore the cache line size if we had modified | 5157 | /* Restore the cache line size if we had modified |
5158 | * it. | 5158 | * it. |
5159 | */ | 5159 | */ |
5160 | pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, | 5160 | pci_write_config_byte(pdev, PCI_CACHE_LINE_SIZE, |
5161 | cp->orig_cacheline_size); | 5161 | cp->orig_cacheline_size); |
5162 | } | 5162 | } |
5163 | #endif | 5163 | #endif |
@@ -5178,7 +5178,7 @@ static int cas_suspend(struct pci_dev *pdev, pm_message_t state) | |||
5178 | unsigned long flags; | 5178 | unsigned long flags; |
5179 | 5179 | ||
5180 | mutex_lock(&cp->pm_mutex); | 5180 | mutex_lock(&cp->pm_mutex); |
5181 | 5181 | ||
5182 | /* If the driver is opened, we stop the DMA */ | 5182 | /* If the driver is opened, we stop the DMA */ |
5183 | if (cp->opened) { | 5183 | if (cp->opened) { |
5184 | netif_device_detach(dev); | 5184 | netif_device_detach(dev); |