aboutsummaryrefslogtreecommitdiffstats
path: root/drivers
diff options
context:
space:
mode:
Diffstat (limited to 'drivers')
-rw-r--r--drivers/net/pcnet32.c4155
1 files changed, 2156 insertions, 1999 deletions
diff --git a/drivers/net/pcnet32.c b/drivers/net/pcnet32.c
index 1bc3f5bffb95..861361018640 100644
--- a/drivers/net/pcnet32.c
+++ b/drivers/net/pcnet32.c
@@ -26,8 +26,8 @@
26#define DRV_RELDATE "18.Mar.2006" 26#define DRV_RELDATE "18.Mar.2006"
27#define PFX DRV_NAME ": " 27#define PFX DRV_NAME ": "
28 28
29static const char * const version = 29static const char *const version =
30DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " tsbogend@alpha.franken.de\n"; 30 DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " tsbogend@alpha.franken.de\n";
31 31
32#include <linux/module.h> 32#include <linux/module.h>
33#include <linux/kernel.h> 33#include <linux/kernel.h>
@@ -58,18 +58,23 @@ DRV_NAME ".c:v" DRV_VERSION " " DRV_RELDATE " tsbogend@alpha.franken.de\n";
58 * PCI device identifiers for "new style" Linux PCI Device Drivers 58 * PCI device identifiers for "new style" Linux PCI Device Drivers
59 */ 59 */
60static struct pci_device_id pcnet32_pci_tbl[] = { 60static struct pci_device_id pcnet32_pci_tbl[] = {
61 { PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE_HOME, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, 61 { PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE_HOME,
62 { PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE, PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, 62 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0},
63 /* 63 { PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_LANCE,
64 * Adapters that were sold with IBM's RS/6000 or pSeries hardware have 64 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0},
65 * the incorrect vendor id. 65
66 */ 66 /*
67 { PCI_VENDOR_ID_TRIDENT, PCI_DEVICE_ID_AMD_LANCE, PCI_ANY_ID, PCI_ANY_ID, 67 * Adapters that were sold with IBM's RS/6000 or pSeries hardware have
68 PCI_CLASS_NETWORK_ETHERNET << 8, 0xffff00, 0 }, 68 * the incorrect vendor id.
69 { 0, } 69 */
70 { PCI_VENDOR_ID_TRIDENT, PCI_DEVICE_ID_AMD_LANCE,
71 PCI_ANY_ID, PCI_ANY_ID,
72 PCI_CLASS_NETWORK_ETHERNET << 8, 0xffff00, 0},
73
74 { } /* terminate list */
70}; 75};
71 76
72MODULE_DEVICE_TABLE (pci, pcnet32_pci_tbl); 77MODULE_DEVICE_TABLE(pci, pcnet32_pci_tbl);
73 78
74static int cards_found; 79static int cards_found;
75 80
@@ -77,13 +82,11 @@ static int cards_found;
77 * VLB I/O addresses 82 * VLB I/O addresses
78 */ 83 */
79static unsigned int pcnet32_portlist[] __initdata = 84static unsigned int pcnet32_portlist[] __initdata =
80 { 0x300, 0x320, 0x340, 0x360, 0 }; 85 { 0x300, 0x320, 0x340, 0x360, 0 };
81
82
83 86
84static int pcnet32_debug = 0; 87static int pcnet32_debug = 0;
85static int tx_start = 1; /* Mapping -- 0:20, 1:64, 2:128, 3:~220 (depends on chip vers) */ 88static int tx_start = 1; /* Mapping -- 0:20, 1:64, 2:128, 3:~220 (depends on chip vers) */
86static int pcnet32vlb; /* check for VLB cards ? */ 89static int pcnet32vlb; /* check for VLB cards ? */
87 90
88static struct net_device *pcnet32_dev; 91static struct net_device *pcnet32_dev;
89 92
@@ -110,32 +113,34 @@ static int rx_copybreak = 200;
110 * to internal options 113 * to internal options
111 */ 114 */
112static const unsigned char options_mapping[] = { 115static const unsigned char options_mapping[] = {
113 PCNET32_PORT_ASEL, /* 0 Auto-select */ 116 PCNET32_PORT_ASEL, /* 0 Auto-select */
114 PCNET32_PORT_AUI, /* 1 BNC/AUI */ 117 PCNET32_PORT_AUI, /* 1 BNC/AUI */
115 PCNET32_PORT_AUI, /* 2 AUI/BNC */ 118 PCNET32_PORT_AUI, /* 2 AUI/BNC */
116 PCNET32_PORT_ASEL, /* 3 not supported */ 119 PCNET32_PORT_ASEL, /* 3 not supported */
117 PCNET32_PORT_10BT | PCNET32_PORT_FD, /* 4 10baseT-FD */ 120 PCNET32_PORT_10BT | PCNET32_PORT_FD, /* 4 10baseT-FD */
118 PCNET32_PORT_ASEL, /* 5 not supported */ 121 PCNET32_PORT_ASEL, /* 5 not supported */
119 PCNET32_PORT_ASEL, /* 6 not supported */ 122 PCNET32_PORT_ASEL, /* 6 not supported */
120 PCNET32_PORT_ASEL, /* 7 not supported */ 123 PCNET32_PORT_ASEL, /* 7 not supported */
121 PCNET32_PORT_ASEL, /* 8 not supported */ 124 PCNET32_PORT_ASEL, /* 8 not supported */
122 PCNET32_PORT_MII, /* 9 MII 10baseT */ 125 PCNET32_PORT_MII, /* 9 MII 10baseT */
123 PCNET32_PORT_MII | PCNET32_PORT_FD, /* 10 MII 10baseT-FD */ 126 PCNET32_PORT_MII | PCNET32_PORT_FD, /* 10 MII 10baseT-FD */
124 PCNET32_PORT_MII, /* 11 MII (autosel) */ 127 PCNET32_PORT_MII, /* 11 MII (autosel) */
125 PCNET32_PORT_10BT, /* 12 10BaseT */ 128 PCNET32_PORT_10BT, /* 12 10BaseT */
126 PCNET32_PORT_MII | PCNET32_PORT_100, /* 13 MII 100BaseTx */ 129 PCNET32_PORT_MII | PCNET32_PORT_100, /* 13 MII 100BaseTx */
127 PCNET32_PORT_MII | PCNET32_PORT_100 | PCNET32_PORT_FD, /* 14 MII 100BaseTx-FD */ 130 /* 14 MII 100BaseTx-FD */
128 PCNET32_PORT_ASEL /* 15 not supported */ 131 PCNET32_PORT_MII | PCNET32_PORT_100 | PCNET32_PORT_FD,
132 PCNET32_PORT_ASEL /* 15 not supported */
129}; 133};
130 134
131static const char pcnet32_gstrings_test[][ETH_GSTRING_LEN] = { 135static const char pcnet32_gstrings_test[][ETH_GSTRING_LEN] = {
132 "Loopback test (offline)" 136 "Loopback test (offline)"
133}; 137};
138
134#define PCNET32_TEST_LEN (sizeof(pcnet32_gstrings_test) / ETH_GSTRING_LEN) 139#define PCNET32_TEST_LEN (sizeof(pcnet32_gstrings_test) / ETH_GSTRING_LEN)
135 140
136#define PCNET32_NUM_REGS 136 141#define PCNET32_NUM_REGS 136
137 142
138#define MAX_UNITS 8 /* More are supported, limit only on options */ 143#define MAX_UNITS 8 /* More are supported, limit only on options */
139static int options[MAX_UNITS]; 144static int options[MAX_UNITS];
140static int full_duplex[MAX_UNITS]; 145static int full_duplex[MAX_UNITS];
141static int homepna[MAX_UNITS]; 146static int homepna[MAX_UNITS];
@@ -270,7 +275,6 @@ static int homepna[MAX_UNITS];
270 * Philippe Seewer assisted with auto negotiation and testing. 275 * Philippe Seewer assisted with auto negotiation and testing.
271 */ 276 */
272 277
273
274/* 278/*
275 * Set the number of Tx and Rx buffers, using Log_2(# buffers). 279 * Set the number of Tx and Rx buffers, using Log_2(# buffers).
276 * Reasonable default values are 4 Tx buffers, and 16 Rx buffers. 280 * Reasonable default values are 4 Tx buffers, and 16 Rx buffers.
@@ -306,42 +310,42 @@ static int homepna[MAX_UNITS];
306 310
307/* The PCNET32 Rx and Tx ring descriptors. */ 311/* The PCNET32 Rx and Tx ring descriptors. */
308struct pcnet32_rx_head { 312struct pcnet32_rx_head {
309 u32 base; 313 u32 base;
310 s16 buf_length; 314 s16 buf_length;
311 s16 status; 315 s16 status;
312 u32 msg_length; 316 u32 msg_length;
313 u32 reserved; 317 u32 reserved;
314}; 318};
315 319
316struct pcnet32_tx_head { 320struct pcnet32_tx_head {
317 u32 base; 321 u32 base;
318 s16 length; 322 s16 length;
319 s16 status; 323 s16 status;
320 u32 misc; 324 u32 misc;
321 u32 reserved; 325 u32 reserved;
322}; 326};
323 327
324/* The PCNET32 32-Bit initialization block, described in databook. */ 328/* The PCNET32 32-Bit initialization block, described in databook. */
325struct pcnet32_init_block { 329struct pcnet32_init_block {
326 u16 mode; 330 u16 mode;
327 u16 tlen_rlen; 331 u16 tlen_rlen;
328 u8 phys_addr[6]; 332 u8 phys_addr[6];
329 u16 reserved; 333 u16 reserved;
330 u32 filter[2]; 334 u32 filter[2];
331 /* Receive and transmit ring base, along with extra bits. */ 335 /* Receive and transmit ring base, along with extra bits. */
332 u32 rx_ring; 336 u32 rx_ring;
333 u32 tx_ring; 337 u32 tx_ring;
334}; 338};
335 339
336/* PCnet32 access functions */ 340/* PCnet32 access functions */
337struct pcnet32_access { 341struct pcnet32_access {
338 u16 (*read_csr)(unsigned long, int); 342 u16 (*read_csr) (unsigned long, int);
339 void (*write_csr)(unsigned long, int, u16); 343 void (*write_csr) (unsigned long, int, u16);
340 u16 (*read_bcr)(unsigned long, int); 344 u16 (*read_bcr) (unsigned long, int);
341 void (*write_bcr)(unsigned long, int, u16); 345 void (*write_bcr) (unsigned long, int, u16);
342 u16 (*read_rap)(unsigned long); 346 u16 (*read_rap) (unsigned long);
343 void (*write_rap)(unsigned long, u16); 347 void (*write_rap) (unsigned long, u16);
344 void (*reset)(unsigned long); 348 void (*reset) (unsigned long);
345}; 349};
346 350
347/* 351/*
@@ -349,771 +353,794 @@ struct pcnet32_access {
349 * so the structure should be allocated using pci_alloc_consistent(). 353 * so the structure should be allocated using pci_alloc_consistent().
350 */ 354 */
351struct pcnet32_private { 355struct pcnet32_private {
352 struct pcnet32_init_block init_block; 356 struct pcnet32_init_block init_block;
353 /* The Tx and Rx ring entries must be aligned on 16-byte boundaries in 32bit mode. */ 357 /* The Tx and Rx ring entries must be aligned on 16-byte boundaries in 32bit mode. */
354 struct pcnet32_rx_head *rx_ring; 358 struct pcnet32_rx_head *rx_ring;
355 struct pcnet32_tx_head *tx_ring; 359 struct pcnet32_tx_head *tx_ring;
356 dma_addr_t dma_addr; /* DMA address of beginning of this 360 dma_addr_t dma_addr; /* DMA address of beginning of this
357 object, returned by 361 object, returned by
358 pci_alloc_consistent */ 362 pci_alloc_consistent */
359 struct pci_dev *pci_dev; /* Pointer to the associated pci device 363 struct pci_dev *pci_dev; /* Pointer to the associated pci device
360 structure */ 364 structure */
361 const char *name; 365 const char *name;
362 /* The saved address of a sent-in-place packet/buffer, for skfree(). */ 366 /* The saved address of a sent-in-place packet/buffer, for skfree(). */
363 struct sk_buff **tx_skbuff; 367 struct sk_buff **tx_skbuff;
364 struct sk_buff **rx_skbuff; 368 struct sk_buff **rx_skbuff;
365 dma_addr_t *tx_dma_addr; 369 dma_addr_t *tx_dma_addr;
366 dma_addr_t *rx_dma_addr; 370 dma_addr_t *rx_dma_addr;
367 struct pcnet32_access a; 371 struct pcnet32_access a;
368 spinlock_t lock; /* Guard lock */ 372 spinlock_t lock; /* Guard lock */
369 unsigned int cur_rx, cur_tx; /* The next free ring entry */ 373 unsigned int cur_rx, cur_tx; /* The next free ring entry */
370 unsigned int rx_ring_size; /* current rx ring size */ 374 unsigned int rx_ring_size; /* current rx ring size */
371 unsigned int tx_ring_size; /* current tx ring size */ 375 unsigned int tx_ring_size; /* current tx ring size */
372 unsigned int rx_mod_mask; /* rx ring modular mask */ 376 unsigned int rx_mod_mask; /* rx ring modular mask */
373 unsigned int tx_mod_mask; /* tx ring modular mask */ 377 unsigned int tx_mod_mask; /* tx ring modular mask */
374 unsigned short rx_len_bits; 378 unsigned short rx_len_bits;
375 unsigned short tx_len_bits; 379 unsigned short tx_len_bits;
376 dma_addr_t rx_ring_dma_addr; 380 dma_addr_t rx_ring_dma_addr;
377 dma_addr_t tx_ring_dma_addr; 381 dma_addr_t tx_ring_dma_addr;
378 unsigned int dirty_rx, dirty_tx; /* The ring entries to be free()ed. */ 382 unsigned int dirty_rx, dirty_tx; /* The ring entries to be free()ed. */
379 struct net_device_stats stats; 383 struct net_device_stats stats;
380 char tx_full; 384 char tx_full;
381 char phycount; /* number of phys found */ 385 char phycount; /* number of phys found */
382 int options; 386 int options;
383 unsigned int shared_irq:1, /* shared irq possible */ 387 unsigned int shared_irq:1, /* shared irq possible */
384 dxsuflo:1, /* disable transmit stop on uflo */ 388 dxsuflo:1, /* disable transmit stop on uflo */
385 mii:1; /* mii port available */ 389 mii:1; /* mii port available */
386 struct net_device *next; 390 struct net_device *next;
387 struct mii_if_info mii_if; 391 struct mii_if_info mii_if;
388 struct timer_list watchdog_timer; 392 struct timer_list watchdog_timer;
389 struct timer_list blink_timer; 393 struct timer_list blink_timer;
390 u32 msg_enable; /* debug message level */ 394 u32 msg_enable; /* debug message level */
391 395
392 /* each bit indicates an available PHY */ 396 /* each bit indicates an available PHY */
393 u32 phymask; 397 u32 phymask;
394}; 398};
395 399
396static void pcnet32_probe_vlbus(void); 400static void pcnet32_probe_vlbus(void);
397static int pcnet32_probe_pci(struct pci_dev *, const struct pci_device_id *); 401static int pcnet32_probe_pci(struct pci_dev *, const struct pci_device_id *);
398static int pcnet32_probe1(unsigned long, int, struct pci_dev *); 402static int pcnet32_probe1(unsigned long, int, struct pci_dev *);
399static int pcnet32_open(struct net_device *); 403static int pcnet32_open(struct net_device *);
400static int pcnet32_init_ring(struct net_device *); 404static int pcnet32_init_ring(struct net_device *);
401static int pcnet32_start_xmit(struct sk_buff *, struct net_device *); 405static int pcnet32_start_xmit(struct sk_buff *, struct net_device *);
402static int pcnet32_rx(struct net_device *); 406static int pcnet32_rx(struct net_device *);
403static void pcnet32_tx_timeout (struct net_device *dev); 407static void pcnet32_tx_timeout(struct net_device *dev);
404static irqreturn_t pcnet32_interrupt(int, void *, struct pt_regs *); 408static irqreturn_t pcnet32_interrupt(int, void *, struct pt_regs *);
405static int pcnet32_close(struct net_device *); 409static int pcnet32_close(struct net_device *);
406static struct net_device_stats *pcnet32_get_stats(struct net_device *); 410static struct net_device_stats *pcnet32_get_stats(struct net_device *);
407static void pcnet32_load_multicast(struct net_device *dev); 411static void pcnet32_load_multicast(struct net_device *dev);
408static void pcnet32_set_multicast_list(struct net_device *); 412static void pcnet32_set_multicast_list(struct net_device *);
409static int pcnet32_ioctl(struct net_device *, struct ifreq *, int); 413static int pcnet32_ioctl(struct net_device *, struct ifreq *, int);
410static void pcnet32_watchdog(struct net_device *); 414static void pcnet32_watchdog(struct net_device *);
411static int mdio_read(struct net_device *dev, int phy_id, int reg_num); 415static int mdio_read(struct net_device *dev, int phy_id, int reg_num);
412static void mdio_write(struct net_device *dev, int phy_id, int reg_num, int val); 416static void mdio_write(struct net_device *dev, int phy_id, int reg_num,
417 int val);
413static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits); 418static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits);
414static void pcnet32_ethtool_test(struct net_device *dev, 419static void pcnet32_ethtool_test(struct net_device *dev,
415 struct ethtool_test *eth_test, u64 *data); 420 struct ethtool_test *eth_test, u64 * data);
416static int pcnet32_loopback_test(struct net_device *dev, uint64_t *data1); 421static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1);
417static int pcnet32_phys_id(struct net_device *dev, u32 data); 422static int pcnet32_phys_id(struct net_device *dev, u32 data);
418static void pcnet32_led_blink_callback(struct net_device *dev); 423static void pcnet32_led_blink_callback(struct net_device *dev);
419static int pcnet32_get_regs_len(struct net_device *dev); 424static int pcnet32_get_regs_len(struct net_device *dev);
420static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs, 425static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs,
421 void *ptr); 426 void *ptr);
422static void pcnet32_purge_tx_ring(struct net_device *dev); 427static void pcnet32_purge_tx_ring(struct net_device *dev);
423static int pcnet32_alloc_ring(struct net_device *dev, char *name); 428static int pcnet32_alloc_ring(struct net_device *dev, char *name);
424static void pcnet32_free_ring(struct net_device *dev); 429static void pcnet32_free_ring(struct net_device *dev);
425static void pcnet32_check_media(struct net_device *dev, int verbose); 430static void pcnet32_check_media(struct net_device *dev, int verbose);
426 431
427
428enum pci_flags_bit { 432enum pci_flags_bit {
429 PCI_USES_IO=1, PCI_USES_MEM=2, PCI_USES_MASTER=4, 433 PCI_USES_IO = 1, PCI_USES_MEM = 2, PCI_USES_MASTER = 4,
430 PCI_ADDR0=0x10<<0, PCI_ADDR1=0x10<<1, PCI_ADDR2=0x10<<2, PCI_ADDR3=0x10<<3, 434 PCI_ADDR0 = 0x10 << 0, PCI_ADDR1 = 0x10 << 1, PCI_ADDR2 =
435 0x10 << 2, PCI_ADDR3 = 0x10 << 3,
431}; 436};
432 437
433 438static u16 pcnet32_wio_read_csr(unsigned long addr, int index)
434static u16 pcnet32_wio_read_csr (unsigned long addr, int index)
435{ 439{
436 outw (index, addr+PCNET32_WIO_RAP); 440 outw(index, addr + PCNET32_WIO_RAP);
437 return inw (addr+PCNET32_WIO_RDP); 441 return inw(addr + PCNET32_WIO_RDP);
438} 442}
439 443
440static void pcnet32_wio_write_csr (unsigned long addr, int index, u16 val) 444static void pcnet32_wio_write_csr(unsigned long addr, int index, u16 val)
441{ 445{
442 outw (index, addr+PCNET32_WIO_RAP); 446 outw(index, addr + PCNET32_WIO_RAP);
443 outw (val, addr+PCNET32_WIO_RDP); 447 outw(val, addr + PCNET32_WIO_RDP);
444} 448}
445 449
446static u16 pcnet32_wio_read_bcr (unsigned long addr, int index) 450static u16 pcnet32_wio_read_bcr(unsigned long addr, int index)
447{ 451{
448 outw (index, addr+PCNET32_WIO_RAP); 452 outw(index, addr + PCNET32_WIO_RAP);
449 return inw (addr+PCNET32_WIO_BDP); 453 return inw(addr + PCNET32_WIO_BDP);
450} 454}
451 455
452static void pcnet32_wio_write_bcr (unsigned long addr, int index, u16 val) 456static void pcnet32_wio_write_bcr(unsigned long addr, int index, u16 val)
453{ 457{
454 outw (index, addr+PCNET32_WIO_RAP); 458 outw(index, addr + PCNET32_WIO_RAP);
455 outw (val, addr+PCNET32_WIO_BDP); 459 outw(val, addr + PCNET32_WIO_BDP);
456} 460}
457 461
458static u16 pcnet32_wio_read_rap (unsigned long addr) 462static u16 pcnet32_wio_read_rap(unsigned long addr)
459{ 463{
460 return inw (addr+PCNET32_WIO_RAP); 464 return inw(addr + PCNET32_WIO_RAP);
461} 465}
462 466
463static void pcnet32_wio_write_rap (unsigned long addr, u16 val) 467static void pcnet32_wio_write_rap(unsigned long addr, u16 val)
464{ 468{
465 outw (val, addr+PCNET32_WIO_RAP); 469 outw(val, addr + PCNET32_WIO_RAP);
466} 470}
467 471
468static void pcnet32_wio_reset (unsigned long addr) 472static void pcnet32_wio_reset(unsigned long addr)
469{ 473{
470 inw (addr+PCNET32_WIO_RESET); 474 inw(addr + PCNET32_WIO_RESET);
471} 475}
472 476
473static int pcnet32_wio_check (unsigned long addr) 477static int pcnet32_wio_check(unsigned long addr)
474{ 478{
475 outw (88, addr+PCNET32_WIO_RAP); 479 outw(88, addr + PCNET32_WIO_RAP);
476 return (inw (addr+PCNET32_WIO_RAP) == 88); 480 return (inw(addr + PCNET32_WIO_RAP) == 88);
477} 481}
478 482
479static struct pcnet32_access pcnet32_wio = { 483static struct pcnet32_access pcnet32_wio = {
480 .read_csr = pcnet32_wio_read_csr, 484 .read_csr = pcnet32_wio_read_csr,
481 .write_csr = pcnet32_wio_write_csr, 485 .write_csr = pcnet32_wio_write_csr,
482 .read_bcr = pcnet32_wio_read_bcr, 486 .read_bcr = pcnet32_wio_read_bcr,
483 .write_bcr = pcnet32_wio_write_bcr, 487 .write_bcr = pcnet32_wio_write_bcr,
484 .read_rap = pcnet32_wio_read_rap, 488 .read_rap = pcnet32_wio_read_rap,
485 .write_rap = pcnet32_wio_write_rap, 489 .write_rap = pcnet32_wio_write_rap,
486 .reset = pcnet32_wio_reset 490 .reset = pcnet32_wio_reset
487}; 491};
488 492
489static u16 pcnet32_dwio_read_csr (unsigned long addr, int index) 493static u16 pcnet32_dwio_read_csr(unsigned long addr, int index)
490{ 494{
491 outl (index, addr+PCNET32_DWIO_RAP); 495 outl(index, addr + PCNET32_DWIO_RAP);
492 return (inl (addr+PCNET32_DWIO_RDP) & 0xffff); 496 return (inl(addr + PCNET32_DWIO_RDP) & 0xffff);
493} 497}
494 498
495static void pcnet32_dwio_write_csr (unsigned long addr, int index, u16 val) 499static void pcnet32_dwio_write_csr(unsigned long addr, int index, u16 val)
496{ 500{
497 outl (index, addr+PCNET32_DWIO_RAP); 501 outl(index, addr + PCNET32_DWIO_RAP);
498 outl (val, addr+PCNET32_DWIO_RDP); 502 outl(val, addr + PCNET32_DWIO_RDP);
499} 503}
500 504
501static u16 pcnet32_dwio_read_bcr (unsigned long addr, int index) 505static u16 pcnet32_dwio_read_bcr(unsigned long addr, int index)
502{ 506{
503 outl (index, addr+PCNET32_DWIO_RAP); 507 outl(index, addr + PCNET32_DWIO_RAP);
504 return (inl (addr+PCNET32_DWIO_BDP) & 0xffff); 508 return (inl(addr + PCNET32_DWIO_BDP) & 0xffff);
505} 509}
506 510
507static void pcnet32_dwio_write_bcr (unsigned long addr, int index, u16 val) 511static void pcnet32_dwio_write_bcr(unsigned long addr, int index, u16 val)
508{ 512{
509 outl (index, addr+PCNET32_DWIO_RAP); 513 outl(index, addr + PCNET32_DWIO_RAP);
510 outl (val, addr+PCNET32_DWIO_BDP); 514 outl(val, addr + PCNET32_DWIO_BDP);
511} 515}
512 516
513static u16 pcnet32_dwio_read_rap (unsigned long addr) 517static u16 pcnet32_dwio_read_rap(unsigned long addr)
514{ 518{
515 return (inl (addr+PCNET32_DWIO_RAP) & 0xffff); 519 return (inl(addr + PCNET32_DWIO_RAP) & 0xffff);
516} 520}
517 521
518static void pcnet32_dwio_write_rap (unsigned long addr, u16 val) 522static void pcnet32_dwio_write_rap(unsigned long addr, u16 val)
519{ 523{
520 outl (val, addr+PCNET32_DWIO_RAP); 524 outl(val, addr + PCNET32_DWIO_RAP);
521} 525}
522 526
523static void pcnet32_dwio_reset (unsigned long addr) 527static void pcnet32_dwio_reset(unsigned long addr)
524{ 528{
525 inl (addr+PCNET32_DWIO_RESET); 529 inl(addr + PCNET32_DWIO_RESET);
526} 530}
527 531
528static int pcnet32_dwio_check (unsigned long addr) 532static int pcnet32_dwio_check(unsigned long addr)
529{ 533{
530 outl (88, addr+PCNET32_DWIO_RAP); 534 outl(88, addr + PCNET32_DWIO_RAP);
531 return ((inl (addr+PCNET32_DWIO_RAP) & 0xffff) == 88); 535 return ((inl(addr + PCNET32_DWIO_RAP) & 0xffff) == 88);
532} 536}
533 537
534static struct pcnet32_access pcnet32_dwio = { 538static struct pcnet32_access pcnet32_dwio = {
535 .read_csr = pcnet32_dwio_read_csr, 539 .read_csr = pcnet32_dwio_read_csr,
536 .write_csr = pcnet32_dwio_write_csr, 540 .write_csr = pcnet32_dwio_write_csr,
537 .read_bcr = pcnet32_dwio_read_bcr, 541 .read_bcr = pcnet32_dwio_read_bcr,
538 .write_bcr = pcnet32_dwio_write_bcr, 542 .write_bcr = pcnet32_dwio_write_bcr,
539 .read_rap = pcnet32_dwio_read_rap, 543 .read_rap = pcnet32_dwio_read_rap,
540 .write_rap = pcnet32_dwio_write_rap, 544 .write_rap = pcnet32_dwio_write_rap,
541 .reset = pcnet32_dwio_reset 545 .reset = pcnet32_dwio_reset
542}; 546};
543 547
544#ifdef CONFIG_NET_POLL_CONTROLLER 548#ifdef CONFIG_NET_POLL_CONTROLLER
545static void pcnet32_poll_controller(struct net_device *dev) 549static void pcnet32_poll_controller(struct net_device *dev)
546{ 550{
547 disable_irq(dev->irq); 551 disable_irq(dev->irq);
548 pcnet32_interrupt(0, dev, NULL); 552 pcnet32_interrupt(0, dev, NULL);
549 enable_irq(dev->irq); 553 enable_irq(dev->irq);
550} 554}
551#endif 555#endif
552 556
553
554static int pcnet32_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) 557static int pcnet32_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
555{ 558{
556 struct pcnet32_private *lp = dev->priv; 559 struct pcnet32_private *lp = dev->priv;
557 unsigned long flags; 560 unsigned long flags;
558 int r = -EOPNOTSUPP; 561 int r = -EOPNOTSUPP;
559 562
560 if (lp->mii) { 563 if (lp->mii) {
561 spin_lock_irqsave(&lp->lock, flags); 564 spin_lock_irqsave(&lp->lock, flags);
562 mii_ethtool_gset(&lp->mii_if, cmd); 565 mii_ethtool_gset(&lp->mii_if, cmd);
563 spin_unlock_irqrestore(&lp->lock, flags); 566 spin_unlock_irqrestore(&lp->lock, flags);
564 r = 0; 567 r = 0;
565 } 568 }
566 return r; 569 return r;
567} 570}
568 571
569static int pcnet32_set_settings(struct net_device *dev, struct ethtool_cmd *cmd) 572static int pcnet32_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
570{ 573{
571 struct pcnet32_private *lp = dev->priv; 574 struct pcnet32_private *lp = dev->priv;
572 unsigned long flags; 575 unsigned long flags;
573 int r = -EOPNOTSUPP; 576 int r = -EOPNOTSUPP;
574 577
575 if (lp->mii) { 578 if (lp->mii) {
576 spin_lock_irqsave(&lp->lock, flags); 579 spin_lock_irqsave(&lp->lock, flags);
577 r = mii_ethtool_sset(&lp->mii_if, cmd); 580 r = mii_ethtool_sset(&lp->mii_if, cmd);
578 spin_unlock_irqrestore(&lp->lock, flags); 581 spin_unlock_irqrestore(&lp->lock, flags);
579 } 582 }
580 return r; 583 return r;
581} 584}
582 585
583static void pcnet32_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) 586static void pcnet32_get_drvinfo(struct net_device *dev,
587 struct ethtool_drvinfo *info)
584{ 588{
585 struct pcnet32_private *lp = dev->priv; 589 struct pcnet32_private *lp = dev->priv;
586 590
587 strcpy (info->driver, DRV_NAME); 591 strcpy(info->driver, DRV_NAME);
588 strcpy (info->version, DRV_VERSION); 592 strcpy(info->version, DRV_VERSION);
589 if (lp->pci_dev) 593 if (lp->pci_dev)
590 strcpy (info->bus_info, pci_name(lp->pci_dev)); 594 strcpy(info->bus_info, pci_name(lp->pci_dev));
591 else 595 else
592 sprintf(info->bus_info, "VLB 0x%lx", dev->base_addr); 596 sprintf(info->bus_info, "VLB 0x%lx", dev->base_addr);
593} 597}
594 598
595static u32 pcnet32_get_link(struct net_device *dev) 599static u32 pcnet32_get_link(struct net_device *dev)
596{ 600{
597 struct pcnet32_private *lp = dev->priv; 601 struct pcnet32_private *lp = dev->priv;
598 unsigned long flags; 602 unsigned long flags;
599 int r; 603 int r;
600
601 spin_lock_irqsave(&lp->lock, flags);
602 if (lp->mii) {
603 r = mii_link_ok(&lp->mii_if);
604 } else {
605 ulong ioaddr = dev->base_addr; /* card base I/O address */
606 r = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
607 }
608 spin_unlock_irqrestore(&lp->lock, flags);
609 604
610 return r; 605 spin_lock_irqsave(&lp->lock, flags);
606 if (lp->mii) {
607 r = mii_link_ok(&lp->mii_if);
608 } else {
609 ulong ioaddr = dev->base_addr; /* card base I/O address */
610 r = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
611 }
612 spin_unlock_irqrestore(&lp->lock, flags);
613
614 return r;
611} 615}
612 616
613static u32 pcnet32_get_msglevel(struct net_device *dev) 617static u32 pcnet32_get_msglevel(struct net_device *dev)
614{ 618{
615 struct pcnet32_private *lp = dev->priv; 619 struct pcnet32_private *lp = dev->priv;
616 return lp->msg_enable; 620 return lp->msg_enable;
617} 621}
618 622
619static void pcnet32_set_msglevel(struct net_device *dev, u32 value) 623static void pcnet32_set_msglevel(struct net_device *dev, u32 value)
620{ 624{
621 struct pcnet32_private *lp = dev->priv; 625 struct pcnet32_private *lp = dev->priv;
622 lp->msg_enable = value; 626 lp->msg_enable = value;
623} 627}
624 628
625static int pcnet32_nway_reset(struct net_device *dev) 629static int pcnet32_nway_reset(struct net_device *dev)
626{ 630{
627 struct pcnet32_private *lp = dev->priv; 631 struct pcnet32_private *lp = dev->priv;
628 unsigned long flags; 632 unsigned long flags;
629 int r = -EOPNOTSUPP; 633 int r = -EOPNOTSUPP;
630 634
631 if (lp->mii) { 635 if (lp->mii) {
632 spin_lock_irqsave(&lp->lock, flags); 636 spin_lock_irqsave(&lp->lock, flags);
633 r = mii_nway_restart(&lp->mii_if); 637 r = mii_nway_restart(&lp->mii_if);
634 spin_unlock_irqrestore(&lp->lock, flags); 638 spin_unlock_irqrestore(&lp->lock, flags);
635 } 639 }
636 return r; 640 return r;
637} 641}
638 642
639static void pcnet32_get_ringparam(struct net_device *dev, struct ethtool_ringparam *ering) 643static void pcnet32_get_ringparam(struct net_device *dev,
644 struct ethtool_ringparam *ering)
640{ 645{
641 struct pcnet32_private *lp = dev->priv; 646 struct pcnet32_private *lp = dev->priv;
642 647
643 ering->tx_max_pending = TX_MAX_RING_SIZE - 1; 648 ering->tx_max_pending = TX_MAX_RING_SIZE - 1;
644 ering->tx_pending = lp->tx_ring_size - 1; 649 ering->tx_pending = lp->tx_ring_size - 1;
645 ering->rx_max_pending = RX_MAX_RING_SIZE - 1; 650 ering->rx_max_pending = RX_MAX_RING_SIZE - 1;
646 ering->rx_pending = lp->rx_ring_size - 1; 651 ering->rx_pending = lp->rx_ring_size - 1;
647} 652}
648 653
649static int pcnet32_set_ringparam(struct net_device *dev, struct ethtool_ringparam *ering) 654static int pcnet32_set_ringparam(struct net_device *dev,
655 struct ethtool_ringparam *ering)
650{ 656{
651 struct pcnet32_private *lp = dev->priv; 657 struct pcnet32_private *lp = dev->priv;
652 unsigned long flags; 658 unsigned long flags;
653 int i; 659 int i;
654 660
655 if (ering->rx_mini_pending || ering->rx_jumbo_pending) 661 if (ering->rx_mini_pending || ering->rx_jumbo_pending)
656 return -EINVAL; 662 return -EINVAL;
657 663
658 if (netif_running(dev)) 664 if (netif_running(dev))
659 pcnet32_close(dev); 665 pcnet32_close(dev);
660 666
661 spin_lock_irqsave(&lp->lock, flags); 667 spin_lock_irqsave(&lp->lock, flags);
662 pcnet32_free_ring(dev);
663 lp->tx_ring_size = min(ering->tx_pending, (unsigned int) TX_MAX_RING_SIZE);
664 lp->rx_ring_size = min(ering->rx_pending, (unsigned int) RX_MAX_RING_SIZE);
665
666 /* set the minimum ring size to 4, to allow the loopback test to work
667 * unchanged.
668 */
669 for (i = 2; i <= PCNET32_LOG_MAX_TX_BUFFERS; i++) {
670 if (lp->tx_ring_size <= (1 << i))
671 break;
672 }
673 lp->tx_ring_size = (1 << i);
674 lp->tx_mod_mask = lp->tx_ring_size - 1;
675 lp->tx_len_bits = (i << 12);
676
677 for (i = 2; i <= PCNET32_LOG_MAX_RX_BUFFERS; i++) {
678 if (lp->rx_ring_size <= (1 << i))
679 break;
680 }
681 lp->rx_ring_size = (1 << i);
682 lp->rx_mod_mask = lp->rx_ring_size - 1;
683 lp->rx_len_bits = (i << 4);
684
685 if (pcnet32_alloc_ring(dev, dev->name)) {
686 pcnet32_free_ring(dev); 668 pcnet32_free_ring(dev);
687 spin_unlock_irqrestore(&lp->lock, flags); 669 lp->tx_ring_size =
688 return -ENOMEM; 670 min(ering->tx_pending, (unsigned int)TX_MAX_RING_SIZE);
689 } 671 lp->rx_ring_size =
672 min(ering->rx_pending, (unsigned int)RX_MAX_RING_SIZE);
673
674 /* set the minimum ring size to 4, to allow the loopback test to work
675 * unchanged.
676 */
677 for (i = 2; i <= PCNET32_LOG_MAX_TX_BUFFERS; i++) {
678 if (lp->tx_ring_size <= (1 << i))
679 break;
680 }
681 lp->tx_ring_size = (1 << i);
682 lp->tx_mod_mask = lp->tx_ring_size - 1;
683 lp->tx_len_bits = (i << 12);
690 684
691 spin_unlock_irqrestore(&lp->lock, flags); 685 for (i = 2; i <= PCNET32_LOG_MAX_RX_BUFFERS; i++) {
686 if (lp->rx_ring_size <= (1 << i))
687 break;
688 }
689 lp->rx_ring_size = (1 << i);
690 lp->rx_mod_mask = lp->rx_ring_size - 1;
691 lp->rx_len_bits = (i << 4);
692
693 if (pcnet32_alloc_ring(dev, dev->name)) {
694 pcnet32_free_ring(dev);
695 spin_unlock_irqrestore(&lp->lock, flags);
696 return -ENOMEM;
697 }
692 698
693 if (pcnet32_debug & NETIF_MSG_DRV) 699 spin_unlock_irqrestore(&lp->lock, flags);
694 printk(KERN_INFO PFX "%s: Ring Param Settings: RX: %d, TX: %d\n",
695 dev->name, lp->rx_ring_size, lp->tx_ring_size);
696 700
697 if (netif_running(dev)) 701 if (pcnet32_debug & NETIF_MSG_DRV)
698 pcnet32_open(dev); 702 printk(KERN_INFO PFX
703 "%s: Ring Param Settings: RX: %d, TX: %d\n", dev->name,
704 lp->rx_ring_size, lp->tx_ring_size);
699 705
700 return 0; 706 if (netif_running(dev))
707 pcnet32_open(dev);
708
709 return 0;
701} 710}
702 711
703static void pcnet32_get_strings(struct net_device *dev, u32 stringset, u8 *data) 712static void pcnet32_get_strings(struct net_device *dev, u32 stringset,
713 u8 * data)
704{ 714{
705 memcpy(data, pcnet32_gstrings_test, sizeof(pcnet32_gstrings_test)); 715 memcpy(data, pcnet32_gstrings_test, sizeof(pcnet32_gstrings_test));
706} 716}
707 717
708static int pcnet32_self_test_count(struct net_device *dev) 718static int pcnet32_self_test_count(struct net_device *dev)
709{ 719{
710 return PCNET32_TEST_LEN; 720 return PCNET32_TEST_LEN;
711} 721}
712 722
713static void pcnet32_ethtool_test(struct net_device *dev, 723static void pcnet32_ethtool_test(struct net_device *dev,
714 struct ethtool_test *test, u64 *data) 724 struct ethtool_test *test, u64 * data)
715{ 725{
716 struct pcnet32_private *lp = dev->priv; 726 struct pcnet32_private *lp = dev->priv;
717 int rc; 727 int rc;
718 728
719 if (test->flags == ETH_TEST_FL_OFFLINE) { 729 if (test->flags == ETH_TEST_FL_OFFLINE) {
720 rc = pcnet32_loopback_test(dev, data); 730 rc = pcnet32_loopback_test(dev, data);
721 if (rc) { 731 if (rc) {
722 if (netif_msg_hw(lp)) 732 if (netif_msg_hw(lp))
723 printk(KERN_DEBUG "%s: Loopback test failed.\n", dev->name); 733 printk(KERN_DEBUG "%s: Loopback test failed.\n",
724 test->flags |= ETH_TEST_FL_FAILED; 734 dev->name);
735 test->flags |= ETH_TEST_FL_FAILED;
736 } else if (netif_msg_hw(lp))
737 printk(KERN_DEBUG "%s: Loopback test passed.\n",
738 dev->name);
725 } else if (netif_msg_hw(lp)) 739 } else if (netif_msg_hw(lp))
726 printk(KERN_DEBUG "%s: Loopback test passed.\n", dev->name); 740 printk(KERN_DEBUG
727 } else if (netif_msg_hw(lp)) 741 "%s: No tests to run (specify 'Offline' on ethtool).",
728 printk(KERN_DEBUG "%s: No tests to run (specify 'Offline' on ethtool).", dev->name); 742 dev->name);
729} /* end pcnet32_ethtool_test */ 743} /* end pcnet32_ethtool_test */
730 744
731static int pcnet32_loopback_test(struct net_device *dev, uint64_t *data1) 745static int pcnet32_loopback_test(struct net_device *dev, uint64_t * data1)
732{ 746{
733 struct pcnet32_private *lp = dev->priv; 747 struct pcnet32_private *lp = dev->priv;
734 struct pcnet32_access *a = &lp->a; /* access to registers */ 748 struct pcnet32_access *a = &lp->a; /* access to registers */
735 ulong ioaddr = dev->base_addr; /* card base I/O address */ 749 ulong ioaddr = dev->base_addr; /* card base I/O address */
736 struct sk_buff *skb; /* sk buff */ 750 struct sk_buff *skb; /* sk buff */
737 int x, i; /* counters */ 751 int x, i; /* counters */
738 int numbuffs = 4; /* number of TX/RX buffers and descs */ 752 int numbuffs = 4; /* number of TX/RX buffers and descs */
739 u16 status = 0x8300; /* TX ring status */ 753 u16 status = 0x8300; /* TX ring status */
740 u16 teststatus; /* test of ring status */ 754 u16 teststatus; /* test of ring status */
741 int rc; /* return code */ 755 int rc; /* return code */
742 int size; /* size of packets */ 756 int size; /* size of packets */
743 unsigned char *packet; /* source packet data */ 757 unsigned char *packet; /* source packet data */
744 static const int data_len = 60; /* length of source packets */ 758 static const int data_len = 60; /* length of source packets */
745 unsigned long flags; 759 unsigned long flags;
746 unsigned long ticks; 760 unsigned long ticks;
747 761
748 *data1 = 1; /* status of test, default to fail */ 762 *data1 = 1; /* status of test, default to fail */
749 rc = 1; /* default to fail */ 763 rc = 1; /* default to fail */
750 764
751 if (netif_running(dev)) 765 if (netif_running(dev))
752 pcnet32_close(dev); 766 pcnet32_close(dev);
753 767
754 spin_lock_irqsave(&lp->lock, flags); 768 spin_lock_irqsave(&lp->lock, flags);
755 769
756 /* Reset the PCNET32 */ 770 /* Reset the PCNET32 */
757 lp->a.reset (ioaddr); 771 lp->a.reset(ioaddr);
758 772
759 /* switch pcnet32 to 32bit mode */ 773 /* switch pcnet32 to 32bit mode */
760 lp->a.write_bcr (ioaddr, 20, 2); 774 lp->a.write_bcr(ioaddr, 20, 2);
761 775
762 lp->init_block.mode = le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7); 776 lp->init_block.mode =
763 lp->init_block.filter[0] = 0; 777 le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7);
764 lp->init_block.filter[1] = 0; 778 lp->init_block.filter[0] = 0;
765 779 lp->init_block.filter[1] = 0;
766 /* purge & init rings but don't actually restart */ 780
767 pcnet32_restart(dev, 0x0000); 781 /* purge & init rings but don't actually restart */
768 782 pcnet32_restart(dev, 0x0000);
769 lp->a.write_csr(ioaddr, 0, 0x0004); /* Set STOP bit */ 783
770 784 lp->a.write_csr(ioaddr, 0, 0x0004); /* Set STOP bit */
771 /* Initialize Transmit buffers. */ 785
772 size = data_len + 15; 786 /* Initialize Transmit buffers. */
773 for (x=0; x<numbuffs; x++) { 787 size = data_len + 15;
774 if (!(skb = dev_alloc_skb(size))) { 788 for (x = 0; x < numbuffs; x++) {
775 if (netif_msg_hw(lp)) 789 if (!(skb = dev_alloc_skb(size))) {
776 printk(KERN_DEBUG "%s: Cannot allocate skb at line: %d!\n", 790 if (netif_msg_hw(lp))
777 dev->name, __LINE__); 791 printk(KERN_DEBUG
778 goto clean_up; 792 "%s: Cannot allocate skb at line: %d!\n",
779 } else { 793 dev->name, __LINE__);
780 packet = skb->data; 794 goto clean_up;
781 skb_put(skb, size); /* create space for data */ 795 } else {
782 lp->tx_skbuff[x] = skb; 796 packet = skb->data;
783 lp->tx_ring[x].length = le16_to_cpu(-skb->len); 797 skb_put(skb, size); /* create space for data */
784 lp->tx_ring[x].misc = 0; 798 lp->tx_skbuff[x] = skb;
785 799 lp->tx_ring[x].length = le16_to_cpu(-skb->len);
786 /* put DA and SA into the skb */ 800 lp->tx_ring[x].misc = 0;
787 for (i=0; i<6; i++) 801
788 *packet++ = dev->dev_addr[i]; 802 /* put DA and SA into the skb */
789 for (i=0; i<6; i++) 803 for (i = 0; i < 6; i++)
790 *packet++ = dev->dev_addr[i]; 804 *packet++ = dev->dev_addr[i];
791 /* type */ 805 for (i = 0; i < 6; i++)
792 *packet++ = 0x08; 806 *packet++ = dev->dev_addr[i];
793 *packet++ = 0x06; 807 /* type */
794 /* packet number */ 808 *packet++ = 0x08;
795 *packet++ = x; 809 *packet++ = 0x06;
796 /* fill packet with data */ 810 /* packet number */
797 for (i=0; i<data_len; i++) 811 *packet++ = x;
798 *packet++ = i; 812 /* fill packet with data */
799 813 for (i = 0; i < data_len; i++)
800 lp->tx_dma_addr[x] = pci_map_single(lp->pci_dev, skb->data, 814 *packet++ = i;
801 skb->len, PCI_DMA_TODEVICE); 815
802 lp->tx_ring[x].base = (u32)le32_to_cpu(lp->tx_dma_addr[x]); 816 lp->tx_dma_addr[x] =
803 wmb(); /* Make sure owner changes after all others are visible */ 817 pci_map_single(lp->pci_dev, skb->data, skb->len,
804 lp->tx_ring[x].status = le16_to_cpu(status); 818 PCI_DMA_TODEVICE);
805 } 819 lp->tx_ring[x].base =
806 } 820 (u32) le32_to_cpu(lp->tx_dma_addr[x]);
807 821 wmb(); /* Make sure owner changes after all others are visible */
808 x = a->read_bcr(ioaddr, 32); /* set internal loopback in BSR32 */ 822 lp->tx_ring[x].status = le16_to_cpu(status);
809 x = x | 0x0002; 823 }
810 a->write_bcr(ioaddr, 32, x);
811
812 lp->a.write_csr (ioaddr, 15, 0x0044); /* set int loopback in CSR15 */
813
814 teststatus = le16_to_cpu(0x8000);
815 lp->a.write_csr(ioaddr, 0, 0x0002); /* Set STRT bit */
816
817 /* Check status of descriptors */
818 for (x=0; x<numbuffs; x++) {
819 ticks = 0;
820 rmb();
821 while ((lp->rx_ring[x].status & teststatus) && (ticks < 200)) {
822 spin_unlock_irqrestore(&lp->lock, flags);
823 mdelay(1);
824 spin_lock_irqsave(&lp->lock, flags);
825 rmb();
826 ticks++;
827 }
828 if (ticks == 200) {
829 if (netif_msg_hw(lp))
830 printk("%s: Desc %d failed to reset!\n",dev->name,x);
831 break;
832 }
833 }
834
835 lp->a.write_csr(ioaddr, 0, 0x0004); /* Set STOP bit */
836 wmb();
837 if (netif_msg_hw(lp) && netif_msg_pktdata(lp)) {
838 printk(KERN_DEBUG "%s: RX loopback packets:\n", dev->name);
839
840 for (x=0; x<numbuffs; x++) {
841 printk(KERN_DEBUG "%s: Packet %d:\n", dev->name, x);
842 skb = lp->rx_skbuff[x];
843 for (i=0; i<size; i++) {
844 printk("%02x ", *(skb->data+i));
845 }
846 printk("\n");
847 }
848 }
849
850 x = 0;
851 rc = 0;
852 while (x<numbuffs && !rc) {
853 skb = lp->rx_skbuff[x];
854 packet = lp->tx_skbuff[x]->data;
855 for (i=0; i<size; i++) {
856 if (*(skb->data+i) != packet[i]) {
857 if (netif_msg_hw(lp))
858 printk(KERN_DEBUG "%s: Error in compare! %2x - %02x %02x\n",
859 dev->name, i, *(skb->data+i), packet[i]);
860 rc = 1;
861 break;
862 }
863 } 824 }
864 x++;
865 }
866 if (!rc) {
867 *data1 = 0;
868 }
869 825
870clean_up: 826 x = a->read_bcr(ioaddr, 32); /* set internal loopback in BSR32 */
871 pcnet32_purge_tx_ring(dev); 827 x = x | 0x0002;
872 x = a->read_csr(ioaddr, 15) & 0xFFFF; 828 a->write_bcr(ioaddr, 32, x);
873 a->write_csr(ioaddr, 15, (x & ~0x0044)); /* reset bits 6 and 2 */ 829
830 lp->a.write_csr(ioaddr, 15, 0x0044); /* set int loopback in CSR15 */
831
832 teststatus = le16_to_cpu(0x8000);
833 lp->a.write_csr(ioaddr, 0, 0x0002); /* Set STRT bit */
834
835 /* Check status of descriptors */
836 for (x = 0; x < numbuffs; x++) {
837 ticks = 0;
838 rmb();
839 while ((lp->rx_ring[x].status & teststatus) && (ticks < 200)) {
840 spin_unlock_irqrestore(&lp->lock, flags);
841 mdelay(1);
842 spin_lock_irqsave(&lp->lock, flags);
843 rmb();
844 ticks++;
845 }
846 if (ticks == 200) {
847 if (netif_msg_hw(lp))
848 printk("%s: Desc %d failed to reset!\n",
849 dev->name, x);
850 break;
851 }
852 }
853
854 lp->a.write_csr(ioaddr, 0, 0x0004); /* Set STOP bit */
855 wmb();
856 if (netif_msg_hw(lp) && netif_msg_pktdata(lp)) {
857 printk(KERN_DEBUG "%s: RX loopback packets:\n", dev->name);
858
859 for (x = 0; x < numbuffs; x++) {
860 printk(KERN_DEBUG "%s: Packet %d:\n", dev->name, x);
861 skb = lp->rx_skbuff[x];
862 for (i = 0; i < size; i++) {
863 printk("%02x ", *(skb->data + i));
864 }
865 printk("\n");
866 }
867 }
874 868
875 x = a->read_bcr(ioaddr, 32); /* reset internal loopback */ 869 x = 0;
876 x = x & ~0x0002; 870 rc = 0;
877 a->write_bcr(ioaddr, 32, x); 871 while (x < numbuffs && !rc) {
872 skb = lp->rx_skbuff[x];
873 packet = lp->tx_skbuff[x]->data;
874 for (i = 0; i < size; i++) {
875 if (*(skb->data + i) != packet[i]) {
876 if (netif_msg_hw(lp))
877 printk(KERN_DEBUG
878 "%s: Error in compare! %2x - %02x %02x\n",
879 dev->name, i, *(skb->data + i),
880 packet[i]);
881 rc = 1;
882 break;
883 }
884 }
885 x++;
886 }
887 if (!rc) {
888 *data1 = 0;
889 }
878 890
879 spin_unlock_irqrestore(&lp->lock, flags); 891 clean_up:
892 pcnet32_purge_tx_ring(dev);
893 x = a->read_csr(ioaddr, 15) & 0xFFFF;
894 a->write_csr(ioaddr, 15, (x & ~0x0044)); /* reset bits 6 and 2 */
880 895
881 if (netif_running(dev)) { 896 x = a->read_bcr(ioaddr, 32); /* reset internal loopback */
882 pcnet32_open(dev); 897 x = x & ~0x0002;
883 } else { 898 a->write_bcr(ioaddr, 32, x);
884 lp->a.write_bcr (ioaddr, 20, 4); /* return to 16bit mode */
885 }
886 899
887 return(rc); 900 spin_unlock_irqrestore(&lp->lock, flags);
888} /* end pcnet32_loopback_test */ 901
902 if (netif_running(dev)) {
903 pcnet32_open(dev);
904 } else {
905 lp->a.write_bcr(ioaddr, 20, 4); /* return to 16bit mode */
906 }
907
908 return (rc);
909} /* end pcnet32_loopback_test */
889 910
890static void pcnet32_led_blink_callback(struct net_device *dev) 911static void pcnet32_led_blink_callback(struct net_device *dev)
891{ 912{
892 struct pcnet32_private *lp = dev->priv; 913 struct pcnet32_private *lp = dev->priv;
893 struct pcnet32_access *a = &lp->a; 914 struct pcnet32_access *a = &lp->a;
894 ulong ioaddr = dev->base_addr; 915 ulong ioaddr = dev->base_addr;
895 unsigned long flags; 916 unsigned long flags;
896 int i; 917 int i;
897 918
898 spin_lock_irqsave(&lp->lock, flags); 919 spin_lock_irqsave(&lp->lock, flags);
899 for (i=4; i<8; i++) { 920 for (i = 4; i < 8; i++) {
900 a->write_bcr(ioaddr, i, a->read_bcr(ioaddr, i) ^ 0x4000); 921 a->write_bcr(ioaddr, i, a->read_bcr(ioaddr, i) ^ 0x4000);
901 } 922 }
902 spin_unlock_irqrestore(&lp->lock, flags); 923 spin_unlock_irqrestore(&lp->lock, flags);
903 924
904 mod_timer(&lp->blink_timer, PCNET32_BLINK_TIMEOUT); 925 mod_timer(&lp->blink_timer, PCNET32_BLINK_TIMEOUT);
905} 926}
906 927
907static int pcnet32_phys_id(struct net_device *dev, u32 data) 928static int pcnet32_phys_id(struct net_device *dev, u32 data)
908{ 929{
909 struct pcnet32_private *lp = dev->priv; 930 struct pcnet32_private *lp = dev->priv;
910 struct pcnet32_access *a = &lp->a; 931 struct pcnet32_access *a = &lp->a;
911 ulong ioaddr = dev->base_addr; 932 ulong ioaddr = dev->base_addr;
912 unsigned long flags; 933 unsigned long flags;
913 int i, regs[4]; 934 int i, regs[4];
914 935
915 if (!lp->blink_timer.function) { 936 if (!lp->blink_timer.function) {
916 init_timer(&lp->blink_timer); 937 init_timer(&lp->blink_timer);
917 lp->blink_timer.function = (void *) pcnet32_led_blink_callback; 938 lp->blink_timer.function = (void *)pcnet32_led_blink_callback;
918 lp->blink_timer.data = (unsigned long) dev; 939 lp->blink_timer.data = (unsigned long)dev;
919 } 940 }
920 941
921 /* Save the current value of the bcrs */ 942 /* Save the current value of the bcrs */
922 spin_lock_irqsave(&lp->lock, flags); 943 spin_lock_irqsave(&lp->lock, flags);
923 for (i=4; i<8; i++) { 944 for (i = 4; i < 8; i++) {
924 regs[i-4] = a->read_bcr(ioaddr, i); 945 regs[i - 4] = a->read_bcr(ioaddr, i);
925 } 946 }
926 spin_unlock_irqrestore(&lp->lock, flags); 947 spin_unlock_irqrestore(&lp->lock, flags);
927 948
928 mod_timer(&lp->blink_timer, jiffies); 949 mod_timer(&lp->blink_timer, jiffies);
929 set_current_state(TASK_INTERRUPTIBLE); 950 set_current_state(TASK_INTERRUPTIBLE);
930 951
931 if ((!data) || (data > (u32)(MAX_SCHEDULE_TIMEOUT / HZ))) 952 if ((!data) || (data > (u32) (MAX_SCHEDULE_TIMEOUT / HZ)))
932 data = (u32)(MAX_SCHEDULE_TIMEOUT / HZ); 953 data = (u32) (MAX_SCHEDULE_TIMEOUT / HZ);
933 954
934 msleep_interruptible(data * 1000); 955 msleep_interruptible(data * 1000);
935 del_timer_sync(&lp->blink_timer); 956 del_timer_sync(&lp->blink_timer);
936 957
937 /* Restore the original value of the bcrs */ 958 /* Restore the original value of the bcrs */
938 spin_lock_irqsave(&lp->lock, flags); 959 spin_lock_irqsave(&lp->lock, flags);
939 for (i=4; i<8; i++) { 960 for (i = 4; i < 8; i++) {
940 a->write_bcr(ioaddr, i, regs[i-4]); 961 a->write_bcr(ioaddr, i, regs[i - 4]);
941 } 962 }
942 spin_unlock_irqrestore(&lp->lock, flags); 963 spin_unlock_irqrestore(&lp->lock, flags);
943 964
944 return 0; 965 return 0;
945} 966}
946 967
947#define PCNET32_REGS_PER_PHY 32 968#define PCNET32_REGS_PER_PHY 32
948#define PCNET32_MAX_PHYS 32 969#define PCNET32_MAX_PHYS 32
949static int pcnet32_get_regs_len(struct net_device *dev) 970static int pcnet32_get_regs_len(struct net_device *dev)
950{ 971{
951 struct pcnet32_private *lp = dev->priv; 972 struct pcnet32_private *lp = dev->priv;
952 int j = lp->phycount * PCNET32_REGS_PER_PHY; 973 int j = lp->phycount * PCNET32_REGS_PER_PHY;
953 974
954 return((PCNET32_NUM_REGS + j) * sizeof(u16)); 975 return ((PCNET32_NUM_REGS + j) * sizeof(u16));
955} 976}
956 977
957static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs, 978static void pcnet32_get_regs(struct net_device *dev, struct ethtool_regs *regs,
958 void *ptr) 979 void *ptr)
959{ 980{
960 int i, csr0; 981 int i, csr0;
961 u16 *buff = ptr; 982 u16 *buff = ptr;
962 struct pcnet32_private *lp = dev->priv; 983 struct pcnet32_private *lp = dev->priv;
963 struct pcnet32_access *a = &lp->a; 984 struct pcnet32_access *a = &lp->a;
964 ulong ioaddr = dev->base_addr; 985 ulong ioaddr = dev->base_addr;
965 int ticks; 986 int ticks;
966 unsigned long flags; 987 unsigned long flags;
967 988
968 spin_lock_irqsave(&lp->lock, flags); 989 spin_lock_irqsave(&lp->lock, flags);
969 990
970 csr0 = a->read_csr(ioaddr, 0); 991 csr0 = a->read_csr(ioaddr, 0);
971 if (!(csr0 & 0x0004)) { /* If not stopped */ 992 if (!(csr0 & 0x0004)) { /* If not stopped */
972 /* set SUSPEND (SPND) - CSR5 bit 0 */ 993 /* set SUSPEND (SPND) - CSR5 bit 0 */
973 a->write_csr(ioaddr, 5, 0x0001); 994 a->write_csr(ioaddr, 5, 0x0001);
974 995
975 /* poll waiting for bit to be set */ 996 /* poll waiting for bit to be set */
976 ticks = 0; 997 ticks = 0;
977 while (!(a->read_csr(ioaddr, 5) & 0x0001)) { 998 while (!(a->read_csr(ioaddr, 5) & 0x0001)) {
978 spin_unlock_irqrestore(&lp->lock, flags); 999 spin_unlock_irqrestore(&lp->lock, flags);
979 mdelay(1); 1000 mdelay(1);
980 spin_lock_irqsave(&lp->lock, flags); 1001 spin_lock_irqsave(&lp->lock, flags);
981 ticks++; 1002 ticks++;
982 if (ticks > 200) { 1003 if (ticks > 200) {
983 if (netif_msg_hw(lp)) 1004 if (netif_msg_hw(lp))
984 printk(KERN_DEBUG "%s: Error getting into suspend!\n", 1005 printk(KERN_DEBUG
985 dev->name); 1006 "%s: Error getting into suspend!\n",
986 break; 1007 dev->name);
987 } 1008 break;
988 } 1009 }
989 }
990
991 /* read address PROM */
992 for (i=0; i<16; i += 2)
993 *buff++ = inw(ioaddr + i);
994
995 /* read control and status registers */
996 for (i=0; i<90; i++) {
997 *buff++ = a->read_csr(ioaddr, i);
998 }
999
1000 *buff++ = a->read_csr(ioaddr, 112);
1001 *buff++ = a->read_csr(ioaddr, 114);
1002
1003 /* read bus configuration registers */
1004 for (i=0; i<30; i++) {
1005 *buff++ = a->read_bcr(ioaddr, i);
1006 }
1007 *buff++ = 0; /* skip bcr30 so as not to hang 79C976 */
1008 for (i=31; i<36; i++) {
1009 *buff++ = a->read_bcr(ioaddr, i);
1010 }
1011
1012 /* read mii phy registers */
1013 if (lp->mii) {
1014 int j;
1015 for (j=0; j<PCNET32_MAX_PHYS; j++) {
1016 if (lp->phymask & (1 << j)) {
1017 for (i=0; i<PCNET32_REGS_PER_PHY; i++) {
1018 lp->a.write_bcr(ioaddr, 33, (j << 5) | i);
1019 *buff++ = lp->a.read_bcr(ioaddr, 34);
1020 } 1010 }
1021 }
1022 } 1011 }
1023 }
1024 1012
1025 if (!(csr0 & 0x0004)) { /* If not stopped */ 1013 /* read address PROM */
1026 /* clear SUSPEND (SPND) - CSR5 bit 0 */ 1014 for (i = 0; i < 16; i += 2)
1027 a->write_csr(ioaddr, 5, 0x0000); 1015 *buff++ = inw(ioaddr + i);
1028 } 1016
1017 /* read control and status registers */
1018 for (i = 0; i < 90; i++) {
1019 *buff++ = a->read_csr(ioaddr, i);
1020 }
1021
1022 *buff++ = a->read_csr(ioaddr, 112);
1023 *buff++ = a->read_csr(ioaddr, 114);
1029 1024
1030 spin_unlock_irqrestore(&lp->lock, flags); 1025 /* read bus configuration registers */
1026 for (i = 0; i < 30; i++) {
1027 *buff++ = a->read_bcr(ioaddr, i);
1028 }
1029 *buff++ = 0; /* skip bcr30 so as not to hang 79C976 */
1030 for (i = 31; i < 36; i++) {
1031 *buff++ = a->read_bcr(ioaddr, i);
1032 }
1033
1034 /* read mii phy registers */
1035 if (lp->mii) {
1036 int j;
1037 for (j = 0; j < PCNET32_MAX_PHYS; j++) {
1038 if (lp->phymask & (1 << j)) {
1039 for (i = 0; i < PCNET32_REGS_PER_PHY; i++) {
1040 lp->a.write_bcr(ioaddr, 33,
1041 (j << 5) | i);
1042 *buff++ = lp->a.read_bcr(ioaddr, 34);
1043 }
1044 }
1045 }
1046 }
1047
1048 if (!(csr0 & 0x0004)) { /* If not stopped */
1049 /* clear SUSPEND (SPND) - CSR5 bit 0 */
1050 a->write_csr(ioaddr, 5, 0x0000);
1051 }
1052
1053 spin_unlock_irqrestore(&lp->lock, flags);
1031} 1054}
1032 1055
1033static struct ethtool_ops pcnet32_ethtool_ops = { 1056static struct ethtool_ops pcnet32_ethtool_ops = {
1034 .get_settings = pcnet32_get_settings, 1057 .get_settings = pcnet32_get_settings,
1035 .set_settings = pcnet32_set_settings, 1058 .set_settings = pcnet32_set_settings,
1036 .get_drvinfo = pcnet32_get_drvinfo, 1059 .get_drvinfo = pcnet32_get_drvinfo,
1037 .get_msglevel = pcnet32_get_msglevel, 1060 .get_msglevel = pcnet32_get_msglevel,
1038 .set_msglevel = pcnet32_set_msglevel, 1061 .set_msglevel = pcnet32_set_msglevel,
1039 .nway_reset = pcnet32_nway_reset, 1062 .nway_reset = pcnet32_nway_reset,
1040 .get_link = pcnet32_get_link, 1063 .get_link = pcnet32_get_link,
1041 .get_ringparam = pcnet32_get_ringparam, 1064 .get_ringparam = pcnet32_get_ringparam,
1042 .set_ringparam = pcnet32_set_ringparam, 1065 .set_ringparam = pcnet32_set_ringparam,
1043 .get_tx_csum = ethtool_op_get_tx_csum, 1066 .get_tx_csum = ethtool_op_get_tx_csum,
1044 .get_sg = ethtool_op_get_sg, 1067 .get_sg = ethtool_op_get_sg,
1045 .get_tso = ethtool_op_get_tso, 1068 .get_tso = ethtool_op_get_tso,
1046 .get_strings = pcnet32_get_strings, 1069 .get_strings = pcnet32_get_strings,
1047 .self_test_count = pcnet32_self_test_count, 1070 .self_test_count = pcnet32_self_test_count,
1048 .self_test = pcnet32_ethtool_test, 1071 .self_test = pcnet32_ethtool_test,
1049 .phys_id = pcnet32_phys_id, 1072 .phys_id = pcnet32_phys_id,
1050 .get_regs_len = pcnet32_get_regs_len, 1073 .get_regs_len = pcnet32_get_regs_len,
1051 .get_regs = pcnet32_get_regs, 1074 .get_regs = pcnet32_get_regs,
1052 .get_perm_addr = ethtool_op_get_perm_addr, 1075 .get_perm_addr = ethtool_op_get_perm_addr,
1053}; 1076};
1054 1077
1055/* only probes for non-PCI devices, the rest are handled by 1078/* only probes for non-PCI devices, the rest are handled by
1056 * pci_register_driver via pcnet32_probe_pci */ 1079 * pci_register_driver via pcnet32_probe_pci */
1057 1080
1058static void __devinit 1081static void __devinit pcnet32_probe_vlbus(void)
1059pcnet32_probe_vlbus(void)
1060{ 1082{
1061 unsigned int *port, ioaddr; 1083 unsigned int *port, ioaddr;
1062 1084
1063 /* search for PCnet32 VLB cards at known addresses */ 1085 /* search for PCnet32 VLB cards at known addresses */
1064 for (port = pcnet32_portlist; (ioaddr = *port); port++) { 1086 for (port = pcnet32_portlist; (ioaddr = *port); port++) {
1065 if (request_region(ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_vlbus")) { 1087 if (request_region
1066 /* check if there is really a pcnet chip on that ioaddr */ 1088 (ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_vlbus")) {
1067 if ((inb(ioaddr + 14) == 0x57) && (inb(ioaddr + 15) == 0x57)) { 1089 /* check if there is really a pcnet chip on that ioaddr */
1068 pcnet32_probe1(ioaddr, 0, NULL); 1090 if ((inb(ioaddr + 14) == 0x57)
1069 } else { 1091 && (inb(ioaddr + 15) == 0x57)) {
1070 release_region(ioaddr, PCNET32_TOTAL_SIZE); 1092 pcnet32_probe1(ioaddr, 0, NULL);
1071 } 1093 } else {
1072 } 1094 release_region(ioaddr, PCNET32_TOTAL_SIZE);
1073 } 1095 }
1096 }
1097 }
1074} 1098}
1075 1099
1076
1077static int __devinit 1100static int __devinit
1078pcnet32_probe_pci(struct pci_dev *pdev, const struct pci_device_id *ent) 1101pcnet32_probe_pci(struct pci_dev *pdev, const struct pci_device_id *ent)
1079{ 1102{
1080 unsigned long ioaddr; 1103 unsigned long ioaddr;
1081 int err; 1104 int err;
1082 1105
1083 err = pci_enable_device(pdev); 1106 err = pci_enable_device(pdev);
1084 if (err < 0) { 1107 if (err < 0) {
1085 if (pcnet32_debug & NETIF_MSG_PROBE) 1108 if (pcnet32_debug & NETIF_MSG_PROBE)
1086 printk(KERN_ERR PFX "failed to enable device -- err=%d\n", err); 1109 printk(KERN_ERR PFX
1087 return err; 1110 "failed to enable device -- err=%d\n", err);
1088 } 1111 return err;
1089 pci_set_master(pdev); 1112 }
1113 pci_set_master(pdev);
1114
1115 ioaddr = pci_resource_start(pdev, 0);
1116 if (!ioaddr) {
1117 if (pcnet32_debug & NETIF_MSG_PROBE)
1118 printk(KERN_ERR PFX
1119 "card has no PCI IO resources, aborting\n");
1120 return -ENODEV;
1121 }
1090 1122
1091 ioaddr = pci_resource_start (pdev, 0); 1123 if (!pci_dma_supported(pdev, PCNET32_DMA_MASK)) {
1092 if (!ioaddr) { 1124 if (pcnet32_debug & NETIF_MSG_PROBE)
1093 if (pcnet32_debug & NETIF_MSG_PROBE) 1125 printk(KERN_ERR PFX
1094 printk (KERN_ERR PFX "card has no PCI IO resources, aborting\n"); 1126 "architecture does not support 32bit PCI busmaster DMA\n");
1095 return -ENODEV; 1127 return -ENODEV;
1096 } 1128 }
1129 if (request_region(ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_pci") ==
1130 NULL) {
1131 if (pcnet32_debug & NETIF_MSG_PROBE)
1132 printk(KERN_ERR PFX
1133 "io address range already allocated\n");
1134 return -EBUSY;
1135 }
1097 1136
1098 if (!pci_dma_supported(pdev, PCNET32_DMA_MASK)) { 1137 err = pcnet32_probe1(ioaddr, 1, pdev);
1099 if (pcnet32_debug & NETIF_MSG_PROBE) 1138 if (err < 0) {
1100 printk(KERN_ERR PFX "architecture does not support 32bit PCI busmaster DMA\n"); 1139 pci_disable_device(pdev);
1101 return -ENODEV; 1140 }
1102 } 1141 return err;
1103 if (request_region(ioaddr, PCNET32_TOTAL_SIZE, "pcnet32_probe_pci") == NULL) {
1104 if (pcnet32_debug & NETIF_MSG_PROBE)
1105 printk(KERN_ERR PFX "io address range already allocated\n");
1106 return -EBUSY;
1107 }
1108
1109 err = pcnet32_probe1(ioaddr, 1, pdev);
1110 if (err < 0) {
1111 pci_disable_device(pdev);
1112 }
1113 return err;
1114} 1142}
1115 1143
1116
1117/* pcnet32_probe1 1144/* pcnet32_probe1
1118 * Called from both pcnet32_probe_vlbus and pcnet_probe_pci. 1145 * Called from both pcnet32_probe_vlbus and pcnet_probe_pci.
1119 * pdev will be NULL when called from pcnet32_probe_vlbus. 1146 * pdev will be NULL when called from pcnet32_probe_vlbus.
@@ -1121,710 +1148,764 @@ pcnet32_probe_pci(struct pci_dev *pdev, const struct pci_device_id *ent)
1121static int __devinit 1148static int __devinit
1122pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev) 1149pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev)
1123{ 1150{
1124 struct pcnet32_private *lp; 1151 struct pcnet32_private *lp;
1125 dma_addr_t lp_dma_addr; 1152 dma_addr_t lp_dma_addr;
1126 int i, media; 1153 int i, media;
1127 int fdx, mii, fset, dxsuflo; 1154 int fdx, mii, fset, dxsuflo;
1128 int chip_version; 1155 int chip_version;
1129 char *chipname; 1156 char *chipname;
1130 struct net_device *dev; 1157 struct net_device *dev;
1131 struct pcnet32_access *a = NULL; 1158 struct pcnet32_access *a = NULL;
1132 u8 promaddr[6]; 1159 u8 promaddr[6];
1133 int ret = -ENODEV; 1160 int ret = -ENODEV;
1134 1161
1135 /* reset the chip */ 1162 /* reset the chip */
1136 pcnet32_wio_reset(ioaddr); 1163 pcnet32_wio_reset(ioaddr);
1137 1164
1138 /* NOTE: 16-bit check is first, otherwise some older PCnet chips fail */ 1165 /* NOTE: 16-bit check is first, otherwise some older PCnet chips fail */
1139 if (pcnet32_wio_read_csr(ioaddr, 0) == 4 && pcnet32_wio_check(ioaddr)) { 1166 if (pcnet32_wio_read_csr(ioaddr, 0) == 4 && pcnet32_wio_check(ioaddr)) {
1140 a = &pcnet32_wio; 1167 a = &pcnet32_wio;
1141 } else { 1168 } else {
1142 pcnet32_dwio_reset(ioaddr); 1169 pcnet32_dwio_reset(ioaddr);
1143 if (pcnet32_dwio_read_csr(ioaddr, 0) == 4 && pcnet32_dwio_check(ioaddr)) { 1170 if (pcnet32_dwio_read_csr(ioaddr, 0) == 4
1144 a = &pcnet32_dwio; 1171 && pcnet32_dwio_check(ioaddr)) {
1145 } else 1172 a = &pcnet32_dwio;
1146 goto err_release_region; 1173 } else
1147 } 1174 goto err_release_region;
1148 1175 }
1149 chip_version = a->read_csr(ioaddr, 88) | (a->read_csr(ioaddr,89) << 16); 1176
1150 if ((pcnet32_debug & NETIF_MSG_PROBE) && (pcnet32_debug & NETIF_MSG_HW)) 1177 chip_version =
1151 printk(KERN_INFO " PCnet chip version is %#x.\n", chip_version); 1178 a->read_csr(ioaddr, 88) | (a->read_csr(ioaddr, 89) << 16);
1152 if ((chip_version & 0xfff) != 0x003) { 1179 if ((pcnet32_debug & NETIF_MSG_PROBE) && (pcnet32_debug & NETIF_MSG_HW))
1153 if (pcnet32_debug & NETIF_MSG_PROBE) 1180 printk(KERN_INFO " PCnet chip version is %#x.\n",
1154 printk(KERN_INFO PFX "Unsupported chip version.\n"); 1181 chip_version);
1155 goto err_release_region; 1182 if ((chip_version & 0xfff) != 0x003) {
1156 } 1183 if (pcnet32_debug & NETIF_MSG_PROBE)
1157 1184 printk(KERN_INFO PFX "Unsupported chip version.\n");
1158 /* initialize variables */ 1185 goto err_release_region;
1159 fdx = mii = fset = dxsuflo = 0; 1186 }
1160 chip_version = (chip_version >> 12) & 0xffff; 1187
1161 1188 /* initialize variables */
1162 switch (chip_version) { 1189 fdx = mii = fset = dxsuflo = 0;
1163 case 0x2420: 1190 chip_version = (chip_version >> 12) & 0xffff;
1164 chipname = "PCnet/PCI 79C970"; /* PCI */ 1191
1165 break; 1192 switch (chip_version) {
1166 case 0x2430: 1193 case 0x2420:
1167 if (shared) 1194 chipname = "PCnet/PCI 79C970"; /* PCI */
1168 chipname = "PCnet/PCI 79C970"; /* 970 gives the wrong chip id back */ 1195 break;
1169 else 1196 case 0x2430:
1170 chipname = "PCnet/32 79C965"; /* 486/VL bus */ 1197 if (shared)
1171 break; 1198 chipname = "PCnet/PCI 79C970"; /* 970 gives the wrong chip id back */
1172 case 0x2621: 1199 else
1173 chipname = "PCnet/PCI II 79C970A"; /* PCI */ 1200 chipname = "PCnet/32 79C965"; /* 486/VL bus */
1174 fdx = 1; 1201 break;
1175 break; 1202 case 0x2621:
1176 case 0x2623: 1203 chipname = "PCnet/PCI II 79C970A"; /* PCI */
1177 chipname = "PCnet/FAST 79C971"; /* PCI */ 1204 fdx = 1;
1178 fdx = 1; mii = 1; fset = 1; 1205 break;
1179 break; 1206 case 0x2623:
1180 case 0x2624: 1207 chipname = "PCnet/FAST 79C971"; /* PCI */
1181 chipname = "PCnet/FAST+ 79C972"; /* PCI */ 1208 fdx = 1;
1182 fdx = 1; mii = 1; fset = 1; 1209 mii = 1;
1183 break; 1210 fset = 1;
1184 case 0x2625: 1211 break;
1185 chipname = "PCnet/FAST III 79C973"; /* PCI */ 1212 case 0x2624:
1186 fdx = 1; mii = 1; 1213 chipname = "PCnet/FAST+ 79C972"; /* PCI */
1187 break; 1214 fdx = 1;
1188 case 0x2626: 1215 mii = 1;
1189 chipname = "PCnet/Home 79C978"; /* PCI */ 1216 fset = 1;
1190 fdx = 1; 1217 break;
1218 case 0x2625:
1219 chipname = "PCnet/FAST III 79C973"; /* PCI */
1220 fdx = 1;
1221 mii = 1;
1222 break;
1223 case 0x2626:
1224 chipname = "PCnet/Home 79C978"; /* PCI */
1225 fdx = 1;
1226 /*
1227 * This is based on specs published at www.amd.com. This section
1228 * assumes that a card with a 79C978 wants to go into standard
1229 * ethernet mode. The 79C978 can also go into 1Mb HomePNA mode,
1230 * and the module option homepna=1 can select this instead.
1231 */
1232 media = a->read_bcr(ioaddr, 49);
1233 media &= ~3; /* default to 10Mb ethernet */
1234 if (cards_found < MAX_UNITS && homepna[cards_found])
1235 media |= 1; /* switch to home wiring mode */
1236 if (pcnet32_debug & NETIF_MSG_PROBE)
1237 printk(KERN_DEBUG PFX "media set to %sMbit mode.\n",
1238 (media & 1) ? "1" : "10");
1239 a->write_bcr(ioaddr, 49, media);
1240 break;
1241 case 0x2627:
1242 chipname = "PCnet/FAST III 79C975"; /* PCI */
1243 fdx = 1;
1244 mii = 1;
1245 break;
1246 case 0x2628:
1247 chipname = "PCnet/PRO 79C976";
1248 fdx = 1;
1249 mii = 1;
1250 break;
1251 default:
1252 if (pcnet32_debug & NETIF_MSG_PROBE)
1253 printk(KERN_INFO PFX
1254 "PCnet version %#x, no PCnet32 chip.\n",
1255 chip_version);
1256 goto err_release_region;
1257 }
1258
1191 /* 1259 /*
1192 * This is based on specs published at www.amd.com. This section 1260 * On selected chips turn on the BCR18:NOUFLO bit. This stops transmit
1193 * assumes that a card with a 79C978 wants to go into standard 1261 * starting until the packet is loaded. Strike one for reliability, lose
1194 * ethernet mode. The 79C978 can also go into 1Mb HomePNA mode, 1262 * one for latency - although on PCI this isnt a big loss. Older chips
1195 * and the module option homepna=1 can select this instead. 1263 * have FIFO's smaller than a packet, so you can't do this.
1264 * Turn on BCR18:BurstRdEn and BCR18:BurstWrEn.
1196 */ 1265 */
1197 media = a->read_bcr(ioaddr, 49); 1266
1198 media &= ~3; /* default to 10Mb ethernet */ 1267 if (fset) {
1199 if (cards_found < MAX_UNITS && homepna[cards_found]) 1268 a->write_bcr(ioaddr, 18, (a->read_bcr(ioaddr, 18) | 0x0860));
1200 media |= 1; /* switch to home wiring mode */ 1269 a->write_csr(ioaddr, 80,
1201 if (pcnet32_debug & NETIF_MSG_PROBE) 1270 (a->read_csr(ioaddr, 80) & 0x0C00) | 0x0c00);
1202 printk(KERN_DEBUG PFX "media set to %sMbit mode.\n", 1271 dxsuflo = 1;
1203 (media & 1) ? "1" : "10"); 1272 }
1204 a->write_bcr(ioaddr, 49, media); 1273
1205 break; 1274 dev = alloc_etherdev(0);
1206 case 0x2627: 1275 if (!dev) {
1207 chipname = "PCnet/FAST III 79C975"; /* PCI */ 1276 if (pcnet32_debug & NETIF_MSG_PROBE)
1208 fdx = 1; mii = 1; 1277 printk(KERN_ERR PFX "Memory allocation failed.\n");
1209 break; 1278 ret = -ENOMEM;
1210 case 0x2628: 1279 goto err_release_region;
1211 chipname = "PCnet/PRO 79C976"; 1280 }
1212 fdx = 1; mii = 1; 1281 SET_NETDEV_DEV(dev, &pdev->dev);
1213 break; 1282
1214 default:
1215 if (pcnet32_debug & NETIF_MSG_PROBE)
1216 printk(KERN_INFO PFX "PCnet version %#x, no PCnet32 chip.\n",
1217 chip_version);
1218 goto err_release_region;
1219 }
1220
1221 /*
1222 * On selected chips turn on the BCR18:NOUFLO bit. This stops transmit
1223 * starting until the packet is loaded. Strike one for reliability, lose
1224 * one for latency - although on PCI this isnt a big loss. Older chips
1225 * have FIFO's smaller than a packet, so you can't do this.
1226 * Turn on BCR18:BurstRdEn and BCR18:BurstWrEn.
1227 */
1228
1229 if (fset) {
1230 a->write_bcr(ioaddr, 18, (a->read_bcr(ioaddr, 18) | 0x0860));
1231 a->write_csr(ioaddr, 80, (a->read_csr(ioaddr, 80) & 0x0C00) | 0x0c00);
1232 dxsuflo = 1;
1233 }
1234
1235 dev = alloc_etherdev(0);
1236 if (!dev) {
1237 if (pcnet32_debug & NETIF_MSG_PROBE) 1283 if (pcnet32_debug & NETIF_MSG_PROBE)
1238 printk(KERN_ERR PFX "Memory allocation failed.\n"); 1284 printk(KERN_INFO PFX "%s at %#3lx,", chipname, ioaddr);
1239 ret = -ENOMEM; 1285
1240 goto err_release_region; 1286 /* In most chips, after a chip reset, the ethernet address is read from the
1241 } 1287 * station address PROM at the base address and programmed into the
1242 SET_NETDEV_DEV(dev, &pdev->dev); 1288 * "Physical Address Registers" CSR12-14.
1243 1289 * As a precautionary measure, we read the PROM values and complain if
1244 if (pcnet32_debug & NETIF_MSG_PROBE) 1290 * they disagree with the CSRs. Either way, we use the CSR values, and
1245 printk(KERN_INFO PFX "%s at %#3lx,", chipname, ioaddr); 1291 * double check that they are valid.
1246 1292 */
1247 /* In most chips, after a chip reset, the ethernet address is read from the 1293 for (i = 0; i < 3; i++) {
1248 * station address PROM at the base address and programmed into the 1294 unsigned int val;
1249 * "Physical Address Registers" CSR12-14. 1295 val = a->read_csr(ioaddr, i + 12) & 0x0ffff;
1250 * As a precautionary measure, we read the PROM values and complain if 1296 /* There may be endianness issues here. */
1251 * they disagree with the CSRs. Either way, we use the CSR values, and 1297 dev->dev_addr[2 * i] = val & 0x0ff;
1252 * double check that they are valid. 1298 dev->dev_addr[2 * i + 1] = (val >> 8) & 0x0ff;
1253 */ 1299 }
1254 for (i = 0; i < 3; i++) { 1300
1255 unsigned int val; 1301 /* read PROM address and compare with CSR address */
1256 val = a->read_csr(ioaddr, i+12) & 0x0ffff;
1257 /* There may be endianness issues here. */
1258 dev->dev_addr[2*i] = val & 0x0ff;
1259 dev->dev_addr[2*i+1] = (val >> 8) & 0x0ff;
1260 }
1261
1262 /* read PROM address and compare with CSR address */
1263 for (i = 0; i < 6; i++)
1264 promaddr[i] = inb(ioaddr + i);
1265
1266 if (memcmp(promaddr, dev->dev_addr, 6)
1267 || !is_valid_ether_addr(dev->dev_addr)) {
1268 if (is_valid_ether_addr(promaddr)) {
1269 if (pcnet32_debug & NETIF_MSG_PROBE) {
1270 printk(" warning: CSR address invalid,\n");
1271 printk(KERN_INFO " using instead PROM address of");
1272 }
1273 memcpy(dev->dev_addr, promaddr, 6);
1274 }
1275 }
1276 memcpy(dev->perm_addr, dev->dev_addr, dev->addr_len);
1277
1278 /* if the ethernet address is not valid, force to 00:00:00:00:00:00 */
1279 if (!is_valid_ether_addr(dev->perm_addr))
1280 memset(dev->dev_addr, 0, sizeof(dev->dev_addr));
1281
1282 if (pcnet32_debug & NETIF_MSG_PROBE) {
1283 for (i = 0; i < 6; i++) 1302 for (i = 0; i < 6; i++)
1284 printk(" %2.2x", dev->dev_addr[i]); 1303 promaddr[i] = inb(ioaddr + i);
1285 1304
1286 /* Version 0x2623 and 0x2624 */ 1305 if (memcmp(promaddr, dev->dev_addr, 6)
1287 if (((chip_version + 1) & 0xfffe) == 0x2624) { 1306 || !is_valid_ether_addr(dev->dev_addr)) {
1288 i = a->read_csr(ioaddr, 80) & 0x0C00; /* Check tx_start_pt */ 1307 if (is_valid_ether_addr(promaddr)) {
1289 printk("\n" KERN_INFO " tx_start_pt(0x%04x):",i); 1308 if (pcnet32_debug & NETIF_MSG_PROBE) {
1290 switch(i>>10) { 1309 printk(" warning: CSR address invalid,\n");
1291 case 0: printk(" 20 bytes,"); break; 1310 printk(KERN_INFO
1292 case 1: printk(" 64 bytes,"); break; 1311 " using instead PROM address of");
1293 case 2: printk(" 128 bytes,"); break; 1312 }
1294 case 3: printk("~220 bytes,"); break; 1313 memcpy(dev->dev_addr, promaddr, 6);
1295 } 1314 }
1296 i = a->read_bcr(ioaddr, 18); /* Check Burst/Bus control */ 1315 }
1297 printk(" BCR18(%x):",i&0xffff); 1316 memcpy(dev->perm_addr, dev->dev_addr, dev->addr_len);
1298 if (i & (1<<5)) printk("BurstWrEn "); 1317
1299 if (i & (1<<6)) printk("BurstRdEn "); 1318 /* if the ethernet address is not valid, force to 00:00:00:00:00:00 */
1300 if (i & (1<<7)) printk("DWordIO "); 1319 if (!is_valid_ether_addr(dev->perm_addr))
1301 if (i & (1<<11)) printk("NoUFlow "); 1320 memset(dev->dev_addr, 0, sizeof(dev->dev_addr));
1302 i = a->read_bcr(ioaddr, 25); 1321
1303 printk("\n" KERN_INFO " SRAMSIZE=0x%04x,",i<<8); 1322 if (pcnet32_debug & NETIF_MSG_PROBE) {
1304 i = a->read_bcr(ioaddr, 26); 1323 for (i = 0; i < 6; i++)
1305 printk(" SRAM_BND=0x%04x,",i<<8); 1324 printk(" %2.2x", dev->dev_addr[i]);
1306 i = a->read_bcr(ioaddr, 27); 1325
1307 if (i & (1<<14)) printk("LowLatRx"); 1326 /* Version 0x2623 and 0x2624 */
1308 } 1327 if (((chip_version + 1) & 0xfffe) == 0x2624) {
1309 } 1328 i = a->read_csr(ioaddr, 80) & 0x0C00; /* Check tx_start_pt */
1310 1329 printk("\n" KERN_INFO " tx_start_pt(0x%04x):", i);
1311 dev->base_addr = ioaddr; 1330 switch (i >> 10) {
1312 /* pci_alloc_consistent returns page-aligned memory, so we do not have to check the alignment */ 1331 case 0:
1313 if ((lp = pci_alloc_consistent(pdev, sizeof(*lp), &lp_dma_addr)) == NULL) { 1332 printk(" 20 bytes,");
1314 if (pcnet32_debug & NETIF_MSG_PROBE) 1333 break;
1315 printk(KERN_ERR PFX "Consistent memory allocation failed.\n"); 1334 case 1:
1316 ret = -ENOMEM; 1335 printk(" 64 bytes,");
1317 goto err_free_netdev; 1336 break;
1318 } 1337 case 2:
1319 1338 printk(" 128 bytes,");
1320 memset(lp, 0, sizeof(*lp)); 1339 break;
1321 lp->dma_addr = lp_dma_addr; 1340 case 3:
1322 lp->pci_dev = pdev; 1341 printk("~220 bytes,");
1323 1342 break;
1324 spin_lock_init(&lp->lock); 1343 }
1325 1344 i = a->read_bcr(ioaddr, 18); /* Check Burst/Bus control */
1326 SET_MODULE_OWNER(dev); 1345 printk(" BCR18(%x):", i & 0xffff);
1327 SET_NETDEV_DEV(dev, &pdev->dev); 1346 if (i & (1 << 5))
1328 dev->priv = lp; 1347 printk("BurstWrEn ");
1329 lp->name = chipname; 1348 if (i & (1 << 6))
1330 lp->shared_irq = shared; 1349 printk("BurstRdEn ");
1331 lp->tx_ring_size = TX_RING_SIZE; /* default tx ring size */ 1350 if (i & (1 << 7))
1332 lp->rx_ring_size = RX_RING_SIZE; /* default rx ring size */ 1351 printk("DWordIO ");
1333 lp->tx_mod_mask = lp->tx_ring_size - 1; 1352 if (i & (1 << 11))
1334 lp->rx_mod_mask = lp->rx_ring_size - 1; 1353 printk("NoUFlow ");
1335 lp->tx_len_bits = (PCNET32_LOG_TX_BUFFERS << 12); 1354 i = a->read_bcr(ioaddr, 25);
1336 lp->rx_len_bits = (PCNET32_LOG_RX_BUFFERS << 4); 1355 printk("\n" KERN_INFO " SRAMSIZE=0x%04x,", i << 8);
1337 lp->mii_if.full_duplex = fdx; 1356 i = a->read_bcr(ioaddr, 26);
1338 lp->mii_if.phy_id_mask = 0x1f; 1357 printk(" SRAM_BND=0x%04x,", i << 8);
1339 lp->mii_if.reg_num_mask = 0x1f; 1358 i = a->read_bcr(ioaddr, 27);
1340 lp->dxsuflo = dxsuflo; 1359 if (i & (1 << 14))
1341 lp->mii = mii; 1360 printk("LowLatRx");
1342 lp->msg_enable = pcnet32_debug; 1361 }
1343 if ((cards_found >= MAX_UNITS) || (options[cards_found] > sizeof(options_mapping))) 1362 }
1344 lp->options = PCNET32_PORT_ASEL; 1363
1345 else 1364 dev->base_addr = ioaddr;
1346 lp->options = options_mapping[options[cards_found]]; 1365 /* pci_alloc_consistent returns page-aligned memory, so we do not have to check the alignment */
1347 lp->mii_if.dev = dev; 1366 if ((lp =
1348 lp->mii_if.mdio_read = mdio_read; 1367 pci_alloc_consistent(pdev, sizeof(*lp), &lp_dma_addr)) == NULL) {
1349 lp->mii_if.mdio_write = mdio_write; 1368 if (pcnet32_debug & NETIF_MSG_PROBE)
1350 1369 printk(KERN_ERR PFX
1351 if (fdx && !(lp->options & PCNET32_PORT_ASEL) && 1370 "Consistent memory allocation failed.\n");
1352 ((cards_found>=MAX_UNITS) || full_duplex[cards_found])) 1371 ret = -ENOMEM;
1353 lp->options |= PCNET32_PORT_FD; 1372 goto err_free_netdev;
1354 1373 }
1355 if (!a) { 1374
1356 if (pcnet32_debug & NETIF_MSG_PROBE) 1375 memset(lp, 0, sizeof(*lp));
1357 printk(KERN_ERR PFX "No access methods\n"); 1376 lp->dma_addr = lp_dma_addr;
1358 ret = -ENODEV; 1377 lp->pci_dev = pdev;
1359 goto err_free_consistent; 1378
1360 } 1379 spin_lock_init(&lp->lock);
1361 lp->a = *a; 1380
1362 1381 SET_MODULE_OWNER(dev);
1363 /* prior to register_netdev, dev->name is not yet correct */ 1382 SET_NETDEV_DEV(dev, &pdev->dev);
1364 if (pcnet32_alloc_ring(dev, pci_name(lp->pci_dev))) { 1383 dev->priv = lp;
1365 ret = -ENOMEM; 1384 lp->name = chipname;
1366 goto err_free_ring; 1385 lp->shared_irq = shared;
1367 } 1386 lp->tx_ring_size = TX_RING_SIZE; /* default tx ring size */
1368 /* detect special T1/E1 WAN card by checking for MAC address */ 1387 lp->rx_ring_size = RX_RING_SIZE; /* default rx ring size */
1369 if (dev->dev_addr[0] == 0x00 && dev->dev_addr[1] == 0xe0 1388 lp->tx_mod_mask = lp->tx_ring_size - 1;
1389 lp->rx_mod_mask = lp->rx_ring_size - 1;
1390 lp->tx_len_bits = (PCNET32_LOG_TX_BUFFERS << 12);
1391 lp->rx_len_bits = (PCNET32_LOG_RX_BUFFERS << 4);
1392 lp->mii_if.full_duplex = fdx;
1393 lp->mii_if.phy_id_mask = 0x1f;
1394 lp->mii_if.reg_num_mask = 0x1f;
1395 lp->dxsuflo = dxsuflo;
1396 lp->mii = mii;
1397 lp->msg_enable = pcnet32_debug;
1398 if ((cards_found >= MAX_UNITS)
1399 || (options[cards_found] > sizeof(options_mapping)))
1400 lp->options = PCNET32_PORT_ASEL;
1401 else
1402 lp->options = options_mapping[options[cards_found]];
1403 lp->mii_if.dev = dev;
1404 lp->mii_if.mdio_read = mdio_read;
1405 lp->mii_if.mdio_write = mdio_write;
1406
1407 if (fdx && !(lp->options & PCNET32_PORT_ASEL) &&
1408 ((cards_found >= MAX_UNITS) || full_duplex[cards_found]))
1409 lp->options |= PCNET32_PORT_FD;
1410
1411 if (!a) {
1412 if (pcnet32_debug & NETIF_MSG_PROBE)
1413 printk(KERN_ERR PFX "No access methods\n");
1414 ret = -ENODEV;
1415 goto err_free_consistent;
1416 }
1417 lp->a = *a;
1418
1419 /* prior to register_netdev, dev->name is not yet correct */
1420 if (pcnet32_alloc_ring(dev, pci_name(lp->pci_dev))) {
1421 ret = -ENOMEM;
1422 goto err_free_ring;
1423 }
1424 /* detect special T1/E1 WAN card by checking for MAC address */
1425 if (dev->dev_addr[0] == 0x00 && dev->dev_addr[1] == 0xe0
1370 && dev->dev_addr[2] == 0x75) 1426 && dev->dev_addr[2] == 0x75)
1371 lp->options = PCNET32_PORT_FD | PCNET32_PORT_GPSI; 1427 lp->options = PCNET32_PORT_FD | PCNET32_PORT_GPSI;
1372
1373 lp->init_block.mode = le16_to_cpu(0x0003); /* Disable Rx and Tx. */
1374 lp->init_block.tlen_rlen = le16_to_cpu(lp->tx_len_bits | lp->rx_len_bits);
1375 for (i = 0; i < 6; i++)
1376 lp->init_block.phys_addr[i] = dev->dev_addr[i];
1377 lp->init_block.filter[0] = 0x00000000;
1378 lp->init_block.filter[1] = 0x00000000;
1379 lp->init_block.rx_ring = (u32)le32_to_cpu(lp->rx_ring_dma_addr);
1380 lp->init_block.tx_ring = (u32)le32_to_cpu(lp->tx_ring_dma_addr);
1381
1382 /* switch pcnet32 to 32bit mode */
1383 a->write_bcr(ioaddr, 20, 2);
1384
1385 a->write_csr(ioaddr, 1, (lp->dma_addr + offsetof(struct pcnet32_private,
1386 init_block)) & 0xffff);
1387 a->write_csr(ioaddr, 2, (lp->dma_addr + offsetof(struct pcnet32_private,
1388 init_block)) >> 16);
1389
1390 if (pdev) { /* use the IRQ provided by PCI */
1391 dev->irq = pdev->irq;
1392 if (pcnet32_debug & NETIF_MSG_PROBE)
1393 printk(" assigned IRQ %d.\n", dev->irq);
1394 } else {
1395 unsigned long irq_mask = probe_irq_on();
1396 1428
1397 /* 1429 lp->init_block.mode = le16_to_cpu(0x0003); /* Disable Rx and Tx. */
1398 * To auto-IRQ we enable the initialization-done and DMA error 1430 lp->init_block.tlen_rlen =
1399 * interrupts. For ISA boards we get a DMA error, but VLB and PCI 1431 le16_to_cpu(lp->tx_len_bits | lp->rx_len_bits);
1400 * boards will work. 1432 for (i = 0; i < 6; i++)
1401 */ 1433 lp->init_block.phys_addr[i] = dev->dev_addr[i];
1402 /* Trigger an initialization just for the interrupt. */ 1434 lp->init_block.filter[0] = 0x00000000;
1403 a->write_csr (ioaddr, 0, 0x41); 1435 lp->init_block.filter[1] = 0x00000000;
1404 mdelay (1); 1436 lp->init_block.rx_ring = (u32) le32_to_cpu(lp->rx_ring_dma_addr);
1437 lp->init_block.tx_ring = (u32) le32_to_cpu(lp->tx_ring_dma_addr);
1438
1439 /* switch pcnet32 to 32bit mode */
1440 a->write_bcr(ioaddr, 20, 2);
1441
1442 a->write_csr(ioaddr, 1, (lp->dma_addr + offsetof(struct pcnet32_private,
1443 init_block)) & 0xffff);
1444 a->write_csr(ioaddr, 2, (lp->dma_addr + offsetof(struct pcnet32_private,
1445 init_block)) >> 16);
1446
1447 if (pdev) { /* use the IRQ provided by PCI */
1448 dev->irq = pdev->irq;
1449 if (pcnet32_debug & NETIF_MSG_PROBE)
1450 printk(" assigned IRQ %d.\n", dev->irq);
1451 } else {
1452 unsigned long irq_mask = probe_irq_on();
1453
1454 /*
1455 * To auto-IRQ we enable the initialization-done and DMA error
1456 * interrupts. For ISA boards we get a DMA error, but VLB and PCI
1457 * boards will work.
1458 */
1459 /* Trigger an initialization just for the interrupt. */
1460 a->write_csr(ioaddr, 0, 0x41);
1461 mdelay(1);
1462
1463 dev->irq = probe_irq_off(irq_mask);
1464 if (!dev->irq) {
1465 if (pcnet32_debug & NETIF_MSG_PROBE)
1466 printk(", failed to detect IRQ line.\n");
1467 ret = -ENODEV;
1468 goto err_free_ring;
1469 }
1470 if (pcnet32_debug & NETIF_MSG_PROBE)
1471 printk(", probed IRQ %d.\n", dev->irq);
1472 }
1405 1473
1406 dev->irq = probe_irq_off (irq_mask); 1474 /* Set the mii phy_id so that we can query the link state */
1407 if (!dev->irq) { 1475 if (lp->mii) {
1408 if (pcnet32_debug & NETIF_MSG_PROBE) 1476 /* lp->phycount and lp->phymask are set to 0 by memset above */
1409 printk(", failed to detect IRQ line.\n"); 1477
1410 ret = -ENODEV; 1478 lp->mii_if.phy_id = ((lp->a.read_bcr(ioaddr, 33)) >> 5) & 0x1f;
1411 goto err_free_ring; 1479 /* scan for PHYs */
1480 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
1481 unsigned short id1, id2;
1482
1483 id1 = mdio_read(dev, i, MII_PHYSID1);
1484 if (id1 == 0xffff)
1485 continue;
1486 id2 = mdio_read(dev, i, MII_PHYSID2);
1487 if (id2 == 0xffff)
1488 continue;
1489 if (i == 31 && ((chip_version + 1) & 0xfffe) == 0x2624)
1490 continue; /* 79C971 & 79C972 have phantom phy at id 31 */
1491 lp->phycount++;
1492 lp->phymask |= (1 << i);
1493 lp->mii_if.phy_id = i;
1494 if (pcnet32_debug & NETIF_MSG_PROBE)
1495 printk(KERN_INFO PFX
1496 "Found PHY %04x:%04x at address %d.\n",
1497 id1, id2, i);
1498 }
1499 lp->a.write_bcr(ioaddr, 33, (lp->mii_if.phy_id) << 5);
1500 if (lp->phycount > 1) {
1501 lp->options |= PCNET32_PORT_MII;
1502 }
1412 } 1503 }
1413 if (pcnet32_debug & NETIF_MSG_PROBE) 1504
1414 printk(", probed IRQ %d.\n", dev->irq); 1505 init_timer(&lp->watchdog_timer);
1415 } 1506 lp->watchdog_timer.data = (unsigned long)dev;
1416 1507 lp->watchdog_timer.function = (void *)&pcnet32_watchdog;
1417 /* Set the mii phy_id so that we can query the link state */ 1508
1418 if (lp->mii) { 1509 /* The PCNET32-specific entries in the device structure. */
1419 /* lp->phycount and lp->phymask are set to 0 by memset above */ 1510 dev->open = &pcnet32_open;
1420 1511 dev->hard_start_xmit = &pcnet32_start_xmit;
1421 lp->mii_if.phy_id = ((lp->a.read_bcr (ioaddr, 33)) >> 5) & 0x1f; 1512 dev->stop = &pcnet32_close;
1422 /* scan for PHYs */ 1513 dev->get_stats = &pcnet32_get_stats;
1423 for (i=0; i<PCNET32_MAX_PHYS; i++) { 1514 dev->set_multicast_list = &pcnet32_set_multicast_list;
1424 unsigned short id1, id2; 1515 dev->do_ioctl = &pcnet32_ioctl;
1425 1516 dev->ethtool_ops = &pcnet32_ethtool_ops;
1426 id1 = mdio_read(dev, i, MII_PHYSID1); 1517 dev->tx_timeout = pcnet32_tx_timeout;
1427 if (id1 == 0xffff) 1518 dev->watchdog_timeo = (5 * HZ);
1428 continue;
1429 id2 = mdio_read(dev, i, MII_PHYSID2);
1430 if (id2 == 0xffff)
1431 continue;
1432 if (i == 31 && ((chip_version + 1) & 0xfffe) == 0x2624)
1433 continue; /* 79C971 & 79C972 have phantom phy at id 31 */
1434 lp->phycount++;
1435 lp->phymask |= (1 << i);
1436 lp->mii_if.phy_id = i;
1437 if (pcnet32_debug & NETIF_MSG_PROBE)
1438 printk(KERN_INFO PFX "Found PHY %04x:%04x at address %d.\n",
1439 id1, id2, i);
1440 }
1441 lp->a.write_bcr(ioaddr, 33, (lp->mii_if.phy_id) << 5);
1442 if (lp->phycount > 1) {
1443 lp->options |= PCNET32_PORT_MII;
1444 }
1445 }
1446
1447 init_timer (&lp->watchdog_timer);
1448 lp->watchdog_timer.data = (unsigned long) dev;
1449 lp->watchdog_timer.function = (void *) &pcnet32_watchdog;
1450
1451 /* The PCNET32-specific entries in the device structure. */
1452 dev->open = &pcnet32_open;
1453 dev->hard_start_xmit = &pcnet32_start_xmit;
1454 dev->stop = &pcnet32_close;
1455 dev->get_stats = &pcnet32_get_stats;
1456 dev->set_multicast_list = &pcnet32_set_multicast_list;
1457 dev->do_ioctl = &pcnet32_ioctl;
1458 dev->ethtool_ops = &pcnet32_ethtool_ops;
1459 dev->tx_timeout = pcnet32_tx_timeout;
1460 dev->watchdog_timeo = (5*HZ);
1461 1519
1462#ifdef CONFIG_NET_POLL_CONTROLLER 1520#ifdef CONFIG_NET_POLL_CONTROLLER
1463 dev->poll_controller = pcnet32_poll_controller; 1521 dev->poll_controller = pcnet32_poll_controller;
1464#endif 1522#endif
1465 1523
1466 /* Fill in the generic fields of the device structure. */ 1524 /* Fill in the generic fields of the device structure. */
1467 if (register_netdev(dev)) 1525 if (register_netdev(dev))
1468 goto err_free_ring; 1526 goto err_free_ring;
1469 1527
1470 if (pdev) { 1528 if (pdev) {
1471 pci_set_drvdata(pdev, dev); 1529 pci_set_drvdata(pdev, dev);
1472 } else { 1530 } else {
1473 lp->next = pcnet32_dev; 1531 lp->next = pcnet32_dev;
1474 pcnet32_dev = dev; 1532 pcnet32_dev = dev;
1475 } 1533 }
1476 1534
1477 if (pcnet32_debug & NETIF_MSG_PROBE) 1535 if (pcnet32_debug & NETIF_MSG_PROBE)
1478 printk(KERN_INFO "%s: registered as %s\n", dev->name, lp->name); 1536 printk(KERN_INFO "%s: registered as %s\n", dev->name, lp->name);
1479 cards_found++; 1537 cards_found++;
1480 1538
1481 /* enable LED writes */ 1539 /* enable LED writes */
1482 a->write_bcr(ioaddr, 2, a->read_bcr(ioaddr, 2) | 0x1000); 1540 a->write_bcr(ioaddr, 2, a->read_bcr(ioaddr, 2) | 0x1000);
1483
1484 return 0;
1485
1486err_free_ring:
1487 pcnet32_free_ring(dev);
1488err_free_consistent:
1489 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr);
1490err_free_netdev:
1491 free_netdev(dev);
1492err_release_region:
1493 release_region(ioaddr, PCNET32_TOTAL_SIZE);
1494 return ret;
1495}
1496 1541
1542 return 0;
1543
1544 err_free_ring:
1545 pcnet32_free_ring(dev);
1546 err_free_consistent:
1547 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr);
1548 err_free_netdev:
1549 free_netdev(dev);
1550 err_release_region:
1551 release_region(ioaddr, PCNET32_TOTAL_SIZE);
1552 return ret;
1553}
1497 1554
1498/* if any allocation fails, caller must also call pcnet32_free_ring */ 1555/* if any allocation fails, caller must also call pcnet32_free_ring */
1499static int pcnet32_alloc_ring(struct net_device *dev, char *name) 1556static int pcnet32_alloc_ring(struct net_device *dev, char *name)
1500{ 1557{
1501 struct pcnet32_private *lp = dev->priv; 1558 struct pcnet32_private *lp = dev->priv;
1502 1559
1503 lp->tx_ring = pci_alloc_consistent(lp->pci_dev, 1560 lp->tx_ring = pci_alloc_consistent(lp->pci_dev,
1504 sizeof(struct pcnet32_tx_head) * lp->tx_ring_size, 1561 sizeof(struct pcnet32_tx_head) *
1505 &lp->tx_ring_dma_addr); 1562 lp->tx_ring_size,
1506 if (lp->tx_ring == NULL) { 1563 &lp->tx_ring_dma_addr);
1507 if (pcnet32_debug & NETIF_MSG_DRV) 1564 if (lp->tx_ring == NULL) {
1508 printk("\n" KERN_ERR PFX "%s: Consistent memory allocation failed.\n", 1565 if (pcnet32_debug & NETIF_MSG_DRV)
1509 name); 1566 printk("\n" KERN_ERR PFX
1510 return -ENOMEM; 1567 "%s: Consistent memory allocation failed.\n",
1511 } 1568 name);
1512 1569 return -ENOMEM;
1513 lp->rx_ring = pci_alloc_consistent(lp->pci_dev, 1570 }
1514 sizeof(struct pcnet32_rx_head) * lp->rx_ring_size,
1515 &lp->rx_ring_dma_addr);
1516 if (lp->rx_ring == NULL) {
1517 if (pcnet32_debug & NETIF_MSG_DRV)
1518 printk("\n" KERN_ERR PFX "%s: Consistent memory allocation failed.\n",
1519 name);
1520 return -ENOMEM;
1521 }
1522
1523 lp->tx_dma_addr = kmalloc(sizeof(dma_addr_t) * lp->tx_ring_size,
1524 GFP_ATOMIC);
1525 if (!lp->tx_dma_addr) {
1526 if (pcnet32_debug & NETIF_MSG_DRV)
1527 printk("\n" KERN_ERR PFX "%s: Memory allocation failed.\n", name);
1528 return -ENOMEM;
1529 }
1530 memset(lp->tx_dma_addr, 0, sizeof(dma_addr_t) * lp->tx_ring_size);
1531
1532 lp->rx_dma_addr = kmalloc(sizeof(dma_addr_t) * lp->rx_ring_size,
1533 GFP_ATOMIC);
1534 if (!lp->rx_dma_addr) {
1535 if (pcnet32_debug & NETIF_MSG_DRV)
1536 printk("\n" KERN_ERR PFX "%s: Memory allocation failed.\n", name);
1537 return -ENOMEM;
1538 }
1539 memset(lp->rx_dma_addr, 0, sizeof(dma_addr_t) * lp->rx_ring_size);
1540
1541 lp->tx_skbuff = kmalloc(sizeof(struct sk_buff *) * lp->tx_ring_size,
1542 GFP_ATOMIC);
1543 if (!lp->tx_skbuff) {
1544 if (pcnet32_debug & NETIF_MSG_DRV)
1545 printk("\n" KERN_ERR PFX "%s: Memory allocation failed.\n", name);
1546 return -ENOMEM;
1547 }
1548 memset(lp->tx_skbuff, 0, sizeof(struct sk_buff *) * lp->tx_ring_size);
1549
1550 lp->rx_skbuff = kmalloc(sizeof(struct sk_buff *) * lp->rx_ring_size,
1551 GFP_ATOMIC);
1552 if (!lp->rx_skbuff) {
1553 if (pcnet32_debug & NETIF_MSG_DRV)
1554 printk("\n" KERN_ERR PFX "%s: Memory allocation failed.\n", name);
1555 return -ENOMEM;
1556 }
1557 memset(lp->rx_skbuff, 0, sizeof(struct sk_buff *) * lp->rx_ring_size);
1558 1571
1559 return 0; 1572 lp->rx_ring = pci_alloc_consistent(lp->pci_dev,
1560} 1573 sizeof(struct pcnet32_rx_head) *
1574 lp->rx_ring_size,
1575 &lp->rx_ring_dma_addr);
1576 if (lp->rx_ring == NULL) {
1577 if (pcnet32_debug & NETIF_MSG_DRV)
1578 printk("\n" KERN_ERR PFX
1579 "%s: Consistent memory allocation failed.\n",
1580 name);
1581 return -ENOMEM;
1582 }
1561 1583
1584 lp->tx_dma_addr = kmalloc(sizeof(dma_addr_t) * lp->tx_ring_size,
1585 GFP_ATOMIC);
1586 if (!lp->tx_dma_addr) {
1587 if (pcnet32_debug & NETIF_MSG_DRV)
1588 printk("\n" KERN_ERR PFX
1589 "%s: Memory allocation failed.\n", name);
1590 return -ENOMEM;
1591 }
1592 memset(lp->tx_dma_addr, 0, sizeof(dma_addr_t) * lp->tx_ring_size);
1593
1594 lp->rx_dma_addr = kmalloc(sizeof(dma_addr_t) * lp->rx_ring_size,
1595 GFP_ATOMIC);
1596 if (!lp->rx_dma_addr) {
1597 if (pcnet32_debug & NETIF_MSG_DRV)
1598 printk("\n" KERN_ERR PFX
1599 "%s: Memory allocation failed.\n", name);
1600 return -ENOMEM;
1601 }
1602 memset(lp->rx_dma_addr, 0, sizeof(dma_addr_t) * lp->rx_ring_size);
1603
1604 lp->tx_skbuff = kmalloc(sizeof(struct sk_buff *) * lp->tx_ring_size,
1605 GFP_ATOMIC);
1606 if (!lp->tx_skbuff) {
1607 if (pcnet32_debug & NETIF_MSG_DRV)
1608 printk("\n" KERN_ERR PFX
1609 "%s: Memory allocation failed.\n", name);
1610 return -ENOMEM;
1611 }
1612 memset(lp->tx_skbuff, 0, sizeof(struct sk_buff *) * lp->tx_ring_size);
1613
1614 lp->rx_skbuff = kmalloc(sizeof(struct sk_buff *) * lp->rx_ring_size,
1615 GFP_ATOMIC);
1616 if (!lp->rx_skbuff) {
1617 if (pcnet32_debug & NETIF_MSG_DRV)
1618 printk("\n" KERN_ERR PFX
1619 "%s: Memory allocation failed.\n", name);
1620 return -ENOMEM;
1621 }
1622 memset(lp->rx_skbuff, 0, sizeof(struct sk_buff *) * lp->rx_ring_size);
1623
1624 return 0;
1625}
1562 1626
1563static void pcnet32_free_ring(struct net_device *dev) 1627static void pcnet32_free_ring(struct net_device *dev)
1564{ 1628{
1565 struct pcnet32_private *lp = dev->priv; 1629 struct pcnet32_private *lp = dev->priv;
1566 1630
1567 kfree(lp->tx_skbuff); 1631 kfree(lp->tx_skbuff);
1568 lp->tx_skbuff = NULL; 1632 lp->tx_skbuff = NULL;
1569 1633
1570 kfree(lp->rx_skbuff); 1634 kfree(lp->rx_skbuff);
1571 lp->rx_skbuff = NULL; 1635 lp->rx_skbuff = NULL;
1572 1636
1573 kfree(lp->tx_dma_addr); 1637 kfree(lp->tx_dma_addr);
1574 lp->tx_dma_addr = NULL; 1638 lp->tx_dma_addr = NULL;
1575 1639
1576 kfree(lp->rx_dma_addr); 1640 kfree(lp->rx_dma_addr);
1577 lp->rx_dma_addr = NULL; 1641 lp->rx_dma_addr = NULL;
1578 1642
1579 if (lp->tx_ring) { 1643 if (lp->tx_ring) {
1580 pci_free_consistent(lp->pci_dev, sizeof(struct pcnet32_tx_head) * lp->tx_ring_size, 1644 pci_free_consistent(lp->pci_dev,
1581 lp->tx_ring, lp->tx_ring_dma_addr); 1645 sizeof(struct pcnet32_tx_head) *
1582 lp->tx_ring = NULL; 1646 lp->tx_ring_size, lp->tx_ring,
1583 } 1647 lp->tx_ring_dma_addr);
1648 lp->tx_ring = NULL;
1649 }
1584 1650
1585 if (lp->rx_ring) { 1651 if (lp->rx_ring) {
1586 pci_free_consistent(lp->pci_dev, sizeof(struct pcnet32_rx_head) * lp->rx_ring_size, 1652 pci_free_consistent(lp->pci_dev,
1587 lp->rx_ring, lp->rx_ring_dma_addr); 1653 sizeof(struct pcnet32_rx_head) *
1588 lp->rx_ring = NULL; 1654 lp->rx_ring_size, lp->rx_ring,
1589 } 1655 lp->rx_ring_dma_addr);
1656 lp->rx_ring = NULL;
1657 }
1590} 1658}
1591 1659
1592 1660static int pcnet32_open(struct net_device *dev)
1593static int
1594pcnet32_open(struct net_device *dev)
1595{ 1661{
1596 struct pcnet32_private *lp = dev->priv; 1662 struct pcnet32_private *lp = dev->priv;
1597 unsigned long ioaddr = dev->base_addr; 1663 unsigned long ioaddr = dev->base_addr;
1598 u16 val; 1664 u16 val;
1599 int i; 1665 int i;
1600 int rc; 1666 int rc;
1601 unsigned long flags; 1667 unsigned long flags;
1602 1668
1603 if (request_irq(dev->irq, &pcnet32_interrupt, 1669 if (request_irq(dev->irq, &pcnet32_interrupt,
1604 lp->shared_irq ? SA_SHIRQ : 0, dev->name, (void *)dev)) { 1670 lp->shared_irq ? SA_SHIRQ : 0, dev->name,
1605 return -EAGAIN; 1671 (void *)dev)) {
1606 } 1672 return -EAGAIN;
1607 1673 }
1608 spin_lock_irqsave(&lp->lock, flags); 1674
1609 /* Check for a valid station address */ 1675 spin_lock_irqsave(&lp->lock, flags);
1610 if (!is_valid_ether_addr(dev->dev_addr)) { 1676 /* Check for a valid station address */
1611 rc = -EINVAL; 1677 if (!is_valid_ether_addr(dev->dev_addr)) {
1612 goto err_free_irq; 1678 rc = -EINVAL;
1613 } 1679 goto err_free_irq;
1614 1680 }
1615 /* Reset the PCNET32 */ 1681
1616 lp->a.reset (ioaddr); 1682 /* Reset the PCNET32 */
1617 1683 lp->a.reset(ioaddr);
1618 /* switch pcnet32 to 32bit mode */ 1684
1619 lp->a.write_bcr (ioaddr, 20, 2); 1685 /* switch pcnet32 to 32bit mode */
1620 1686 lp->a.write_bcr(ioaddr, 20, 2);
1621 if (netif_msg_ifup(lp)) 1687
1622 printk(KERN_DEBUG "%s: pcnet32_open() irq %d tx/rx rings %#x/%#x init %#x.\n", 1688 if (netif_msg_ifup(lp))
1623 dev->name, dev->irq, 1689 printk(KERN_DEBUG
1624 (u32) (lp->tx_ring_dma_addr), 1690 "%s: pcnet32_open() irq %d tx/rx rings %#x/%#x init %#x.\n",
1625 (u32) (lp->rx_ring_dma_addr), 1691 dev->name, dev->irq, (u32) (lp->tx_ring_dma_addr),
1626 (u32) (lp->dma_addr + offsetof(struct pcnet32_private, init_block))); 1692 (u32) (lp->rx_ring_dma_addr),
1627 1693 (u32) (lp->dma_addr +
1628 /* set/reset autoselect bit */ 1694 offsetof(struct pcnet32_private, init_block)));
1629 val = lp->a.read_bcr (ioaddr, 2) & ~2; 1695
1630 if (lp->options & PCNET32_PORT_ASEL) 1696 /* set/reset autoselect bit */
1631 val |= 2; 1697 val = lp->a.read_bcr(ioaddr, 2) & ~2;
1632 lp->a.write_bcr (ioaddr, 2, val); 1698 if (lp->options & PCNET32_PORT_ASEL)
1633
1634 /* handle full duplex setting */
1635 if (lp->mii_if.full_duplex) {
1636 val = lp->a.read_bcr (ioaddr, 9) & ~3;
1637 if (lp->options & PCNET32_PORT_FD) {
1638 val |= 1;
1639 if (lp->options == (PCNET32_PORT_FD | PCNET32_PORT_AUI))
1640 val |= 2; 1699 val |= 2;
1641 } else if (lp->options & PCNET32_PORT_ASEL) { 1700 lp->a.write_bcr(ioaddr, 2, val);
1642 /* workaround of xSeries250, turn on for 79C975 only */ 1701
1643 i = ((lp->a.read_csr(ioaddr, 88) | 1702 /* handle full duplex setting */
1644 (lp->a.read_csr(ioaddr,89) << 16)) >> 12) & 0xffff; 1703 if (lp->mii_if.full_duplex) {
1645 if (i == 0x2627) 1704 val = lp->a.read_bcr(ioaddr, 9) & ~3;
1646 val |= 3; 1705 if (lp->options & PCNET32_PORT_FD) {
1647 } 1706 val |= 1;
1648 lp->a.write_bcr (ioaddr, 9, val); 1707 if (lp->options == (PCNET32_PORT_FD | PCNET32_PORT_AUI))
1649 } 1708 val |= 2;
1650 1709 } else if (lp->options & PCNET32_PORT_ASEL) {
1651 /* set/reset GPSI bit in test register */ 1710 /* workaround of xSeries250, turn on for 79C975 only */
1652 val = lp->a.read_csr (ioaddr, 124) & ~0x10; 1711 i = ((lp->a.read_csr(ioaddr, 88) |
1653 if ((lp->options & PCNET32_PORT_PORTSEL) == PCNET32_PORT_GPSI) 1712 (lp->a.
1654 val |= 0x10; 1713 read_csr(ioaddr, 89) << 16)) >> 12) & 0xffff;
1655 lp->a.write_csr (ioaddr, 124, val); 1714 if (i == 0x2627)
1656 1715 val |= 3;
1657 /* Allied Telesyn AT 2700/2701 FX are 100Mbit only and do not negotiate */ 1716 }
1658 if (lp->pci_dev->subsystem_vendor == PCI_VENDOR_ID_AT && 1717 lp->a.write_bcr(ioaddr, 9, val);
1718 }
1719
1720 /* set/reset GPSI bit in test register */
1721 val = lp->a.read_csr(ioaddr, 124) & ~0x10;
1722 if ((lp->options & PCNET32_PORT_PORTSEL) == PCNET32_PORT_GPSI)
1723 val |= 0x10;
1724 lp->a.write_csr(ioaddr, 124, val);
1725
1726 /* Allied Telesyn AT 2700/2701 FX are 100Mbit only and do not negotiate */
1727 if (lp->pci_dev->subsystem_vendor == PCI_VENDOR_ID_AT &&
1659 (lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2700FX || 1728 (lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2700FX ||
1660 lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2701FX)) { 1729 lp->pci_dev->subsystem_device == PCI_SUBDEVICE_ID_AT_2701FX)) {
1661 if (lp->options & PCNET32_PORT_ASEL) {
1662 lp->options = PCNET32_PORT_FD | PCNET32_PORT_100;
1663 if (netif_msg_link(lp))
1664 printk(KERN_DEBUG "%s: Setting 100Mb-Full Duplex.\n",
1665 dev->name);
1666 }
1667 }
1668 if (lp->phycount < 2) {
1669 /*
1670 * 24 Jun 2004 according AMD, in order to change the PHY,
1671 * DANAS (or DISPM for 79C976) must be set; then select the speed,
1672 * duplex, and/or enable auto negotiation, and clear DANAS
1673 */
1674 if (lp->mii && !(lp->options & PCNET32_PORT_ASEL)) {
1675 lp->a.write_bcr(ioaddr, 32,
1676 lp->a.read_bcr(ioaddr, 32) | 0x0080);
1677 /* disable Auto Negotiation, set 10Mpbs, HD */
1678 val = lp->a.read_bcr(ioaddr, 32) & ~0xb8;
1679 if (lp->options & PCNET32_PORT_FD)
1680 val |= 0x10;
1681 if (lp->options & PCNET32_PORT_100)
1682 val |= 0x08;
1683 lp->a.write_bcr (ioaddr, 32, val);
1684 } else {
1685 if (lp->options & PCNET32_PORT_ASEL) {
1686 lp->a.write_bcr(ioaddr, 32,
1687 lp->a.read_bcr(ioaddr, 32) | 0x0080);
1688 /* enable auto negotiate, setup, disable fd */
1689 val = lp->a.read_bcr(ioaddr, 32) & ~0x98;
1690 val |= 0x20;
1691 lp->a.write_bcr(ioaddr, 32, val);
1692 }
1693 }
1694 } else {
1695 int first_phy = -1;
1696 u16 bmcr;
1697 u32 bcr9;
1698 struct ethtool_cmd ecmd;
1699
1700 /*
1701 * There is really no good other way to handle multiple PHYs
1702 * other than turning off all automatics
1703 */
1704 val = lp->a.read_bcr(ioaddr, 2);
1705 lp->a.write_bcr(ioaddr, 2, val & ~2);
1706 val = lp->a.read_bcr(ioaddr, 32);
1707 lp->a.write_bcr(ioaddr, 32, val & ~(1 << 7)); /* stop MII manager */
1708
1709 if (!(lp->options & PCNET32_PORT_ASEL)) {
1710 /* setup ecmd */
1711 ecmd.port = PORT_MII;
1712 ecmd.transceiver = XCVR_INTERNAL;
1713 ecmd.autoneg = AUTONEG_DISABLE;
1714 ecmd.speed = lp->options & PCNET32_PORT_100 ? SPEED_100 : SPEED_10;
1715 bcr9 = lp->a.read_bcr(ioaddr, 9);
1716
1717 if (lp->options & PCNET32_PORT_FD) {
1718 ecmd.duplex = DUPLEX_FULL;
1719 bcr9 |= (1 << 0);
1720 } else {
1721 ecmd.duplex = DUPLEX_HALF;
1722 bcr9 |= ~(1 << 0);
1723 }
1724 lp->a.write_bcr(ioaddr, 9, bcr9);
1725 }
1726
1727 for (i=0; i<PCNET32_MAX_PHYS; i++) {
1728 if (lp->phymask & (1 << i)) {
1729 /* isolate all but the first PHY */
1730 bmcr = mdio_read(dev, i, MII_BMCR);
1731 if (first_phy == -1) {
1732 first_phy = i;
1733 mdio_write(dev, i, MII_BMCR, bmcr & ~BMCR_ISOLATE);
1734 } else {
1735 mdio_write(dev, i, MII_BMCR, bmcr | BMCR_ISOLATE);
1736 }
1737 /* use mii_ethtool_sset to setup PHY */
1738 lp->mii_if.phy_id = i;
1739 ecmd.phy_address = i;
1740 if (lp->options & PCNET32_PORT_ASEL) { 1730 if (lp->options & PCNET32_PORT_ASEL) {
1741 mii_ethtool_gset(&lp->mii_if, &ecmd); 1731 lp->options = PCNET32_PORT_FD | PCNET32_PORT_100;
1742 ecmd.autoneg = AUTONEG_ENABLE; 1732 if (netif_msg_link(lp))
1733 printk(KERN_DEBUG
1734 "%s: Setting 100Mb-Full Duplex.\n",
1735 dev->name);
1736 }
1737 }
1738 if (lp->phycount < 2) {
1739 /*
1740 * 24 Jun 2004 according AMD, in order to change the PHY,
1741 * DANAS (or DISPM for 79C976) must be set; then select the speed,
1742 * duplex, and/or enable auto negotiation, and clear DANAS
1743 */
1744 if (lp->mii && !(lp->options & PCNET32_PORT_ASEL)) {
1745 lp->a.write_bcr(ioaddr, 32,
1746 lp->a.read_bcr(ioaddr, 32) | 0x0080);
1747 /* disable Auto Negotiation, set 10Mpbs, HD */
1748 val = lp->a.read_bcr(ioaddr, 32) & ~0xb8;
1749 if (lp->options & PCNET32_PORT_FD)
1750 val |= 0x10;
1751 if (lp->options & PCNET32_PORT_100)
1752 val |= 0x08;
1753 lp->a.write_bcr(ioaddr, 32, val);
1754 } else {
1755 if (lp->options & PCNET32_PORT_ASEL) {
1756 lp->a.write_bcr(ioaddr, 32,
1757 lp->a.read_bcr(ioaddr,
1758 32) | 0x0080);
1759 /* enable auto negotiate, setup, disable fd */
1760 val = lp->a.read_bcr(ioaddr, 32) & ~0x98;
1761 val |= 0x20;
1762 lp->a.write_bcr(ioaddr, 32, val);
1763 }
1764 }
1765 } else {
1766 int first_phy = -1;
1767 u16 bmcr;
1768 u32 bcr9;
1769 struct ethtool_cmd ecmd;
1770
1771 /*
1772 * There is really no good other way to handle multiple PHYs
1773 * other than turning off all automatics
1774 */
1775 val = lp->a.read_bcr(ioaddr, 2);
1776 lp->a.write_bcr(ioaddr, 2, val & ~2);
1777 val = lp->a.read_bcr(ioaddr, 32);
1778 lp->a.write_bcr(ioaddr, 32, val & ~(1 << 7)); /* stop MII manager */
1779
1780 if (!(lp->options & PCNET32_PORT_ASEL)) {
1781 /* setup ecmd */
1782 ecmd.port = PORT_MII;
1783 ecmd.transceiver = XCVR_INTERNAL;
1784 ecmd.autoneg = AUTONEG_DISABLE;
1785 ecmd.speed =
1786 lp->
1787 options & PCNET32_PORT_100 ? SPEED_100 : SPEED_10;
1788 bcr9 = lp->a.read_bcr(ioaddr, 9);
1789
1790 if (lp->options & PCNET32_PORT_FD) {
1791 ecmd.duplex = DUPLEX_FULL;
1792 bcr9 |= (1 << 0);
1793 } else {
1794 ecmd.duplex = DUPLEX_HALF;
1795 bcr9 |= ~(1 << 0);
1796 }
1797 lp->a.write_bcr(ioaddr, 9, bcr9);
1743 } 1798 }
1744 mii_ethtool_sset(&lp->mii_if, &ecmd); 1799
1745 } 1800 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
1746 } 1801 if (lp->phymask & (1 << i)) {
1747 lp->mii_if.phy_id = first_phy; 1802 /* isolate all but the first PHY */
1748 if (netif_msg_link(lp)) 1803 bmcr = mdio_read(dev, i, MII_BMCR);
1749 printk(KERN_INFO "%s: Using PHY number %d.\n", dev->name, first_phy); 1804 if (first_phy == -1) {
1750 } 1805 first_phy = i;
1806 mdio_write(dev, i, MII_BMCR,
1807 bmcr & ~BMCR_ISOLATE);
1808 } else {
1809 mdio_write(dev, i, MII_BMCR,
1810 bmcr | BMCR_ISOLATE);
1811 }
1812 /* use mii_ethtool_sset to setup PHY */
1813 lp->mii_if.phy_id = i;
1814 ecmd.phy_address = i;
1815 if (lp->options & PCNET32_PORT_ASEL) {
1816 mii_ethtool_gset(&lp->mii_if, &ecmd);
1817 ecmd.autoneg = AUTONEG_ENABLE;
1818 }
1819 mii_ethtool_sset(&lp->mii_if, &ecmd);
1820 }
1821 }
1822 lp->mii_if.phy_id = first_phy;
1823 if (netif_msg_link(lp))
1824 printk(KERN_INFO "%s: Using PHY number %d.\n",
1825 dev->name, first_phy);
1826 }
1751 1827
1752#ifdef DO_DXSUFLO 1828#ifdef DO_DXSUFLO
1753 if (lp->dxsuflo) { /* Disable transmit stop on underflow */ 1829 if (lp->dxsuflo) { /* Disable transmit stop on underflow */
1754 val = lp->a.read_csr (ioaddr, 3); 1830 val = lp->a.read_csr(ioaddr, 3);
1755 val |= 0x40; 1831 val |= 0x40;
1756 lp->a.write_csr (ioaddr, 3, val); 1832 lp->a.write_csr(ioaddr, 3, val);
1757 } 1833 }
1758#endif 1834#endif
1759 1835
1760 lp->init_block.mode = le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7); 1836 lp->init_block.mode =
1761 pcnet32_load_multicast(dev); 1837 le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7);
1762 1838 pcnet32_load_multicast(dev);
1763 if (pcnet32_init_ring(dev)) { 1839
1764 rc = -ENOMEM; 1840 if (pcnet32_init_ring(dev)) {
1765 goto err_free_ring; 1841 rc = -ENOMEM;
1766 } 1842 goto err_free_ring;
1767 1843 }
1768 /* Re-initialize the PCNET32, and start it when done. */ 1844
1769 lp->a.write_csr (ioaddr, 1, (lp->dma_addr + 1845 /* Re-initialize the PCNET32, and start it when done. */
1770 offsetof(struct pcnet32_private, init_block)) & 0xffff); 1846 lp->a.write_csr(ioaddr, 1, (lp->dma_addr +
1771 lp->a.write_csr (ioaddr, 2, (lp->dma_addr + 1847 offsetof(struct pcnet32_private,
1772 offsetof(struct pcnet32_private, init_block)) >> 16); 1848 init_block)) & 0xffff);
1773 1849 lp->a.write_csr(ioaddr, 2,
1774 lp->a.write_csr (ioaddr, 4, 0x0915); 1850 (lp->dma_addr +
1775 lp->a.write_csr (ioaddr, 0, 0x0001); 1851 offsetof(struct pcnet32_private, init_block)) >> 16);
1776 1852
1777 netif_start_queue(dev); 1853 lp->a.write_csr(ioaddr, 4, 0x0915);
1778 1854 lp->a.write_csr(ioaddr, 0, 0x0001);
1779 /* Print the link status and start the watchdog */ 1855
1780 pcnet32_check_media (dev, 1); 1856 netif_start_queue(dev);
1781 mod_timer (&(lp->watchdog_timer), PCNET32_WATCHDOG_TIMEOUT); 1857
1782 1858 /* Print the link status and start the watchdog */
1783 i = 0; 1859 pcnet32_check_media(dev, 1);
1784 while (i++ < 100) 1860 mod_timer(&(lp->watchdog_timer), PCNET32_WATCHDOG_TIMEOUT);
1785 if (lp->a.read_csr (ioaddr, 0) & 0x0100) 1861
1786 break; 1862 i = 0;
1787 /* 1863 while (i++ < 100)
1788 * We used to clear the InitDone bit, 0x0100, here but Mark Stockton 1864 if (lp->a.read_csr(ioaddr, 0) & 0x0100)
1789 * reports that doing so triggers a bug in the '974. 1865 break;
1790 */ 1866 /*
1791 lp->a.write_csr (ioaddr, 0, 0x0042); 1867 * We used to clear the InitDone bit, 0x0100, here but Mark Stockton
1792 1868 * reports that doing so triggers a bug in the '974.
1793 if (netif_msg_ifup(lp)) 1869 */
1794 printk(KERN_DEBUG "%s: pcnet32 open after %d ticks, init block %#x csr0 %4.4x.\n", 1870 lp->a.write_csr(ioaddr, 0, 0x0042);
1795 dev->name, i, (u32) (lp->dma_addr + 1871
1796 offsetof(struct pcnet32_private, init_block)), 1872 if (netif_msg_ifup(lp))
1797 lp->a.read_csr(ioaddr, 0)); 1873 printk(KERN_DEBUG
1798 1874 "%s: pcnet32 open after %d ticks, init block %#x csr0 %4.4x.\n",
1799 spin_unlock_irqrestore(&lp->lock, flags); 1875 dev->name, i,
1800 1876 (u32) (lp->dma_addr +
1801 return 0; /* Always succeed */ 1877 offsetof(struct pcnet32_private, init_block)),
1802 1878 lp->a.read_csr(ioaddr, 0));
1803err_free_ring: 1879
1804 /* free any allocated skbuffs */ 1880 spin_unlock_irqrestore(&lp->lock, flags);
1805 for (i = 0; i < lp->rx_ring_size; i++) { 1881
1806 lp->rx_ring[i].status = 0; 1882 return 0; /* Always succeed */
1807 if (lp->rx_skbuff[i]) { 1883
1808 pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[i], PKT_BUF_SZ-2, 1884 err_free_ring:
1809 PCI_DMA_FROMDEVICE); 1885 /* free any allocated skbuffs */
1810 dev_kfree_skb(lp->rx_skbuff[i]); 1886 for (i = 0; i < lp->rx_ring_size; i++) {
1811 } 1887 lp->rx_ring[i].status = 0;
1812 lp->rx_skbuff[i] = NULL; 1888 if (lp->rx_skbuff[i]) {
1813 lp->rx_dma_addr[i] = 0; 1889 pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[i],
1814 } 1890 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
1815 1891 dev_kfree_skb(lp->rx_skbuff[i]);
1816 pcnet32_free_ring(dev); 1892 }
1817 1893 lp->rx_skbuff[i] = NULL;
1818 /* 1894 lp->rx_dma_addr[i] = 0;
1819 * Switch back to 16bit mode to avoid problems with dumb 1895 }
1820 * DOS packet driver after a warm reboot 1896
1821 */ 1897 pcnet32_free_ring(dev);
1822 lp->a.write_bcr (ioaddr, 20, 4); 1898
1823 1899 /*
1824err_free_irq: 1900 * Switch back to 16bit mode to avoid problems with dumb
1825 spin_unlock_irqrestore(&lp->lock, flags); 1901 * DOS packet driver after a warm reboot
1826 free_irq(dev->irq, dev); 1902 */
1827 return rc; 1903 lp->a.write_bcr(ioaddr, 20, 4);
1904
1905 err_free_irq:
1906 spin_unlock_irqrestore(&lp->lock, flags);
1907 free_irq(dev->irq, dev);
1908 return rc;
1828} 1909}
1829 1910
1830/* 1911/*
@@ -1840,722 +1921,792 @@ err_free_irq:
1840 * restarting the chip, but I'm too lazy to do so right now. dplatt@3do.com 1921 * restarting the chip, but I'm too lazy to do so right now. dplatt@3do.com
1841 */ 1922 */
1842 1923
1843static void 1924static void pcnet32_purge_tx_ring(struct net_device *dev)
1844pcnet32_purge_tx_ring(struct net_device *dev)
1845{ 1925{
1846 struct pcnet32_private *lp = dev->priv; 1926 struct pcnet32_private *lp = dev->priv;
1847 int i; 1927 int i;
1848
1849 for (i = 0; i < lp->tx_ring_size; i++) {
1850 lp->tx_ring[i].status = 0; /* CPU owns buffer */
1851 wmb(); /* Make sure adapter sees owner change */
1852 if (lp->tx_skbuff[i]) {
1853 pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[i],
1854 lp->tx_skbuff[i]->len, PCI_DMA_TODEVICE);
1855 dev_kfree_skb_any(lp->tx_skbuff[i]);
1856 }
1857 lp->tx_skbuff[i] = NULL;
1858 lp->tx_dma_addr[i] = 0;
1859 }
1860}
1861 1928
1929 for (i = 0; i < lp->tx_ring_size; i++) {
1930 lp->tx_ring[i].status = 0; /* CPU owns buffer */
1931 wmb(); /* Make sure adapter sees owner change */
1932 if (lp->tx_skbuff[i]) {
1933 pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[i],
1934 lp->tx_skbuff[i]->len,
1935 PCI_DMA_TODEVICE);
1936 dev_kfree_skb_any(lp->tx_skbuff[i]);
1937 }
1938 lp->tx_skbuff[i] = NULL;
1939 lp->tx_dma_addr[i] = 0;
1940 }
1941}
1862 1942
1863/* Initialize the PCNET32 Rx and Tx rings. */ 1943/* Initialize the PCNET32 Rx and Tx rings. */
1864static int 1944static int pcnet32_init_ring(struct net_device *dev)
1865pcnet32_init_ring(struct net_device *dev)
1866{ 1945{
1867 struct pcnet32_private *lp = dev->priv; 1946 struct pcnet32_private *lp = dev->priv;
1868 int i; 1947 int i;
1869 1948
1870 lp->tx_full = 0; 1949 lp->tx_full = 0;
1871 lp->cur_rx = lp->cur_tx = 0; 1950 lp->cur_rx = lp->cur_tx = 0;
1872 lp->dirty_rx = lp->dirty_tx = 0; 1951 lp->dirty_rx = lp->dirty_tx = 0;
1873 1952
1874 for (i = 0; i < lp->rx_ring_size; i++) { 1953 for (i = 0; i < lp->rx_ring_size; i++) {
1875 struct sk_buff *rx_skbuff = lp->rx_skbuff[i]; 1954 struct sk_buff *rx_skbuff = lp->rx_skbuff[i];
1876 if (rx_skbuff == NULL) { 1955 if (rx_skbuff == NULL) {
1877 if (!(rx_skbuff = lp->rx_skbuff[i] = dev_alloc_skb (PKT_BUF_SZ))) { 1956 if (!
1878 /* there is not much, we can do at this point */ 1957 (rx_skbuff = lp->rx_skbuff[i] =
1879 if (pcnet32_debug & NETIF_MSG_DRV) 1958 dev_alloc_skb(PKT_BUF_SZ))) {
1880 printk(KERN_ERR "%s: pcnet32_init_ring dev_alloc_skb failed.\n", 1959 /* there is not much, we can do at this point */
1881 dev->name); 1960 if (pcnet32_debug & NETIF_MSG_DRV)
1882 return -1; 1961 printk(KERN_ERR
1883 } 1962 "%s: pcnet32_init_ring dev_alloc_skb failed.\n",
1884 skb_reserve (rx_skbuff, 2); 1963 dev->name);
1885 } 1964 return -1;
1886 1965 }
1887 rmb(); 1966 skb_reserve(rx_skbuff, 2);
1888 if (lp->rx_dma_addr[i] == 0) 1967 }
1889 lp->rx_dma_addr[i] = pci_map_single(lp->pci_dev, rx_skbuff->data, 1968
1890 PKT_BUF_SZ-2, PCI_DMA_FROMDEVICE); 1969 rmb();
1891 lp->rx_ring[i].base = (u32)le32_to_cpu(lp->rx_dma_addr[i]); 1970 if (lp->rx_dma_addr[i] == 0)
1892 lp->rx_ring[i].buf_length = le16_to_cpu(2-PKT_BUF_SZ); 1971 lp->rx_dma_addr[i] =
1893 wmb(); /* Make sure owner changes after all others are visible */ 1972 pci_map_single(lp->pci_dev, rx_skbuff->data,
1894 lp->rx_ring[i].status = le16_to_cpu(0x8000); 1973 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
1895 } 1974 lp->rx_ring[i].base = (u32) le32_to_cpu(lp->rx_dma_addr[i]);
1896 /* The Tx buffer address is filled in as needed, but we do need to clear 1975 lp->rx_ring[i].buf_length = le16_to_cpu(2 - PKT_BUF_SZ);
1897 * the upper ownership bit. */ 1976 wmb(); /* Make sure owner changes after all others are visible */
1898 for (i = 0; i < lp->tx_ring_size; i++) { 1977 lp->rx_ring[i].status = le16_to_cpu(0x8000);
1899 lp->tx_ring[i].status = 0; /* CPU owns buffer */ 1978 }
1900 wmb(); /* Make sure adapter sees owner change */ 1979 /* The Tx buffer address is filled in as needed, but we do need to clear
1901 lp->tx_ring[i].base = 0; 1980 * the upper ownership bit. */
1902 lp->tx_dma_addr[i] = 0; 1981 for (i = 0; i < lp->tx_ring_size; i++) {
1903 } 1982 lp->tx_ring[i].status = 0; /* CPU owns buffer */
1904 1983 wmb(); /* Make sure adapter sees owner change */
1905 lp->init_block.tlen_rlen = le16_to_cpu(lp->tx_len_bits | lp->rx_len_bits); 1984 lp->tx_ring[i].base = 0;
1906 for (i = 0; i < 6; i++) 1985 lp->tx_dma_addr[i] = 0;
1907 lp->init_block.phys_addr[i] = dev->dev_addr[i]; 1986 }
1908 lp->init_block.rx_ring = (u32)le32_to_cpu(lp->rx_ring_dma_addr); 1987
1909 lp->init_block.tx_ring = (u32)le32_to_cpu(lp->tx_ring_dma_addr); 1988 lp->init_block.tlen_rlen =
1910 wmb(); /* Make sure all changes are visible */ 1989 le16_to_cpu(lp->tx_len_bits | lp->rx_len_bits);
1911 return 0; 1990 for (i = 0; i < 6; i++)
1991 lp->init_block.phys_addr[i] = dev->dev_addr[i];
1992 lp->init_block.rx_ring = (u32) le32_to_cpu(lp->rx_ring_dma_addr);
1993 lp->init_block.tx_ring = (u32) le32_to_cpu(lp->tx_ring_dma_addr);
1994 wmb(); /* Make sure all changes are visible */
1995 return 0;
1912} 1996}
1913 1997
1914/* the pcnet32 has been issued a stop or reset. Wait for the stop bit 1998/* the pcnet32 has been issued a stop or reset. Wait for the stop bit
1915 * then flush the pending transmit operations, re-initialize the ring, 1999 * then flush the pending transmit operations, re-initialize the ring,
1916 * and tell the chip to initialize. 2000 * and tell the chip to initialize.
1917 */ 2001 */
1918static void 2002static void pcnet32_restart(struct net_device *dev, unsigned int csr0_bits)
1919pcnet32_restart(struct net_device *dev, unsigned int csr0_bits)
1920{ 2003{
1921 struct pcnet32_private *lp = dev->priv; 2004 struct pcnet32_private *lp = dev->priv;
1922 unsigned long ioaddr = dev->base_addr; 2005 unsigned long ioaddr = dev->base_addr;
1923 int i; 2006 int i;
1924 2007
1925 /* wait for stop */ 2008 /* wait for stop */
1926 for (i=0; i<100; i++) 2009 for (i = 0; i < 100; i++)
1927 if (lp->a.read_csr(ioaddr, 0) & 0x0004) 2010 if (lp->a.read_csr(ioaddr, 0) & 0x0004)
1928 break; 2011 break;
1929 2012
1930 if (i >= 100 && netif_msg_drv(lp)) 2013 if (i >= 100 && netif_msg_drv(lp))
1931 printk(KERN_ERR "%s: pcnet32_restart timed out waiting for stop.\n", 2014 printk(KERN_ERR
1932 dev->name); 2015 "%s: pcnet32_restart timed out waiting for stop.\n",
2016 dev->name);
1933 2017
1934 pcnet32_purge_tx_ring(dev); 2018 pcnet32_purge_tx_ring(dev);
1935 if (pcnet32_init_ring(dev)) 2019 if (pcnet32_init_ring(dev))
1936 return; 2020 return;
1937 2021
1938 /* ReInit Ring */ 2022 /* ReInit Ring */
1939 lp->a.write_csr (ioaddr, 0, 1); 2023 lp->a.write_csr(ioaddr, 0, 1);
1940 i = 0; 2024 i = 0;
1941 while (i++ < 1000) 2025 while (i++ < 1000)
1942 if (lp->a.read_csr (ioaddr, 0) & 0x0100) 2026 if (lp->a.read_csr(ioaddr, 0) & 0x0100)
1943 break; 2027 break;
1944 2028
1945 lp->a.write_csr (ioaddr, 0, csr0_bits); 2029 lp->a.write_csr(ioaddr, 0, csr0_bits);
1946} 2030}
1947 2031
1948 2032static void pcnet32_tx_timeout(struct net_device *dev)
1949static void
1950pcnet32_tx_timeout (struct net_device *dev)
1951{ 2033{
1952 struct pcnet32_private *lp = dev->priv; 2034 struct pcnet32_private *lp = dev->priv;
1953 unsigned long ioaddr = dev->base_addr, flags; 2035 unsigned long ioaddr = dev->base_addr, flags;
1954 2036
1955 spin_lock_irqsave(&lp->lock, flags); 2037 spin_lock_irqsave(&lp->lock, flags);
1956 /* Transmitter timeout, serious problems. */ 2038 /* Transmitter timeout, serious problems. */
1957 if (pcnet32_debug & NETIF_MSG_DRV) 2039 if (pcnet32_debug & NETIF_MSG_DRV)
1958 printk(KERN_ERR "%s: transmit timed out, status %4.4x, resetting.\n", 2040 printk(KERN_ERR
1959 dev->name, lp->a.read_csr(ioaddr, 0)); 2041 "%s: transmit timed out, status %4.4x, resetting.\n",
1960 lp->a.write_csr (ioaddr, 0, 0x0004); 2042 dev->name, lp->a.read_csr(ioaddr, 0));
1961 lp->stats.tx_errors++; 2043 lp->a.write_csr(ioaddr, 0, 0x0004);
1962 if (netif_msg_tx_err(lp)) { 2044 lp->stats.tx_errors++;
1963 int i; 2045 if (netif_msg_tx_err(lp)) {
1964 printk(KERN_DEBUG " Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.", 2046 int i;
1965 lp->dirty_tx, lp->cur_tx, lp->tx_full ? " (full)" : "", 2047 printk(KERN_DEBUG
1966 lp->cur_rx); 2048 " Ring data dump: dirty_tx %d cur_tx %d%s cur_rx %d.",
1967 for (i = 0 ; i < lp->rx_ring_size; i++) 2049 lp->dirty_tx, lp->cur_tx, lp->tx_full ? " (full)" : "",
1968 printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ", 2050 lp->cur_rx);
1969 le32_to_cpu(lp->rx_ring[i].base), 2051 for (i = 0; i < lp->rx_ring_size; i++)
1970 (-le16_to_cpu(lp->rx_ring[i].buf_length)) & 0xffff, 2052 printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ",
1971 le32_to_cpu(lp->rx_ring[i].msg_length), 2053 le32_to_cpu(lp->rx_ring[i].base),
1972 le16_to_cpu(lp->rx_ring[i].status)); 2054 (-le16_to_cpu(lp->rx_ring[i].buf_length)) &
1973 for (i = 0 ; i < lp->tx_ring_size; i++) 2055 0xffff, le32_to_cpu(lp->rx_ring[i].msg_length),
1974 printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ", 2056 le16_to_cpu(lp->rx_ring[i].status));
1975 le32_to_cpu(lp->tx_ring[i].base), 2057 for (i = 0; i < lp->tx_ring_size; i++)
1976 (-le16_to_cpu(lp->tx_ring[i].length)) & 0xffff, 2058 printk("%s %08x %04x %08x %04x", i & 1 ? "" : "\n ",
1977 le32_to_cpu(lp->tx_ring[i].misc), 2059 le32_to_cpu(lp->tx_ring[i].base),
1978 le16_to_cpu(lp->tx_ring[i].status)); 2060 (-le16_to_cpu(lp->tx_ring[i].length)) & 0xffff,
1979 printk("\n"); 2061 le32_to_cpu(lp->tx_ring[i].misc),
1980 } 2062 le16_to_cpu(lp->tx_ring[i].status));
1981 pcnet32_restart(dev, 0x0042); 2063 printk("\n");
1982 2064 }
1983 dev->trans_start = jiffies; 2065 pcnet32_restart(dev, 0x0042);
1984 netif_wake_queue(dev);
1985
1986 spin_unlock_irqrestore(&lp->lock, flags);
1987}
1988 2066
2067 dev->trans_start = jiffies;
2068 netif_wake_queue(dev);
1989 2069
1990static int 2070 spin_unlock_irqrestore(&lp->lock, flags);
1991pcnet32_start_xmit(struct sk_buff *skb, struct net_device *dev) 2071}
2072
2073static int pcnet32_start_xmit(struct sk_buff *skb, struct net_device *dev)
1992{ 2074{
1993 struct pcnet32_private *lp = dev->priv; 2075 struct pcnet32_private *lp = dev->priv;
1994 unsigned long ioaddr = dev->base_addr; 2076 unsigned long ioaddr = dev->base_addr;
1995 u16 status; 2077 u16 status;
1996 int entry; 2078 int entry;
1997 unsigned long flags; 2079 unsigned long flags;
1998 2080
1999 spin_lock_irqsave(&lp->lock, flags); 2081 spin_lock_irqsave(&lp->lock, flags);
2000 2082
2001 if (netif_msg_tx_queued(lp)) { 2083 if (netif_msg_tx_queued(lp)) {
2002 printk(KERN_DEBUG "%s: pcnet32_start_xmit() called, csr0 %4.4x.\n", 2084 printk(KERN_DEBUG
2003 dev->name, lp->a.read_csr(ioaddr, 0)); 2085 "%s: pcnet32_start_xmit() called, csr0 %4.4x.\n",
2004 } 2086 dev->name, lp->a.read_csr(ioaddr, 0));
2087 }
2005 2088
2006 /* Default status -- will not enable Successful-TxDone 2089 /* Default status -- will not enable Successful-TxDone
2007 * interrupt when that option is available to us. 2090 * interrupt when that option is available to us.
2008 */ 2091 */
2009 status = 0x8300; 2092 status = 0x8300;
2010 2093
2011 /* Fill in a Tx ring entry */ 2094 /* Fill in a Tx ring entry */
2012 2095
2013 /* Mask to ring buffer boundary. */ 2096 /* Mask to ring buffer boundary. */
2014 entry = lp->cur_tx & lp->tx_mod_mask; 2097 entry = lp->cur_tx & lp->tx_mod_mask;
2015 2098
2016 /* Caution: the write order is important here, set the status 2099 /* Caution: the write order is important here, set the status
2017 * with the "ownership" bits last. */ 2100 * with the "ownership" bits last. */
2018 2101
2019 lp->tx_ring[entry].length = le16_to_cpu(-skb->len); 2102 lp->tx_ring[entry].length = le16_to_cpu(-skb->len);
2020 2103
2021 lp->tx_ring[entry].misc = 0x00000000; 2104 lp->tx_ring[entry].misc = 0x00000000;
2022 2105
2023 lp->tx_skbuff[entry] = skb; 2106 lp->tx_skbuff[entry] = skb;
2024 lp->tx_dma_addr[entry] = pci_map_single(lp->pci_dev, skb->data, skb->len, 2107 lp->tx_dma_addr[entry] =
2025 PCI_DMA_TODEVICE); 2108 pci_map_single(lp->pci_dev, skb->data, skb->len, PCI_DMA_TODEVICE);
2026 lp->tx_ring[entry].base = (u32)le32_to_cpu(lp->tx_dma_addr[entry]); 2109 lp->tx_ring[entry].base = (u32) le32_to_cpu(lp->tx_dma_addr[entry]);
2027 wmb(); /* Make sure owner changes after all others are visible */ 2110 wmb(); /* Make sure owner changes after all others are visible */
2028 lp->tx_ring[entry].status = le16_to_cpu(status); 2111 lp->tx_ring[entry].status = le16_to_cpu(status);
2029 2112
2030 lp->cur_tx++; 2113 lp->cur_tx++;
2031 lp->stats.tx_bytes += skb->len; 2114 lp->stats.tx_bytes += skb->len;
2032 2115
2033 /* Trigger an immediate send poll. */ 2116 /* Trigger an immediate send poll. */
2034 lp->a.write_csr (ioaddr, 0, 0x0048); 2117 lp->a.write_csr(ioaddr, 0, 0x0048);
2035 2118
2036 dev->trans_start = jiffies; 2119 dev->trans_start = jiffies;
2037 2120
2038 if (lp->tx_ring[(entry+1) & lp->tx_mod_mask].base != 0) { 2121 if (lp->tx_ring[(entry + 1) & lp->tx_mod_mask].base != 0) {
2039 lp->tx_full = 1; 2122 lp->tx_full = 1;
2040 netif_stop_queue(dev); 2123 netif_stop_queue(dev);
2041 } 2124 }
2042 spin_unlock_irqrestore(&lp->lock, flags); 2125 spin_unlock_irqrestore(&lp->lock, flags);
2043 return 0; 2126 return 0;
2044} 2127}
2045 2128
2046/* The PCNET32 interrupt handler. */ 2129/* The PCNET32 interrupt handler. */
2047static irqreturn_t 2130static irqreturn_t
2048pcnet32_interrupt(int irq, void *dev_id, struct pt_regs * regs) 2131pcnet32_interrupt(int irq, void *dev_id, struct pt_regs *regs)
2049{ 2132{
2050 struct net_device *dev = dev_id; 2133 struct net_device *dev = dev_id;
2051 struct pcnet32_private *lp; 2134 struct pcnet32_private *lp;
2052 unsigned long ioaddr; 2135 unsigned long ioaddr;
2053 u16 csr0,rap; 2136 u16 csr0, rap;
2054 int boguscnt = max_interrupt_work; 2137 int boguscnt = max_interrupt_work;
2055 int must_restart; 2138 int must_restart;
2056 2139
2057 if (!dev) { 2140 if (!dev) {
2058 if (pcnet32_debug & NETIF_MSG_INTR) 2141 if (pcnet32_debug & NETIF_MSG_INTR)
2059 printk (KERN_DEBUG "%s(): irq %d for unknown device\n", 2142 printk(KERN_DEBUG "%s(): irq %d for unknown device\n",
2060 __FUNCTION__, irq); 2143 __FUNCTION__, irq);
2061 return IRQ_NONE; 2144 return IRQ_NONE;
2062 }
2063
2064 ioaddr = dev->base_addr;
2065 lp = dev->priv;
2066
2067 spin_lock(&lp->lock);
2068
2069 rap = lp->a.read_rap(ioaddr);
2070 while ((csr0 = lp->a.read_csr (ioaddr, 0)) & 0x8f00 && --boguscnt >= 0) {
2071 if (csr0 == 0xffff) {
2072 break; /* PCMCIA remove happened */
2073 } 2145 }
2074 /* Acknowledge all of the current interrupt sources ASAP. */
2075 lp->a.write_csr (ioaddr, 0, csr0 & ~0x004f);
2076 2146
2077 must_restart = 0; 2147 ioaddr = dev->base_addr;
2148 lp = dev->priv;
2078 2149
2079 if (netif_msg_intr(lp)) 2150 spin_lock(&lp->lock);
2080 printk(KERN_DEBUG "%s: interrupt csr0=%#2.2x new csr=%#2.2x.\n", 2151
2081 dev->name, csr0, lp->a.read_csr (ioaddr, 0)); 2152 rap = lp->a.read_rap(ioaddr);
2082 2153 while ((csr0 = lp->a.read_csr(ioaddr, 0)) & 0x8f00 && --boguscnt >= 0) {
2083 if (csr0 & 0x0400) /* Rx interrupt */ 2154 if (csr0 == 0xffff) {
2084 pcnet32_rx(dev); 2155 break; /* PCMCIA remove happened */
2085 2156 }
2086 if (csr0 & 0x0200) { /* Tx-done interrupt */ 2157 /* Acknowledge all of the current interrupt sources ASAP. */
2087 unsigned int dirty_tx = lp->dirty_tx; 2158 lp->a.write_csr(ioaddr, 0, csr0 & ~0x004f);
2088 int delta; 2159
2089 2160 must_restart = 0;
2090 while (dirty_tx != lp->cur_tx) { 2161
2091 int entry = dirty_tx & lp->tx_mod_mask; 2162 if (netif_msg_intr(lp))
2092 int status = (short)le16_to_cpu(lp->tx_ring[entry].status); 2163 printk(KERN_DEBUG
2093 2164 "%s: interrupt csr0=%#2.2x new csr=%#2.2x.\n",
2094 if (status < 0) 2165 dev->name, csr0, lp->a.read_csr(ioaddr, 0));
2095 break; /* It still hasn't been Txed */ 2166
2096 2167 if (csr0 & 0x0400) /* Rx interrupt */
2097 lp->tx_ring[entry].base = 0; 2168 pcnet32_rx(dev);
2098 2169
2099 if (status & 0x4000) { 2170 if (csr0 & 0x0200) { /* Tx-done interrupt */
2100 /* There was an major error, log it. */ 2171 unsigned int dirty_tx = lp->dirty_tx;
2101 int err_status = le32_to_cpu(lp->tx_ring[entry].misc); 2172 int delta;
2102 lp->stats.tx_errors++; 2173
2103 if (netif_msg_tx_err(lp)) 2174 while (dirty_tx != lp->cur_tx) {
2104 printk(KERN_ERR "%s: Tx error status=%04x err_status=%08x\n", 2175 int entry = dirty_tx & lp->tx_mod_mask;
2105 dev->name, status, err_status); 2176 int status =
2106 if (err_status & 0x04000000) lp->stats.tx_aborted_errors++; 2177 (short)le16_to_cpu(lp->tx_ring[entry].
2107 if (err_status & 0x08000000) lp->stats.tx_carrier_errors++; 2178 status);
2108 if (err_status & 0x10000000) lp->stats.tx_window_errors++; 2179
2180 if (status < 0)
2181 break; /* It still hasn't been Txed */
2182
2183 lp->tx_ring[entry].base = 0;
2184
2185 if (status & 0x4000) {
2186 /* There was an major error, log it. */
2187 int err_status =
2188 le32_to_cpu(lp->tx_ring[entry].
2189 misc);
2190 lp->stats.tx_errors++;
2191 if (netif_msg_tx_err(lp))
2192 printk(KERN_ERR
2193 "%s: Tx error status=%04x err_status=%08x\n",
2194 dev->name, status,
2195 err_status);
2196 if (err_status & 0x04000000)
2197 lp->stats.tx_aborted_errors++;
2198 if (err_status & 0x08000000)
2199 lp->stats.tx_carrier_errors++;
2200 if (err_status & 0x10000000)
2201 lp->stats.tx_window_errors++;
2109#ifndef DO_DXSUFLO 2202#ifndef DO_DXSUFLO
2110 if (err_status & 0x40000000) { 2203 if (err_status & 0x40000000) {
2111 lp->stats.tx_fifo_errors++; 2204 lp->stats.tx_fifo_errors++;
2112 /* Ackk! On FIFO errors the Tx unit is turned off! */ 2205 /* Ackk! On FIFO errors the Tx unit is turned off! */
2113 /* Remove this verbosity later! */ 2206 /* Remove this verbosity later! */
2114 if (netif_msg_tx_err(lp)) 2207 if (netif_msg_tx_err(lp))
2115 printk(KERN_ERR "%s: Tx FIFO error! CSR0=%4.4x\n", 2208 printk(KERN_ERR
2116 dev->name, csr0); 2209 "%s: Tx FIFO error! CSR0=%4.4x\n",
2117 must_restart = 1; 2210 dev->name, csr0);
2118 } 2211 must_restart = 1;
2212 }
2119#else 2213#else
2120 if (err_status & 0x40000000) { 2214 if (err_status & 0x40000000) {
2121 lp->stats.tx_fifo_errors++; 2215 lp->stats.tx_fifo_errors++;
2122 if (! lp->dxsuflo) { /* If controller doesn't recover ... */ 2216 if (!lp->dxsuflo) { /* If controller doesn't recover ... */
2123 /* Ackk! On FIFO errors the Tx unit is turned off! */ 2217 /* Ackk! On FIFO errors the Tx unit is turned off! */
2124 /* Remove this verbosity later! */ 2218 /* Remove this verbosity later! */
2125 if (netif_msg_tx_err(lp)) 2219 if (netif_msg_tx_err
2126 printk(KERN_ERR "%s: Tx FIFO error! CSR0=%4.4x\n", 2220 (lp))
2127 dev->name, csr0); 2221 printk(KERN_ERR
2128 must_restart = 1; 2222 "%s: Tx FIFO error! CSR0=%4.4x\n",
2129 } 2223 dev->
2130 } 2224 name,
2225 csr0);
2226 must_restart = 1;
2227 }
2228 }
2131#endif 2229#endif
2132 } else { 2230 } else {
2133 if (status & 0x1800) 2231 if (status & 0x1800)
2134 lp->stats.collisions++; 2232 lp->stats.collisions++;
2135 lp->stats.tx_packets++; 2233 lp->stats.tx_packets++;
2234 }
2235
2236 /* We must free the original skb */
2237 if (lp->tx_skbuff[entry]) {
2238 pci_unmap_single(lp->pci_dev,
2239 lp->tx_dma_addr[entry],
2240 lp->tx_skbuff[entry]->
2241 len, PCI_DMA_TODEVICE);
2242 dev_kfree_skb_irq(lp->tx_skbuff[entry]);
2243 lp->tx_skbuff[entry] = NULL;
2244 lp->tx_dma_addr[entry] = 0;
2245 }
2246 dirty_tx++;
2247 }
2248
2249 delta =
2250 (lp->cur_tx - dirty_tx) & (lp->tx_mod_mask +
2251 lp->tx_ring_size);
2252 if (delta > lp->tx_ring_size) {
2253 if (netif_msg_drv(lp))
2254 printk(KERN_ERR
2255 "%s: out-of-sync dirty pointer, %d vs. %d, full=%d.\n",
2256 dev->name, dirty_tx, lp->cur_tx,
2257 lp->tx_full);
2258 dirty_tx += lp->tx_ring_size;
2259 delta -= lp->tx_ring_size;
2260 }
2261
2262 if (lp->tx_full &&
2263 netif_queue_stopped(dev) &&
2264 delta < lp->tx_ring_size - 2) {
2265 /* The ring is no longer full, clear tbusy. */
2266 lp->tx_full = 0;
2267 netif_wake_queue(dev);
2268 }
2269 lp->dirty_tx = dirty_tx;
2270 }
2271
2272 /* Log misc errors. */
2273 if (csr0 & 0x4000)
2274 lp->stats.tx_errors++; /* Tx babble. */
2275 if (csr0 & 0x1000) {
2276 /*
2277 * this happens when our receive ring is full. This shouldn't
2278 * be a problem as we will see normal rx interrupts for the frames
2279 * in the receive ring. But there are some PCI chipsets (I can
2280 * reproduce this on SP3G with Intel saturn chipset) which have
2281 * sometimes problems and will fill up the receive ring with
2282 * error descriptors. In this situation we don't get a rx
2283 * interrupt, but a missed frame interrupt sooner or later.
2284 * So we try to clean up our receive ring here.
2285 */
2286 pcnet32_rx(dev);
2287 lp->stats.rx_errors++; /* Missed a Rx frame. */
2288 }
2289 if (csr0 & 0x0800) {
2290 if (netif_msg_drv(lp))
2291 printk(KERN_ERR
2292 "%s: Bus master arbitration failure, status %4.4x.\n",
2293 dev->name, csr0);
2294 /* unlike for the lance, there is no restart needed */
2136 } 2295 }
2137 2296
2138 /* We must free the original skb */ 2297 if (must_restart) {
2139 if (lp->tx_skbuff[entry]) { 2298 /* reset the chip to clear the error condition, then restart */
2140 pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[entry], 2299 lp->a.reset(ioaddr);
2141 lp->tx_skbuff[entry]->len, PCI_DMA_TODEVICE); 2300 lp->a.write_csr(ioaddr, 4, 0x0915);
2142 dev_kfree_skb_irq(lp->tx_skbuff[entry]); 2301 pcnet32_restart(dev, 0x0002);
2143 lp->tx_skbuff[entry] = NULL; 2302 netif_wake_queue(dev);
2144 lp->tx_dma_addr[entry] = 0;
2145 } 2303 }
2146 dirty_tx++; 2304 }
2147 } 2305
2148 2306 /* Set interrupt enable. */
2149 delta = (lp->cur_tx - dirty_tx) & (lp->tx_mod_mask + lp->tx_ring_size); 2307 lp->a.write_csr(ioaddr, 0, 0x0040);
2150 if (delta > lp->tx_ring_size) { 2308 lp->a.write_rap(ioaddr, rap);
2151 if (netif_msg_drv(lp)) 2309
2152 printk(KERN_ERR "%s: out-of-sync dirty pointer, %d vs. %d, full=%d.\n", 2310 if (netif_msg_intr(lp))
2153 dev->name, dirty_tx, lp->cur_tx, lp->tx_full); 2311 printk(KERN_DEBUG "%s: exiting interrupt, csr0=%#4.4x.\n",
2154 dirty_tx += lp->tx_ring_size; 2312 dev->name, lp->a.read_csr(ioaddr, 0));
2155 delta -= lp->tx_ring_size; 2313
2156 } 2314 spin_unlock(&lp->lock);
2157 2315
2158 if (lp->tx_full && 2316 return IRQ_HANDLED;
2159 netif_queue_stopped(dev) &&
2160 delta < lp->tx_ring_size - 2) {
2161 /* The ring is no longer full, clear tbusy. */
2162 lp->tx_full = 0;
2163 netif_wake_queue (dev);
2164 }
2165 lp->dirty_tx = dirty_tx;
2166 }
2167
2168 /* Log misc errors. */
2169 if (csr0 & 0x4000) lp->stats.tx_errors++; /* Tx babble. */
2170 if (csr0 & 0x1000) {
2171 /*
2172 * this happens when our receive ring is full. This shouldn't
2173 * be a problem as we will see normal rx interrupts for the frames
2174 * in the receive ring. But there are some PCI chipsets (I can
2175 * reproduce this on SP3G with Intel saturn chipset) which have
2176 * sometimes problems and will fill up the receive ring with
2177 * error descriptors. In this situation we don't get a rx
2178 * interrupt, but a missed frame interrupt sooner or later.
2179 * So we try to clean up our receive ring here.
2180 */
2181 pcnet32_rx(dev);
2182 lp->stats.rx_errors++; /* Missed a Rx frame. */
2183 }
2184 if (csr0 & 0x0800) {
2185 if (netif_msg_drv(lp))
2186 printk(KERN_ERR "%s: Bus master arbitration failure, status %4.4x.\n",
2187 dev->name, csr0);
2188 /* unlike for the lance, there is no restart needed */
2189 }
2190
2191 if (must_restart) {
2192 /* reset the chip to clear the error condition, then restart */
2193 lp->a.reset(ioaddr);
2194 lp->a.write_csr(ioaddr, 4, 0x0915);
2195 pcnet32_restart(dev, 0x0002);
2196 netif_wake_queue(dev);
2197 }
2198 }
2199
2200 /* Set interrupt enable. */
2201 lp->a.write_csr (ioaddr, 0, 0x0040);
2202 lp->a.write_rap (ioaddr,rap);
2203
2204 if (netif_msg_intr(lp))
2205 printk(KERN_DEBUG "%s: exiting interrupt, csr0=%#4.4x.\n",
2206 dev->name, lp->a.read_csr (ioaddr, 0));
2207
2208 spin_unlock(&lp->lock);
2209
2210 return IRQ_HANDLED;
2211} 2317}
2212 2318
2213static int 2319static int pcnet32_rx(struct net_device *dev)
2214pcnet32_rx(struct net_device *dev)
2215{ 2320{
2216 struct pcnet32_private *lp = dev->priv; 2321 struct pcnet32_private *lp = dev->priv;
2217 int entry = lp->cur_rx & lp->rx_mod_mask; 2322 int entry = lp->cur_rx & lp->rx_mod_mask;
2218 int boguscnt = lp->rx_ring_size / 2; 2323 int boguscnt = lp->rx_ring_size / 2;
2219 2324
2220 /* If we own the next entry, it's a new packet. Send it up. */ 2325 /* If we own the next entry, it's a new packet. Send it up. */
2221 while ((short)le16_to_cpu(lp->rx_ring[entry].status) >= 0) { 2326 while ((short)le16_to_cpu(lp->rx_ring[entry].status) >= 0) {
2222 int status = (short)le16_to_cpu(lp->rx_ring[entry].status) >> 8; 2327 int status = (short)le16_to_cpu(lp->rx_ring[entry].status) >> 8;
2223 2328
2224 if (status != 0x03) { /* There was an error. */ 2329 if (status != 0x03) { /* There was an error. */
2225 /* 2330 /*
2226 * There is a tricky error noted by John Murphy, 2331 * There is a tricky error noted by John Murphy,
2227 * <murf@perftech.com> to Russ Nelson: Even with full-sized 2332 * <murf@perftech.com> to Russ Nelson: Even with full-sized
2228 * buffers it's possible for a jabber packet to use two 2333 * buffers it's possible for a jabber packet to use two
2229 * buffers, with only the last correctly noting the error. 2334 * buffers, with only the last correctly noting the error.
2230 */ 2335 */
2231 if (status & 0x01) /* Only count a general error at the */ 2336 if (status & 0x01) /* Only count a general error at the */
2232 lp->stats.rx_errors++; /* end of a packet.*/ 2337 lp->stats.rx_errors++; /* end of a packet. */
2233 if (status & 0x20) lp->stats.rx_frame_errors++; 2338 if (status & 0x20)
2234 if (status & 0x10) lp->stats.rx_over_errors++; 2339 lp->stats.rx_frame_errors++;
2235 if (status & 0x08) lp->stats.rx_crc_errors++; 2340 if (status & 0x10)
2236 if (status & 0x04) lp->stats.rx_fifo_errors++; 2341 lp->stats.rx_over_errors++;
2237 lp->rx_ring[entry].status &= le16_to_cpu(0x03ff); 2342 if (status & 0x08)
2238 } else { 2343 lp->stats.rx_crc_errors++;
2239 /* Malloc up new buffer, compatible with net-2e. */ 2344 if (status & 0x04)
2240 short pkt_len = (le32_to_cpu(lp->rx_ring[entry].msg_length) & 0xfff)-4; 2345 lp->stats.rx_fifo_errors++;
2241 struct sk_buff *skb; 2346 lp->rx_ring[entry].status &= le16_to_cpu(0x03ff);
2242
2243 /* Discard oversize frames. */
2244 if (unlikely(pkt_len > PKT_BUF_SZ - 2)) {
2245 if (netif_msg_drv(lp))
2246 printk(KERN_ERR "%s: Impossible packet size %d!\n",
2247 dev->name, pkt_len);
2248 lp->stats.rx_errors++;
2249 } else if (pkt_len < 60) {
2250 if (netif_msg_rx_err(lp))
2251 printk(KERN_ERR "%s: Runt packet!\n", dev->name);
2252 lp->stats.rx_errors++;
2253 } else {
2254 int rx_in_place = 0;
2255
2256 if (pkt_len > rx_copybreak) {
2257 struct sk_buff *newskb;
2258
2259 if ((newskb = dev_alloc_skb(PKT_BUF_SZ))) {
2260 skb_reserve (newskb, 2);
2261 skb = lp->rx_skbuff[entry];
2262 pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[entry],
2263 PKT_BUF_SZ-2, PCI_DMA_FROMDEVICE);
2264 skb_put (skb, pkt_len);
2265 lp->rx_skbuff[entry] = newskb;
2266 newskb->dev = dev;
2267 lp->rx_dma_addr[entry] =
2268 pci_map_single(lp->pci_dev, newskb->data,
2269 PKT_BUF_SZ-2, PCI_DMA_FROMDEVICE);
2270 lp->rx_ring[entry].base = le32_to_cpu(lp->rx_dma_addr[entry]);
2271 rx_in_place = 1;
2272 } else
2273 skb = NULL;
2274 } else { 2347 } else {
2275 skb = dev_alloc_skb(pkt_len+2); 2348 /* Malloc up new buffer, compatible with net-2e. */
2276 } 2349 short pkt_len =
2277 2350 (le32_to_cpu(lp->rx_ring[entry].msg_length) & 0xfff)
2278 if (skb == NULL) { 2351 - 4;
2279 int i; 2352 struct sk_buff *skb;
2280 if (netif_msg_drv(lp)) 2353
2281 printk(KERN_ERR "%s: Memory squeeze, deferring packet.\n", 2354 /* Discard oversize frames. */
2282 dev->name); 2355 if (unlikely(pkt_len > PKT_BUF_SZ - 2)) {
2283 for (i = 0; i < lp->rx_ring_size; i++) 2356 if (netif_msg_drv(lp))
2284 if ((short)le16_to_cpu(lp->rx_ring[(entry+i) 2357 printk(KERN_ERR
2285 & lp->rx_mod_mask].status) < 0) 2358 "%s: Impossible packet size %d!\n",
2286 break; 2359 dev->name, pkt_len);
2287 2360 lp->stats.rx_errors++;
2288 if (i > lp->rx_ring_size -2) { 2361 } else if (pkt_len < 60) {
2289 lp->stats.rx_dropped++; 2362 if (netif_msg_rx_err(lp))
2290 lp->rx_ring[entry].status |= le16_to_cpu(0x8000); 2363 printk(KERN_ERR "%s: Runt packet!\n",
2291 wmb(); /* Make sure adapter sees owner change */ 2364 dev->name);
2292 lp->cur_rx++; 2365 lp->stats.rx_errors++;
2293 } 2366 } else {
2294 break; 2367 int rx_in_place = 0;
2295 } 2368
2296 skb->dev = dev; 2369 if (pkt_len > rx_copybreak) {
2297 if (!rx_in_place) { 2370 struct sk_buff *newskb;
2298 skb_reserve(skb,2); /* 16 byte align */ 2371
2299 skb_put(skb,pkt_len); /* Make room */ 2372 if ((newskb =
2300 pci_dma_sync_single_for_cpu(lp->pci_dev, 2373 dev_alloc_skb(PKT_BUF_SZ))) {
2301 lp->rx_dma_addr[entry], 2374 skb_reserve(newskb, 2);
2302 PKT_BUF_SZ-2, 2375 skb = lp->rx_skbuff[entry];
2303 PCI_DMA_FROMDEVICE); 2376 pci_unmap_single(lp->pci_dev,
2304 eth_copy_and_sum(skb, 2377 lp->
2305 (unsigned char *)(lp->rx_skbuff[entry]->data), 2378 rx_dma_addr
2306 pkt_len,0); 2379 [entry],
2307 pci_dma_sync_single_for_device(lp->pci_dev, 2380 PKT_BUF_SZ - 2,
2308 lp->rx_dma_addr[entry], 2381 PCI_DMA_FROMDEVICE);
2309 PKT_BUF_SZ-2, 2382 skb_put(skb, pkt_len);
2310 PCI_DMA_FROMDEVICE); 2383 lp->rx_skbuff[entry] = newskb;
2384 newskb->dev = dev;
2385 lp->rx_dma_addr[entry] =
2386 pci_map_single(lp->pci_dev,
2387 newskb->data,
2388 PKT_BUF_SZ -
2389 2,
2390 PCI_DMA_FROMDEVICE);
2391 lp->rx_ring[entry].base =
2392 le32_to_cpu(lp->
2393 rx_dma_addr
2394 [entry]);
2395 rx_in_place = 1;
2396 } else
2397 skb = NULL;
2398 } else {
2399 skb = dev_alloc_skb(pkt_len + 2);
2400 }
2401
2402 if (skb == NULL) {
2403 int i;
2404 if (netif_msg_drv(lp))
2405 printk(KERN_ERR
2406 "%s: Memory squeeze, deferring packet.\n",
2407 dev->name);
2408 for (i = 0; i < lp->rx_ring_size; i++)
2409 if ((short)
2410 le16_to_cpu(lp->
2411 rx_ring[(entry +
2412 i)
2413 & lp->
2414 rx_mod_mask].
2415 status) < 0)
2416 break;
2417
2418 if (i > lp->rx_ring_size - 2) {
2419 lp->stats.rx_dropped++;
2420 lp->rx_ring[entry].status |=
2421 le16_to_cpu(0x8000);
2422 wmb(); /* Make sure adapter sees owner change */
2423 lp->cur_rx++;
2424 }
2425 break;
2426 }
2427 skb->dev = dev;
2428 if (!rx_in_place) {
2429 skb_reserve(skb, 2); /* 16 byte align */
2430 skb_put(skb, pkt_len); /* Make room */
2431 pci_dma_sync_single_for_cpu(lp->pci_dev,
2432 lp->
2433 rx_dma_addr
2434 [entry],
2435 PKT_BUF_SZ -
2436 2,
2437 PCI_DMA_FROMDEVICE);
2438 eth_copy_and_sum(skb,
2439 (unsigned char *)(lp->
2440 rx_skbuff
2441 [entry]->
2442 data),
2443 pkt_len, 0);
2444 pci_dma_sync_single_for_device(lp->
2445 pci_dev,
2446 lp->
2447 rx_dma_addr
2448 [entry],
2449 PKT_BUF_SZ
2450 - 2,
2451 PCI_DMA_FROMDEVICE);
2452 }
2453 lp->stats.rx_bytes += skb->len;
2454 skb->protocol = eth_type_trans(skb, dev);
2455 netif_rx(skb);
2456 dev->last_rx = jiffies;
2457 lp->stats.rx_packets++;
2458 }
2311 } 2459 }
2312 lp->stats.rx_bytes += skb->len; 2460 /*
2313 skb->protocol=eth_type_trans(skb,dev); 2461 * The docs say that the buffer length isn't touched, but Andrew Boyd
2314 netif_rx(skb); 2462 * of QNX reports that some revs of the 79C965 clear it.
2315 dev->last_rx = jiffies; 2463 */
2316 lp->stats.rx_packets++; 2464 lp->rx_ring[entry].buf_length = le16_to_cpu(2 - PKT_BUF_SZ);
2317 } 2465 wmb(); /* Make sure owner changes after all others are visible */
2466 lp->rx_ring[entry].status |= le16_to_cpu(0x8000);
2467 entry = (++lp->cur_rx) & lp->rx_mod_mask;
2468 if (--boguscnt <= 0)
2469 break; /* don't stay in loop forever */
2318 } 2470 }
2319 /* 2471
2320 * The docs say that the buffer length isn't touched, but Andrew Boyd 2472 return 0;
2321 * of QNX reports that some revs of the 79C965 clear it.
2322 */
2323 lp->rx_ring[entry].buf_length = le16_to_cpu(2-PKT_BUF_SZ);
2324 wmb(); /* Make sure owner changes after all others are visible */
2325 lp->rx_ring[entry].status |= le16_to_cpu(0x8000);
2326 entry = (++lp->cur_rx) & lp->rx_mod_mask;
2327 if (--boguscnt <= 0) break; /* don't stay in loop forever */
2328 }
2329
2330 return 0;
2331} 2473}
2332 2474
2333static int 2475static int pcnet32_close(struct net_device *dev)
2334pcnet32_close(struct net_device *dev)
2335{ 2476{
2336 unsigned long ioaddr = dev->base_addr; 2477 unsigned long ioaddr = dev->base_addr;
2337 struct pcnet32_private *lp = dev->priv; 2478 struct pcnet32_private *lp = dev->priv;
2338 int i; 2479 int i;
2339 unsigned long flags; 2480 unsigned long flags;
2340 2481
2341 del_timer_sync(&lp->watchdog_timer); 2482 del_timer_sync(&lp->watchdog_timer);
2342 2483
2343 netif_stop_queue(dev); 2484 netif_stop_queue(dev);
2344 2485
2345 spin_lock_irqsave(&lp->lock, flags); 2486 spin_lock_irqsave(&lp->lock, flags);
2346 2487
2347 lp->stats.rx_missed_errors = lp->a.read_csr (ioaddr, 112); 2488 lp->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112);
2348 2489
2349 if (netif_msg_ifdown(lp)) 2490 if (netif_msg_ifdown(lp))
2350 printk(KERN_DEBUG "%s: Shutting down ethercard, status was %2.2x.\n", 2491 printk(KERN_DEBUG
2351 dev->name, lp->a.read_csr (ioaddr, 0)); 2492 "%s: Shutting down ethercard, status was %2.2x.\n",
2493 dev->name, lp->a.read_csr(ioaddr, 0));
2352 2494
2353 /* We stop the PCNET32 here -- it occasionally polls memory if we don't. */ 2495 /* We stop the PCNET32 here -- it occasionally polls memory if we don't. */
2354 lp->a.write_csr (ioaddr, 0, 0x0004); 2496 lp->a.write_csr(ioaddr, 0, 0x0004);
2355 2497
2356 /* 2498 /*
2357 * Switch back to 16bit mode to avoid problems with dumb 2499 * Switch back to 16bit mode to avoid problems with dumb
2358 * DOS packet driver after a warm reboot 2500 * DOS packet driver after a warm reboot
2359 */ 2501 */
2360 lp->a.write_bcr (ioaddr, 20, 4); 2502 lp->a.write_bcr(ioaddr, 20, 4);
2361 2503
2362 spin_unlock_irqrestore(&lp->lock, flags); 2504 spin_unlock_irqrestore(&lp->lock, flags);
2363 2505
2364 free_irq(dev->irq, dev); 2506 free_irq(dev->irq, dev);
2365 2507
2366 spin_lock_irqsave(&lp->lock, flags); 2508 spin_lock_irqsave(&lp->lock, flags);
2367 2509
2368 /* free all allocated skbuffs */ 2510 /* free all allocated skbuffs */
2369 for (i = 0; i < lp->rx_ring_size; i++) { 2511 for (i = 0; i < lp->rx_ring_size; i++) {
2370 lp->rx_ring[i].status = 0; 2512 lp->rx_ring[i].status = 0;
2371 wmb(); /* Make sure adapter sees owner change */ 2513 wmb(); /* Make sure adapter sees owner change */
2372 if (lp->rx_skbuff[i]) { 2514 if (lp->rx_skbuff[i]) {
2373 pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[i], PKT_BUF_SZ-2, 2515 pci_unmap_single(lp->pci_dev, lp->rx_dma_addr[i],
2374 PCI_DMA_FROMDEVICE); 2516 PKT_BUF_SZ - 2, PCI_DMA_FROMDEVICE);
2375 dev_kfree_skb(lp->rx_skbuff[i]); 2517 dev_kfree_skb(lp->rx_skbuff[i]);
2518 }
2519 lp->rx_skbuff[i] = NULL;
2520 lp->rx_dma_addr[i] = 0;
2376 } 2521 }
2377 lp->rx_skbuff[i] = NULL;
2378 lp->rx_dma_addr[i] = 0;
2379 }
2380 2522
2381 for (i = 0; i < lp->tx_ring_size; i++) { 2523 for (i = 0; i < lp->tx_ring_size; i++) {
2382 lp->tx_ring[i].status = 0; /* CPU owns buffer */ 2524 lp->tx_ring[i].status = 0; /* CPU owns buffer */
2383 wmb(); /* Make sure adapter sees owner change */ 2525 wmb(); /* Make sure adapter sees owner change */
2384 if (lp->tx_skbuff[i]) { 2526 if (lp->tx_skbuff[i]) {
2385 pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[i], 2527 pci_unmap_single(lp->pci_dev, lp->tx_dma_addr[i],
2386 lp->tx_skbuff[i]->len, PCI_DMA_TODEVICE); 2528 lp->tx_skbuff[i]->len,
2387 dev_kfree_skb(lp->tx_skbuff[i]); 2529 PCI_DMA_TODEVICE);
2530 dev_kfree_skb(lp->tx_skbuff[i]);
2531 }
2532 lp->tx_skbuff[i] = NULL;
2533 lp->tx_dma_addr[i] = 0;
2388 } 2534 }
2389 lp->tx_skbuff[i] = NULL;
2390 lp->tx_dma_addr[i] = 0;
2391 }
2392 2535
2393 spin_unlock_irqrestore(&lp->lock, flags); 2536 spin_unlock_irqrestore(&lp->lock, flags);
2394 2537
2395 return 0; 2538 return 0;
2396} 2539}
2397 2540
2398static struct net_device_stats * 2541static struct net_device_stats *pcnet32_get_stats(struct net_device *dev)
2399pcnet32_get_stats(struct net_device *dev)
2400{ 2542{
2401 struct pcnet32_private *lp = dev->priv; 2543 struct pcnet32_private *lp = dev->priv;
2402 unsigned long ioaddr = dev->base_addr; 2544 unsigned long ioaddr = dev->base_addr;
2403 u16 saved_addr; 2545 u16 saved_addr;
2404 unsigned long flags; 2546 unsigned long flags;
2405 2547
2406 spin_lock_irqsave(&lp->lock, flags); 2548 spin_lock_irqsave(&lp->lock, flags);
2407 saved_addr = lp->a.read_rap(ioaddr); 2549 saved_addr = lp->a.read_rap(ioaddr);
2408 lp->stats.rx_missed_errors = lp->a.read_csr (ioaddr, 112); 2550 lp->stats.rx_missed_errors = lp->a.read_csr(ioaddr, 112);
2409 lp->a.write_rap(ioaddr, saved_addr); 2551 lp->a.write_rap(ioaddr, saved_addr);
2410 spin_unlock_irqrestore(&lp->lock, flags); 2552 spin_unlock_irqrestore(&lp->lock, flags);
2411 2553
2412 return &lp->stats; 2554 return &lp->stats;
2413} 2555}
2414 2556
2415/* taken from the sunlance driver, which it took from the depca driver */ 2557/* taken from the sunlance driver, which it took from the depca driver */
2416static void pcnet32_load_multicast (struct net_device *dev) 2558static void pcnet32_load_multicast(struct net_device *dev)
2417{ 2559{
2418 struct pcnet32_private *lp = dev->priv; 2560 struct pcnet32_private *lp = dev->priv;
2419 volatile struct pcnet32_init_block *ib = &lp->init_block; 2561 volatile struct pcnet32_init_block *ib = &lp->init_block;
2420 volatile u16 *mcast_table = (u16 *)&ib->filter; 2562 volatile u16 *mcast_table = (u16 *) & ib->filter;
2421 struct dev_mc_list *dmi=dev->mc_list; 2563 struct dev_mc_list *dmi = dev->mc_list;
2422 char *addrs; 2564 char *addrs;
2423 int i; 2565 int i;
2424 u32 crc; 2566 u32 crc;
2425 2567
2426 /* set all multicast bits */ 2568 /* set all multicast bits */
2427 if (dev->flags & IFF_ALLMULTI) { 2569 if (dev->flags & IFF_ALLMULTI) {
2428 ib->filter[0] = 0xffffffff; 2570 ib->filter[0] = 0xffffffff;
2429 ib->filter[1] = 0xffffffff; 2571 ib->filter[1] = 0xffffffff;
2572 return;
2573 }
2574 /* clear the multicast filter */
2575 ib->filter[0] = 0;
2576 ib->filter[1] = 0;
2577
2578 /* Add addresses */
2579 for (i = 0; i < dev->mc_count; i++) {
2580 addrs = dmi->dmi_addr;
2581 dmi = dmi->next;
2582
2583 /* multicast address? */
2584 if (!(*addrs & 1))
2585 continue;
2586
2587 crc = ether_crc_le(6, addrs);
2588 crc = crc >> 26;
2589 mcast_table[crc >> 4] =
2590 le16_to_cpu(le16_to_cpu(mcast_table[crc >> 4]) |
2591 (1 << (crc & 0xf)));
2592 }
2430 return; 2593 return;
2431 }
2432 /* clear the multicast filter */
2433 ib->filter[0] = 0;
2434 ib->filter[1] = 0;
2435
2436 /* Add addresses */
2437 for (i = 0; i < dev->mc_count; i++) {
2438 addrs = dmi->dmi_addr;
2439 dmi = dmi->next;
2440
2441 /* multicast address? */
2442 if (!(*addrs & 1))
2443 continue;
2444
2445 crc = ether_crc_le(6, addrs);
2446 crc = crc >> 26;
2447 mcast_table [crc >> 4] = le16_to_cpu(
2448 le16_to_cpu(mcast_table [crc >> 4]) | (1 << (crc & 0xf)));
2449 }
2450 return;
2451} 2594}
2452 2595
2453
2454/* 2596/*
2455 * Set or clear the multicast filter for this adaptor. 2597 * Set or clear the multicast filter for this adaptor.
2456 */ 2598 */
2457static void pcnet32_set_multicast_list(struct net_device *dev) 2599static void pcnet32_set_multicast_list(struct net_device *dev)
2458{ 2600{
2459 unsigned long ioaddr = dev->base_addr, flags; 2601 unsigned long ioaddr = dev->base_addr, flags;
2460 struct pcnet32_private *lp = dev->priv; 2602 struct pcnet32_private *lp = dev->priv;
2461 2603
2462 spin_lock_irqsave(&lp->lock, flags); 2604 spin_lock_irqsave(&lp->lock, flags);
2463 if (dev->flags&IFF_PROMISC) { 2605 if (dev->flags & IFF_PROMISC) {
2464 /* Log any net taps. */ 2606 /* Log any net taps. */
2465 if (netif_msg_hw(lp)) 2607 if (netif_msg_hw(lp))
2466 printk(KERN_INFO "%s: Promiscuous mode enabled.\n", dev->name); 2608 printk(KERN_INFO "%s: Promiscuous mode enabled.\n",
2467 lp->init_block.mode = le16_to_cpu(0x8000 | (lp->options & PCNET32_PORT_PORTSEL) << 7); 2609 dev->name);
2468 } else { 2610 lp->init_block.mode =
2469 lp->init_block.mode = le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7); 2611 le16_to_cpu(0x8000 | (lp->options & PCNET32_PORT_PORTSEL) <<
2470 pcnet32_load_multicast (dev); 2612 7);
2471 } 2613 } else {
2472 2614 lp->init_block.mode =
2473 lp->a.write_csr (ioaddr, 0, 0x0004); /* Temporarily stop the lance. */ 2615 le16_to_cpu((lp->options & PCNET32_PORT_PORTSEL) << 7);
2474 pcnet32_restart(dev, 0x0042); /* Resume normal operation */ 2616 pcnet32_load_multicast(dev);
2475 netif_wake_queue(dev); 2617 }
2476 2618
2477 spin_unlock_irqrestore(&lp->lock, flags); 2619 lp->a.write_csr(ioaddr, 0, 0x0004); /* Temporarily stop the lance. */
2620 pcnet32_restart(dev, 0x0042); /* Resume normal operation */
2621 netif_wake_queue(dev);
2622
2623 spin_unlock_irqrestore(&lp->lock, flags);
2478} 2624}
2479 2625
2480/* This routine assumes that the lp->lock is held */ 2626/* This routine assumes that the lp->lock is held */
2481static int mdio_read(struct net_device *dev, int phy_id, int reg_num) 2627static int mdio_read(struct net_device *dev, int phy_id, int reg_num)
2482{ 2628{
2483 struct pcnet32_private *lp = dev->priv; 2629 struct pcnet32_private *lp = dev->priv;
2484 unsigned long ioaddr = dev->base_addr; 2630 unsigned long ioaddr = dev->base_addr;
2485 u16 val_out; 2631 u16 val_out;
2486 2632
2487 if (!lp->mii) 2633 if (!lp->mii)
2488 return 0; 2634 return 0;
2489 2635
2490 lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); 2636 lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f));
2491 val_out = lp->a.read_bcr(ioaddr, 34); 2637 val_out = lp->a.read_bcr(ioaddr, 34);
2492 2638
2493 return val_out; 2639 return val_out;
2494} 2640}
2495 2641
2496/* This routine assumes that the lp->lock is held */ 2642/* This routine assumes that the lp->lock is held */
2497static void mdio_write(struct net_device *dev, int phy_id, int reg_num, int val) 2643static void mdio_write(struct net_device *dev, int phy_id, int reg_num, int val)
2498{ 2644{
2499 struct pcnet32_private *lp = dev->priv; 2645 struct pcnet32_private *lp = dev->priv;
2500 unsigned long ioaddr = dev->base_addr; 2646 unsigned long ioaddr = dev->base_addr;
2501 2647
2502 if (!lp->mii) 2648 if (!lp->mii)
2503 return; 2649 return;
2504 2650
2505 lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f)); 2651 lp->a.write_bcr(ioaddr, 33, ((phy_id & 0x1f) << 5) | (reg_num & 0x1f));
2506 lp->a.write_bcr(ioaddr, 34, val); 2652 lp->a.write_bcr(ioaddr, 34, val);
2507} 2653}
2508 2654
2509static int pcnet32_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) 2655static int pcnet32_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
2510{ 2656{
2511 struct pcnet32_private *lp = dev->priv; 2657 struct pcnet32_private *lp = dev->priv;
2512 int rc; 2658 int rc;
2513 unsigned long flags; 2659 unsigned long flags;
2514 2660
2515 /* SIOC[GS]MIIxxx ioctls */ 2661 /* SIOC[GS]MIIxxx ioctls */
2516 if (lp->mii) { 2662 if (lp->mii) {
2517 spin_lock_irqsave(&lp->lock, flags); 2663 spin_lock_irqsave(&lp->lock, flags);
2518 rc = generic_mii_ioctl(&lp->mii_if, if_mii(rq), cmd, NULL); 2664 rc = generic_mii_ioctl(&lp->mii_if, if_mii(rq), cmd, NULL);
2519 spin_unlock_irqrestore(&lp->lock, flags); 2665 spin_unlock_irqrestore(&lp->lock, flags);
2520 } else { 2666 } else {
2521 rc = -EOPNOTSUPP; 2667 rc = -EOPNOTSUPP;
2522 } 2668 }
2523 2669
2524 return rc; 2670 return rc;
2525} 2671}
2526 2672
2527static int pcnet32_check_otherphy(struct net_device *dev) 2673static int pcnet32_check_otherphy(struct net_device *dev)
2528{ 2674{
2529 struct pcnet32_private *lp = dev->priv; 2675 struct pcnet32_private *lp = dev->priv;
2530 struct mii_if_info mii = lp->mii_if; 2676 struct mii_if_info mii = lp->mii_if;
2531 u16 bmcr; 2677 u16 bmcr;
2532 int i; 2678 int i;
2533
2534 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
2535 if (i == lp->mii_if.phy_id)
2536 continue; /* skip active phy */
2537 if (lp->phymask & (1 << i)) {
2538 mii.phy_id = i;
2539 if (mii_link_ok(&mii)) {
2540 /* found PHY with active link */
2541 if (netif_msg_link(lp))
2542 printk(KERN_INFO "%s: Using PHY number %d.\n", dev->name, i);
2543
2544 /* isolate inactive phy */
2545 bmcr = mdio_read(dev, lp->mii_if.phy_id, MII_BMCR);
2546 mdio_write(dev, lp->mii_if.phy_id, MII_BMCR, bmcr | BMCR_ISOLATE);
2547
2548 /* de-isolate new phy */
2549 bmcr = mdio_read(dev, i, MII_BMCR);
2550 mdio_write(dev, i, MII_BMCR, bmcr & ~BMCR_ISOLATE);
2551 2679
2552 /* set new phy address */ 2680 for (i = 0; i < PCNET32_MAX_PHYS; i++) {
2553 lp->mii_if.phy_id = i; 2681 if (i == lp->mii_if.phy_id)
2554 return 1; 2682 continue; /* skip active phy */
2555 } 2683 if (lp->phymask & (1 << i)) {
2684 mii.phy_id = i;
2685 if (mii_link_ok(&mii)) {
2686 /* found PHY with active link */
2687 if (netif_msg_link(lp))
2688 printk(KERN_INFO
2689 "%s: Using PHY number %d.\n",
2690 dev->name, i);
2691
2692 /* isolate inactive phy */
2693 bmcr =
2694 mdio_read(dev, lp->mii_if.phy_id, MII_BMCR);
2695 mdio_write(dev, lp->mii_if.phy_id, MII_BMCR,
2696 bmcr | BMCR_ISOLATE);
2697
2698 /* de-isolate new phy */
2699 bmcr = mdio_read(dev, i, MII_BMCR);
2700 mdio_write(dev, i, MII_BMCR,
2701 bmcr & ~BMCR_ISOLATE);
2702
2703 /* set new phy address */
2704 lp->mii_if.phy_id = i;
2705 return 1;
2706 }
2707 }
2556 } 2708 }
2557 } 2709 return 0;
2558 return 0;
2559} 2710}
2560 2711
2561/* 2712/*
@@ -2568,51 +2719,53 @@ static int pcnet32_check_otherphy(struct net_device *dev)
2568 2719
2569static void pcnet32_check_media(struct net_device *dev, int verbose) 2720static void pcnet32_check_media(struct net_device *dev, int verbose)
2570{ 2721{
2571 struct pcnet32_private *lp = dev->priv; 2722 struct pcnet32_private *lp = dev->priv;
2572 int curr_link; 2723 int curr_link;
2573 int prev_link = netif_carrier_ok(dev) ? 1 : 0; 2724 int prev_link = netif_carrier_ok(dev) ? 1 : 0;
2574 u32 bcr9; 2725 u32 bcr9;
2575 2726
2576 if (lp->mii) {
2577 curr_link = mii_link_ok(&lp->mii_if);
2578 } else {
2579 ulong ioaddr = dev->base_addr; /* card base I/O address */
2580 curr_link = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
2581 }
2582 if (!curr_link) {
2583 if (prev_link || verbose) {
2584 netif_carrier_off(dev);
2585 if (netif_msg_link(lp))
2586 printk(KERN_INFO "%s: link down\n", dev->name);
2587 }
2588 if (lp->phycount > 1) {
2589 curr_link = pcnet32_check_otherphy(dev);
2590 prev_link = 0;
2591 }
2592 } else if (verbose || !prev_link) {
2593 netif_carrier_on(dev);
2594 if (lp->mii) { 2727 if (lp->mii) {
2595 if (netif_msg_link(lp)) { 2728 curr_link = mii_link_ok(&lp->mii_if);
2596 struct ethtool_cmd ecmd;
2597 mii_ethtool_gset(&lp->mii_if, &ecmd);
2598 printk(KERN_INFO "%s: link up, %sMbps, %s-duplex\n",
2599 dev->name,
2600 (ecmd.speed == SPEED_100) ? "100" : "10",
2601 (ecmd.duplex == DUPLEX_FULL) ? "full" : "half");
2602 }
2603 bcr9 = lp->a.read_bcr(dev->base_addr, 9);
2604 if ((bcr9 & (1 << 0)) != lp->mii_if.full_duplex) {
2605 if (lp->mii_if.full_duplex)
2606 bcr9 |= (1 << 0);
2607 else
2608 bcr9 &= ~(1 << 0);
2609 lp->a.write_bcr(dev->base_addr, 9, bcr9);
2610 }
2611 } else { 2729 } else {
2612 if (netif_msg_link(lp)) 2730 ulong ioaddr = dev->base_addr; /* card base I/O address */
2613 printk(KERN_INFO "%s: link up\n", dev->name); 2731 curr_link = (lp->a.read_bcr(ioaddr, 4) != 0xc0);
2732 }
2733 if (!curr_link) {
2734 if (prev_link || verbose) {
2735 netif_carrier_off(dev);
2736 if (netif_msg_link(lp))
2737 printk(KERN_INFO "%s: link down\n", dev->name);
2738 }
2739 if (lp->phycount > 1) {
2740 curr_link = pcnet32_check_otherphy(dev);
2741 prev_link = 0;
2742 }
2743 } else if (verbose || !prev_link) {
2744 netif_carrier_on(dev);
2745 if (lp->mii) {
2746 if (netif_msg_link(lp)) {
2747 struct ethtool_cmd ecmd;
2748 mii_ethtool_gset(&lp->mii_if, &ecmd);
2749 printk(KERN_INFO
2750 "%s: link up, %sMbps, %s-duplex\n",
2751 dev->name,
2752 (ecmd.speed == SPEED_100) ? "100" : "10",
2753 (ecmd.duplex ==
2754 DUPLEX_FULL) ? "full" : "half");
2755 }
2756 bcr9 = lp->a.read_bcr(dev->base_addr, 9);
2757 if ((bcr9 & (1 << 0)) != lp->mii_if.full_duplex) {
2758 if (lp->mii_if.full_duplex)
2759 bcr9 |= (1 << 0);
2760 else
2761 bcr9 &= ~(1 << 0);
2762 lp->a.write_bcr(dev->base_addr, 9, bcr9);
2763 }
2764 } else {
2765 if (netif_msg_link(lp))
2766 printk(KERN_INFO "%s: link up\n", dev->name);
2767 }
2614 } 2768 }
2615 }
2616} 2769}
2617 2770
2618/* 2771/*
@@ -2622,39 +2775,39 @@ static void pcnet32_check_media(struct net_device *dev, int verbose)
2622 2775
2623static void pcnet32_watchdog(struct net_device *dev) 2776static void pcnet32_watchdog(struct net_device *dev)
2624{ 2777{
2625 struct pcnet32_private *lp = dev->priv; 2778 struct pcnet32_private *lp = dev->priv;
2626 unsigned long flags; 2779 unsigned long flags;
2627 2780
2628 /* Print the link status if it has changed */ 2781 /* Print the link status if it has changed */
2629 spin_lock_irqsave(&lp->lock, flags); 2782 spin_lock_irqsave(&lp->lock, flags);
2630 pcnet32_check_media(dev, 0); 2783 pcnet32_check_media(dev, 0);
2631 spin_unlock_irqrestore(&lp->lock, flags); 2784 spin_unlock_irqrestore(&lp->lock, flags);
2632 2785
2633 mod_timer (&(lp->watchdog_timer), PCNET32_WATCHDOG_TIMEOUT); 2786 mod_timer(&(lp->watchdog_timer), PCNET32_WATCHDOG_TIMEOUT);
2634} 2787}
2635 2788
2636static void __devexit pcnet32_remove_one(struct pci_dev *pdev) 2789static void __devexit pcnet32_remove_one(struct pci_dev *pdev)
2637{ 2790{
2638 struct net_device *dev = pci_get_drvdata(pdev); 2791 struct net_device *dev = pci_get_drvdata(pdev);
2639 2792
2640 if (dev) { 2793 if (dev) {
2641 struct pcnet32_private *lp = dev->priv; 2794 struct pcnet32_private *lp = dev->priv;
2642 2795
2643 unregister_netdev(dev); 2796 unregister_netdev(dev);
2644 pcnet32_free_ring(dev); 2797 pcnet32_free_ring(dev);
2645 release_region(dev->base_addr, PCNET32_TOTAL_SIZE); 2798 release_region(dev->base_addr, PCNET32_TOTAL_SIZE);
2646 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr); 2799 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr);
2647 free_netdev(dev); 2800 free_netdev(dev);
2648 pci_disable_device(pdev); 2801 pci_disable_device(pdev);
2649 pci_set_drvdata(pdev, NULL); 2802 pci_set_drvdata(pdev, NULL);
2650 } 2803 }
2651} 2804}
2652 2805
2653static struct pci_driver pcnet32_driver = { 2806static struct pci_driver pcnet32_driver = {
2654 .name = DRV_NAME, 2807 .name = DRV_NAME,
2655 .probe = pcnet32_probe_pci, 2808 .probe = pcnet32_probe_pci,
2656 .remove = __devexit_p(pcnet32_remove_one), 2809 .remove = __devexit_p(pcnet32_remove_one),
2657 .id_table = pcnet32_pci_tbl, 2810 .id_table = pcnet32_pci_tbl,
2658}; 2811};
2659 2812
2660/* An additional parameter that may be passed in... */ 2813/* An additional parameter that may be passed in... */
@@ -2665,9 +2818,11 @@ static int pcnet32_have_pci;
2665module_param(debug, int, 0); 2818module_param(debug, int, 0);
2666MODULE_PARM_DESC(debug, DRV_NAME " debug level"); 2819MODULE_PARM_DESC(debug, DRV_NAME " debug level");
2667module_param(max_interrupt_work, int, 0); 2820module_param(max_interrupt_work, int, 0);
2668MODULE_PARM_DESC(max_interrupt_work, DRV_NAME " maximum events handled per interrupt"); 2821MODULE_PARM_DESC(max_interrupt_work,
2822 DRV_NAME " maximum events handled per interrupt");
2669module_param(rx_copybreak, int, 0); 2823module_param(rx_copybreak, int, 0);
2670MODULE_PARM_DESC(rx_copybreak, DRV_NAME " copy breakpoint for copy-only-tiny-frames"); 2824MODULE_PARM_DESC(rx_copybreak,
2825 DRV_NAME " copy breakpoint for copy-only-tiny-frames");
2671module_param(tx_start_pt, int, 0); 2826module_param(tx_start_pt, int, 0);
2672MODULE_PARM_DESC(tx_start_pt, DRV_NAME " transmit start point (0-3)"); 2827MODULE_PARM_DESC(tx_start_pt, DRV_NAME " transmit start point (0-3)");
2673module_param(pcnet32vlb, int, 0); 2828module_param(pcnet32vlb, int, 0);
@@ -2678,7 +2833,9 @@ module_param_array(full_duplex, int, NULL, 0);
2678MODULE_PARM_DESC(full_duplex, DRV_NAME " full duplex setting(s) (1)"); 2833MODULE_PARM_DESC(full_duplex, DRV_NAME " full duplex setting(s) (1)");
2679/* Module Parameter for HomePNA cards added by Patrick Simmons, 2004 */ 2834/* Module Parameter for HomePNA cards added by Patrick Simmons, 2004 */
2680module_param_array(homepna, int, NULL, 0); 2835module_param_array(homepna, int, NULL, 0);
2681MODULE_PARM_DESC(homepna, DRV_NAME " mode for 79C978 cards (1 for HomePNA, 0 for Ethernet, default Ethernet"); 2836MODULE_PARM_DESC(homepna,
2837 DRV_NAME
2838 " mode for 79C978 cards (1 for HomePNA, 0 for Ethernet, default Ethernet");
2682 2839
2683MODULE_AUTHOR("Thomas Bogendoerfer"); 2840MODULE_AUTHOR("Thomas Bogendoerfer");
2684MODULE_DESCRIPTION("Driver for PCnet32 and PCnetPCI based ethercards"); 2841MODULE_DESCRIPTION("Driver for PCnet32 and PCnetPCI based ethercards");
@@ -2688,44 +2845,44 @@ MODULE_LICENSE("GPL");
2688 2845
2689static int __init pcnet32_init_module(void) 2846static int __init pcnet32_init_module(void)
2690{ 2847{
2691 printk(KERN_INFO "%s", version); 2848 printk(KERN_INFO "%s", version);
2692 2849
2693 pcnet32_debug = netif_msg_init(debug, PCNET32_MSG_DEFAULT); 2850 pcnet32_debug = netif_msg_init(debug, PCNET32_MSG_DEFAULT);
2694 2851
2695 if ((tx_start_pt >= 0) && (tx_start_pt <= 3)) 2852 if ((tx_start_pt >= 0) && (tx_start_pt <= 3))
2696 tx_start = tx_start_pt; 2853 tx_start = tx_start_pt;
2697 2854
2698 /* find the PCI devices */ 2855 /* find the PCI devices */
2699 if (!pci_module_init(&pcnet32_driver)) 2856 if (!pci_module_init(&pcnet32_driver))
2700 pcnet32_have_pci = 1; 2857 pcnet32_have_pci = 1;
2701 2858
2702 /* should we find any remaining VLbus devices ? */ 2859 /* should we find any remaining VLbus devices ? */
2703 if (pcnet32vlb) 2860 if (pcnet32vlb)
2704 pcnet32_probe_vlbus(); 2861 pcnet32_probe_vlbus();
2705 2862
2706 if (cards_found && (pcnet32_debug & NETIF_MSG_PROBE)) 2863 if (cards_found && (pcnet32_debug & NETIF_MSG_PROBE))
2707 printk(KERN_INFO PFX "%d cards_found.\n", cards_found); 2864 printk(KERN_INFO PFX "%d cards_found.\n", cards_found);
2708 2865
2709 return (pcnet32_have_pci + cards_found) ? 0 : -ENODEV; 2866 return (pcnet32_have_pci + cards_found) ? 0 : -ENODEV;
2710} 2867}
2711 2868
2712static void __exit pcnet32_cleanup_module(void) 2869static void __exit pcnet32_cleanup_module(void)
2713{ 2870{
2714 struct net_device *next_dev; 2871 struct net_device *next_dev;
2715 2872
2716 while (pcnet32_dev) { 2873 while (pcnet32_dev) {
2717 struct pcnet32_private *lp = pcnet32_dev->priv; 2874 struct pcnet32_private *lp = pcnet32_dev->priv;
2718 next_dev = lp->next; 2875 next_dev = lp->next;
2719 unregister_netdev(pcnet32_dev); 2876 unregister_netdev(pcnet32_dev);
2720 pcnet32_free_ring(pcnet32_dev); 2877 pcnet32_free_ring(pcnet32_dev);
2721 release_region(pcnet32_dev->base_addr, PCNET32_TOTAL_SIZE); 2878 release_region(pcnet32_dev->base_addr, PCNET32_TOTAL_SIZE);
2722 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr); 2879 pci_free_consistent(lp->pci_dev, sizeof(*lp), lp, lp->dma_addr);
2723 free_netdev(pcnet32_dev); 2880 free_netdev(pcnet32_dev);
2724 pcnet32_dev = next_dev; 2881 pcnet32_dev = next_dev;
2725 } 2882 }
2726 2883
2727 if (pcnet32_have_pci) 2884 if (pcnet32_have_pci)
2728 pci_unregister_driver(&pcnet32_driver); 2885 pci_unregister_driver(&pcnet32_driver);
2729} 2886}
2730 2887
2731module_init(pcnet32_init_module); 2888module_init(pcnet32_init_module);