diff options
44 files changed, 2400 insertions, 1800 deletions
diff --git a/drivers/net/8139cp.c b/drivers/net/8139cp.c index 807e6992e614..e970e64bf966 100644 --- a/drivers/net/8139cp.c +++ b/drivers/net/8139cp.c | |||
| @@ -26,7 +26,6 @@ | |||
| 26 | 26 | ||
| 27 | TODO: | 27 | TODO: |
| 28 | * Test Tx checksumming thoroughly | 28 | * Test Tx checksumming thoroughly |
| 29 | * Implement dev->tx_timeout | ||
| 30 | 29 | ||
| 31 | Low priority TODO: | 30 | Low priority TODO: |
| 32 | * Complete reset on PciErr | 31 | * Complete reset on PciErr |
| @@ -1218,6 +1217,30 @@ static int cp_close (struct net_device *dev) | |||
| 1218 | return 0; | 1217 | return 0; |
| 1219 | } | 1218 | } |
| 1220 | 1219 | ||
| 1220 | static void cp_tx_timeout(struct net_device *dev) | ||
| 1221 | { | ||
| 1222 | struct cp_private *cp = netdev_priv(dev); | ||
| 1223 | unsigned long flags; | ||
| 1224 | int rc; | ||
| 1225 | |||
| 1226 | printk(KERN_WARNING "%s: Transmit timeout, status %2x %4x %4x %4x\n", | ||
| 1227 | dev->name, cpr8(Cmd), cpr16(CpCmd), | ||
| 1228 | cpr16(IntrStatus), cpr16(IntrMask)); | ||
| 1229 | |||
| 1230 | spin_lock_irqsave(&cp->lock, flags); | ||
| 1231 | |||
| 1232 | cp_stop_hw(cp); | ||
| 1233 | cp_clean_rings(cp); | ||
| 1234 | rc = cp_init_rings(cp); | ||
| 1235 | cp_start_hw(cp); | ||
| 1236 | |||
| 1237 | netif_wake_queue(dev); | ||
| 1238 | |||
| 1239 | spin_unlock_irqrestore(&cp->lock, flags); | ||
| 1240 | |||
| 1241 | return; | ||
| 1242 | } | ||
| 1243 | |||
| 1221 | #ifdef BROKEN | 1244 | #ifdef BROKEN |
| 1222 | static int cp_change_mtu(struct net_device *dev, int new_mtu) | 1245 | static int cp_change_mtu(struct net_device *dev, int new_mtu) |
| 1223 | { | 1246 | { |
| @@ -1920,10 +1943,8 @@ static int cp_init_one (struct pci_dev *pdev, const struct pci_device_id *ent) | |||
| 1920 | dev->change_mtu = cp_change_mtu; | 1943 | dev->change_mtu = cp_change_mtu; |
| 1921 | #endif | 1944 | #endif |
| 1922 | dev->ethtool_ops = &cp_ethtool_ops; | 1945 | dev->ethtool_ops = &cp_ethtool_ops; |
| 1923 | #if 0 | ||
| 1924 | dev->tx_timeout = cp_tx_timeout; | 1946 | dev->tx_timeout = cp_tx_timeout; |
| 1925 | dev->watchdog_timeo = TX_TIMEOUT; | 1947 | dev->watchdog_timeo = TX_TIMEOUT; |
| 1926 | #endif | ||
| 1927 | 1948 | ||
| 1928 | #if CP_VLAN_TAG_USED | 1949 | #if CP_VLAN_TAG_USED |
| 1929 | dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX; | 1950 | dev->features |= NETIF_F_HW_VLAN_TX | NETIF_F_HW_VLAN_RX; |
diff --git a/drivers/net/Kconfig b/drivers/net/Kconfig index 713ab05a87c0..43d03178064d 100644 --- a/drivers/net/Kconfig +++ b/drivers/net/Kconfig | |||
| @@ -205,7 +205,7 @@ config MII | |||
| 205 | config MACB | 205 | config MACB |
| 206 | tristate "Atmel MACB support" | 206 | tristate "Atmel MACB support" |
| 207 | depends on AVR32 || ARCH_AT91SAM9260 || ARCH_AT91SAM9263 | 207 | depends on AVR32 || ARCH_AT91SAM9260 || ARCH_AT91SAM9263 |
| 208 | select MII | 208 | select PHYLIB |
| 209 | help | 209 | help |
| 210 | The Atmel MACB ethernet interface is found on many AT32 and AT91 | 210 | The Atmel MACB ethernet interface is found on many AT32 and AT91 |
| 211 | parts. Say Y to include support for the MACB chip. | 211 | parts. Say Y to include support for the MACB chip. |
diff --git a/drivers/net/arm/Kconfig b/drivers/net/arm/Kconfig index 5bf2d33887ac..f9cc2b621fe2 100644 --- a/drivers/net/arm/Kconfig +++ b/drivers/net/arm/Kconfig | |||
| @@ -43,6 +43,7 @@ config ARM_AT91_ETHER | |||
| 43 | config EP93XX_ETH | 43 | config EP93XX_ETH |
| 44 | tristate "EP93xx Ethernet support" | 44 | tristate "EP93xx Ethernet support" |
| 45 | depends on ARM && ARCH_EP93XX | 45 | depends on ARM && ARCH_EP93XX |
| 46 | select MII | ||
| 46 | help | 47 | help |
| 47 | This is a driver for the ethernet hardware included in EP93xx CPUs. | 48 | This is a driver for the ethernet hardware included in EP93xx CPUs. |
| 48 | Say Y if you are building a kernel for EP93xx based devices. | 49 | Say Y if you are building a kernel for EP93xx based devices. |
diff --git a/drivers/net/atl1/atl1.h b/drivers/net/atl1/atl1.h index df4c1a0071aa..ff4765f6c3de 100644 --- a/drivers/net/atl1/atl1.h +++ b/drivers/net/atl1/atl1.h | |||
| @@ -43,6 +43,7 @@ extern const struct ethtool_ops atl1_ethtool_ops; | |||
| 43 | struct atl1_adapter; | 43 | struct atl1_adapter; |
| 44 | 44 | ||
| 45 | #define ATL1_MAX_INTR 3 | 45 | #define ATL1_MAX_INTR 3 |
| 46 | #define ATL1_MAX_TX_BUF_LEN 0x3000 /* 12288 bytes */ | ||
| 46 | 47 | ||
| 47 | #define ATL1_DEFAULT_TPD 256 | 48 | #define ATL1_DEFAULT_TPD 256 |
| 48 | #define ATL1_MAX_TPD 1024 | 49 | #define ATL1_MAX_TPD 1024 |
| @@ -57,29 +58,45 @@ struct atl1_adapter; | |||
| 57 | #define ATL1_RRD_DESC(R, i) ATL1_GET_DESC(R, i, struct rx_return_desc) | 58 | #define ATL1_RRD_DESC(R, i) ATL1_GET_DESC(R, i, struct rx_return_desc) |
| 58 | 59 | ||
| 59 | /* | 60 | /* |
| 61 | * This detached comment is preserved for documentation purposes only. | ||
| 62 | * It was originally attached to some code that got deleted, but seems | ||
| 63 | * important enough to keep around... | ||
| 64 | * | ||
| 65 | * <begin detached comment> | ||
| 60 | * Some workarounds require millisecond delays and are run during interrupt | 66 | * Some workarounds require millisecond delays and are run during interrupt |
| 61 | * context. Most notably, when establishing link, the phy may need tweaking | 67 | * context. Most notably, when establishing link, the phy may need tweaking |
| 62 | * but cannot process phy register reads/writes faster than millisecond | 68 | * but cannot process phy register reads/writes faster than millisecond |
| 63 | * intervals...and we establish link due to a "link status change" interrupt. | 69 | * intervals...and we establish link due to a "link status change" interrupt. |
| 70 | * <end detached comment> | ||
| 71 | */ | ||
| 72 | |||
| 73 | /* | ||
| 74 | * atl1_ring_header represents a single, contiguous block of DMA space | ||
| 75 | * mapped for the three descriptor rings (tpd, rfd, rrd) and the two | ||
| 76 | * message blocks (cmb, smb) described below | ||
| 64 | */ | 77 | */ |
| 78 | struct atl1_ring_header { | ||
| 79 | void *desc; /* virtual address */ | ||
| 80 | dma_addr_t dma; /* physical address*/ | ||
| 81 | unsigned int size; /* length in bytes */ | ||
| 82 | }; | ||
| 65 | 83 | ||
| 66 | /* | 84 | /* |
| 67 | * wrapper around a pointer to a socket buffer, | 85 | * atl1_buffer is wrapper around a pointer to a socket buffer |
| 68 | * so a DMA handle can be stored along with the buffer | 86 | * so a DMA handle can be stored along with the skb |
| 69 | */ | 87 | */ |
| 70 | struct atl1_buffer { | 88 | struct atl1_buffer { |
| 71 | struct sk_buff *skb; | 89 | struct sk_buff *skb; /* socket buffer */ |
| 72 | u16 length; | 90 | u16 length; /* rx buffer length */ |
| 73 | u16 alloced; | 91 | u16 alloced; /* 1 if skb allocated */ |
| 74 | dma_addr_t dma; | 92 | dma_addr_t dma; |
| 75 | }; | 93 | }; |
| 76 | 94 | ||
| 77 | #define MAX_TX_BUF_LEN 0x3000 /* 12KB */ | 95 | /* transmit packet descriptor (tpd) ring */ |
| 78 | |||
| 79 | struct atl1_tpd_ring { | 96 | struct atl1_tpd_ring { |
| 80 | void *desc; /* pointer to the descriptor ring memory */ | 97 | void *desc; /* descriptor ring virtual address */ |
| 81 | dma_addr_t dma; /* physical adress of the descriptor ring */ | 98 | dma_addr_t dma; /* descriptor ring physical address */ |
| 82 | u16 size; /* length of descriptor ring in bytes */ | 99 | u16 size; /* descriptor ring length in bytes */ |
| 83 | u16 count; /* number of descriptors in the ring */ | 100 | u16 count; /* number of descriptors in the ring */ |
| 84 | u16 hw_idx; /* hardware index */ | 101 | u16 hw_idx; /* hardware index */ |
| 85 | atomic_t next_to_clean; | 102 | atomic_t next_to_clean; |
| @@ -87,36 +104,34 @@ struct atl1_tpd_ring { | |||
| 87 | struct atl1_buffer *buffer_info; | 104 | struct atl1_buffer *buffer_info; |
| 88 | }; | 105 | }; |
| 89 | 106 | ||
| 107 | /* receive free descriptor (rfd) ring */ | ||
| 90 | struct atl1_rfd_ring { | 108 | struct atl1_rfd_ring { |
| 91 | void *desc; | 109 | void *desc; /* descriptor ring virtual address */ |
| 92 | dma_addr_t dma; | 110 | dma_addr_t dma; /* descriptor ring physical address */ |
| 93 | u16 size; | 111 | u16 size; /* descriptor ring length in bytes */ |
| 94 | u16 count; | 112 | u16 count; /* number of descriptors in the ring */ |
| 95 | atomic_t next_to_use; | 113 | atomic_t next_to_use; |
| 96 | u16 next_to_clean; | 114 | u16 next_to_clean; |
| 97 | struct atl1_buffer *buffer_info; | 115 | struct atl1_buffer *buffer_info; |
| 98 | }; | 116 | }; |
| 99 | 117 | ||
| 118 | /* receive return descriptor (rrd) ring */ | ||
| 100 | struct atl1_rrd_ring { | 119 | struct atl1_rrd_ring { |
| 101 | void *desc; | 120 | void *desc; /* descriptor ring virtual address */ |
| 102 | dma_addr_t dma; | 121 | dma_addr_t dma; /* descriptor ring physical address */ |
| 103 | unsigned int size; | 122 | unsigned int size; /* descriptor ring length in bytes */ |
| 104 | u16 count; | 123 | u16 count; /* number of descriptors in the ring */ |
| 105 | u16 next_to_use; | 124 | u16 next_to_use; |
| 106 | atomic_t next_to_clean; | 125 | atomic_t next_to_clean; |
| 107 | }; | 126 | }; |
| 108 | 127 | ||
| 109 | struct atl1_ring_header { | 128 | /* coalescing message block (cmb) */ |
| 110 | void *desc; /* pointer to the descriptor ring memory */ | ||
| 111 | dma_addr_t dma; /* physical adress of the descriptor ring */ | ||
| 112 | unsigned int size; /* length of descriptor ring in bytes */ | ||
| 113 | }; | ||
| 114 | |||
| 115 | struct atl1_cmb { | 129 | struct atl1_cmb { |
| 116 | struct coals_msg_block *cmb; | 130 | struct coals_msg_block *cmb; |
| 117 | dma_addr_t dma; | 131 | dma_addr_t dma; |
| 118 | }; | 132 | }; |
| 119 | 133 | ||
| 134 | /* statistics message block (smb) */ | ||
| 120 | struct atl1_smb { | 135 | struct atl1_smb { |
| 121 | struct stats_msg_block *smb; | 136 | struct stats_msg_block *smb; |
| 122 | dma_addr_t dma; | 137 | dma_addr_t dma; |
| @@ -141,24 +156,26 @@ struct atl1_sft_stats { | |||
| 141 | u64 tx_aborted_errors; | 156 | u64 tx_aborted_errors; |
| 142 | u64 tx_window_errors; | 157 | u64 tx_window_errors; |
| 143 | u64 tx_carrier_errors; | 158 | u64 tx_carrier_errors; |
| 144 | 159 | u64 tx_pause; /* num pause packets transmitted. */ | |
| 145 | u64 tx_pause; /* num Pause packet transmitted. */ | 160 | u64 excecol; /* num tx packets w/ excessive collisions. */ |
| 146 | u64 excecol; /* num tx packets aborted due to excessive collisions. */ | 161 | u64 deffer; /* num tx packets deferred */ |
| 147 | u64 deffer; /* num deferred tx packets */ | 162 | u64 scc; /* num packets subsequently transmitted |
| 148 | u64 scc; /* num packets subsequently transmitted successfully w/ single prior collision. */ | 163 | * successfully w/ single prior collision. */ |
| 149 | u64 mcc; /* num packets subsequently transmitted successfully w/ multiple prior collisions. */ | 164 | u64 mcc; /* num packets subsequently transmitted |
| 165 | * successfully w/ multiple prior collisions. */ | ||
| 150 | u64 latecol; /* num tx packets w/ late collisions. */ | 166 | u64 latecol; /* num tx packets w/ late collisions. */ |
| 151 | u64 tx_underun; /* num tx packets aborted due to transmit FIFO underrun, or TRD FIFO underrun */ | 167 | u64 tx_underun; /* num tx packets aborted due to transmit |
| 152 | u64 tx_trunc; /* num tx packets truncated due to size exceeding MTU, regardless whether truncated by Selene or not. (The name doesn't really reflect the meaning in this case.) */ | 168 | * FIFO underrun, or TRD FIFO underrun */ |
| 169 | u64 tx_trunc; /* num tx packets truncated due to size | ||
| 170 | * exceeding MTU, regardless whether truncated | ||
| 171 | * by the chip or not. (The name doesn't really | ||
| 172 | * reflect the meaning in this case.) */ | ||
| 153 | u64 rx_pause; /* num Pause packets received. */ | 173 | u64 rx_pause; /* num Pause packets received. */ |
| 154 | u64 rx_rrd_ov; | 174 | u64 rx_rrd_ov; |
| 155 | u64 rx_trunc; | 175 | u64 rx_trunc; |
| 156 | }; | 176 | }; |
| 157 | 177 | ||
| 158 | /* board specific private data structure */ | 178 | /* hardware structure */ |
| 159 | #define ATL1_REGS_LEN 8 | ||
| 160 | |||
| 161 | /* Structure containing variables used by the shared code */ | ||
| 162 | struct atl1_hw { | 179 | struct atl1_hw { |
| 163 | u8 __iomem *hw_addr; | 180 | u8 __iomem *hw_addr; |
| 164 | struct atl1_adapter *back; | 181 | struct atl1_adapter *back; |
| @@ -167,24 +184,35 @@ struct atl1_hw { | |||
| 167 | enum atl1_dma_req_block dmar_block; | 184 | enum atl1_dma_req_block dmar_block; |
| 168 | enum atl1_dma_req_block dmaw_block; | 185 | enum atl1_dma_req_block dmaw_block; |
| 169 | u8 preamble_len; | 186 | u8 preamble_len; |
| 170 | u8 max_retry; /* Retransmission maximum, after which the packet will be discarded */ | 187 | u8 max_retry; /* Retransmission maximum, after which the |
| 171 | u8 jam_ipg; /* IPG to start JAM for collision based flow control in half-duplex mode. In units of 8-bit time */ | 188 | * packet will be discarded */ |
| 172 | u8 ipgt; /* Desired back to back inter-packet gap. The default is 96-bit time */ | 189 | u8 jam_ipg; /* IPG to start JAM for collision based flow |
| 173 | u8 min_ifg; /* Minimum number of IFG to enforce in between RX frames. Frame gap below such IFP is dropped */ | 190 | * control in half-duplex mode. In units of |
| 191 | * 8-bit time */ | ||
| 192 | u8 ipgt; /* Desired back to back inter-packet gap. | ||
| 193 | * The default is 96-bit time */ | ||
| 194 | u8 min_ifg; /* Minimum number of IFG to enforce in between | ||
| 195 | * receive frames. Frame gap below such IFP | ||
| 196 | * is dropped */ | ||
| 174 | u8 ipgr1; /* 64bit Carrier-Sense window */ | 197 | u8 ipgr1; /* 64bit Carrier-Sense window */ |
| 175 | u8 ipgr2; /* 96-bit IPG window */ | 198 | u8 ipgr2; /* 96-bit IPG window */ |
| 176 | u8 tpd_burst; /* Number of TPD to prefetch in cache-aligned burst. Each TPD is 16 bytes long */ | 199 | u8 tpd_burst; /* Number of TPD to prefetch in cache-aligned |
| 177 | u8 rfd_burst; /* Number of RFD to prefetch in cache-aligned burst. Each RFD is 12 bytes long */ | 200 | * burst. Each TPD is 16 bytes long */ |
| 201 | u8 rfd_burst; /* Number of RFD to prefetch in cache-aligned | ||
| 202 | * burst. Each RFD is 12 bytes long */ | ||
| 178 | u8 rfd_fetch_gap; | 203 | u8 rfd_fetch_gap; |
| 179 | u8 rrd_burst; /* Threshold number of RRDs that can be retired in a burst. Each RRD is 16 bytes long */ | 204 | u8 rrd_burst; /* Threshold number of RRDs that can be retired |
| 205 | * in a burst. Each RRD is 16 bytes long */ | ||
| 180 | u8 tpd_fetch_th; | 206 | u8 tpd_fetch_th; |
| 181 | u8 tpd_fetch_gap; | 207 | u8 tpd_fetch_gap; |
| 182 | u16 tx_jumbo_task_th; | 208 | u16 tx_jumbo_task_th; |
| 183 | u16 txf_burst; /* Number of data bytes to read in a cache-aligned burst. Each SRAM entry is | 209 | u16 txf_burst; /* Number of data bytes to read in a cache- |
| 184 | 8 bytes long */ | 210 | * aligned burst. Each SRAM entry is 8 bytes */ |
| 185 | u16 rx_jumbo_th; /* Jumbo packet size for non-VLAN packet. VLAN packets should add 4 bytes */ | 211 | u16 rx_jumbo_th; /* Jumbo packet size for non-VLAN packet. VLAN |
| 212 | * packets should add 4 bytes */ | ||
| 186 | u16 rx_jumbo_lkah; | 213 | u16 rx_jumbo_lkah; |
| 187 | u16 rrd_ret_timer; /* RRD retirement timer. Decrement by 1 after every 512ns passes. */ | 214 | u16 rrd_ret_timer; /* RRD retirement timer. Decrement by 1 after |
| 215 | * every 512ns passes. */ | ||
| 188 | u16 lcol; /* Collision Window */ | 216 | u16 lcol; /* Collision Window */ |
| 189 | 217 | ||
| 190 | u16 cmb_tpd; | 218 | u16 cmb_tpd; |
| @@ -194,48 +222,35 @@ struct atl1_hw { | |||
| 194 | u32 smb_timer; | 222 | u32 smb_timer; |
| 195 | u16 media_type; | 223 | u16 media_type; |
| 196 | u16 autoneg_advertised; | 224 | u16 autoneg_advertised; |
| 197 | u16 pci_cmd_word; | ||
| 198 | 225 | ||
| 199 | u16 mii_autoneg_adv_reg; | 226 | u16 mii_autoneg_adv_reg; |
| 200 | u16 mii_1000t_ctrl_reg; | 227 | u16 mii_1000t_ctrl_reg; |
| 201 | 228 | ||
| 202 | u32 mem_rang; | ||
| 203 | u32 txcw; | ||
| 204 | u32 max_frame_size; | 229 | u32 max_frame_size; |
| 205 | u32 min_frame_size; | 230 | u32 min_frame_size; |
| 206 | u32 mc_filter_type; | ||
| 207 | u32 num_mc_addrs; | ||
| 208 | u32 collision_delta; | ||
| 209 | u32 tx_packet_delta; | ||
| 210 | u16 phy_spd_default; | ||
| 211 | 231 | ||
| 212 | u16 dev_rev; | 232 | u16 dev_rev; |
| 213 | 233 | ||
| 214 | /* spi flash */ | 234 | /* spi flash */ |
| 215 | u8 flash_vendor; | 235 | u8 flash_vendor; |
| 216 | 236 | ||
| 217 | u8 dma_fairness; | ||
| 218 | u8 mac_addr[ETH_ALEN]; | 237 | u8 mac_addr[ETH_ALEN]; |
| 219 | u8 perm_mac_addr[ETH_ALEN]; | 238 | u8 perm_mac_addr[ETH_ALEN]; |
| 220 | 239 | ||
| 221 | /* bool phy_preamble_sup; */ | ||
| 222 | bool phy_configured; | 240 | bool phy_configured; |
| 223 | }; | 241 | }; |
| 224 | 242 | ||
| 225 | struct atl1_adapter { | 243 | struct atl1_adapter { |
| 226 | /* OS defined structs */ | ||
| 227 | struct net_device *netdev; | 244 | struct net_device *netdev; |
| 228 | struct pci_dev *pdev; | 245 | struct pci_dev *pdev; |
| 229 | struct net_device_stats net_stats; | 246 | struct net_device_stats net_stats; |
| 230 | struct atl1_sft_stats soft_stats; | 247 | struct atl1_sft_stats soft_stats; |
| 231 | |||
| 232 | struct vlan_group *vlgrp; | 248 | struct vlan_group *vlgrp; |
| 233 | u32 rx_buffer_len; | 249 | u32 rx_buffer_len; |
| 234 | u32 wol; | 250 | u32 wol; |
| 235 | u16 link_speed; | 251 | u16 link_speed; |
| 236 | u16 link_duplex; | 252 | u16 link_duplex; |
| 237 | spinlock_t lock; | 253 | spinlock_t lock; |
| 238 | atomic_t irq_sem; | ||
| 239 | struct work_struct tx_timeout_task; | 254 | struct work_struct tx_timeout_task; |
| 240 | struct work_struct link_chg_task; | 255 | struct work_struct link_chg_task; |
| 241 | struct work_struct pcie_dma_to_rst_task; | 256 | struct work_struct pcie_dma_to_rst_task; |
| @@ -243,9 +258,7 @@ struct atl1_adapter { | |||
| 243 | struct timer_list phy_config_timer; | 258 | struct timer_list phy_config_timer; |
| 244 | bool phy_timer_pending; | 259 | bool phy_timer_pending; |
| 245 | 260 | ||
| 246 | bool mac_disabled; | 261 | /* all descriptor rings' memory */ |
| 247 | |||
| 248 | /* All descriptor rings' memory */ | ||
| 249 | struct atl1_ring_header ring_header; | 262 | struct atl1_ring_header ring_header; |
| 250 | 263 | ||
| 251 | /* TX */ | 264 | /* TX */ |
| @@ -258,25 +271,16 @@ struct atl1_adapter { | |||
| 258 | u64 hw_csum_err; | 271 | u64 hw_csum_err; |
| 259 | u64 hw_csum_good; | 272 | u64 hw_csum_good; |
| 260 | 273 | ||
| 261 | u32 gorcl; | 274 | u16 imt; /* interrupt moderator timer (2us resolution */ |
| 262 | u64 gorcl_old; | 275 | u16 ict; /* interrupt clear timer (2us resolution */ |
| 263 | 276 | struct mii_if_info mii; /* MII interface info */ | |
| 264 | /* Interrupt Moderator timer ( 2us resolution) */ | ||
| 265 | u16 imt; | ||
| 266 | /* Interrupt Clear timer (2us resolution) */ | ||
| 267 | u16 ict; | ||
| 268 | |||
| 269 | /* MII interface info */ | ||
| 270 | struct mii_if_info mii; | ||
| 271 | 277 | ||
| 272 | /* structs defined in atl1_hw.h */ | 278 | /* structs defined in atl1_hw.h */ |
| 273 | u32 bd_number; /* board number */ | 279 | u32 bd_number; /* board number */ |
| 274 | bool pci_using_64; | 280 | bool pci_using_64; |
| 275 | struct atl1_hw hw; | 281 | struct atl1_hw hw; |
| 276 | struct atl1_smb smb; | 282 | struct atl1_smb smb; |
| 277 | struct atl1_cmb cmb; | 283 | struct atl1_cmb cmb; |
| 278 | |||
| 279 | u32 pci_state[16]; | ||
| 280 | }; | 284 | }; |
| 281 | 285 | ||
| 282 | #endif /* _ATL1_H_ */ | 286 | #endif /* _ATL1_H_ */ |
diff --git a/drivers/net/atl1/atl1_main.c b/drivers/net/atl1/atl1_main.c index 501919eb7f5e..4a18b881ae9a 100644 --- a/drivers/net/atl1/atl1_main.c +++ b/drivers/net/atl1/atl1_main.c | |||
| @@ -38,7 +38,7 @@ | |||
| 38 | * TODO: | 38 | * TODO: |
| 39 | * Fix TSO; tx performance is horrible with TSO enabled. | 39 | * Fix TSO; tx performance is horrible with TSO enabled. |
| 40 | * Wake on LAN. | 40 | * Wake on LAN. |
| 41 | * Add more ethtool functions, including set ring parameters. | 41 | * Add more ethtool functions. |
| 42 | * Fix abstruse irq enable/disable condition described here: | 42 | * Fix abstruse irq enable/disable condition described here: |
| 43 | * http://marc.theaimsgroup.com/?l=linux-netdev&m=116398508500553&w=2 | 43 | * http://marc.theaimsgroup.com/?l=linux-netdev&m=116398508500553&w=2 |
| 44 | * | 44 | * |
| @@ -158,13 +158,70 @@ static int __devinit atl1_sw_init(struct atl1_adapter *adapter) | |||
| 158 | hw->cmb_tx_timer = 1; /* about 2us */ | 158 | hw->cmb_tx_timer = 1; /* about 2us */ |
| 159 | hw->smb_timer = 100000; /* about 200ms */ | 159 | hw->smb_timer = 100000; /* about 200ms */ |
| 160 | 160 | ||
| 161 | atomic_set(&adapter->irq_sem, 0); | ||
| 162 | spin_lock_init(&adapter->lock); | 161 | spin_lock_init(&adapter->lock); |
| 163 | spin_lock_init(&adapter->mb_lock); | 162 | spin_lock_init(&adapter->mb_lock); |
| 164 | 163 | ||
| 165 | return 0; | 164 | return 0; |
| 166 | } | 165 | } |
| 167 | 166 | ||
| 167 | static int mdio_read(struct net_device *netdev, int phy_id, int reg_num) | ||
| 168 | { | ||
| 169 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 170 | u16 result; | ||
| 171 | |||
| 172 | atl1_read_phy_reg(&adapter->hw, reg_num & 0x1f, &result); | ||
| 173 | |||
| 174 | return result; | ||
| 175 | } | ||
| 176 | |||
| 177 | static void mdio_write(struct net_device *netdev, int phy_id, int reg_num, | ||
| 178 | int val) | ||
| 179 | { | ||
| 180 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 181 | |||
| 182 | atl1_write_phy_reg(&adapter->hw, reg_num, val); | ||
| 183 | } | ||
| 184 | |||
| 185 | /* | ||
| 186 | * atl1_mii_ioctl - | ||
| 187 | * @netdev: | ||
| 188 | * @ifreq: | ||
| 189 | * @cmd: | ||
| 190 | */ | ||
| 191 | static int atl1_mii_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) | ||
| 192 | { | ||
| 193 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 194 | unsigned long flags; | ||
| 195 | int retval; | ||
| 196 | |||
| 197 | if (!netif_running(netdev)) | ||
| 198 | return -EINVAL; | ||
| 199 | |||
| 200 | spin_lock_irqsave(&adapter->lock, flags); | ||
| 201 | retval = generic_mii_ioctl(&adapter->mii, if_mii(ifr), cmd, NULL); | ||
| 202 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 203 | |||
| 204 | return retval; | ||
| 205 | } | ||
| 206 | |||
| 207 | /* | ||
| 208 | * atl1_ioctl - | ||
| 209 | * @netdev: | ||
| 210 | * @ifreq: | ||
| 211 | * @cmd: | ||
| 212 | */ | ||
| 213 | static int atl1_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) | ||
| 214 | { | ||
| 215 | switch (cmd) { | ||
| 216 | case SIOCGMIIPHY: | ||
| 217 | case SIOCGMIIREG: | ||
| 218 | case SIOCSMIIREG: | ||
| 219 | return atl1_mii_ioctl(netdev, ifr, cmd); | ||
| 220 | default: | ||
| 221 | return -EOPNOTSUPP; | ||
| 222 | } | ||
| 223 | } | ||
| 224 | |||
| 168 | /* | 225 | /* |
| 169 | * atl1_setup_mem_resources - allocate Tx / RX descriptor resources | 226 | * atl1_setup_mem_resources - allocate Tx / RX descriptor resources |
| 170 | * @adapter: board private structure | 227 | * @adapter: board private structure |
| @@ -188,19 +245,22 @@ s32 atl1_setup_ring_resources(struct atl1_adapter *adapter) | |||
| 188 | goto err_nomem; | 245 | goto err_nomem; |
| 189 | } | 246 | } |
| 190 | rfd_ring->buffer_info = | 247 | rfd_ring->buffer_info = |
| 191 | (struct atl1_buffer *)(tpd_ring->buffer_info + tpd_ring->count); | 248 | (struct atl1_buffer *)(tpd_ring->buffer_info + tpd_ring->count); |
| 192 | 249 | ||
| 193 | /* real ring DMA buffer */ | 250 | /* real ring DMA buffer |
| 194 | ring_header->size = size = sizeof(struct tx_packet_desc) * | 251 | * each ring/block may need up to 8 bytes for alignment, hence the |
| 195 | tpd_ring->count | 252 | * additional 40 bytes tacked onto the end. |
| 196 | + sizeof(struct rx_free_desc) * rfd_ring->count | 253 | */ |
| 197 | + sizeof(struct rx_return_desc) * rrd_ring->count | 254 | ring_header->size = size = |
| 198 | + sizeof(struct coals_msg_block) | 255 | sizeof(struct tx_packet_desc) * tpd_ring->count |
| 199 | + sizeof(struct stats_msg_block) | 256 | + sizeof(struct rx_free_desc) * rfd_ring->count |
| 200 | + 40; /* "40: for 8 bytes align" huh? -- CHS */ | 257 | + sizeof(struct rx_return_desc) * rrd_ring->count |
| 258 | + sizeof(struct coals_msg_block) | ||
| 259 | + sizeof(struct stats_msg_block) | ||
| 260 | + 40; | ||
| 201 | 261 | ||
| 202 | ring_header->desc = pci_alloc_consistent(pdev, ring_header->size, | 262 | ring_header->desc = pci_alloc_consistent(pdev, ring_header->size, |
| 203 | &ring_header->dma); | 263 | &ring_header->dma); |
| 204 | if (unlikely(!ring_header->desc)) { | 264 | if (unlikely(!ring_header->desc)) { |
| 205 | dev_err(&pdev->dev, "pci_alloc_consistent failed\n"); | 265 | dev_err(&pdev->dev, "pci_alloc_consistent failed\n"); |
| 206 | goto err_nomem; | 266 | goto err_nomem; |
| @@ -214,8 +274,6 @@ s32 atl1_setup_ring_resources(struct atl1_adapter *adapter) | |||
| 214 | tpd_ring->dma += offset; | 274 | tpd_ring->dma += offset; |
| 215 | tpd_ring->desc = (u8 *) ring_header->desc + offset; | 275 | tpd_ring->desc = (u8 *) ring_header->desc + offset; |
| 216 | tpd_ring->size = sizeof(struct tx_packet_desc) * tpd_ring->count; | 276 | tpd_ring->size = sizeof(struct tx_packet_desc) * tpd_ring->count; |
| 217 | atomic_set(&tpd_ring->next_to_use, 0); | ||
| 218 | atomic_set(&tpd_ring->next_to_clean, 0); | ||
| 219 | 277 | ||
| 220 | /* init RFD ring */ | 278 | /* init RFD ring */ |
| 221 | rfd_ring->dma = tpd_ring->dma + tpd_ring->size; | 279 | rfd_ring->dma = tpd_ring->dma + tpd_ring->size; |
| @@ -223,9 +281,7 @@ s32 atl1_setup_ring_resources(struct atl1_adapter *adapter) | |||
| 223 | rfd_ring->dma += offset; | 281 | rfd_ring->dma += offset; |
| 224 | rfd_ring->desc = (u8 *) tpd_ring->desc + (tpd_ring->size + offset); | 282 | rfd_ring->desc = (u8 *) tpd_ring->desc + (tpd_ring->size + offset); |
| 225 | rfd_ring->size = sizeof(struct rx_free_desc) * rfd_ring->count; | 283 | rfd_ring->size = sizeof(struct rx_free_desc) * rfd_ring->count; |
| 226 | rfd_ring->next_to_clean = 0; | 284 | |
| 227 | /* rfd_ring->next_to_use = rfd_ring->count - 1; */ | ||
| 228 | atomic_set(&rfd_ring->next_to_use, 0); | ||
| 229 | 285 | ||
| 230 | /* init RRD ring */ | 286 | /* init RRD ring */ |
| 231 | rrd_ring->dma = rfd_ring->dma + rfd_ring->size; | 287 | rrd_ring->dma = rfd_ring->dma + rfd_ring->size; |
| @@ -233,23 +289,22 @@ s32 atl1_setup_ring_resources(struct atl1_adapter *adapter) | |||
| 233 | rrd_ring->dma += offset; | 289 | rrd_ring->dma += offset; |
| 234 | rrd_ring->desc = (u8 *) rfd_ring->desc + (rfd_ring->size + offset); | 290 | rrd_ring->desc = (u8 *) rfd_ring->desc + (rfd_ring->size + offset); |
| 235 | rrd_ring->size = sizeof(struct rx_return_desc) * rrd_ring->count; | 291 | rrd_ring->size = sizeof(struct rx_return_desc) * rrd_ring->count; |
| 236 | rrd_ring->next_to_use = 0; | 292 | |
| 237 | atomic_set(&rrd_ring->next_to_clean, 0); | ||
| 238 | 293 | ||
| 239 | /* init CMB */ | 294 | /* init CMB */ |
| 240 | adapter->cmb.dma = rrd_ring->dma + rrd_ring->size; | 295 | adapter->cmb.dma = rrd_ring->dma + rrd_ring->size; |
| 241 | offset = (adapter->cmb.dma & 0x7) ? (8 - (adapter->cmb.dma & 0x7)) : 0; | 296 | offset = (adapter->cmb.dma & 0x7) ? (8 - (adapter->cmb.dma & 0x7)) : 0; |
| 242 | adapter->cmb.dma += offset; | 297 | adapter->cmb.dma += offset; |
| 243 | adapter->cmb.cmb = | 298 | adapter->cmb.cmb = (struct coals_msg_block *) |
| 244 | (struct coals_msg_block *) ((u8 *) rrd_ring->desc + | 299 | ((u8 *) rrd_ring->desc + (rrd_ring->size + offset)); |
| 245 | (rrd_ring->size + offset)); | ||
| 246 | 300 | ||
| 247 | /* init SMB */ | 301 | /* init SMB */ |
| 248 | adapter->smb.dma = adapter->cmb.dma + sizeof(struct coals_msg_block); | 302 | adapter->smb.dma = adapter->cmb.dma + sizeof(struct coals_msg_block); |
| 249 | offset = (adapter->smb.dma & 0x7) ? (8 - (adapter->smb.dma & 0x7)) : 0; | 303 | offset = (adapter->smb.dma & 0x7) ? (8 - (adapter->smb.dma & 0x7)) : 0; |
| 250 | adapter->smb.dma += offset; | 304 | adapter->smb.dma += offset; |
| 251 | adapter->smb.smb = (struct stats_msg_block *) | 305 | adapter->smb.smb = (struct stats_msg_block *) |
| 252 | ((u8 *) adapter->cmb.cmb + (sizeof(struct coals_msg_block) + offset)); | 306 | ((u8 *) adapter->cmb.cmb + |
| 307 | (sizeof(struct coals_msg_block) + offset)); | ||
| 253 | 308 | ||
| 254 | return ATL1_SUCCESS; | 309 | return ATL1_SUCCESS; |
| 255 | 310 | ||
| @@ -258,559 +313,133 @@ err_nomem: | |||
| 258 | return -ENOMEM; | 313 | return -ENOMEM; |
| 259 | } | 314 | } |
| 260 | 315 | ||
| 261 | /* | 316 | void atl1_init_ring_ptrs(struct atl1_adapter *adapter) |
| 262 | * atl1_irq_enable - Enable default interrupt generation settings | ||
| 263 | * @adapter: board private structure | ||
| 264 | */ | ||
| 265 | static void atl1_irq_enable(struct atl1_adapter *adapter) | ||
| 266 | { | ||
| 267 | if (likely(!atomic_dec_and_test(&adapter->irq_sem))) | ||
| 268 | iowrite32(IMR_NORMAL_MASK, adapter->hw.hw_addr + REG_IMR); | ||
| 269 | } | ||
| 270 | |||
| 271 | static void atl1_clear_phy_int(struct atl1_adapter *adapter) | ||
| 272 | { | ||
| 273 | u16 phy_data; | ||
| 274 | unsigned long flags; | ||
| 275 | |||
| 276 | spin_lock_irqsave(&adapter->lock, flags); | ||
| 277 | atl1_read_phy_reg(&adapter->hw, 19, &phy_data); | ||
| 278 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 279 | } | ||
| 280 | |||
| 281 | static void atl1_inc_smb(struct atl1_adapter *adapter) | ||
| 282 | { | ||
| 283 | struct stats_msg_block *smb = adapter->smb.smb; | ||
| 284 | |||
| 285 | /* Fill out the OS statistics structure */ | ||
| 286 | adapter->soft_stats.rx_packets += smb->rx_ok; | ||
| 287 | adapter->soft_stats.tx_packets += smb->tx_ok; | ||
| 288 | adapter->soft_stats.rx_bytes += smb->rx_byte_cnt; | ||
| 289 | adapter->soft_stats.tx_bytes += smb->tx_byte_cnt; | ||
| 290 | adapter->soft_stats.multicast += smb->rx_mcast; | ||
| 291 | adapter->soft_stats.collisions += (smb->tx_1_col + | ||
| 292 | smb->tx_2_col * 2 + | ||
| 293 | smb->tx_late_col + | ||
| 294 | smb->tx_abort_col * | ||
| 295 | adapter->hw.max_retry); | ||
| 296 | |||
| 297 | /* Rx Errors */ | ||
| 298 | adapter->soft_stats.rx_errors += (smb->rx_frag + | ||
| 299 | smb->rx_fcs_err + | ||
| 300 | smb->rx_len_err + | ||
| 301 | smb->rx_sz_ov + | ||
| 302 | smb->rx_rxf_ov + | ||
| 303 | smb->rx_rrd_ov + smb->rx_align_err); | ||
| 304 | adapter->soft_stats.rx_fifo_errors += smb->rx_rxf_ov; | ||
| 305 | adapter->soft_stats.rx_length_errors += smb->rx_len_err; | ||
| 306 | adapter->soft_stats.rx_crc_errors += smb->rx_fcs_err; | ||
| 307 | adapter->soft_stats.rx_frame_errors += smb->rx_align_err; | ||
| 308 | adapter->soft_stats.rx_missed_errors += (smb->rx_rrd_ov + | ||
| 309 | smb->rx_rxf_ov); | ||
| 310 | |||
| 311 | adapter->soft_stats.rx_pause += smb->rx_pause; | ||
| 312 | adapter->soft_stats.rx_rrd_ov += smb->rx_rrd_ov; | ||
| 313 | adapter->soft_stats.rx_trunc += smb->rx_sz_ov; | ||
| 314 | |||
| 315 | /* Tx Errors */ | ||
| 316 | adapter->soft_stats.tx_errors += (smb->tx_late_col + | ||
| 317 | smb->tx_abort_col + | ||
| 318 | smb->tx_underrun + smb->tx_trunc); | ||
| 319 | adapter->soft_stats.tx_fifo_errors += smb->tx_underrun; | ||
| 320 | adapter->soft_stats.tx_aborted_errors += smb->tx_abort_col; | ||
| 321 | adapter->soft_stats.tx_window_errors += smb->tx_late_col; | ||
| 322 | |||
| 323 | adapter->soft_stats.excecol += smb->tx_abort_col; | ||
| 324 | adapter->soft_stats.deffer += smb->tx_defer; | ||
| 325 | adapter->soft_stats.scc += smb->tx_1_col; | ||
| 326 | adapter->soft_stats.mcc += smb->tx_2_col; | ||
| 327 | adapter->soft_stats.latecol += smb->tx_late_col; | ||
| 328 | adapter->soft_stats.tx_underun += smb->tx_underrun; | ||
| 329 | adapter->soft_stats.tx_trunc += smb->tx_trunc; | ||
| 330 | adapter->soft_stats.tx_pause += smb->tx_pause; | ||
| 331 | |||
| 332 | adapter->net_stats.rx_packets = adapter->soft_stats.rx_packets; | ||
| 333 | adapter->net_stats.tx_packets = adapter->soft_stats.tx_packets; | ||
| 334 | adapter->net_stats.rx_bytes = adapter->soft_stats.rx_bytes; | ||
| 335 | adapter->net_stats.tx_bytes = adapter->soft_stats.tx_bytes; | ||
| 336 | adapter->net_stats.multicast = adapter->soft_stats.multicast; | ||
| 337 | adapter->net_stats.collisions = adapter->soft_stats.collisions; | ||
| 338 | adapter->net_stats.rx_errors = adapter->soft_stats.rx_errors; | ||
| 339 | adapter->net_stats.rx_over_errors = | ||
| 340 | adapter->soft_stats.rx_missed_errors; | ||
| 341 | adapter->net_stats.rx_length_errors = | ||
| 342 | adapter->soft_stats.rx_length_errors; | ||
| 343 | adapter->net_stats.rx_crc_errors = adapter->soft_stats.rx_crc_errors; | ||
| 344 | adapter->net_stats.rx_frame_errors = | ||
| 345 | adapter->soft_stats.rx_frame_errors; | ||
| 346 | adapter->net_stats.rx_fifo_errors = adapter->soft_stats.rx_fifo_errors; | ||
| 347 | adapter->net_stats.rx_missed_errors = | ||
| 348 | adapter->soft_stats.rx_missed_errors; | ||
| 349 | adapter->net_stats.tx_errors = adapter->soft_stats.tx_errors; | ||
| 350 | adapter->net_stats.tx_fifo_errors = adapter->soft_stats.tx_fifo_errors; | ||
| 351 | adapter->net_stats.tx_aborted_errors = | ||
| 352 | adapter->soft_stats.tx_aborted_errors; | ||
| 353 | adapter->net_stats.tx_window_errors = | ||
| 354 | adapter->soft_stats.tx_window_errors; | ||
| 355 | adapter->net_stats.tx_carrier_errors = | ||
| 356 | adapter->soft_stats.tx_carrier_errors; | ||
| 357 | } | ||
| 358 | |||
| 359 | static void atl1_rx_checksum(struct atl1_adapter *adapter, | ||
| 360 | struct rx_return_desc *rrd, | ||
| 361 | struct sk_buff *skb) | ||
| 362 | { | 317 | { |
| 363 | skb->ip_summed = CHECKSUM_NONE; | 318 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; |
| 364 | 319 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | |
| 365 | if (unlikely(rrd->pkt_flg & PACKET_FLAG_ERR)) { | 320 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; |
| 366 | if (rrd->err_flg & (ERR_FLAG_CRC | ERR_FLAG_TRUNC | | ||
| 367 | ERR_FLAG_CODE | ERR_FLAG_OV)) { | ||
| 368 | adapter->hw_csum_err++; | ||
| 369 | dev_dbg(&adapter->pdev->dev, "rx checksum error\n"); | ||
| 370 | return; | ||
| 371 | } | ||
| 372 | } | ||
| 373 | 321 | ||
| 374 | /* not IPv4 */ | 322 | atomic_set(&tpd_ring->next_to_use, 0); |
| 375 | if (!(rrd->pkt_flg & PACKET_FLAG_IPV4)) | 323 | atomic_set(&tpd_ring->next_to_clean, 0); |
| 376 | /* checksum is invalid, but it's not an IPv4 pkt, so ok */ | ||
| 377 | return; | ||
| 378 | 324 | ||
| 379 | /* IPv4 packet */ | 325 | rfd_ring->next_to_clean = 0; |
| 380 | if (likely(!(rrd->err_flg & | 326 | atomic_set(&rfd_ring->next_to_use, 0); |
| 381 | (ERR_FLAG_IP_CHKSUM | ERR_FLAG_L4_CHKSUM)))) { | ||
| 382 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
| 383 | adapter->hw_csum_good++; | ||
| 384 | return; | ||
| 385 | } | ||
| 386 | 327 | ||
| 387 | /* IPv4, but hardware thinks its checksum is wrong */ | 328 | rrd_ring->next_to_use = 0; |
| 388 | dev_dbg(&adapter->pdev->dev, | 329 | atomic_set(&rrd_ring->next_to_clean, 0); |
| 389 | "hw csum wrong, pkt_flag:%x, err_flag:%x\n", | ||
| 390 | rrd->pkt_flg, rrd->err_flg); | ||
| 391 | skb->ip_summed = CHECKSUM_COMPLETE; | ||
| 392 | skb->csum = htons(rrd->xsz.xsum_sz.rx_chksum); | ||
| 393 | adapter->hw_csum_err++; | ||
| 394 | return; | ||
| 395 | } | 330 | } |
| 396 | 331 | ||
| 397 | /* | 332 | /* |
| 398 | * atl1_alloc_rx_buffers - Replace used receive buffers | 333 | * atl1_clean_rx_ring - Free RFD Buffers |
| 399 | * @adapter: address of board private structure | 334 | * @adapter: board private structure |
| 400 | */ | 335 | */ |
| 401 | static u16 atl1_alloc_rx_buffers(struct atl1_adapter *adapter) | 336 | static void atl1_clean_rx_ring(struct atl1_adapter *adapter) |
| 402 | { | ||
| 403 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | ||
| 404 | struct pci_dev *pdev = adapter->pdev; | ||
| 405 | struct page *page; | ||
| 406 | unsigned long offset; | ||
| 407 | struct atl1_buffer *buffer_info, *next_info; | ||
| 408 | struct sk_buff *skb; | ||
| 409 | u16 num_alloc = 0; | ||
| 410 | u16 rfd_next_to_use, next_next; | ||
| 411 | struct rx_free_desc *rfd_desc; | ||
| 412 | |||
| 413 | next_next = rfd_next_to_use = atomic_read(&rfd_ring->next_to_use); | ||
| 414 | if (++next_next == rfd_ring->count) | ||
| 415 | next_next = 0; | ||
| 416 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; | ||
| 417 | next_info = &rfd_ring->buffer_info[next_next]; | ||
| 418 | |||
| 419 | while (!buffer_info->alloced && !next_info->alloced) { | ||
| 420 | if (buffer_info->skb) { | ||
| 421 | buffer_info->alloced = 1; | ||
| 422 | goto next; | ||
| 423 | } | ||
| 424 | |||
| 425 | rfd_desc = ATL1_RFD_DESC(rfd_ring, rfd_next_to_use); | ||
| 426 | |||
| 427 | skb = dev_alloc_skb(adapter->rx_buffer_len + NET_IP_ALIGN); | ||
| 428 | if (unlikely(!skb)) { /* Better luck next round */ | ||
| 429 | adapter->net_stats.rx_dropped++; | ||
| 430 | break; | ||
| 431 | } | ||
| 432 | |||
| 433 | /* | ||
| 434 | * Make buffer alignment 2 beyond a 16 byte boundary | ||
| 435 | * this will result in a 16 byte aligned IP header after | ||
| 436 | * the 14 byte MAC header is removed | ||
| 437 | */ | ||
| 438 | skb_reserve(skb, NET_IP_ALIGN); | ||
| 439 | |||
| 440 | buffer_info->alloced = 1; | ||
| 441 | buffer_info->skb = skb; | ||
| 442 | buffer_info->length = (u16) adapter->rx_buffer_len; | ||
| 443 | page = virt_to_page(skb->data); | ||
| 444 | offset = (unsigned long)skb->data & ~PAGE_MASK; | ||
| 445 | buffer_info->dma = pci_map_page(pdev, page, offset, | ||
| 446 | adapter->rx_buffer_len, | ||
| 447 | PCI_DMA_FROMDEVICE); | ||
| 448 | rfd_desc->buffer_addr = cpu_to_le64(buffer_info->dma); | ||
| 449 | rfd_desc->buf_len = cpu_to_le16(adapter->rx_buffer_len); | ||
| 450 | rfd_desc->coalese = 0; | ||
| 451 | |||
| 452 | next: | ||
| 453 | rfd_next_to_use = next_next; | ||
| 454 | if (unlikely(++next_next == rfd_ring->count)) | ||
| 455 | next_next = 0; | ||
| 456 | |||
| 457 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; | ||
| 458 | next_info = &rfd_ring->buffer_info[next_next]; | ||
| 459 | num_alloc++; | ||
| 460 | } | ||
| 461 | |||
| 462 | if (num_alloc) { | ||
| 463 | /* | ||
| 464 | * Force memory writes to complete before letting h/w | ||
| 465 | * know there are new descriptors to fetch. (Only | ||
| 466 | * applicable for weak-ordered memory model archs, | ||
| 467 | * such as IA-64). | ||
| 468 | */ | ||
| 469 | wmb(); | ||
| 470 | atomic_set(&rfd_ring->next_to_use, (int)rfd_next_to_use); | ||
| 471 | } | ||
| 472 | return num_alloc; | ||
| 473 | } | ||
| 474 | |||
| 475 | static void atl1_intr_rx(struct atl1_adapter *adapter) | ||
| 476 | { | 337 | { |
| 477 | int i, count; | ||
| 478 | u16 length; | ||
| 479 | u16 rrd_next_to_clean; | ||
| 480 | u32 value; | ||
| 481 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | 338 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; |
| 482 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; | 339 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; |
| 483 | struct atl1_buffer *buffer_info; | 340 | struct atl1_buffer *buffer_info; |
| 484 | struct rx_return_desc *rrd; | 341 | struct pci_dev *pdev = adapter->pdev; |
| 485 | struct sk_buff *skb; | 342 | unsigned long size; |
| 486 | 343 | unsigned int i; | |
| 487 | count = 0; | ||
| 488 | |||
| 489 | rrd_next_to_clean = atomic_read(&rrd_ring->next_to_clean); | ||
| 490 | |||
| 491 | while (1) { | ||
| 492 | rrd = ATL1_RRD_DESC(rrd_ring, rrd_next_to_clean); | ||
| 493 | i = 1; | ||
| 494 | if (likely(rrd->xsz.valid)) { /* packet valid */ | ||
| 495 | chk_rrd: | ||
| 496 | /* check rrd status */ | ||
| 497 | if (likely(rrd->num_buf == 1)) | ||
| 498 | goto rrd_ok; | ||
| 499 | |||
| 500 | /* rrd seems to be bad */ | ||
| 501 | if (unlikely(i-- > 0)) { | ||
| 502 | /* rrd may not be DMAed completely */ | ||
| 503 | dev_dbg(&adapter->pdev->dev, | ||
| 504 | "incomplete RRD DMA transfer\n"); | ||
| 505 | udelay(1); | ||
| 506 | goto chk_rrd; | ||
| 507 | } | ||
| 508 | /* bad rrd */ | ||
| 509 | dev_dbg(&adapter->pdev->dev, "bad RRD\n"); | ||
| 510 | /* see if update RFD index */ | ||
| 511 | if (rrd->num_buf > 1) { | ||
| 512 | u16 num_buf; | ||
| 513 | num_buf = | ||
| 514 | (rrd->xsz.xsum_sz.pkt_size + | ||
| 515 | adapter->rx_buffer_len - | ||
| 516 | 1) / adapter->rx_buffer_len; | ||
| 517 | if (rrd->num_buf == num_buf) { | ||
| 518 | /* clean alloc flag for bad rrd */ | ||
| 519 | while (rfd_ring->next_to_clean != | ||
| 520 | (rrd->buf_indx + num_buf)) { | ||
| 521 | rfd_ring->buffer_info[rfd_ring-> | ||
| 522 | next_to_clean].alloced = 0; | ||
| 523 | if (++rfd_ring->next_to_clean == | ||
| 524 | rfd_ring->count) { | ||
| 525 | rfd_ring-> | ||
| 526 | next_to_clean = 0; | ||
| 527 | } | ||
| 528 | } | ||
| 529 | } | ||
| 530 | } | ||
| 531 | |||
| 532 | /* update rrd */ | ||
| 533 | rrd->xsz.valid = 0; | ||
| 534 | if (++rrd_next_to_clean == rrd_ring->count) | ||
| 535 | rrd_next_to_clean = 0; | ||
| 536 | count++; | ||
| 537 | continue; | ||
| 538 | } else { /* current rrd still not be updated */ | ||
| 539 | 344 | ||
| 540 | break; | 345 | /* Free all the Rx ring sk_buffs */ |
| 541 | } | 346 | for (i = 0; i < rfd_ring->count; i++) { |
| 542 | rrd_ok: | 347 | buffer_info = &rfd_ring->buffer_info[i]; |
| 543 | /* clean alloc flag for bad rrd */ | 348 | if (buffer_info->dma) { |
| 544 | while (rfd_ring->next_to_clean != rrd->buf_indx) { | 349 | pci_unmap_page(pdev, buffer_info->dma, |
| 545 | rfd_ring->buffer_info[rfd_ring->next_to_clean].alloced = | 350 | buffer_info->length, PCI_DMA_FROMDEVICE); |
| 546 | 0; | 351 | buffer_info->dma = 0; |
| 547 | if (++rfd_ring->next_to_clean == rfd_ring->count) | ||
| 548 | rfd_ring->next_to_clean = 0; | ||
| 549 | } | 352 | } |
| 550 | 353 | if (buffer_info->skb) { | |
| 551 | buffer_info = &rfd_ring->buffer_info[rrd->buf_indx]; | 354 | dev_kfree_skb(buffer_info->skb); |
| 552 | if (++rfd_ring->next_to_clean == rfd_ring->count) | 355 | buffer_info->skb = NULL; |
| 553 | rfd_ring->next_to_clean = 0; | ||
| 554 | |||
| 555 | /* update rrd next to clean */ | ||
| 556 | if (++rrd_next_to_clean == rrd_ring->count) | ||
| 557 | rrd_next_to_clean = 0; | ||
| 558 | count++; | ||
| 559 | |||
| 560 | if (unlikely(rrd->pkt_flg & PACKET_FLAG_ERR)) { | ||
| 561 | if (!(rrd->err_flg & | ||
| 562 | (ERR_FLAG_IP_CHKSUM | ERR_FLAG_L4_CHKSUM | ||
| 563 | | ERR_FLAG_LEN))) { | ||
| 564 | /* packet error, don't need upstream */ | ||
| 565 | buffer_info->alloced = 0; | ||
| 566 | rrd->xsz.valid = 0; | ||
| 567 | continue; | ||
| 568 | } | ||
| 569 | } | 356 | } |
| 570 | |||
| 571 | /* Good Receive */ | ||
| 572 | pci_unmap_page(adapter->pdev, buffer_info->dma, | ||
| 573 | buffer_info->length, PCI_DMA_FROMDEVICE); | ||
| 574 | skb = buffer_info->skb; | ||
| 575 | length = le16_to_cpu(rrd->xsz.xsum_sz.pkt_size); | ||
| 576 | |||
| 577 | skb_put(skb, length - ETHERNET_FCS_SIZE); | ||
| 578 | |||
| 579 | /* Receive Checksum Offload */ | ||
| 580 | atl1_rx_checksum(adapter, rrd, skb); | ||
| 581 | skb->protocol = eth_type_trans(skb, adapter->netdev); | ||
| 582 | |||
| 583 | if (adapter->vlgrp && (rrd->pkt_flg & PACKET_FLAG_VLAN_INS)) { | ||
| 584 | u16 vlan_tag = (rrd->vlan_tag >> 4) | | ||
| 585 | ((rrd->vlan_tag & 7) << 13) | | ||
| 586 | ((rrd->vlan_tag & 8) << 9); | ||
| 587 | vlan_hwaccel_rx(skb, adapter->vlgrp, vlan_tag); | ||
| 588 | } else | ||
| 589 | netif_rx(skb); | ||
| 590 | |||
| 591 | /* let protocol layer free skb */ | ||
| 592 | buffer_info->skb = NULL; | ||
| 593 | buffer_info->alloced = 0; | ||
| 594 | rrd->xsz.valid = 0; | ||
| 595 | |||
| 596 | adapter->netdev->last_rx = jiffies; | ||
| 597 | } | 357 | } |
| 598 | 358 | ||
| 599 | atomic_set(&rrd_ring->next_to_clean, rrd_next_to_clean); | 359 | size = sizeof(struct atl1_buffer) * rfd_ring->count; |
| 600 | 360 | memset(rfd_ring->buffer_info, 0, size); | |
| 601 | atl1_alloc_rx_buffers(adapter); | ||
| 602 | 361 | ||
| 603 | /* update mailbox ? */ | 362 | /* Zero out the descriptor ring */ |
| 604 | if (count) { | 363 | memset(rfd_ring->desc, 0, rfd_ring->size); |
| 605 | u32 tpd_next_to_use; | ||
| 606 | u32 rfd_next_to_use; | ||
| 607 | u32 rrd_next_to_clean; | ||
| 608 | 364 | ||
| 609 | spin_lock(&adapter->mb_lock); | 365 | rfd_ring->next_to_clean = 0; |
| 366 | atomic_set(&rfd_ring->next_to_use, 0); | ||
| 610 | 367 | ||
| 611 | tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); | 368 | rrd_ring->next_to_use = 0; |
| 612 | rfd_next_to_use = | 369 | atomic_set(&rrd_ring->next_to_clean, 0); |
| 613 | atomic_read(&adapter->rfd_ring.next_to_use); | ||
| 614 | rrd_next_to_clean = | ||
| 615 | atomic_read(&adapter->rrd_ring.next_to_clean); | ||
| 616 | value = ((rfd_next_to_use & MB_RFD_PROD_INDX_MASK) << | ||
| 617 | MB_RFD_PROD_INDX_SHIFT) | | ||
| 618 | ((rrd_next_to_clean & MB_RRD_CONS_INDX_MASK) << | ||
| 619 | MB_RRD_CONS_INDX_SHIFT) | | ||
| 620 | ((tpd_next_to_use & MB_TPD_PROD_INDX_MASK) << | ||
| 621 | MB_TPD_PROD_INDX_SHIFT); | ||
| 622 | iowrite32(value, adapter->hw.hw_addr + REG_MAILBOX); | ||
| 623 | spin_unlock(&adapter->mb_lock); | ||
| 624 | } | ||
| 625 | } | 370 | } |
| 626 | 371 | ||
| 627 | static void atl1_intr_tx(struct atl1_adapter *adapter) | 372 | /* |
| 373 | * atl1_clean_tx_ring - Free Tx Buffers | ||
| 374 | * @adapter: board private structure | ||
| 375 | */ | ||
| 376 | static void atl1_clean_tx_ring(struct atl1_adapter *adapter) | ||
| 628 | { | 377 | { |
| 629 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | 378 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; |
| 630 | struct atl1_buffer *buffer_info; | 379 | struct atl1_buffer *buffer_info; |
| 631 | u16 sw_tpd_next_to_clean; | 380 | struct pci_dev *pdev = adapter->pdev; |
| 632 | u16 cmb_tpd_next_to_clean; | 381 | unsigned long size; |
| 633 | 382 | unsigned int i; | |
| 634 | sw_tpd_next_to_clean = atomic_read(&tpd_ring->next_to_clean); | ||
| 635 | cmb_tpd_next_to_clean = le16_to_cpu(adapter->cmb.cmb->tpd_cons_idx); | ||
| 636 | |||
| 637 | while (cmb_tpd_next_to_clean != sw_tpd_next_to_clean) { | ||
| 638 | struct tx_packet_desc *tpd; | ||
| 639 | 383 | ||
| 640 | tpd = ATL1_TPD_DESC(tpd_ring, sw_tpd_next_to_clean); | 384 | /* Free all the Tx ring sk_buffs */ |
| 641 | buffer_info = &tpd_ring->buffer_info[sw_tpd_next_to_clean]; | 385 | for (i = 0; i < tpd_ring->count; i++) { |
| 386 | buffer_info = &tpd_ring->buffer_info[i]; | ||
| 642 | if (buffer_info->dma) { | 387 | if (buffer_info->dma) { |
| 643 | pci_unmap_page(adapter->pdev, buffer_info->dma, | 388 | pci_unmap_page(pdev, buffer_info->dma, |
| 644 | buffer_info->length, PCI_DMA_TODEVICE); | 389 | buffer_info->length, PCI_DMA_TODEVICE); |
| 645 | buffer_info->dma = 0; | 390 | buffer_info->dma = 0; |
| 646 | } | 391 | } |
| 392 | } | ||
| 647 | 393 | ||
| 394 | for (i = 0; i < tpd_ring->count; i++) { | ||
| 395 | buffer_info = &tpd_ring->buffer_info[i]; | ||
| 648 | if (buffer_info->skb) { | 396 | if (buffer_info->skb) { |
| 649 | dev_kfree_skb_irq(buffer_info->skb); | 397 | dev_kfree_skb_any(buffer_info->skb); |
| 650 | buffer_info->skb = NULL; | 398 | buffer_info->skb = NULL; |
| 651 | } | 399 | } |
| 652 | tpd->buffer_addr = 0; | ||
| 653 | tpd->desc.data = 0; | ||
| 654 | |||
| 655 | if (++sw_tpd_next_to_clean == tpd_ring->count) | ||
| 656 | sw_tpd_next_to_clean = 0; | ||
| 657 | } | 400 | } |
| 658 | atomic_set(&tpd_ring->next_to_clean, sw_tpd_next_to_clean); | ||
| 659 | |||
| 660 | if (netif_queue_stopped(adapter->netdev) | ||
| 661 | && netif_carrier_ok(adapter->netdev)) | ||
| 662 | netif_wake_queue(adapter->netdev); | ||
| 663 | } | ||
| 664 | 401 | ||
| 665 | static void atl1_check_for_link(struct atl1_adapter *adapter) | 402 | size = sizeof(struct atl1_buffer) * tpd_ring->count; |
| 666 | { | 403 | memset(tpd_ring->buffer_info, 0, size); |
| 667 | struct net_device *netdev = adapter->netdev; | ||
| 668 | u16 phy_data = 0; | ||
| 669 | |||
| 670 | spin_lock(&adapter->lock); | ||
| 671 | adapter->phy_timer_pending = false; | ||
| 672 | atl1_read_phy_reg(&adapter->hw, MII_BMSR, &phy_data); | ||
| 673 | atl1_read_phy_reg(&adapter->hw, MII_BMSR, &phy_data); | ||
| 674 | spin_unlock(&adapter->lock); | ||
| 675 | |||
| 676 | /* notify upper layer link down ASAP */ | ||
| 677 | if (!(phy_data & BMSR_LSTATUS)) { /* Link Down */ | ||
| 678 | if (netif_carrier_ok(netdev)) { /* old link state: Up */ | ||
| 679 | dev_info(&adapter->pdev->dev, "%s link is down\n", | ||
| 680 | netdev->name); | ||
| 681 | adapter->link_speed = SPEED_0; | ||
| 682 | netif_carrier_off(netdev); | ||
| 683 | netif_stop_queue(netdev); | ||
| 684 | } | ||
| 685 | } | ||
| 686 | schedule_work(&adapter->link_chg_task); | ||
| 687 | } | ||
| 688 | |||
| 689 | /* | ||
| 690 | * atl1_intr - Interrupt Handler | ||
| 691 | * @irq: interrupt number | ||
| 692 | * @data: pointer to a network interface device structure | ||
| 693 | * @pt_regs: CPU registers structure | ||
| 694 | */ | ||
| 695 | static irqreturn_t atl1_intr(int irq, void *data) | ||
| 696 | { | ||
| 697 | /*struct atl1_adapter *adapter = ((struct net_device *)data)->priv;*/ | ||
| 698 | struct atl1_adapter *adapter = netdev_priv(data); | ||
| 699 | u32 status; | ||
| 700 | u8 update_rx; | ||
| 701 | int max_ints = 10; | ||
| 702 | |||
| 703 | status = adapter->cmb.cmb->int_stats; | ||
| 704 | if (!status) | ||
| 705 | return IRQ_NONE; | ||
| 706 | |||
| 707 | update_rx = 0; | ||
| 708 | |||
| 709 | do { | ||
| 710 | /* clear CMB interrupt status at once */ | ||
| 711 | adapter->cmb.cmb->int_stats = 0; | ||
| 712 | |||
| 713 | if (status & ISR_GPHY) /* clear phy status */ | ||
| 714 | atl1_clear_phy_int(adapter); | ||
| 715 | |||
| 716 | /* clear ISR status, and Enable CMB DMA/Disable Interrupt */ | ||
| 717 | iowrite32(status | ISR_DIS_INT, adapter->hw.hw_addr + REG_ISR); | ||
| 718 | |||
| 719 | /* check if SMB intr */ | ||
| 720 | if (status & ISR_SMB) | ||
| 721 | atl1_inc_smb(adapter); | ||
| 722 | |||
| 723 | /* check if PCIE PHY Link down */ | ||
| 724 | if (status & ISR_PHY_LINKDOWN) { | ||
| 725 | dev_dbg(&adapter->pdev->dev, "pcie phy link down %x\n", | ||
| 726 | status); | ||
| 727 | if (netif_running(adapter->netdev)) { /* reset MAC */ | ||
| 728 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | ||
| 729 | schedule_work(&adapter->pcie_dma_to_rst_task); | ||
| 730 | return IRQ_HANDLED; | ||
| 731 | } | ||
| 732 | } | ||
| 733 | |||
| 734 | /* check if DMA read/write error ? */ | ||
| 735 | if (status & (ISR_DMAR_TO_RST | ISR_DMAW_TO_RST)) { | ||
| 736 | dev_dbg(&adapter->pdev->dev, | ||
| 737 | "pcie DMA r/w error (status = 0x%x)\n", | ||
| 738 | status); | ||
| 739 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | ||
| 740 | schedule_work(&adapter->pcie_dma_to_rst_task); | ||
| 741 | return IRQ_HANDLED; | ||
| 742 | } | ||
| 743 | |||
| 744 | /* link event */ | ||
| 745 | if (status & ISR_GPHY) { | ||
| 746 | adapter->soft_stats.tx_carrier_errors++; | ||
| 747 | atl1_check_for_link(adapter); | ||
| 748 | } | ||
| 749 | |||
| 750 | /* transmit event */ | ||
| 751 | if (status & ISR_CMB_TX) | ||
| 752 | atl1_intr_tx(adapter); | ||
| 753 | |||
| 754 | /* rx exception */ | ||
| 755 | if (unlikely(status & (ISR_RXF_OV | ISR_RFD_UNRUN | | ||
| 756 | ISR_RRD_OV | ISR_HOST_RFD_UNRUN | | ||
| 757 | ISR_HOST_RRD_OV | ISR_CMB_RX))) { | ||
| 758 | if (status & (ISR_RXF_OV | ISR_RFD_UNRUN | | ||
| 759 | ISR_RRD_OV | ISR_HOST_RFD_UNRUN | | ||
| 760 | ISR_HOST_RRD_OV)) | ||
| 761 | dev_dbg(&adapter->pdev->dev, | ||
| 762 | "rx exception, ISR = 0x%x\n", status); | ||
| 763 | atl1_intr_rx(adapter); | ||
| 764 | } | ||
| 765 | |||
| 766 | if (--max_ints < 0) | ||
| 767 | break; | ||
| 768 | 404 | ||
| 769 | } while ((status = adapter->cmb.cmb->int_stats)); | 405 | /* Zero out the descriptor ring */ |
| 406 | memset(tpd_ring->desc, 0, tpd_ring->size); | ||
| 770 | 407 | ||
| 771 | /* re-enable Interrupt */ | 408 | atomic_set(&tpd_ring->next_to_use, 0); |
| 772 | iowrite32(ISR_DIS_SMB | ISR_DIS_DMA, adapter->hw.hw_addr + REG_ISR); | 409 | atomic_set(&tpd_ring->next_to_clean, 0); |
| 773 | return IRQ_HANDLED; | ||
| 774 | } | 410 | } |
| 775 | 411 | ||
| 776 | /* | 412 | /* |
| 777 | * atl1_set_multi - Multicast and Promiscuous mode set | 413 | * atl1_free_ring_resources - Free Tx / RX descriptor Resources |
| 778 | * @netdev: network interface device structure | 414 | * @adapter: board private structure |
| 779 | * | 415 | * |
| 780 | * The set_multi entry point is called whenever the multicast address | 416 | * Free all transmit software resources |
| 781 | * list or the network interface flags are updated. This routine is | ||
| 782 | * responsible for configuring the hardware for proper multicast, | ||
| 783 | * promiscuous mode, and all-multi behavior. | ||
| 784 | */ | 417 | */ |
| 785 | static void atl1_set_multi(struct net_device *netdev) | 418 | void atl1_free_ring_resources(struct atl1_adapter *adapter) |
| 786 | { | 419 | { |
| 787 | struct atl1_adapter *adapter = netdev_priv(netdev); | 420 | struct pci_dev *pdev = adapter->pdev; |
| 788 | struct atl1_hw *hw = &adapter->hw; | 421 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; |
| 789 | struct dev_mc_list *mc_ptr; | 422 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; |
| 790 | u32 rctl; | 423 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; |
| 791 | u32 hash_value; | 424 | struct atl1_ring_header *ring_header = &adapter->ring_header; |
| 792 | 425 | ||
| 793 | /* Check for Promiscuous and All Multicast modes */ | 426 | atl1_clean_tx_ring(adapter); |
| 794 | rctl = ioread32(hw->hw_addr + REG_MAC_CTRL); | 427 | atl1_clean_rx_ring(adapter); |
| 795 | if (netdev->flags & IFF_PROMISC) | ||
| 796 | rctl |= MAC_CTRL_PROMIS_EN; | ||
| 797 | else if (netdev->flags & IFF_ALLMULTI) { | ||
| 798 | rctl |= MAC_CTRL_MC_ALL_EN; | ||
| 799 | rctl &= ~MAC_CTRL_PROMIS_EN; | ||
| 800 | } else | ||
| 801 | rctl &= ~(MAC_CTRL_PROMIS_EN | MAC_CTRL_MC_ALL_EN); | ||
| 802 | 428 | ||
| 803 | iowrite32(rctl, hw->hw_addr + REG_MAC_CTRL); | 429 | kfree(tpd_ring->buffer_info); |
| 430 | pci_free_consistent(pdev, ring_header->size, ring_header->desc, | ||
| 431 | ring_header->dma); | ||
| 804 | 432 | ||
| 805 | /* clear the old settings from the multicast hash table */ | 433 | tpd_ring->buffer_info = NULL; |
| 806 | iowrite32(0, hw->hw_addr + REG_RX_HASH_TABLE); | 434 | tpd_ring->desc = NULL; |
| 807 | iowrite32(0, (hw->hw_addr + REG_RX_HASH_TABLE) + (1 << 2)); | 435 | tpd_ring->dma = 0; |
| 808 | 436 | ||
| 809 | /* compute mc addresses' hash value ,and put it into hash table */ | 437 | rfd_ring->buffer_info = NULL; |
| 810 | for (mc_ptr = netdev->mc_list; mc_ptr; mc_ptr = mc_ptr->next) { | 438 | rfd_ring->desc = NULL; |
| 811 | hash_value = atl1_hash_mc_addr(hw, mc_ptr->dmi_addr); | 439 | rfd_ring->dma = 0; |
| 812 | atl1_hash_set(hw, hash_value); | 440 | |
| 813 | } | 441 | rrd_ring->desc = NULL; |
| 442 | rrd_ring->dma = 0; | ||
| 814 | } | 443 | } |
| 815 | 444 | ||
| 816 | static void atl1_setup_mac_ctrl(struct atl1_adapter *adapter) | 445 | static void atl1_setup_mac_ctrl(struct atl1_adapter *adapter) |
| @@ -851,6 +480,31 @@ static void atl1_setup_mac_ctrl(struct atl1_adapter *adapter) | |||
| 851 | iowrite32(value, hw->hw_addr + REG_MAC_CTRL); | 480 | iowrite32(value, hw->hw_addr + REG_MAC_CTRL); |
| 852 | } | 481 | } |
| 853 | 482 | ||
| 483 | /* | ||
| 484 | * atl1_set_mac - Change the Ethernet Address of the NIC | ||
| 485 | * @netdev: network interface device structure | ||
| 486 | * @p: pointer to an address structure | ||
| 487 | * | ||
| 488 | * Returns 0 on success, negative on failure | ||
| 489 | */ | ||
| 490 | static int atl1_set_mac(struct net_device *netdev, void *p) | ||
| 491 | { | ||
| 492 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 493 | struct sockaddr *addr = p; | ||
| 494 | |||
| 495 | if (netif_running(netdev)) | ||
| 496 | return -EBUSY; | ||
| 497 | |||
| 498 | if (!is_valid_ether_addr(addr->sa_data)) | ||
| 499 | return -EADDRNOTAVAIL; | ||
| 500 | |||
| 501 | memcpy(netdev->dev_addr, addr->sa_data, netdev->addr_len); | ||
| 502 | memcpy(adapter->hw.mac_addr, addr->sa_data, netdev->addr_len); | ||
| 503 | |||
| 504 | atl1_set_mac_addr(&adapter->hw); | ||
| 505 | return 0; | ||
| 506 | } | ||
| 507 | |||
| 854 | static u32 atl1_check_link(struct atl1_adapter *adapter) | 508 | static u32 atl1_check_link(struct atl1_adapter *adapter) |
| 855 | { | 509 | { |
| 856 | struct atl1_hw *hw = &adapter->hw; | 510 | struct atl1_hw *hw = &adapter->hw; |
| @@ -958,6 +612,103 @@ static u32 atl1_check_link(struct atl1_adapter *adapter) | |||
| 958 | return ATL1_SUCCESS; | 612 | return ATL1_SUCCESS; |
| 959 | } | 613 | } |
| 960 | 614 | ||
| 615 | static void atl1_check_for_link(struct atl1_adapter *adapter) | ||
| 616 | { | ||
| 617 | struct net_device *netdev = adapter->netdev; | ||
| 618 | u16 phy_data = 0; | ||
| 619 | |||
| 620 | spin_lock(&adapter->lock); | ||
| 621 | adapter->phy_timer_pending = false; | ||
| 622 | atl1_read_phy_reg(&adapter->hw, MII_BMSR, &phy_data); | ||
| 623 | atl1_read_phy_reg(&adapter->hw, MII_BMSR, &phy_data); | ||
| 624 | spin_unlock(&adapter->lock); | ||
| 625 | |||
| 626 | /* notify upper layer link down ASAP */ | ||
| 627 | if (!(phy_data & BMSR_LSTATUS)) { /* Link Down */ | ||
| 628 | if (netif_carrier_ok(netdev)) { /* old link state: Up */ | ||
| 629 | dev_info(&adapter->pdev->dev, "%s link is down\n", | ||
| 630 | netdev->name); | ||
| 631 | adapter->link_speed = SPEED_0; | ||
| 632 | netif_carrier_off(netdev); | ||
| 633 | netif_stop_queue(netdev); | ||
| 634 | } | ||
| 635 | } | ||
| 636 | schedule_work(&adapter->link_chg_task); | ||
| 637 | } | ||
| 638 | |||
| 639 | /* | ||
| 640 | * atl1_set_multi - Multicast and Promiscuous mode set | ||
| 641 | * @netdev: network interface device structure | ||
| 642 | * | ||
| 643 | * The set_multi entry point is called whenever the multicast address | ||
| 644 | * list or the network interface flags are updated. This routine is | ||
| 645 | * responsible for configuring the hardware for proper multicast, | ||
| 646 | * promiscuous mode, and all-multi behavior. | ||
| 647 | */ | ||
| 648 | static void atl1_set_multi(struct net_device *netdev) | ||
| 649 | { | ||
| 650 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 651 | struct atl1_hw *hw = &adapter->hw; | ||
| 652 | struct dev_mc_list *mc_ptr; | ||
| 653 | u32 rctl; | ||
| 654 | u32 hash_value; | ||
| 655 | |||
| 656 | /* Check for Promiscuous and All Multicast modes */ | ||
| 657 | rctl = ioread32(hw->hw_addr + REG_MAC_CTRL); | ||
| 658 | if (netdev->flags & IFF_PROMISC) | ||
| 659 | rctl |= MAC_CTRL_PROMIS_EN; | ||
| 660 | else if (netdev->flags & IFF_ALLMULTI) { | ||
| 661 | rctl |= MAC_CTRL_MC_ALL_EN; | ||
| 662 | rctl &= ~MAC_CTRL_PROMIS_EN; | ||
| 663 | } else | ||
| 664 | rctl &= ~(MAC_CTRL_PROMIS_EN | MAC_CTRL_MC_ALL_EN); | ||
| 665 | |||
| 666 | iowrite32(rctl, hw->hw_addr + REG_MAC_CTRL); | ||
| 667 | |||
| 668 | /* clear the old settings from the multicast hash table */ | ||
| 669 | iowrite32(0, hw->hw_addr + REG_RX_HASH_TABLE); | ||
| 670 | iowrite32(0, (hw->hw_addr + REG_RX_HASH_TABLE) + (1 << 2)); | ||
| 671 | |||
| 672 | /* compute mc addresses' hash value ,and put it into hash table */ | ||
| 673 | for (mc_ptr = netdev->mc_list; mc_ptr; mc_ptr = mc_ptr->next) { | ||
| 674 | hash_value = atl1_hash_mc_addr(hw, mc_ptr->dmi_addr); | ||
| 675 | atl1_hash_set(hw, hash_value); | ||
| 676 | } | ||
| 677 | } | ||
| 678 | |||
| 679 | /* | ||
| 680 | * atl1_change_mtu - Change the Maximum Transfer Unit | ||
| 681 | * @netdev: network interface device structure | ||
| 682 | * @new_mtu: new value for maximum frame size | ||
| 683 | * | ||
| 684 | * Returns 0 on success, negative on failure | ||
| 685 | */ | ||
| 686 | static int atl1_change_mtu(struct net_device *netdev, int new_mtu) | ||
| 687 | { | ||
| 688 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 689 | int old_mtu = netdev->mtu; | ||
| 690 | int max_frame = new_mtu + ENET_HEADER_SIZE + ETHERNET_FCS_SIZE; | ||
| 691 | |||
| 692 | if ((max_frame < MINIMUM_ETHERNET_FRAME_SIZE) || | ||
| 693 | (max_frame > MAX_JUMBO_FRAME_SIZE)) { | ||
| 694 | dev_warn(&adapter->pdev->dev, "invalid MTU setting\n"); | ||
| 695 | return -EINVAL; | ||
| 696 | } | ||
| 697 | |||
| 698 | adapter->hw.max_frame_size = max_frame; | ||
| 699 | adapter->hw.tx_jumbo_task_th = (max_frame + 7) >> 3; | ||
| 700 | adapter->rx_buffer_len = (max_frame + 7) & ~7; | ||
| 701 | adapter->hw.rx_jumbo_th = adapter->rx_buffer_len / 8; | ||
| 702 | |||
| 703 | netdev->mtu = new_mtu; | ||
| 704 | if ((old_mtu != new_mtu) && netif_running(netdev)) { | ||
| 705 | atl1_down(adapter); | ||
| 706 | atl1_up(adapter); | ||
| 707 | } | ||
| 708 | |||
| 709 | return 0; | ||
| 710 | } | ||
| 711 | |||
| 961 | static void set_flow_ctrl_old(struct atl1_adapter *adapter) | 712 | static void set_flow_ctrl_old(struct atl1_adapter *adapter) |
| 962 | { | 713 | { |
| 963 | u32 hi, lo, value; | 714 | u32 hi, lo, value; |
| @@ -970,7 +721,7 @@ static void set_flow_ctrl_old(struct atl1_adapter *adapter) | |||
| 970 | lo = value * 7 / 8; | 721 | lo = value * 7 / 8; |
| 971 | 722 | ||
| 972 | value = ((hi & RXQ_RXF_PAUSE_TH_HI_MASK) << RXQ_RXF_PAUSE_TH_HI_SHIFT) | | 723 | value = ((hi & RXQ_RXF_PAUSE_TH_HI_MASK) << RXQ_RXF_PAUSE_TH_HI_SHIFT) | |
| 973 | ((lo & RXQ_RXF_PAUSE_TH_LO_MASK) << RXQ_RXF_PAUSE_TH_LO_SHIFT); | 724 | ((lo & RXQ_RXF_PAUSE_TH_LO_MASK) << RXQ_RXF_PAUSE_TH_LO_SHIFT); |
| 974 | iowrite32(value, adapter->hw.hw_addr + REG_RXQ_RXF_PAUSE_THRESH); | 725 | iowrite32(value, adapter->hw.hw_addr + REG_RXQ_RXF_PAUSE_THRESH); |
| 975 | 726 | ||
| 976 | /* RRD Flow Control */ | 727 | /* RRD Flow Control */ |
| @@ -980,7 +731,7 @@ static void set_flow_ctrl_old(struct atl1_adapter *adapter) | |||
| 980 | if (lo < 2) | 731 | if (lo < 2) |
| 981 | lo = 2; | 732 | lo = 2; |
| 982 | value = ((hi & RXQ_RRD_PAUSE_TH_HI_MASK) << RXQ_RRD_PAUSE_TH_HI_SHIFT) | | 733 | value = ((hi & RXQ_RRD_PAUSE_TH_HI_MASK) << RXQ_RRD_PAUSE_TH_HI_SHIFT) | |
| 983 | ((lo & RXQ_RRD_PAUSE_TH_LO_MASK) << RXQ_RRD_PAUSE_TH_LO_SHIFT); | 734 | ((lo & RXQ_RRD_PAUSE_TH_LO_MASK) << RXQ_RRD_PAUSE_TH_LO_SHIFT); |
| 984 | iowrite32(value, adapter->hw.hw_addr + REG_RXQ_RRD_PAUSE_THRESH); | 735 | iowrite32(value, adapter->hw.hw_addr + REG_RXQ_RRD_PAUSE_THRESH); |
| 985 | } | 736 | } |
| 986 | 737 | ||
| @@ -997,7 +748,7 @@ static void set_flow_ctrl_new(struct atl1_hw *hw) | |||
| 997 | if (hi < lo) | 748 | if (hi < lo) |
| 998 | hi = lo + 16; | 749 | hi = lo + 16; |
| 999 | value = ((hi & RXQ_RXF_PAUSE_TH_HI_MASK) << RXQ_RXF_PAUSE_TH_HI_SHIFT) | | 750 | value = ((hi & RXQ_RXF_PAUSE_TH_HI_MASK) << RXQ_RXF_PAUSE_TH_HI_SHIFT) | |
| 1000 | ((lo & RXQ_RXF_PAUSE_TH_LO_MASK) << RXQ_RXF_PAUSE_TH_LO_SHIFT); | 751 | ((lo & RXQ_RXF_PAUSE_TH_LO_MASK) << RXQ_RXF_PAUSE_TH_LO_SHIFT); |
| 1001 | iowrite32(value, hw->hw_addr + REG_RXQ_RXF_PAUSE_THRESH); | 752 | iowrite32(value, hw->hw_addr + REG_RXQ_RXF_PAUSE_THRESH); |
| 1002 | 753 | ||
| 1003 | /* RRD Flow Control */ | 754 | /* RRD Flow Control */ |
| @@ -1009,7 +760,7 @@ static void set_flow_ctrl_new(struct atl1_hw *hw) | |||
| 1009 | if (hi < lo) | 760 | if (hi < lo) |
| 1010 | hi = lo + 3; | 761 | hi = lo + 3; |
| 1011 | value = ((hi & RXQ_RRD_PAUSE_TH_HI_MASK) << RXQ_RRD_PAUSE_TH_HI_SHIFT) | | 762 | value = ((hi & RXQ_RRD_PAUSE_TH_HI_MASK) << RXQ_RRD_PAUSE_TH_HI_SHIFT) | |
| 1012 | ((lo & RXQ_RRD_PAUSE_TH_LO_MASK) << RXQ_RRD_PAUSE_TH_LO_SHIFT); | 763 | ((lo & RXQ_RRD_PAUSE_TH_LO_MASK) << RXQ_RRD_PAUSE_TH_LO_SHIFT); |
| 1013 | iowrite32(value, hw->hw_addr + REG_RXQ_RRD_PAUSE_THRESH); | 764 | iowrite32(value, hw->hw_addr + REG_RXQ_RRD_PAUSE_THRESH); |
| 1014 | } | 765 | } |
| 1015 | 766 | ||
| @@ -1058,7 +809,8 @@ static u32 atl1_configure(struct atl1_adapter *adapter) | |||
| 1058 | value <<= 16; | 809 | value <<= 16; |
| 1059 | value += adapter->rfd_ring.count; | 810 | value += adapter->rfd_ring.count; |
| 1060 | iowrite32(value, hw->hw_addr + REG_DESC_RFD_RRD_RING_SIZE); | 811 | iowrite32(value, hw->hw_addr + REG_DESC_RFD_RRD_RING_SIZE); |
| 1061 | iowrite32(adapter->tpd_ring.count, hw->hw_addr + REG_DESC_TPD_RING_SIZE); | 812 | iowrite32(adapter->tpd_ring.count, hw->hw_addr + |
| 813 | REG_DESC_TPD_RING_SIZE); | ||
| 1062 | 814 | ||
| 1063 | /* Load Ptr */ | 815 | /* Load Ptr */ |
| 1064 | iowrite32(1, hw->hw_addr + REG_LOAD_PTR); | 816 | iowrite32(1, hw->hw_addr + REG_LOAD_PTR); |
| @@ -1066,31 +818,31 @@ static u32 atl1_configure(struct atl1_adapter *adapter) | |||
| 1066 | /* config Mailbox */ | 818 | /* config Mailbox */ |
| 1067 | value = ((atomic_read(&adapter->tpd_ring.next_to_use) | 819 | value = ((atomic_read(&adapter->tpd_ring.next_to_use) |
| 1068 | & MB_TPD_PROD_INDX_MASK) << MB_TPD_PROD_INDX_SHIFT) | | 820 | & MB_TPD_PROD_INDX_MASK) << MB_TPD_PROD_INDX_SHIFT) | |
| 1069 | ((atomic_read(&adapter->rrd_ring.next_to_clean) | 821 | ((atomic_read(&adapter->rrd_ring.next_to_clean) |
| 1070 | & MB_RRD_CONS_INDX_MASK) << MB_RRD_CONS_INDX_SHIFT) | | 822 | & MB_RRD_CONS_INDX_MASK) << MB_RRD_CONS_INDX_SHIFT) | |
| 1071 | ((atomic_read(&adapter->rfd_ring.next_to_use) | 823 | ((atomic_read(&adapter->rfd_ring.next_to_use) |
| 1072 | & MB_RFD_PROD_INDX_MASK) << MB_RFD_PROD_INDX_SHIFT); | 824 | & MB_RFD_PROD_INDX_MASK) << MB_RFD_PROD_INDX_SHIFT); |
| 1073 | iowrite32(value, hw->hw_addr + REG_MAILBOX); | 825 | iowrite32(value, hw->hw_addr + REG_MAILBOX); |
| 1074 | 826 | ||
| 1075 | /* config IPG/IFG */ | 827 | /* config IPG/IFG */ |
| 1076 | value = (((u32) hw->ipgt & MAC_IPG_IFG_IPGT_MASK) | 828 | value = (((u32) hw->ipgt & MAC_IPG_IFG_IPGT_MASK) |
| 1077 | << MAC_IPG_IFG_IPGT_SHIFT) | | 829 | << MAC_IPG_IFG_IPGT_SHIFT) | |
| 1078 | (((u32) hw->min_ifg & MAC_IPG_IFG_MIFG_MASK) | 830 | (((u32) hw->min_ifg & MAC_IPG_IFG_MIFG_MASK) |
| 1079 | << MAC_IPG_IFG_MIFG_SHIFT) | | 831 | << MAC_IPG_IFG_MIFG_SHIFT) | |
| 1080 | (((u32) hw->ipgr1 & MAC_IPG_IFG_IPGR1_MASK) | 832 | (((u32) hw->ipgr1 & MAC_IPG_IFG_IPGR1_MASK) |
| 1081 | << MAC_IPG_IFG_IPGR1_SHIFT) | | 833 | << MAC_IPG_IFG_IPGR1_SHIFT) | |
| 1082 | (((u32) hw->ipgr2 & MAC_IPG_IFG_IPGR2_MASK) | 834 | (((u32) hw->ipgr2 & MAC_IPG_IFG_IPGR2_MASK) |
| 1083 | << MAC_IPG_IFG_IPGR2_SHIFT); | 835 | << MAC_IPG_IFG_IPGR2_SHIFT); |
| 1084 | iowrite32(value, hw->hw_addr + REG_MAC_IPG_IFG); | 836 | iowrite32(value, hw->hw_addr + REG_MAC_IPG_IFG); |
| 1085 | 837 | ||
| 1086 | /* config Half-Duplex Control */ | 838 | /* config Half-Duplex Control */ |
| 1087 | value = ((u32) hw->lcol & MAC_HALF_DUPLX_CTRL_LCOL_MASK) | | 839 | value = ((u32) hw->lcol & MAC_HALF_DUPLX_CTRL_LCOL_MASK) | |
| 1088 | (((u32) hw->max_retry & MAC_HALF_DUPLX_CTRL_RETRY_MASK) | 840 | (((u32) hw->max_retry & MAC_HALF_DUPLX_CTRL_RETRY_MASK) |
| 1089 | << MAC_HALF_DUPLX_CTRL_RETRY_SHIFT) | | 841 | << MAC_HALF_DUPLX_CTRL_RETRY_SHIFT) | |
| 1090 | MAC_HALF_DUPLX_CTRL_EXC_DEF_EN | | 842 | MAC_HALF_DUPLX_CTRL_EXC_DEF_EN | |
| 1091 | (0xa << MAC_HALF_DUPLX_CTRL_ABEBT_SHIFT) | | 843 | (0xa << MAC_HALF_DUPLX_CTRL_ABEBT_SHIFT) | |
| 1092 | (((u32) hw->jam_ipg & MAC_HALF_DUPLX_CTRL_JAMIPG_MASK) | 844 | (((u32) hw->jam_ipg & MAC_HALF_DUPLX_CTRL_JAMIPG_MASK) |
| 1093 | << MAC_HALF_DUPLX_CTRL_JAMIPG_SHIFT); | 845 | << MAC_HALF_DUPLX_CTRL_JAMIPG_SHIFT); |
| 1094 | iowrite32(value, hw->hw_addr + REG_MAC_HALF_DUPLX_CTRL); | 846 | iowrite32(value, hw->hw_addr + REG_MAC_HALF_DUPLX_CTRL); |
| 1095 | 847 | ||
| 1096 | /* set Interrupt Moderator Timer */ | 848 | /* set Interrupt Moderator Timer */ |
| @@ -1106,10 +858,10 @@ static u32 atl1_configure(struct atl1_adapter *adapter) | |||
| 1106 | /* jumbo size & rrd retirement timer */ | 858 | /* jumbo size & rrd retirement timer */ |
| 1107 | value = (((u32) hw->rx_jumbo_th & RXQ_JMBOSZ_TH_MASK) | 859 | value = (((u32) hw->rx_jumbo_th & RXQ_JMBOSZ_TH_MASK) |
| 1108 | << RXQ_JMBOSZ_TH_SHIFT) | | 860 | << RXQ_JMBOSZ_TH_SHIFT) | |
| 1109 | (((u32) hw->rx_jumbo_lkah & RXQ_JMBO_LKAH_MASK) | 861 | (((u32) hw->rx_jumbo_lkah & RXQ_JMBO_LKAH_MASK) |
| 1110 | << RXQ_JMBO_LKAH_SHIFT) | | 862 | << RXQ_JMBO_LKAH_SHIFT) | |
| 1111 | (((u32) hw->rrd_ret_timer & RXQ_RRD_TIMER_MASK) | 863 | (((u32) hw->rrd_ret_timer & RXQ_RRD_TIMER_MASK) |
| 1112 | << RXQ_RRD_TIMER_SHIFT); | 864 | << RXQ_RRD_TIMER_SHIFT); |
| 1113 | iowrite32(value, hw->hw_addr + REG_RXQ_JMBOSZ_RRDTIM); | 865 | iowrite32(value, hw->hw_addr + REG_RXQ_JMBOSZ_RRDTIM); |
| 1114 | 866 | ||
| 1115 | /* Flow Control */ | 867 | /* Flow Control */ |
| @@ -1128,35 +880,36 @@ static u32 atl1_configure(struct atl1_adapter *adapter) | |||
| 1128 | /* config TXQ */ | 880 | /* config TXQ */ |
| 1129 | value = (((u32) hw->tpd_burst & TXQ_CTRL_TPD_BURST_NUM_MASK) | 881 | value = (((u32) hw->tpd_burst & TXQ_CTRL_TPD_BURST_NUM_MASK) |
| 1130 | << TXQ_CTRL_TPD_BURST_NUM_SHIFT) | | 882 | << TXQ_CTRL_TPD_BURST_NUM_SHIFT) | |
| 1131 | (((u32) hw->txf_burst & TXQ_CTRL_TXF_BURST_NUM_MASK) | 883 | (((u32) hw->txf_burst & TXQ_CTRL_TXF_BURST_NUM_MASK) |
| 1132 | << TXQ_CTRL_TXF_BURST_NUM_SHIFT) | | 884 | << TXQ_CTRL_TXF_BURST_NUM_SHIFT) | |
| 1133 | (((u32) hw->tpd_fetch_th & TXQ_CTRL_TPD_FETCH_TH_MASK) | 885 | (((u32) hw->tpd_fetch_th & TXQ_CTRL_TPD_FETCH_TH_MASK) |
| 1134 | << TXQ_CTRL_TPD_FETCH_TH_SHIFT) | TXQ_CTRL_ENH_MODE | TXQ_CTRL_EN; | 886 | << TXQ_CTRL_TPD_FETCH_TH_SHIFT) | TXQ_CTRL_ENH_MODE | |
| 887 | TXQ_CTRL_EN; | ||
| 1135 | iowrite32(value, hw->hw_addr + REG_TXQ_CTRL); | 888 | iowrite32(value, hw->hw_addr + REG_TXQ_CTRL); |
| 1136 | 889 | ||
| 1137 | /* min tpd fetch gap & tx jumbo packet size threshold for taskoffload */ | 890 | /* min tpd fetch gap & tx jumbo packet size threshold for taskoffload */ |
| 1138 | value = (((u32) hw->tx_jumbo_task_th & TX_JUMBO_TASK_TH_MASK) | 891 | value = (((u32) hw->tx_jumbo_task_th & TX_JUMBO_TASK_TH_MASK) |
| 1139 | << TX_JUMBO_TASK_TH_SHIFT) | | 892 | << TX_JUMBO_TASK_TH_SHIFT) | |
| 1140 | (((u32) hw->tpd_fetch_gap & TX_TPD_MIN_IPG_MASK) | 893 | (((u32) hw->tpd_fetch_gap & TX_TPD_MIN_IPG_MASK) |
| 1141 | << TX_TPD_MIN_IPG_SHIFT); | 894 | << TX_TPD_MIN_IPG_SHIFT); |
| 1142 | iowrite32(value, hw->hw_addr + REG_TX_JUMBO_TASK_TH_TPD_IPG); | 895 | iowrite32(value, hw->hw_addr + REG_TX_JUMBO_TASK_TH_TPD_IPG); |
| 1143 | 896 | ||
| 1144 | /* config RXQ */ | 897 | /* config RXQ */ |
| 1145 | value = (((u32) hw->rfd_burst & RXQ_CTRL_RFD_BURST_NUM_MASK) | 898 | value = (((u32) hw->rfd_burst & RXQ_CTRL_RFD_BURST_NUM_MASK) |
| 1146 | << RXQ_CTRL_RFD_BURST_NUM_SHIFT) | | 899 | << RXQ_CTRL_RFD_BURST_NUM_SHIFT) | |
| 1147 | (((u32) hw->rrd_burst & RXQ_CTRL_RRD_BURST_THRESH_MASK) | 900 | (((u32) hw->rrd_burst & RXQ_CTRL_RRD_BURST_THRESH_MASK) |
| 1148 | << RXQ_CTRL_RRD_BURST_THRESH_SHIFT) | | 901 | << RXQ_CTRL_RRD_BURST_THRESH_SHIFT) | |
| 1149 | (((u32) hw->rfd_fetch_gap & RXQ_CTRL_RFD_PREF_MIN_IPG_MASK) | 902 | (((u32) hw->rfd_fetch_gap & RXQ_CTRL_RFD_PREF_MIN_IPG_MASK) |
| 1150 | << RXQ_CTRL_RFD_PREF_MIN_IPG_SHIFT) | | 903 | << RXQ_CTRL_RFD_PREF_MIN_IPG_SHIFT) | RXQ_CTRL_CUT_THRU_EN | |
| 1151 | RXQ_CTRL_CUT_THRU_EN | RXQ_CTRL_EN; | 904 | RXQ_CTRL_EN; |
| 1152 | iowrite32(value, hw->hw_addr + REG_RXQ_CTRL); | 905 | iowrite32(value, hw->hw_addr + REG_RXQ_CTRL); |
| 1153 | 906 | ||
| 1154 | /* config DMA Engine */ | 907 | /* config DMA Engine */ |
| 1155 | value = ((((u32) hw->dmar_block) & DMA_CTRL_DMAR_BURST_LEN_MASK) | 908 | value = ((((u32) hw->dmar_block) & DMA_CTRL_DMAR_BURST_LEN_MASK) |
| 1156 | << DMA_CTRL_DMAR_BURST_LEN_SHIFT) | | 909 | << DMA_CTRL_DMAR_BURST_LEN_SHIFT) | |
| 1157 | ((((u32) hw->dmaw_block) & DMA_CTRL_DMAR_BURST_LEN_MASK) | 910 | ((((u32) hw->dmaw_block) & DMA_CTRL_DMAR_BURST_LEN_MASK) |
| 1158 | << DMA_CTRL_DMAR_BURST_LEN_SHIFT) | | 911 | << DMA_CTRL_DMAR_BURST_LEN_SHIFT) | DMA_CTRL_DMAR_EN | |
| 1159 | DMA_CTRL_DMAR_EN | DMA_CTRL_DMAW_EN; | 912 | DMA_CTRL_DMAW_EN; |
| 1160 | value |= (u32) hw->dma_ord; | 913 | value |= (u32) hw->dma_ord; |
| 1161 | if (atl1_rcb_128 == hw->rcb_value) | 914 | if (atl1_rcb_128 == hw->rcb_value) |
| 1162 | value |= DMA_CTRL_RCB_VALUE; | 915 | value |= DMA_CTRL_RCB_VALUE; |
| @@ -1186,56 +939,495 @@ static u32 atl1_configure(struct atl1_adapter *adapter) | |||
| 1186 | } | 939 | } |
| 1187 | 940 | ||
| 1188 | /* | 941 | /* |
| 942 | * atl1_pcie_patch - Patch for PCIE module | ||
| 943 | */ | ||
| 944 | static void atl1_pcie_patch(struct atl1_adapter *adapter) | ||
| 945 | { | ||
| 946 | u32 value; | ||
| 947 | |||
| 948 | /* much vendor magic here */ | ||
| 949 | value = 0x6500; | ||
| 950 | iowrite32(value, adapter->hw.hw_addr + 0x12FC); | ||
| 951 | /* pcie flow control mode change */ | ||
| 952 | value = ioread32(adapter->hw.hw_addr + 0x1008); | ||
| 953 | value |= 0x8000; | ||
| 954 | iowrite32(value, adapter->hw.hw_addr + 0x1008); | ||
| 955 | } | ||
| 956 | |||
| 957 | /* | ||
| 958 | * When ACPI resume on some VIA MotherBoard, the Interrupt Disable bit/0x400 | ||
| 959 | * on PCI Command register is disable. | ||
| 960 | * The function enable this bit. | ||
| 961 | * Brackett, 2006/03/15 | ||
| 962 | */ | ||
| 963 | static void atl1_via_workaround(struct atl1_adapter *adapter) | ||
| 964 | { | ||
| 965 | unsigned long value; | ||
| 966 | |||
| 967 | value = ioread16(adapter->hw.hw_addr + PCI_COMMAND); | ||
| 968 | if (value & PCI_COMMAND_INTX_DISABLE) | ||
| 969 | value &= ~PCI_COMMAND_INTX_DISABLE; | ||
| 970 | iowrite32(value, adapter->hw.hw_addr + PCI_COMMAND); | ||
| 971 | } | ||
| 972 | |||
| 973 | /* | ||
| 974 | * atl1_irq_enable - Enable default interrupt generation settings | ||
| 975 | * @adapter: board private structure | ||
| 976 | */ | ||
| 977 | static void atl1_irq_enable(struct atl1_adapter *adapter) | ||
| 978 | { | ||
| 979 | iowrite32(IMR_NORMAL_MASK, adapter->hw.hw_addr + REG_IMR); | ||
| 980 | ioread32(adapter->hw.hw_addr + REG_IMR); | ||
| 981 | } | ||
| 982 | |||
| 983 | /* | ||
| 1189 | * atl1_irq_disable - Mask off interrupt generation on the NIC | 984 | * atl1_irq_disable - Mask off interrupt generation on the NIC |
| 1190 | * @adapter: board private structure | 985 | * @adapter: board private structure |
| 1191 | */ | 986 | */ |
| 1192 | static void atl1_irq_disable(struct atl1_adapter *adapter) | 987 | static void atl1_irq_disable(struct atl1_adapter *adapter) |
| 1193 | { | 988 | { |
| 1194 | atomic_inc(&adapter->irq_sem); | ||
| 1195 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | 989 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); |
| 1196 | ioread32(adapter->hw.hw_addr + REG_IMR); | 990 | ioread32(adapter->hw.hw_addr + REG_IMR); |
| 1197 | synchronize_irq(adapter->pdev->irq); | 991 | synchronize_irq(adapter->pdev->irq); |
| 1198 | } | 992 | } |
| 1199 | 993 | ||
| 1200 | static void atl1_vlan_rx_register(struct net_device *netdev, | 994 | static void atl1_clear_phy_int(struct atl1_adapter *adapter) |
| 1201 | struct vlan_group *grp) | ||
| 1202 | { | 995 | { |
| 1203 | struct atl1_adapter *adapter = netdev_priv(netdev); | 996 | u16 phy_data; |
| 1204 | unsigned long flags; | 997 | unsigned long flags; |
| 1205 | u32 ctrl; | ||
| 1206 | 998 | ||
| 1207 | spin_lock_irqsave(&adapter->lock, flags); | 999 | spin_lock_irqsave(&adapter->lock, flags); |
| 1208 | /* atl1_irq_disable(adapter); */ | 1000 | atl1_read_phy_reg(&adapter->hw, 19, &phy_data); |
| 1209 | adapter->vlgrp = grp; | 1001 | spin_unlock_irqrestore(&adapter->lock, flags); |
| 1002 | } | ||
| 1210 | 1003 | ||
| 1211 | if (grp) { | 1004 | static void atl1_inc_smb(struct atl1_adapter *adapter) |
| 1212 | /* enable VLAN tag insert/strip */ | 1005 | { |
| 1213 | ctrl = ioread32(adapter->hw.hw_addr + REG_MAC_CTRL); | 1006 | struct stats_msg_block *smb = adapter->smb.smb; |
| 1214 | ctrl |= MAC_CTRL_RMV_VLAN; | 1007 | |
| 1215 | iowrite32(ctrl, adapter->hw.hw_addr + REG_MAC_CTRL); | 1008 | /* Fill out the OS statistics structure */ |
| 1216 | } else { | 1009 | adapter->soft_stats.rx_packets += smb->rx_ok; |
| 1217 | /* disable VLAN tag insert/strip */ | 1010 | adapter->soft_stats.tx_packets += smb->tx_ok; |
| 1218 | ctrl = ioread32(adapter->hw.hw_addr + REG_MAC_CTRL); | 1011 | adapter->soft_stats.rx_bytes += smb->rx_byte_cnt; |
| 1219 | ctrl &= ~MAC_CTRL_RMV_VLAN; | 1012 | adapter->soft_stats.tx_bytes += smb->tx_byte_cnt; |
| 1220 | iowrite32(ctrl, adapter->hw.hw_addr + REG_MAC_CTRL); | 1013 | adapter->soft_stats.multicast += smb->rx_mcast; |
| 1014 | adapter->soft_stats.collisions += (smb->tx_1_col + smb->tx_2_col * 2 + | ||
| 1015 | smb->tx_late_col + smb->tx_abort_col * adapter->hw.max_retry); | ||
| 1016 | |||
| 1017 | /* Rx Errors */ | ||
| 1018 | adapter->soft_stats.rx_errors += (smb->rx_frag + smb->rx_fcs_err + | ||
| 1019 | smb->rx_len_err + smb->rx_sz_ov + smb->rx_rxf_ov + | ||
| 1020 | smb->rx_rrd_ov + smb->rx_align_err); | ||
| 1021 | adapter->soft_stats.rx_fifo_errors += smb->rx_rxf_ov; | ||
| 1022 | adapter->soft_stats.rx_length_errors += smb->rx_len_err; | ||
| 1023 | adapter->soft_stats.rx_crc_errors += smb->rx_fcs_err; | ||
| 1024 | adapter->soft_stats.rx_frame_errors += smb->rx_align_err; | ||
| 1025 | adapter->soft_stats.rx_missed_errors += (smb->rx_rrd_ov + | ||
| 1026 | smb->rx_rxf_ov); | ||
| 1027 | |||
| 1028 | adapter->soft_stats.rx_pause += smb->rx_pause; | ||
| 1029 | adapter->soft_stats.rx_rrd_ov += smb->rx_rrd_ov; | ||
| 1030 | adapter->soft_stats.rx_trunc += smb->rx_sz_ov; | ||
| 1031 | |||
| 1032 | /* Tx Errors */ | ||
| 1033 | adapter->soft_stats.tx_errors += (smb->tx_late_col + | ||
| 1034 | smb->tx_abort_col + smb->tx_underrun + smb->tx_trunc); | ||
| 1035 | adapter->soft_stats.tx_fifo_errors += smb->tx_underrun; | ||
| 1036 | adapter->soft_stats.tx_aborted_errors += smb->tx_abort_col; | ||
| 1037 | adapter->soft_stats.tx_window_errors += smb->tx_late_col; | ||
| 1038 | |||
| 1039 | adapter->soft_stats.excecol += smb->tx_abort_col; | ||
| 1040 | adapter->soft_stats.deffer += smb->tx_defer; | ||
| 1041 | adapter->soft_stats.scc += smb->tx_1_col; | ||
| 1042 | adapter->soft_stats.mcc += smb->tx_2_col; | ||
| 1043 | adapter->soft_stats.latecol += smb->tx_late_col; | ||
| 1044 | adapter->soft_stats.tx_underun += smb->tx_underrun; | ||
| 1045 | adapter->soft_stats.tx_trunc += smb->tx_trunc; | ||
| 1046 | adapter->soft_stats.tx_pause += smb->tx_pause; | ||
| 1047 | |||
| 1048 | adapter->net_stats.rx_packets = adapter->soft_stats.rx_packets; | ||
| 1049 | adapter->net_stats.tx_packets = adapter->soft_stats.tx_packets; | ||
| 1050 | adapter->net_stats.rx_bytes = adapter->soft_stats.rx_bytes; | ||
| 1051 | adapter->net_stats.tx_bytes = adapter->soft_stats.tx_bytes; | ||
| 1052 | adapter->net_stats.multicast = adapter->soft_stats.multicast; | ||
| 1053 | adapter->net_stats.collisions = adapter->soft_stats.collisions; | ||
| 1054 | adapter->net_stats.rx_errors = adapter->soft_stats.rx_errors; | ||
| 1055 | adapter->net_stats.rx_over_errors = | ||
| 1056 | adapter->soft_stats.rx_missed_errors; | ||
| 1057 | adapter->net_stats.rx_length_errors = | ||
| 1058 | adapter->soft_stats.rx_length_errors; | ||
| 1059 | adapter->net_stats.rx_crc_errors = adapter->soft_stats.rx_crc_errors; | ||
| 1060 | adapter->net_stats.rx_frame_errors = | ||
| 1061 | adapter->soft_stats.rx_frame_errors; | ||
| 1062 | adapter->net_stats.rx_fifo_errors = adapter->soft_stats.rx_fifo_errors; | ||
| 1063 | adapter->net_stats.rx_missed_errors = | ||
| 1064 | adapter->soft_stats.rx_missed_errors; | ||
| 1065 | adapter->net_stats.tx_errors = adapter->soft_stats.tx_errors; | ||
| 1066 | adapter->net_stats.tx_fifo_errors = adapter->soft_stats.tx_fifo_errors; | ||
| 1067 | adapter->net_stats.tx_aborted_errors = | ||
| 1068 | adapter->soft_stats.tx_aborted_errors; | ||
| 1069 | adapter->net_stats.tx_window_errors = | ||
| 1070 | adapter->soft_stats.tx_window_errors; | ||
| 1071 | adapter->net_stats.tx_carrier_errors = | ||
| 1072 | adapter->soft_stats.tx_carrier_errors; | ||
| 1073 | } | ||
| 1074 | |||
| 1075 | /* | ||
| 1076 | * atl1_get_stats - Get System Network Statistics | ||
| 1077 | * @netdev: network interface device structure | ||
| 1078 | * | ||
| 1079 | * Returns the address of the device statistics structure. | ||
| 1080 | * The statistics are actually updated from the timer callback. | ||
| 1081 | */ | ||
| 1082 | static struct net_device_stats *atl1_get_stats(struct net_device *netdev) | ||
| 1083 | { | ||
| 1084 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1085 | return &adapter->net_stats; | ||
| 1086 | } | ||
| 1087 | |||
| 1088 | static void atl1_update_mailbox(struct atl1_adapter *adapter) | ||
| 1089 | { | ||
| 1090 | unsigned long flags; | ||
| 1091 | u32 tpd_next_to_use; | ||
| 1092 | u32 rfd_next_to_use; | ||
| 1093 | u32 rrd_next_to_clean; | ||
| 1094 | u32 value; | ||
| 1095 | |||
| 1096 | spin_lock_irqsave(&adapter->mb_lock, flags); | ||
| 1097 | |||
| 1098 | tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); | ||
| 1099 | rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use); | ||
| 1100 | rrd_next_to_clean = atomic_read(&adapter->rrd_ring.next_to_clean); | ||
| 1101 | |||
| 1102 | value = ((rfd_next_to_use & MB_RFD_PROD_INDX_MASK) << | ||
| 1103 | MB_RFD_PROD_INDX_SHIFT) | | ||
| 1104 | ((rrd_next_to_clean & MB_RRD_CONS_INDX_MASK) << | ||
| 1105 | MB_RRD_CONS_INDX_SHIFT) | | ||
| 1106 | ((tpd_next_to_use & MB_TPD_PROD_INDX_MASK) << | ||
| 1107 | MB_TPD_PROD_INDX_SHIFT); | ||
| 1108 | iowrite32(value, adapter->hw.hw_addr + REG_MAILBOX); | ||
| 1109 | |||
| 1110 | spin_unlock_irqrestore(&adapter->mb_lock, flags); | ||
| 1111 | } | ||
| 1112 | |||
| 1113 | static void atl1_clean_alloc_flag(struct atl1_adapter *adapter, | ||
| 1114 | struct rx_return_desc *rrd, u16 offset) | ||
| 1115 | { | ||
| 1116 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | ||
| 1117 | |||
| 1118 | while (rfd_ring->next_to_clean != (rrd->buf_indx + offset)) { | ||
| 1119 | rfd_ring->buffer_info[rfd_ring->next_to_clean].alloced = 0; | ||
| 1120 | if (++rfd_ring->next_to_clean == rfd_ring->count) { | ||
| 1121 | rfd_ring->next_to_clean = 0; | ||
| 1122 | } | ||
| 1221 | } | 1123 | } |
| 1124 | } | ||
| 1222 | 1125 | ||
| 1223 | /* atl1_irq_enable(adapter); */ | 1126 | static void atl1_update_rfd_index(struct atl1_adapter *adapter, |
| 1224 | spin_unlock_irqrestore(&adapter->lock, flags); | 1127 | struct rx_return_desc *rrd) |
| 1128 | { | ||
| 1129 | u16 num_buf; | ||
| 1130 | |||
| 1131 | num_buf = (rrd->xsz.xsum_sz.pkt_size + adapter->rx_buffer_len - 1) / | ||
| 1132 | adapter->rx_buffer_len; | ||
| 1133 | if (rrd->num_buf == num_buf) | ||
| 1134 | /* clean alloc flag for bad rrd */ | ||
| 1135 | atl1_clean_alloc_flag(adapter, rrd, num_buf); | ||
| 1225 | } | 1136 | } |
| 1226 | 1137 | ||
| 1227 | static void atl1_restore_vlan(struct atl1_adapter *adapter) | 1138 | static void atl1_rx_checksum(struct atl1_adapter *adapter, |
| 1139 | struct rx_return_desc *rrd, struct sk_buff *skb) | ||
| 1228 | { | 1140 | { |
| 1229 | atl1_vlan_rx_register(adapter->netdev, adapter->vlgrp); | 1141 | struct pci_dev *pdev = adapter->pdev; |
| 1142 | |||
| 1143 | skb->ip_summed = CHECKSUM_NONE; | ||
| 1144 | |||
| 1145 | if (unlikely(rrd->pkt_flg & PACKET_FLAG_ERR)) { | ||
| 1146 | if (rrd->err_flg & (ERR_FLAG_CRC | ERR_FLAG_TRUNC | | ||
| 1147 | ERR_FLAG_CODE | ERR_FLAG_OV)) { | ||
| 1148 | adapter->hw_csum_err++; | ||
| 1149 | dev_printk(KERN_DEBUG, &pdev->dev, | ||
| 1150 | "rx checksum error\n"); | ||
| 1151 | return; | ||
| 1152 | } | ||
| 1153 | } | ||
| 1154 | |||
| 1155 | /* not IPv4 */ | ||
| 1156 | if (!(rrd->pkt_flg & PACKET_FLAG_IPV4)) | ||
| 1157 | /* checksum is invalid, but it's not an IPv4 pkt, so ok */ | ||
| 1158 | return; | ||
| 1159 | |||
| 1160 | /* IPv4 packet */ | ||
| 1161 | if (likely(!(rrd->err_flg & | ||
| 1162 | (ERR_FLAG_IP_CHKSUM | ERR_FLAG_L4_CHKSUM)))) { | ||
| 1163 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
| 1164 | adapter->hw_csum_good++; | ||
| 1165 | return; | ||
| 1166 | } | ||
| 1167 | |||
| 1168 | /* IPv4, but hardware thinks its checksum is wrong */ | ||
| 1169 | dev_printk(KERN_DEBUG, &pdev->dev, | ||
| 1170 | "hw csum wrong, pkt_flag:%x, err_flag:%x\n", | ||
| 1171 | rrd->pkt_flg, rrd->err_flg); | ||
| 1172 | skb->ip_summed = CHECKSUM_COMPLETE; | ||
| 1173 | skb->csum = htons(rrd->xsz.xsum_sz.rx_chksum); | ||
| 1174 | adapter->hw_csum_err++; | ||
| 1175 | return; | ||
| 1176 | } | ||
| 1177 | |||
| 1178 | /* | ||
| 1179 | * atl1_alloc_rx_buffers - Replace used receive buffers | ||
| 1180 | * @adapter: address of board private structure | ||
| 1181 | */ | ||
| 1182 | static u16 atl1_alloc_rx_buffers(struct atl1_adapter *adapter) | ||
| 1183 | { | ||
| 1184 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | ||
| 1185 | struct pci_dev *pdev = adapter->pdev; | ||
| 1186 | struct page *page; | ||
| 1187 | unsigned long offset; | ||
| 1188 | struct atl1_buffer *buffer_info, *next_info; | ||
| 1189 | struct sk_buff *skb; | ||
| 1190 | u16 num_alloc = 0; | ||
| 1191 | u16 rfd_next_to_use, next_next; | ||
| 1192 | struct rx_free_desc *rfd_desc; | ||
| 1193 | |||
| 1194 | next_next = rfd_next_to_use = atomic_read(&rfd_ring->next_to_use); | ||
| 1195 | if (++next_next == rfd_ring->count) | ||
| 1196 | next_next = 0; | ||
| 1197 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; | ||
| 1198 | next_info = &rfd_ring->buffer_info[next_next]; | ||
| 1199 | |||
| 1200 | while (!buffer_info->alloced && !next_info->alloced) { | ||
| 1201 | if (buffer_info->skb) { | ||
| 1202 | buffer_info->alloced = 1; | ||
| 1203 | goto next; | ||
| 1204 | } | ||
| 1205 | |||
| 1206 | rfd_desc = ATL1_RFD_DESC(rfd_ring, rfd_next_to_use); | ||
| 1207 | |||
| 1208 | skb = dev_alloc_skb(adapter->rx_buffer_len + NET_IP_ALIGN); | ||
| 1209 | if (unlikely(!skb)) { /* Better luck next round */ | ||
| 1210 | adapter->net_stats.rx_dropped++; | ||
| 1211 | break; | ||
| 1212 | } | ||
| 1213 | |||
| 1214 | /* | ||
| 1215 | * Make buffer alignment 2 beyond a 16 byte boundary | ||
| 1216 | * this will result in a 16 byte aligned IP header after | ||
| 1217 | * the 14 byte MAC header is removed | ||
| 1218 | */ | ||
| 1219 | skb_reserve(skb, NET_IP_ALIGN); | ||
| 1220 | |||
| 1221 | buffer_info->alloced = 1; | ||
| 1222 | buffer_info->skb = skb; | ||
| 1223 | buffer_info->length = (u16) adapter->rx_buffer_len; | ||
| 1224 | page = virt_to_page(skb->data); | ||
| 1225 | offset = (unsigned long)skb->data & ~PAGE_MASK; | ||
| 1226 | buffer_info->dma = pci_map_page(pdev, page, offset, | ||
| 1227 | adapter->rx_buffer_len, | ||
| 1228 | PCI_DMA_FROMDEVICE); | ||
| 1229 | rfd_desc->buffer_addr = cpu_to_le64(buffer_info->dma); | ||
| 1230 | rfd_desc->buf_len = cpu_to_le16(adapter->rx_buffer_len); | ||
| 1231 | rfd_desc->coalese = 0; | ||
| 1232 | |||
| 1233 | next: | ||
| 1234 | rfd_next_to_use = next_next; | ||
| 1235 | if (unlikely(++next_next == rfd_ring->count)) | ||
| 1236 | next_next = 0; | ||
| 1237 | |||
| 1238 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; | ||
| 1239 | next_info = &rfd_ring->buffer_info[next_next]; | ||
| 1240 | num_alloc++; | ||
| 1241 | } | ||
| 1242 | |||
| 1243 | if (num_alloc) { | ||
| 1244 | /* | ||
| 1245 | * Force memory writes to complete before letting h/w | ||
| 1246 | * know there are new descriptors to fetch. (Only | ||
| 1247 | * applicable for weak-ordered memory model archs, | ||
| 1248 | * such as IA-64). | ||
| 1249 | */ | ||
| 1250 | wmb(); | ||
| 1251 | atomic_set(&rfd_ring->next_to_use, (int)rfd_next_to_use); | ||
| 1252 | } | ||
| 1253 | return num_alloc; | ||
| 1254 | } | ||
| 1255 | |||
| 1256 | static void atl1_intr_rx(struct atl1_adapter *adapter) | ||
| 1257 | { | ||
| 1258 | int i, count; | ||
| 1259 | u16 length; | ||
| 1260 | u16 rrd_next_to_clean; | ||
| 1261 | u32 value; | ||
| 1262 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | ||
| 1263 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; | ||
| 1264 | struct atl1_buffer *buffer_info; | ||
| 1265 | struct rx_return_desc *rrd; | ||
| 1266 | struct sk_buff *skb; | ||
| 1267 | |||
| 1268 | count = 0; | ||
| 1269 | |||
| 1270 | rrd_next_to_clean = atomic_read(&rrd_ring->next_to_clean); | ||
| 1271 | |||
| 1272 | while (1) { | ||
| 1273 | rrd = ATL1_RRD_DESC(rrd_ring, rrd_next_to_clean); | ||
| 1274 | i = 1; | ||
| 1275 | if (likely(rrd->xsz.valid)) { /* packet valid */ | ||
| 1276 | chk_rrd: | ||
| 1277 | /* check rrd status */ | ||
| 1278 | if (likely(rrd->num_buf == 1)) | ||
| 1279 | goto rrd_ok; | ||
| 1280 | |||
| 1281 | /* rrd seems to be bad */ | ||
| 1282 | if (unlikely(i-- > 0)) { | ||
| 1283 | /* rrd may not be DMAed completely */ | ||
| 1284 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, | ||
| 1285 | "incomplete RRD DMA transfer\n"); | ||
| 1286 | udelay(1); | ||
| 1287 | goto chk_rrd; | ||
| 1288 | } | ||
| 1289 | /* bad rrd */ | ||
| 1290 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, | ||
| 1291 | "bad RRD\n"); | ||
| 1292 | /* see if update RFD index */ | ||
| 1293 | if (rrd->num_buf > 1) | ||
| 1294 | atl1_update_rfd_index(adapter, rrd); | ||
| 1295 | |||
| 1296 | /* update rrd */ | ||
| 1297 | rrd->xsz.valid = 0; | ||
| 1298 | if (++rrd_next_to_clean == rrd_ring->count) | ||
| 1299 | rrd_next_to_clean = 0; | ||
| 1300 | count++; | ||
| 1301 | continue; | ||
| 1302 | } else { /* current rrd still not be updated */ | ||
| 1303 | |||
| 1304 | break; | ||
| 1305 | } | ||
| 1306 | rrd_ok: | ||
| 1307 | /* clean alloc flag for bad rrd */ | ||
| 1308 | atl1_clean_alloc_flag(adapter, rrd, 0); | ||
| 1309 | |||
| 1310 | buffer_info = &rfd_ring->buffer_info[rrd->buf_indx]; | ||
| 1311 | if (++rfd_ring->next_to_clean == rfd_ring->count) | ||
| 1312 | rfd_ring->next_to_clean = 0; | ||
| 1313 | |||
| 1314 | /* update rrd next to clean */ | ||
| 1315 | if (++rrd_next_to_clean == rrd_ring->count) | ||
| 1316 | rrd_next_to_clean = 0; | ||
| 1317 | count++; | ||
| 1318 | |||
| 1319 | if (unlikely(rrd->pkt_flg & PACKET_FLAG_ERR)) { | ||
| 1320 | if (!(rrd->err_flg & | ||
| 1321 | (ERR_FLAG_IP_CHKSUM | ERR_FLAG_L4_CHKSUM | ||
| 1322 | | ERR_FLAG_LEN))) { | ||
| 1323 | /* packet error, don't need upstream */ | ||
| 1324 | buffer_info->alloced = 0; | ||
| 1325 | rrd->xsz.valid = 0; | ||
| 1326 | continue; | ||
| 1327 | } | ||
| 1328 | } | ||
| 1329 | |||
| 1330 | /* Good Receive */ | ||
| 1331 | pci_unmap_page(adapter->pdev, buffer_info->dma, | ||
| 1332 | buffer_info->length, PCI_DMA_FROMDEVICE); | ||
| 1333 | skb = buffer_info->skb; | ||
| 1334 | length = le16_to_cpu(rrd->xsz.xsum_sz.pkt_size); | ||
| 1335 | |||
| 1336 | skb_put(skb, length - ETHERNET_FCS_SIZE); | ||
| 1337 | |||
| 1338 | /* Receive Checksum Offload */ | ||
| 1339 | atl1_rx_checksum(adapter, rrd, skb); | ||
| 1340 | skb->protocol = eth_type_trans(skb, adapter->netdev); | ||
| 1341 | |||
| 1342 | if (adapter->vlgrp && (rrd->pkt_flg & PACKET_FLAG_VLAN_INS)) { | ||
| 1343 | u16 vlan_tag = (rrd->vlan_tag >> 4) | | ||
| 1344 | ((rrd->vlan_tag & 7) << 13) | | ||
| 1345 | ((rrd->vlan_tag & 8) << 9); | ||
| 1346 | vlan_hwaccel_rx(skb, adapter->vlgrp, vlan_tag); | ||
| 1347 | } else | ||
| 1348 | netif_rx(skb); | ||
| 1349 | |||
| 1350 | /* let protocol layer free skb */ | ||
| 1351 | buffer_info->skb = NULL; | ||
| 1352 | buffer_info->alloced = 0; | ||
| 1353 | rrd->xsz.valid = 0; | ||
| 1354 | |||
| 1355 | adapter->netdev->last_rx = jiffies; | ||
| 1356 | } | ||
| 1357 | |||
| 1358 | atomic_set(&rrd_ring->next_to_clean, rrd_next_to_clean); | ||
| 1359 | |||
| 1360 | atl1_alloc_rx_buffers(adapter); | ||
| 1361 | |||
| 1362 | /* update mailbox ? */ | ||
| 1363 | if (count) { | ||
| 1364 | u32 tpd_next_to_use; | ||
| 1365 | u32 rfd_next_to_use; | ||
| 1366 | u32 rrd_next_to_clean; | ||
| 1367 | |||
| 1368 | spin_lock(&adapter->mb_lock); | ||
| 1369 | |||
| 1370 | tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); | ||
| 1371 | rfd_next_to_use = | ||
| 1372 | atomic_read(&adapter->rfd_ring.next_to_use); | ||
| 1373 | rrd_next_to_clean = | ||
| 1374 | atomic_read(&adapter->rrd_ring.next_to_clean); | ||
| 1375 | value = ((rfd_next_to_use & MB_RFD_PROD_INDX_MASK) << | ||
| 1376 | MB_RFD_PROD_INDX_SHIFT) | | ||
| 1377 | ((rrd_next_to_clean & MB_RRD_CONS_INDX_MASK) << | ||
| 1378 | MB_RRD_CONS_INDX_SHIFT) | | ||
| 1379 | ((tpd_next_to_use & MB_TPD_PROD_INDX_MASK) << | ||
| 1380 | MB_TPD_PROD_INDX_SHIFT); | ||
| 1381 | iowrite32(value, adapter->hw.hw_addr + REG_MAILBOX); | ||
| 1382 | spin_unlock(&adapter->mb_lock); | ||
| 1383 | } | ||
| 1384 | } | ||
| 1385 | |||
| 1386 | static void atl1_intr_tx(struct atl1_adapter *adapter) | ||
| 1387 | { | ||
| 1388 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | ||
| 1389 | struct atl1_buffer *buffer_info; | ||
| 1390 | u16 sw_tpd_next_to_clean; | ||
| 1391 | u16 cmb_tpd_next_to_clean; | ||
| 1392 | |||
| 1393 | sw_tpd_next_to_clean = atomic_read(&tpd_ring->next_to_clean); | ||
| 1394 | cmb_tpd_next_to_clean = le16_to_cpu(adapter->cmb.cmb->tpd_cons_idx); | ||
| 1395 | |||
| 1396 | while (cmb_tpd_next_to_clean != sw_tpd_next_to_clean) { | ||
| 1397 | struct tx_packet_desc *tpd; | ||
| 1398 | |||
| 1399 | tpd = ATL1_TPD_DESC(tpd_ring, sw_tpd_next_to_clean); | ||
| 1400 | buffer_info = &tpd_ring->buffer_info[sw_tpd_next_to_clean]; | ||
| 1401 | if (buffer_info->dma) { | ||
| 1402 | pci_unmap_page(adapter->pdev, buffer_info->dma, | ||
| 1403 | buffer_info->length, PCI_DMA_TODEVICE); | ||
| 1404 | buffer_info->dma = 0; | ||
| 1405 | } | ||
| 1406 | |||
| 1407 | if (buffer_info->skb) { | ||
| 1408 | dev_kfree_skb_irq(buffer_info->skb); | ||
| 1409 | buffer_info->skb = NULL; | ||
| 1410 | } | ||
| 1411 | tpd->buffer_addr = 0; | ||
| 1412 | tpd->desc.data = 0; | ||
| 1413 | |||
| 1414 | if (++sw_tpd_next_to_clean == tpd_ring->count) | ||
| 1415 | sw_tpd_next_to_clean = 0; | ||
| 1416 | } | ||
| 1417 | atomic_set(&tpd_ring->next_to_clean, sw_tpd_next_to_clean); | ||
| 1418 | |||
| 1419 | if (netif_queue_stopped(adapter->netdev) | ||
| 1420 | && netif_carrier_ok(adapter->netdev)) | ||
| 1421 | netif_wake_queue(adapter->netdev); | ||
| 1230 | } | 1422 | } |
| 1231 | 1423 | ||
| 1232 | static u16 tpd_avail(struct atl1_tpd_ring *tpd_ring) | 1424 | static u16 tpd_avail(struct atl1_tpd_ring *tpd_ring) |
| 1233 | { | 1425 | { |
| 1234 | u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean); | 1426 | u16 next_to_clean = atomic_read(&tpd_ring->next_to_clean); |
| 1235 | u16 next_to_use = atomic_read(&tpd_ring->next_to_use); | 1427 | u16 next_to_use = atomic_read(&tpd_ring->next_to_use); |
| 1236 | return ((next_to_clean > | 1428 | return ((next_to_clean > next_to_use) ? |
| 1237 | next_to_use) ? next_to_clean - next_to_use - | 1429 | next_to_clean - next_to_use - 1 : |
| 1238 | 1 : tpd_ring->count + next_to_clean - next_to_use - 1); | 1430 | tpd_ring->count + next_to_clean - next_to_use - 1); |
| 1239 | } | 1431 | } |
| 1240 | 1432 | ||
| 1241 | static int atl1_tso(struct atl1_adapter *adapter, struct sk_buff *skb, | 1433 | static int atl1_tso(struct atl1_adapter *adapter, struct sk_buff *skb, |
| @@ -1258,9 +1450,7 @@ static int atl1_tso(struct atl1_adapter *adapter, struct sk_buff *skb, | |||
| 1258 | iph->tot_len = 0; | 1450 | iph->tot_len = 0; |
| 1259 | iph->check = 0; | 1451 | iph->check = 0; |
| 1260 | tcp_hdr(skb)->check = ~csum_tcpudp_magic(iph->saddr, | 1452 | tcp_hdr(skb)->check = ~csum_tcpudp_magic(iph->saddr, |
| 1261 | iph->daddr, 0, | 1453 | iph->daddr, 0, IPPROTO_TCP, 0); |
| 1262 | IPPROTO_TCP, | ||
| 1263 | 0); | ||
| 1264 | ipofst = skb_network_offset(skb); | 1454 | ipofst = skb_network_offset(skb); |
| 1265 | if (ipofst != ENET_HEADER_SIZE) /* 802.3 frame */ | 1455 | if (ipofst != ENET_HEADER_SIZE) /* 802.3 frame */ |
| 1266 | tso->tsopl |= 1 << TSO_PARAM_ETHTYPE_SHIFT; | 1456 | tso->tsopl |= 1 << TSO_PARAM_ETHTYPE_SHIFT; |
| @@ -1268,7 +1458,8 @@ static int atl1_tso(struct atl1_adapter *adapter, struct sk_buff *skb, | |||
| 1268 | tso->tsopl |= (iph->ihl & | 1458 | tso->tsopl |= (iph->ihl & |
| 1269 | CSUM_PARAM_IPHL_MASK) << CSUM_PARAM_IPHL_SHIFT; | 1459 | CSUM_PARAM_IPHL_MASK) << CSUM_PARAM_IPHL_SHIFT; |
| 1270 | tso->tsopl |= (tcp_hdrlen(skb) & | 1460 | tso->tsopl |= (tcp_hdrlen(skb) & |
| 1271 | TSO_PARAM_TCPHDRLEN_MASK) << TSO_PARAM_TCPHDRLEN_SHIFT; | 1461 | TSO_PARAM_TCPHDRLEN_MASK) << |
| 1462 | TSO_PARAM_TCPHDRLEN_SHIFT; | ||
| 1272 | tso->tsopl |= (skb_shinfo(skb)->gso_size & | 1463 | tso->tsopl |= (skb_shinfo(skb)->gso_size & |
| 1273 | TSO_PARAM_MSS_MASK) << TSO_PARAM_MSS_SHIFT; | 1464 | TSO_PARAM_MSS_MASK) << TSO_PARAM_MSS_SHIFT; |
| 1274 | tso->tsopl |= 1 << TSO_PARAM_IPCKSUM_SHIFT; | 1465 | tso->tsopl |= 1 << TSO_PARAM_IPCKSUM_SHIFT; |
| @@ -1281,7 +1472,7 @@ static int atl1_tso(struct atl1_adapter *adapter, struct sk_buff *skb, | |||
| 1281 | } | 1472 | } |
| 1282 | 1473 | ||
| 1283 | static int atl1_tx_csum(struct atl1_adapter *adapter, struct sk_buff *skb, | 1474 | static int atl1_tx_csum(struct atl1_adapter *adapter, struct sk_buff *skb, |
| 1284 | struct csum_param *csum) | 1475 | struct csum_param *csum) |
| 1285 | { | 1476 | { |
| 1286 | u8 css, cso; | 1477 | u8 css, cso; |
| 1287 | 1478 | ||
| @@ -1289,7 +1480,7 @@ static int atl1_tx_csum(struct atl1_adapter *adapter, struct sk_buff *skb, | |||
| 1289 | cso = skb_transport_offset(skb); | 1480 | cso = skb_transport_offset(skb); |
| 1290 | css = cso + skb->csum_offset; | 1481 | css = cso + skb->csum_offset; |
| 1291 | if (unlikely(cso & 0x1)) { | 1482 | if (unlikely(cso & 0x1)) { |
| 1292 | dev_dbg(&adapter->pdev->dev, | 1483 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, |
| 1293 | "payload offset not an even number\n"); | 1484 | "payload offset not an even number\n"); |
| 1294 | return -1; | 1485 | return -1; |
| 1295 | } | 1486 | } |
| @@ -1304,8 +1495,8 @@ static int atl1_tx_csum(struct atl1_adapter *adapter, struct sk_buff *skb, | |||
| 1304 | return true; | 1495 | return true; |
| 1305 | } | 1496 | } |
| 1306 | 1497 | ||
| 1307 | static void atl1_tx_map(struct atl1_adapter *adapter, | 1498 | static void atl1_tx_map(struct atl1_adapter *adapter, struct sk_buff *skb, |
| 1308 | struct sk_buff *skb, bool tcp_seg) | 1499 | bool tcp_seg) |
| 1309 | { | 1500 | { |
| 1310 | /* We enter this function holding a spinlock. */ | 1501 | /* We enter this function holding a spinlock. */ |
| 1311 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | 1502 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; |
| @@ -1342,26 +1533,25 @@ static void atl1_tx_map(struct atl1_adapter *adapter, | |||
| 1342 | 1533 | ||
| 1343 | if (first_buf_len > proto_hdr_len) { | 1534 | if (first_buf_len > proto_hdr_len) { |
| 1344 | len12 = first_buf_len - proto_hdr_len; | 1535 | len12 = first_buf_len - proto_hdr_len; |
| 1345 | m = (len12 + MAX_TX_BUF_LEN - 1) / MAX_TX_BUF_LEN; | 1536 | m = (len12 + ATL1_MAX_TX_BUF_LEN - 1) / |
| 1537 | ATL1_MAX_TX_BUF_LEN; | ||
| 1346 | for (i = 0; i < m; i++) { | 1538 | for (i = 0; i < m; i++) { |
| 1347 | buffer_info = | 1539 | buffer_info = |
| 1348 | &tpd_ring->buffer_info[tpd_next_to_use]; | 1540 | &tpd_ring->buffer_info[tpd_next_to_use]; |
| 1349 | buffer_info->skb = NULL; | 1541 | buffer_info->skb = NULL; |
| 1350 | buffer_info->length = | 1542 | buffer_info->length = |
| 1351 | (MAX_TX_BUF_LEN >= | 1543 | (ATL1_MAX_TX_BUF_LEN >= |
| 1352 | len12) ? MAX_TX_BUF_LEN : len12; | 1544 | len12) ? ATL1_MAX_TX_BUF_LEN : len12; |
| 1353 | len12 -= buffer_info->length; | 1545 | len12 -= buffer_info->length; |
| 1354 | page = virt_to_page(skb->data + | 1546 | page = virt_to_page(skb->data + |
| 1355 | (proto_hdr_len + | 1547 | (proto_hdr_len + |
| 1356 | i * MAX_TX_BUF_LEN)); | 1548 | i * ATL1_MAX_TX_BUF_LEN)); |
| 1357 | offset = (unsigned long)(skb->data + | 1549 | offset = (unsigned long)(skb->data + |
| 1358 | (proto_hdr_len + | 1550 | (proto_hdr_len + |
| 1359 | i * MAX_TX_BUF_LEN)) & | 1551 | i * ATL1_MAX_TX_BUF_LEN)) & ~PAGE_MASK; |
| 1360 | ~PAGE_MASK; | 1552 | buffer_info->dma = pci_map_page(adapter->pdev, |
| 1361 | buffer_info->dma = | 1553 | page, offset, buffer_info->length, |
| 1362 | pci_map_page(adapter->pdev, page, offset, | 1554 | PCI_DMA_TODEVICE); |
| 1363 | buffer_info->length, | ||
| 1364 | PCI_DMA_TODEVICE); | ||
| 1365 | if (++tpd_next_to_use == tpd_ring->count) | 1555 | if (++tpd_next_to_use == tpd_ring->count) |
| 1366 | tpd_next_to_use = 0; | 1556 | tpd_next_to_use = 0; |
| 1367 | } | 1557 | } |
| @@ -1372,8 +1562,7 @@ static void atl1_tx_map(struct atl1_adapter *adapter, | |||
| 1372 | page = virt_to_page(skb->data); | 1562 | page = virt_to_page(skb->data); |
| 1373 | offset = (unsigned long)skb->data & ~PAGE_MASK; | 1563 | offset = (unsigned long)skb->data & ~PAGE_MASK; |
| 1374 | buffer_info->dma = pci_map_page(adapter->pdev, page, | 1564 | buffer_info->dma = pci_map_page(adapter->pdev, page, |
| 1375 | offset, first_buf_len, | 1565 | offset, first_buf_len, PCI_DMA_TODEVICE); |
| 1376 | PCI_DMA_TODEVICE); | ||
| 1377 | if (++tpd_next_to_use == tpd_ring->count) | 1566 | if (++tpd_next_to_use == tpd_ring->count) |
| 1378 | tpd_next_to_use = 0; | 1567 | tpd_next_to_use = 0; |
| 1379 | } | 1568 | } |
| @@ -1385,19 +1574,19 @@ static void atl1_tx_map(struct atl1_adapter *adapter, | |||
| 1385 | frag = &skb_shinfo(skb)->frags[f]; | 1574 | frag = &skb_shinfo(skb)->frags[f]; |
| 1386 | lenf = frag->size; | 1575 | lenf = frag->size; |
| 1387 | 1576 | ||
| 1388 | m = (lenf + MAX_TX_BUF_LEN - 1) / MAX_TX_BUF_LEN; | 1577 | m = (lenf + ATL1_MAX_TX_BUF_LEN - 1) / ATL1_MAX_TX_BUF_LEN; |
| 1389 | for (i = 0; i < m; i++) { | 1578 | for (i = 0; i < m; i++) { |
| 1390 | buffer_info = &tpd_ring->buffer_info[tpd_next_to_use]; | 1579 | buffer_info = &tpd_ring->buffer_info[tpd_next_to_use]; |
| 1391 | if (unlikely(buffer_info->skb)) | 1580 | if (unlikely(buffer_info->skb)) |
| 1392 | BUG(); | 1581 | BUG(); |
| 1393 | buffer_info->skb = NULL; | 1582 | buffer_info->skb = NULL; |
| 1394 | buffer_info->length = | 1583 | buffer_info->length = (lenf > ATL1_MAX_TX_BUF_LEN) ? |
| 1395 | (lenf > MAX_TX_BUF_LEN) ? MAX_TX_BUF_LEN : lenf; | 1584 | ATL1_MAX_TX_BUF_LEN : lenf; |
| 1396 | lenf -= buffer_info->length; | 1585 | lenf -= buffer_info->length; |
| 1397 | buffer_info->dma = | 1586 | buffer_info->dma = pci_map_page(adapter->pdev, |
| 1398 | pci_map_page(adapter->pdev, frag->page, | 1587 | frag->page, |
| 1399 | frag->page_offset + i * MAX_TX_BUF_LEN, | 1588 | frag->page_offset + (i * ATL1_MAX_TX_BUF_LEN), |
| 1400 | buffer_info->length, PCI_DMA_TODEVICE); | 1589 | buffer_info->length, PCI_DMA_TODEVICE); |
| 1401 | 1590 | ||
| 1402 | if (++tpd_next_to_use == tpd_ring->count) | 1591 | if (++tpd_next_to_use == tpd_ring->count) |
| 1403 | tpd_next_to_use = 0; | 1592 | tpd_next_to_use = 0; |
| @@ -1409,7 +1598,7 @@ static void atl1_tx_map(struct atl1_adapter *adapter, | |||
| 1409 | } | 1598 | } |
| 1410 | 1599 | ||
| 1411 | static void atl1_tx_queue(struct atl1_adapter *adapter, int count, | 1600 | static void atl1_tx_queue(struct atl1_adapter *adapter, int count, |
| 1412 | union tpd_descr *descr) | 1601 | union tpd_descr *descr) |
| 1413 | { | 1602 | { |
| 1414 | /* We enter this function holding a spinlock. */ | 1603 | /* We enter this function holding a spinlock. */ |
| 1415 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | 1604 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; |
| @@ -1453,31 +1642,6 @@ static void atl1_tx_queue(struct atl1_adapter *adapter, int count, | |||
| 1453 | atomic_set(&tpd_ring->next_to_use, (int)tpd_next_to_use); | 1642 | atomic_set(&tpd_ring->next_to_use, (int)tpd_next_to_use); |
| 1454 | } | 1643 | } |
| 1455 | 1644 | ||
| 1456 | static void atl1_update_mailbox(struct atl1_adapter *adapter) | ||
| 1457 | { | ||
| 1458 | unsigned long flags; | ||
| 1459 | u32 tpd_next_to_use; | ||
| 1460 | u32 rfd_next_to_use; | ||
| 1461 | u32 rrd_next_to_clean; | ||
| 1462 | u32 value; | ||
| 1463 | |||
| 1464 | spin_lock_irqsave(&adapter->mb_lock, flags); | ||
| 1465 | |||
| 1466 | tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); | ||
| 1467 | rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use); | ||
| 1468 | rrd_next_to_clean = atomic_read(&adapter->rrd_ring.next_to_clean); | ||
| 1469 | |||
| 1470 | value = ((rfd_next_to_use & MB_RFD_PROD_INDX_MASK) << | ||
| 1471 | MB_RFD_PROD_INDX_SHIFT) | | ||
| 1472 | ((rrd_next_to_clean & MB_RRD_CONS_INDX_MASK) << | ||
| 1473 | MB_RRD_CONS_INDX_SHIFT) | | ||
| 1474 | ((tpd_next_to_use & MB_TPD_PROD_INDX_MASK) << | ||
| 1475 | MB_TPD_PROD_INDX_SHIFT); | ||
| 1476 | iowrite32(value, adapter->hw.hw_addr + REG_MAILBOX); | ||
| 1477 | |||
| 1478 | spin_unlock_irqrestore(&adapter->mb_lock, flags); | ||
| 1479 | } | ||
| 1480 | |||
| 1481 | static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | 1645 | static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) |
| 1482 | { | 1646 | { |
| 1483 | struct atl1_adapter *adapter = netdev_priv(netdev); | 1647 | struct atl1_adapter *adapter = netdev_priv(netdev); |
| @@ -1513,8 +1677,8 @@ static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | |||
| 1513 | for (f = 0; f < nr_frags; f++) { | 1677 | for (f = 0; f < nr_frags; f++) { |
| 1514 | frag_size = skb_shinfo(skb)->frags[f].size; | 1678 | frag_size = skb_shinfo(skb)->frags[f].size; |
| 1515 | if (frag_size) | 1679 | if (frag_size) |
| 1516 | count += | 1680 | count += (frag_size + ATL1_MAX_TX_BUF_LEN - 1) / |
| 1517 | (frag_size + MAX_TX_BUF_LEN - 1) / MAX_TX_BUF_LEN; | 1681 | ATL1_MAX_TX_BUF_LEN; |
| 1518 | } | 1682 | } |
| 1519 | 1683 | ||
| 1520 | /* mss will be nonzero if we're doing segment offload (TSO/GSO) */ | 1684 | /* mss will be nonzero if we're doing segment offload (TSO/GSO) */ |
| @@ -1530,7 +1694,8 @@ static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | |||
| 1530 | /* need additional TPD ? */ | 1694 | /* need additional TPD ? */ |
| 1531 | if (proto_hdr_len != len) | 1695 | if (proto_hdr_len != len) |
| 1532 | count += (len - proto_hdr_len + | 1696 | count += (len - proto_hdr_len + |
| 1533 | MAX_TX_BUF_LEN - 1) / MAX_TX_BUF_LEN; | 1697 | ATL1_MAX_TX_BUF_LEN - 1) / |
| 1698 | ATL1_MAX_TX_BUF_LEN; | ||
| 1534 | } | 1699 | } |
| 1535 | } | 1700 | } |
| 1536 | 1701 | ||
| @@ -1538,7 +1703,7 @@ static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | |||
| 1538 | if (!spin_trylock(&adapter->lock)) { | 1703 | if (!spin_trylock(&adapter->lock)) { |
| 1539 | /* Can't get lock - tell upper layer to requeue */ | 1704 | /* Can't get lock - tell upper layer to requeue */ |
| 1540 | local_irq_restore(flags); | 1705 | local_irq_restore(flags); |
| 1541 | dev_dbg(&adapter->pdev->dev, "tx locked\n"); | 1706 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, "tx locked\n"); |
| 1542 | return NETDEV_TX_LOCKED; | 1707 | return NETDEV_TX_LOCKED; |
| 1543 | } | 1708 | } |
| 1544 | 1709 | ||
| @@ -1546,7 +1711,7 @@ static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | |||
| 1546 | /* not enough descriptors */ | 1711 | /* not enough descriptors */ |
| 1547 | netif_stop_queue(netdev); | 1712 | netif_stop_queue(netdev); |
| 1548 | spin_unlock_irqrestore(&adapter->lock, flags); | 1713 | spin_unlock_irqrestore(&adapter->lock, flags); |
| 1549 | dev_dbg(&adapter->pdev->dev, "tx busy\n"); | 1714 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, "tx busy\n"); |
| 1550 | return NETDEV_TX_BUSY; | 1715 | return NETDEV_TX_BUSY; |
| 1551 | } | 1716 | } |
| 1552 | 1717 | ||
| @@ -1588,131 +1753,208 @@ static int atl1_xmit_frame(struct sk_buff *skb, struct net_device *netdev) | |||
| 1588 | } | 1753 | } |
| 1589 | 1754 | ||
| 1590 | /* | 1755 | /* |
| 1591 | * atl1_get_stats - Get System Network Statistics | 1756 | * atl1_intr - Interrupt Handler |
| 1592 | * @netdev: network interface device structure | 1757 | * @irq: interrupt number |
| 1593 | * | 1758 | * @data: pointer to a network interface device structure |
| 1594 | * Returns the address of the device statistics structure. | 1759 | * @pt_regs: CPU registers structure |
| 1595 | * The statistics are actually updated from the timer callback. | ||
| 1596 | */ | 1760 | */ |
| 1597 | static struct net_device_stats *atl1_get_stats(struct net_device *netdev) | 1761 | static irqreturn_t atl1_intr(int irq, void *data) |
| 1598 | { | 1762 | { |
| 1599 | struct atl1_adapter *adapter = netdev_priv(netdev); | 1763 | struct atl1_adapter *adapter = netdev_priv(data); |
| 1600 | return &adapter->net_stats; | 1764 | u32 status; |
| 1601 | } | 1765 | u8 update_rx; |
| 1766 | int max_ints = 10; | ||
| 1602 | 1767 | ||
| 1603 | /* | 1768 | status = adapter->cmb.cmb->int_stats; |
| 1604 | * atl1_clean_rx_ring - Free RFD Buffers | 1769 | if (!status) |
| 1605 | * @adapter: board private structure | 1770 | return IRQ_NONE; |
| 1606 | */ | ||
| 1607 | static void atl1_clean_rx_ring(struct atl1_adapter *adapter) | ||
| 1608 | { | ||
| 1609 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | ||
| 1610 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; | ||
| 1611 | struct atl1_buffer *buffer_info; | ||
| 1612 | struct pci_dev *pdev = adapter->pdev; | ||
| 1613 | unsigned long size; | ||
| 1614 | unsigned int i; | ||
| 1615 | 1771 | ||
| 1616 | /* Free all the Rx ring sk_buffs */ | 1772 | update_rx = 0; |
| 1617 | for (i = 0; i < rfd_ring->count; i++) { | 1773 | |
| 1618 | buffer_info = &rfd_ring->buffer_info[i]; | 1774 | do { |
| 1619 | if (buffer_info->dma) { | 1775 | /* clear CMB interrupt status at once */ |
| 1620 | pci_unmap_page(pdev, | 1776 | adapter->cmb.cmb->int_stats = 0; |
| 1621 | buffer_info->dma, | 1777 | |
| 1622 | buffer_info->length, | 1778 | if (status & ISR_GPHY) /* clear phy status */ |
| 1623 | PCI_DMA_FROMDEVICE); | 1779 | atl1_clear_phy_int(adapter); |
| 1624 | buffer_info->dma = 0; | 1780 | |
| 1781 | /* clear ISR status, and Enable CMB DMA/Disable Interrupt */ | ||
| 1782 | iowrite32(status | ISR_DIS_INT, adapter->hw.hw_addr + REG_ISR); | ||
| 1783 | |||
| 1784 | /* check if SMB intr */ | ||
| 1785 | if (status & ISR_SMB) | ||
| 1786 | atl1_inc_smb(adapter); | ||
| 1787 | |||
| 1788 | /* check if PCIE PHY Link down */ | ||
| 1789 | if (status & ISR_PHY_LINKDOWN) { | ||
| 1790 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, | ||
| 1791 | "pcie phy link down %x\n", status); | ||
| 1792 | if (netif_running(adapter->netdev)) { /* reset MAC */ | ||
| 1793 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | ||
| 1794 | schedule_work(&adapter->pcie_dma_to_rst_task); | ||
| 1795 | return IRQ_HANDLED; | ||
| 1796 | } | ||
| 1625 | } | 1797 | } |
| 1626 | if (buffer_info->skb) { | 1798 | |
| 1627 | dev_kfree_skb(buffer_info->skb); | 1799 | /* check if DMA read/write error ? */ |
| 1628 | buffer_info->skb = NULL; | 1800 | if (status & (ISR_DMAR_TO_RST | ISR_DMAW_TO_RST)) { |
| 1801 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, | ||
| 1802 | "pcie DMA r/w error (status = 0x%x)\n", | ||
| 1803 | status); | ||
| 1804 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | ||
| 1805 | schedule_work(&adapter->pcie_dma_to_rst_task); | ||
| 1806 | return IRQ_HANDLED; | ||
| 1629 | } | 1807 | } |
| 1630 | } | ||
| 1631 | 1808 | ||
| 1632 | size = sizeof(struct atl1_buffer) * rfd_ring->count; | 1809 | /* link event */ |
| 1633 | memset(rfd_ring->buffer_info, 0, size); | 1810 | if (status & ISR_GPHY) { |
| 1811 | adapter->soft_stats.tx_carrier_errors++; | ||
| 1812 | atl1_check_for_link(adapter); | ||
| 1813 | } | ||
| 1634 | 1814 | ||
| 1635 | /* Zero out the descriptor ring */ | 1815 | /* transmit event */ |
| 1636 | memset(rfd_ring->desc, 0, rfd_ring->size); | 1816 | if (status & ISR_CMB_TX) |
| 1817 | atl1_intr_tx(adapter); | ||
| 1637 | 1818 | ||
| 1638 | rfd_ring->next_to_clean = 0; | 1819 | /* rx exception */ |
| 1639 | atomic_set(&rfd_ring->next_to_use, 0); | 1820 | if (unlikely(status & (ISR_RXF_OV | ISR_RFD_UNRUN | |
| 1821 | ISR_RRD_OV | ISR_HOST_RFD_UNRUN | | ||
| 1822 | ISR_HOST_RRD_OV | ISR_CMB_RX))) { | ||
| 1823 | if (status & (ISR_RXF_OV | ISR_RFD_UNRUN | | ||
| 1824 | ISR_RRD_OV | ISR_HOST_RFD_UNRUN | | ||
| 1825 | ISR_HOST_RRD_OV)) | ||
| 1826 | dev_printk(KERN_DEBUG, &adapter->pdev->dev, | ||
| 1827 | "rx exception, ISR = 0x%x\n", status); | ||
| 1828 | atl1_intr_rx(adapter); | ||
| 1829 | } | ||
| 1640 | 1830 | ||
| 1641 | rrd_ring->next_to_use = 0; | 1831 | if (--max_ints < 0) |
| 1642 | atomic_set(&rrd_ring->next_to_clean, 0); | 1832 | break; |
| 1833 | |||
| 1834 | } while ((status = adapter->cmb.cmb->int_stats)); | ||
| 1835 | |||
| 1836 | /* re-enable Interrupt */ | ||
| 1837 | iowrite32(ISR_DIS_SMB | ISR_DIS_DMA, adapter->hw.hw_addr + REG_ISR); | ||
| 1838 | return IRQ_HANDLED; | ||
| 1643 | } | 1839 | } |
| 1644 | 1840 | ||
| 1645 | /* | 1841 | /* |
| 1646 | * atl1_clean_tx_ring - Free Tx Buffers | 1842 | * atl1_watchdog - Timer Call-back |
| 1647 | * @adapter: board private structure | 1843 | * @data: pointer to netdev cast into an unsigned long |
| 1648 | */ | 1844 | */ |
| 1649 | static void atl1_clean_tx_ring(struct atl1_adapter *adapter) | 1845 | static void atl1_watchdog(unsigned long data) |
| 1650 | { | 1846 | { |
| 1651 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | 1847 | struct atl1_adapter *adapter = (struct atl1_adapter *)data; |
| 1652 | struct atl1_buffer *buffer_info; | ||
| 1653 | struct pci_dev *pdev = adapter->pdev; | ||
| 1654 | unsigned long size; | ||
| 1655 | unsigned int i; | ||
| 1656 | 1848 | ||
| 1657 | /* Free all the Tx ring sk_buffs */ | 1849 | /* Reset the timer */ |
| 1658 | for (i = 0; i < tpd_ring->count; i++) { | 1850 | mod_timer(&adapter->watchdog_timer, jiffies + 2 * HZ); |
| 1659 | buffer_info = &tpd_ring->buffer_info[i]; | 1851 | } |
| 1660 | if (buffer_info->dma) { | ||
| 1661 | pci_unmap_page(pdev, buffer_info->dma, | ||
| 1662 | buffer_info->length, PCI_DMA_TODEVICE); | ||
| 1663 | buffer_info->dma = 0; | ||
| 1664 | } | ||
| 1665 | } | ||
| 1666 | 1852 | ||
| 1667 | for (i = 0; i < tpd_ring->count; i++) { | 1853 | /* |
| 1668 | buffer_info = &tpd_ring->buffer_info[i]; | 1854 | * atl1_phy_config - Timer Call-back |
| 1669 | if (buffer_info->skb) { | 1855 | * @data: pointer to netdev cast into an unsigned long |
| 1670 | dev_kfree_skb_any(buffer_info->skb); | 1856 | */ |
| 1671 | buffer_info->skb = NULL; | 1857 | static void atl1_phy_config(unsigned long data) |
| 1672 | } | 1858 | { |
| 1673 | } | 1859 | struct atl1_adapter *adapter = (struct atl1_adapter *)data; |
| 1860 | struct atl1_hw *hw = &adapter->hw; | ||
| 1861 | unsigned long flags; | ||
| 1674 | 1862 | ||
| 1675 | size = sizeof(struct atl1_buffer) * tpd_ring->count; | 1863 | spin_lock_irqsave(&adapter->lock, flags); |
| 1676 | memset(tpd_ring->buffer_info, 0, size); | 1864 | adapter->phy_timer_pending = false; |
| 1865 | atl1_write_phy_reg(hw, MII_ADVERTISE, hw->mii_autoneg_adv_reg); | ||
| 1866 | atl1_write_phy_reg(hw, MII_AT001_CR, hw->mii_1000t_ctrl_reg); | ||
| 1867 | atl1_write_phy_reg(hw, MII_BMCR, MII_CR_RESET | MII_CR_AUTO_NEG_EN); | ||
| 1868 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 1869 | } | ||
| 1677 | 1870 | ||
| 1678 | /* Zero out the descriptor ring */ | 1871 | /* |
| 1679 | memset(tpd_ring->desc, 0, tpd_ring->size); | 1872 | * atl1_tx_timeout - Respond to a Tx Hang |
| 1873 | * @netdev: network interface device structure | ||
| 1874 | */ | ||
| 1875 | static void atl1_tx_timeout(struct net_device *netdev) | ||
| 1876 | { | ||
| 1877 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1878 | /* Do the reset outside of interrupt context */ | ||
| 1879 | schedule_work(&adapter->tx_timeout_task); | ||
| 1880 | } | ||
| 1680 | 1881 | ||
| 1681 | atomic_set(&tpd_ring->next_to_use, 0); | 1882 | /* |
| 1682 | atomic_set(&tpd_ring->next_to_clean, 0); | 1883 | * Orphaned vendor comment left intact here: |
| 1884 | * <vendor comment> | ||
| 1885 | * If TPD Buffer size equal to 0, PCIE DMAR_TO_INT | ||
| 1886 | * will assert. We do soft reset <0x1400=1> according | ||
| 1887 | * with the SPEC. BUT, it seemes that PCIE or DMA | ||
| 1888 | * state-machine will not be reset. DMAR_TO_INT will | ||
| 1889 | * assert again and again. | ||
| 1890 | * </vendor comment> | ||
| 1891 | */ | ||
| 1892 | static void atl1_tx_timeout_task(struct work_struct *work) | ||
| 1893 | { | ||
| 1894 | struct atl1_adapter *adapter = | ||
| 1895 | container_of(work, struct atl1_adapter, tx_timeout_task); | ||
| 1896 | struct net_device *netdev = adapter->netdev; | ||
| 1897 | |||
| 1898 | netif_device_detach(netdev); | ||
| 1899 | atl1_down(adapter); | ||
| 1900 | atl1_up(adapter); | ||
| 1901 | netif_device_attach(netdev); | ||
| 1683 | } | 1902 | } |
| 1684 | 1903 | ||
| 1685 | /* | 1904 | /* |
| 1686 | * atl1_free_ring_resources - Free Tx / RX descriptor Resources | 1905 | * atl1_link_chg_task - deal with link change event Out of interrupt context |
| 1687 | * @adapter: board private structure | ||
| 1688 | * | ||
| 1689 | * Free all transmit software resources | ||
| 1690 | */ | 1906 | */ |
| 1691 | void atl1_free_ring_resources(struct atl1_adapter *adapter) | 1907 | static void atl1_link_chg_task(struct work_struct *work) |
| 1692 | { | 1908 | { |
| 1693 | struct pci_dev *pdev = adapter->pdev; | 1909 | struct atl1_adapter *adapter = |
| 1694 | struct atl1_tpd_ring *tpd_ring = &adapter->tpd_ring; | 1910 | container_of(work, struct atl1_adapter, link_chg_task); |
| 1695 | struct atl1_rfd_ring *rfd_ring = &adapter->rfd_ring; | 1911 | unsigned long flags; |
| 1696 | struct atl1_rrd_ring *rrd_ring = &adapter->rrd_ring; | ||
| 1697 | struct atl1_ring_header *ring_header = &adapter->ring_header; | ||
| 1698 | 1912 | ||
| 1699 | atl1_clean_tx_ring(adapter); | 1913 | spin_lock_irqsave(&adapter->lock, flags); |
| 1700 | atl1_clean_rx_ring(adapter); | 1914 | atl1_check_link(adapter); |
| 1915 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 1916 | } | ||
| 1701 | 1917 | ||
| 1702 | kfree(tpd_ring->buffer_info); | 1918 | static void atl1_vlan_rx_register(struct net_device *netdev, |
| 1703 | pci_free_consistent(pdev, ring_header->size, ring_header->desc, | 1919 | struct vlan_group *grp) |
| 1704 | ring_header->dma); | 1920 | { |
| 1921 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1922 | unsigned long flags; | ||
| 1923 | u32 ctrl; | ||
| 1705 | 1924 | ||
| 1706 | tpd_ring->buffer_info = NULL; | 1925 | spin_lock_irqsave(&adapter->lock, flags); |
| 1707 | tpd_ring->desc = NULL; | 1926 | /* atl1_irq_disable(adapter); */ |
| 1708 | tpd_ring->dma = 0; | 1927 | adapter->vlgrp = grp; |
| 1709 | 1928 | ||
| 1710 | rfd_ring->buffer_info = NULL; | 1929 | if (grp) { |
| 1711 | rfd_ring->desc = NULL; | 1930 | /* enable VLAN tag insert/strip */ |
| 1712 | rfd_ring->dma = 0; | 1931 | ctrl = ioread32(adapter->hw.hw_addr + REG_MAC_CTRL); |
| 1932 | ctrl |= MAC_CTRL_RMV_VLAN; | ||
| 1933 | iowrite32(ctrl, adapter->hw.hw_addr + REG_MAC_CTRL); | ||
| 1934 | } else { | ||
| 1935 | /* disable VLAN tag insert/strip */ | ||
| 1936 | ctrl = ioread32(adapter->hw.hw_addr + REG_MAC_CTRL); | ||
| 1937 | ctrl &= ~MAC_CTRL_RMV_VLAN; | ||
| 1938 | iowrite32(ctrl, adapter->hw.hw_addr + REG_MAC_CTRL); | ||
| 1939 | } | ||
| 1713 | 1940 | ||
| 1714 | rrd_ring->desc = NULL; | 1941 | /* atl1_irq_enable(adapter); */ |
| 1715 | rrd_ring->dma = 0; | 1942 | spin_unlock_irqrestore(&adapter->lock, flags); |
| 1943 | } | ||
| 1944 | |||
| 1945 | static void atl1_restore_vlan(struct atl1_adapter *adapter) | ||
| 1946 | { | ||
| 1947 | atl1_vlan_rx_register(adapter->netdev, adapter->vlgrp); | ||
| 1948 | } | ||
| 1949 | |||
| 1950 | int atl1_reset(struct atl1_adapter *adapter) | ||
| 1951 | { | ||
| 1952 | int ret; | ||
| 1953 | |||
| 1954 | ret = atl1_reset_hw(&adapter->hw); | ||
| 1955 | if (ret != ATL1_SUCCESS) | ||
| 1956 | return ret; | ||
| 1957 | return atl1_init_hw(&adapter->hw); | ||
| 1716 | } | 1958 | } |
| 1717 | 1959 | ||
| 1718 | s32 atl1_up(struct atl1_adapter *adapter) | 1960 | s32 atl1_up(struct atl1_adapter *adapter) |
| @@ -1723,6 +1965,7 @@ s32 atl1_up(struct atl1_adapter *adapter) | |||
| 1723 | 1965 | ||
| 1724 | /* hardware has been reset, we need to reload some things */ | 1966 | /* hardware has been reset, we need to reload some things */ |
| 1725 | atl1_set_multi(netdev); | 1967 | atl1_set_multi(netdev); |
| 1968 | atl1_init_ring_ptrs(adapter); | ||
| 1726 | atl1_restore_vlan(adapter); | 1969 | atl1_restore_vlan(adapter); |
| 1727 | err = atl1_alloc_rx_buffers(adapter); | 1970 | err = atl1_alloc_rx_buffers(adapter); |
| 1728 | if (unlikely(!err)) /* no RX BUFFER allocated */ | 1971 | if (unlikely(!err)) /* no RX BUFFER allocated */ |
| @@ -1750,11 +1993,6 @@ s32 atl1_up(struct atl1_adapter *adapter) | |||
| 1750 | atl1_check_link(adapter); | 1993 | atl1_check_link(adapter); |
| 1751 | return 0; | 1994 | return 0; |
| 1752 | 1995 | ||
| 1753 | /* FIXME: unreachable code! -- CHS */ | ||
| 1754 | /* free irq disable any interrupt */ | ||
| 1755 | iowrite32(0, adapter->hw.hw_addr + REG_IMR); | ||
| 1756 | free_irq(adapter->pdev->irq, netdev); | ||
| 1757 | |||
| 1758 | err_up: | 1996 | err_up: |
| 1759 | pci_disable_msi(adapter->pdev); | 1997 | pci_disable_msi(adapter->pdev); |
| 1760 | /* free rx_buffers */ | 1998 | /* free rx_buffers */ |
| @@ -1786,172 +2024,6 @@ void atl1_down(struct atl1_adapter *adapter) | |||
| 1786 | } | 2024 | } |
| 1787 | 2025 | ||
| 1788 | /* | 2026 | /* |
| 1789 | * atl1_change_mtu - Change the Maximum Transfer Unit | ||
| 1790 | * @netdev: network interface device structure | ||
| 1791 | * @new_mtu: new value for maximum frame size | ||
| 1792 | * | ||
| 1793 | * Returns 0 on success, negative on failure | ||
| 1794 | */ | ||
| 1795 | static int atl1_change_mtu(struct net_device *netdev, int new_mtu) | ||
| 1796 | { | ||
| 1797 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1798 | int old_mtu = netdev->mtu; | ||
| 1799 | int max_frame = new_mtu + ENET_HEADER_SIZE + ETHERNET_FCS_SIZE; | ||
| 1800 | |||
| 1801 | if ((max_frame < MINIMUM_ETHERNET_FRAME_SIZE) || | ||
| 1802 | (max_frame > MAX_JUMBO_FRAME_SIZE)) { | ||
| 1803 | dev_warn(&adapter->pdev->dev, "invalid MTU setting\n"); | ||
| 1804 | return -EINVAL; | ||
| 1805 | } | ||
| 1806 | |||
| 1807 | adapter->hw.max_frame_size = max_frame; | ||
| 1808 | adapter->hw.tx_jumbo_task_th = (max_frame + 7) >> 3; | ||
| 1809 | adapter->rx_buffer_len = (max_frame + 7) & ~7; | ||
| 1810 | adapter->hw.rx_jumbo_th = adapter->rx_buffer_len / 8; | ||
| 1811 | |||
| 1812 | netdev->mtu = new_mtu; | ||
| 1813 | if ((old_mtu != new_mtu) && netif_running(netdev)) { | ||
| 1814 | atl1_down(adapter); | ||
| 1815 | atl1_up(adapter); | ||
| 1816 | } | ||
| 1817 | |||
| 1818 | return 0; | ||
| 1819 | } | ||
| 1820 | |||
| 1821 | /* | ||
| 1822 | * atl1_set_mac - Change the Ethernet Address of the NIC | ||
| 1823 | * @netdev: network interface device structure | ||
| 1824 | * @p: pointer to an address structure | ||
| 1825 | * | ||
| 1826 | * Returns 0 on success, negative on failure | ||
| 1827 | */ | ||
| 1828 | static int atl1_set_mac(struct net_device *netdev, void *p) | ||
| 1829 | { | ||
| 1830 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1831 | struct sockaddr *addr = p; | ||
| 1832 | |||
| 1833 | if (netif_running(netdev)) | ||
| 1834 | return -EBUSY; | ||
| 1835 | |||
| 1836 | if (!is_valid_ether_addr(addr->sa_data)) | ||
| 1837 | return -EADDRNOTAVAIL; | ||
| 1838 | |||
| 1839 | memcpy(netdev->dev_addr, addr->sa_data, netdev->addr_len); | ||
| 1840 | memcpy(adapter->hw.mac_addr, addr->sa_data, netdev->addr_len); | ||
| 1841 | |||
| 1842 | atl1_set_mac_addr(&adapter->hw); | ||
| 1843 | return 0; | ||
| 1844 | } | ||
| 1845 | |||
| 1846 | /* | ||
| 1847 | * atl1_watchdog - Timer Call-back | ||
| 1848 | * @data: pointer to netdev cast into an unsigned long | ||
| 1849 | */ | ||
| 1850 | static void atl1_watchdog(unsigned long data) | ||
| 1851 | { | ||
| 1852 | struct atl1_adapter *adapter = (struct atl1_adapter *)data; | ||
| 1853 | |||
| 1854 | /* Reset the timer */ | ||
| 1855 | mod_timer(&adapter->watchdog_timer, jiffies + 2 * HZ); | ||
| 1856 | } | ||
| 1857 | |||
| 1858 | static int mdio_read(struct net_device *netdev, int phy_id, int reg_num) | ||
| 1859 | { | ||
| 1860 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1861 | u16 result; | ||
| 1862 | |||
| 1863 | atl1_read_phy_reg(&adapter->hw, reg_num & 0x1f, &result); | ||
| 1864 | |||
| 1865 | return result; | ||
| 1866 | } | ||
| 1867 | |||
| 1868 | static void mdio_write(struct net_device *netdev, int phy_id, int reg_num, int val) | ||
| 1869 | { | ||
| 1870 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1871 | |||
| 1872 | atl1_write_phy_reg(&adapter->hw, reg_num, val); | ||
| 1873 | } | ||
| 1874 | |||
| 1875 | /* | ||
| 1876 | * atl1_mii_ioctl - | ||
| 1877 | * @netdev: | ||
| 1878 | * @ifreq: | ||
| 1879 | * @cmd: | ||
| 1880 | */ | ||
| 1881 | static int atl1_mii_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) | ||
| 1882 | { | ||
| 1883 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1884 | unsigned long flags; | ||
| 1885 | int retval; | ||
| 1886 | |||
| 1887 | if (!netif_running(netdev)) | ||
| 1888 | return -EINVAL; | ||
| 1889 | |||
| 1890 | spin_lock_irqsave(&adapter->lock, flags); | ||
| 1891 | retval = generic_mii_ioctl(&adapter->mii, if_mii(ifr), cmd, NULL); | ||
| 1892 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 1893 | |||
| 1894 | return retval; | ||
| 1895 | } | ||
| 1896 | |||
| 1897 | /* | ||
| 1898 | * atl1_ioctl - | ||
| 1899 | * @netdev: | ||
| 1900 | * @ifreq: | ||
| 1901 | * @cmd: | ||
| 1902 | */ | ||
| 1903 | static int atl1_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) | ||
| 1904 | { | ||
| 1905 | switch (cmd) { | ||
| 1906 | case SIOCGMIIPHY: | ||
| 1907 | case SIOCGMIIREG: | ||
| 1908 | case SIOCSMIIREG: | ||
| 1909 | return atl1_mii_ioctl(netdev, ifr, cmd); | ||
| 1910 | default: | ||
| 1911 | return -EOPNOTSUPP; | ||
| 1912 | } | ||
| 1913 | } | ||
| 1914 | |||
| 1915 | /* | ||
| 1916 | * atl1_tx_timeout - Respond to a Tx Hang | ||
| 1917 | * @netdev: network interface device structure | ||
| 1918 | */ | ||
| 1919 | static void atl1_tx_timeout(struct net_device *netdev) | ||
| 1920 | { | ||
| 1921 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 1922 | /* Do the reset outside of interrupt context */ | ||
| 1923 | schedule_work(&adapter->tx_timeout_task); | ||
| 1924 | } | ||
| 1925 | |||
| 1926 | /* | ||
| 1927 | * atl1_phy_config - Timer Call-back | ||
| 1928 | * @data: pointer to netdev cast into an unsigned long | ||
| 1929 | */ | ||
| 1930 | static void atl1_phy_config(unsigned long data) | ||
| 1931 | { | ||
| 1932 | struct atl1_adapter *adapter = (struct atl1_adapter *)data; | ||
| 1933 | struct atl1_hw *hw = &adapter->hw; | ||
| 1934 | unsigned long flags; | ||
| 1935 | |||
| 1936 | spin_lock_irqsave(&adapter->lock, flags); | ||
| 1937 | adapter->phy_timer_pending = false; | ||
| 1938 | atl1_write_phy_reg(hw, MII_ADVERTISE, hw->mii_autoneg_adv_reg); | ||
| 1939 | atl1_write_phy_reg(hw, MII_AT001_CR, hw->mii_1000t_ctrl_reg); | ||
| 1940 | atl1_write_phy_reg(hw, MII_BMCR, MII_CR_RESET | MII_CR_AUTO_NEG_EN); | ||
| 1941 | spin_unlock_irqrestore(&adapter->lock, flags); | ||
| 1942 | } | ||
| 1943 | |||
| 1944 | int atl1_reset(struct atl1_adapter *adapter) | ||
| 1945 | { | ||
| 1946 | int ret; | ||
| 1947 | |||
| 1948 | ret = atl1_reset_hw(&adapter->hw); | ||
| 1949 | if (ret != ATL1_SUCCESS) | ||
| 1950 | return ret; | ||
| 1951 | return atl1_init_hw(&adapter->hw); | ||
| 1952 | } | ||
| 1953 | |||
| 1954 | /* | ||
| 1955 | * atl1_open - Called when a network interface is made active | 2027 | * atl1_open - Called when a network interface is made active |
| 1956 | * @netdev: network interface device structure | 2028 | * @netdev: network interface device structure |
| 1957 | * | 2029 | * |
| @@ -2003,77 +2075,113 @@ static int atl1_close(struct net_device *netdev) | |||
| 2003 | return 0; | 2075 | return 0; |
| 2004 | } | 2076 | } |
| 2005 | 2077 | ||
| 2006 | #ifdef CONFIG_NET_POLL_CONTROLLER | 2078 | #ifdef CONFIG_PM |
| 2007 | static void atl1_poll_controller(struct net_device *netdev) | 2079 | static int atl1_suspend(struct pci_dev *pdev, pm_message_t state) |
| 2008 | { | ||
| 2009 | disable_irq(netdev->irq); | ||
| 2010 | atl1_intr(netdev->irq, netdev); | ||
| 2011 | enable_irq(netdev->irq); | ||
| 2012 | } | ||
| 2013 | #endif | ||
| 2014 | |||
| 2015 | /* | ||
| 2016 | * If TPD Buffer size equal to 0, PCIE DMAR_TO_INT | ||
| 2017 | * will assert. We do soft reset <0x1400=1> according | ||
| 2018 | * with the SPEC. BUT, it seemes that PCIE or DMA | ||
| 2019 | * state-machine will not be reset. DMAR_TO_INT will | ||
| 2020 | * assert again and again. | ||
| 2021 | */ | ||
| 2022 | static void atl1_tx_timeout_task(struct work_struct *work) | ||
| 2023 | { | 2080 | { |
| 2024 | struct atl1_adapter *adapter = | 2081 | struct net_device *netdev = pci_get_drvdata(pdev); |
| 2025 | container_of(work, struct atl1_adapter, tx_timeout_task); | 2082 | struct atl1_adapter *adapter = netdev_priv(netdev); |
| 2026 | struct net_device *netdev = adapter->netdev; | 2083 | struct atl1_hw *hw = &adapter->hw; |
| 2084 | u32 ctrl = 0; | ||
| 2085 | u32 wufc = adapter->wol; | ||
| 2027 | 2086 | ||
| 2028 | netif_device_detach(netdev); | 2087 | netif_device_detach(netdev); |
| 2029 | atl1_down(adapter); | 2088 | if (netif_running(netdev)) |
| 2030 | atl1_up(adapter); | 2089 | atl1_down(adapter); |
| 2031 | netif_device_attach(netdev); | ||
| 2032 | } | ||
| 2033 | 2090 | ||
| 2034 | /* | 2091 | atl1_read_phy_reg(hw, MII_BMSR, (u16 *) & ctrl); |
| 2035 | * atl1_link_chg_task - deal with link change event Out of interrupt context | 2092 | atl1_read_phy_reg(hw, MII_BMSR, (u16 *) & ctrl); |
| 2036 | */ | 2093 | if (ctrl & BMSR_LSTATUS) |
| 2037 | static void atl1_link_chg_task(struct work_struct *work) | 2094 | wufc &= ~ATL1_WUFC_LNKC; |
| 2038 | { | ||
| 2039 | struct atl1_adapter *adapter = | ||
| 2040 | container_of(work, struct atl1_adapter, link_chg_task); | ||
| 2041 | unsigned long flags; | ||
| 2042 | 2095 | ||
| 2043 | spin_lock_irqsave(&adapter->lock, flags); | 2096 | /* reduce speed to 10/100M */ |
| 2044 | atl1_check_link(adapter); | 2097 | if (wufc) { |
| 2045 | spin_unlock_irqrestore(&adapter->lock, flags); | 2098 | atl1_phy_enter_power_saving(hw); |
| 2099 | /* if resume, let driver to re- setup link */ | ||
| 2100 | hw->phy_configured = false; | ||
| 2101 | atl1_set_mac_addr(hw); | ||
| 2102 | atl1_set_multi(netdev); | ||
| 2103 | |||
| 2104 | ctrl = 0; | ||
| 2105 | /* turn on magic packet wol */ | ||
| 2106 | if (wufc & ATL1_WUFC_MAG) | ||
| 2107 | ctrl = WOL_MAGIC_EN | WOL_MAGIC_PME_EN; | ||
| 2108 | |||
| 2109 | /* turn on Link change WOL */ | ||
| 2110 | if (wufc & ATL1_WUFC_LNKC) | ||
| 2111 | ctrl |= (WOL_LINK_CHG_EN | WOL_LINK_CHG_PME_EN); | ||
| 2112 | iowrite32(ctrl, hw->hw_addr + REG_WOL_CTRL); | ||
| 2113 | |||
| 2114 | /* turn on all-multi mode if wake on multicast is enabled */ | ||
| 2115 | ctrl = ioread32(hw->hw_addr + REG_MAC_CTRL); | ||
| 2116 | ctrl &= ~MAC_CTRL_DBG; | ||
| 2117 | ctrl &= ~MAC_CTRL_PROMIS_EN; | ||
| 2118 | if (wufc & ATL1_WUFC_MC) | ||
| 2119 | ctrl |= MAC_CTRL_MC_ALL_EN; | ||
| 2120 | else | ||
| 2121 | ctrl &= ~MAC_CTRL_MC_ALL_EN; | ||
| 2122 | |||
| 2123 | /* turn on broadcast mode if wake on-BC is enabled */ | ||
| 2124 | if (wufc & ATL1_WUFC_BC) | ||
| 2125 | ctrl |= MAC_CTRL_BC_EN; | ||
| 2126 | else | ||
| 2127 | ctrl &= ~MAC_CTRL_BC_EN; | ||
| 2128 | |||
| 2129 | /* enable RX */ | ||
| 2130 | ctrl |= MAC_CTRL_RX_EN; | ||
| 2131 | iowrite32(ctrl, hw->hw_addr + REG_MAC_CTRL); | ||
| 2132 | pci_enable_wake(pdev, PCI_D3hot, 1); | ||
| 2133 | pci_enable_wake(pdev, PCI_D3cold, 1); | ||
| 2134 | } else { | ||
| 2135 | iowrite32(0, hw->hw_addr + REG_WOL_CTRL); | ||
| 2136 | pci_enable_wake(pdev, PCI_D3hot, 0); | ||
| 2137 | pci_enable_wake(pdev, PCI_D3cold, 0); | ||
| 2138 | } | ||
| 2139 | |||
| 2140 | pci_save_state(pdev); | ||
| 2141 | pci_disable_device(pdev); | ||
| 2142 | |||
| 2143 | pci_set_power_state(pdev, PCI_D3hot); | ||
| 2144 | |||
| 2145 | return 0; | ||
| 2046 | } | 2146 | } |
| 2047 | 2147 | ||
| 2048 | /* | 2148 | static int atl1_resume(struct pci_dev *pdev) |
| 2049 | * atl1_pcie_patch - Patch for PCIE module | ||
| 2050 | */ | ||
| 2051 | static void atl1_pcie_patch(struct atl1_adapter *adapter) | ||
| 2052 | { | 2149 | { |
| 2053 | u32 value; | 2150 | struct net_device *netdev = pci_get_drvdata(pdev); |
| 2054 | value = 0x6500; | 2151 | struct atl1_adapter *adapter = netdev_priv(netdev); |
| 2055 | iowrite32(value, adapter->hw.hw_addr + 0x12FC); | 2152 | u32 ret_val; |
| 2056 | /* pcie flow control mode change */ | 2153 | |
| 2057 | value = ioread32(adapter->hw.hw_addr + 0x1008); | 2154 | pci_set_power_state(pdev, 0); |
| 2058 | value |= 0x8000; | 2155 | pci_restore_state(pdev); |
| 2059 | iowrite32(value, adapter->hw.hw_addr + 0x1008); | 2156 | |
| 2157 | ret_val = pci_enable_device(pdev); | ||
| 2158 | pci_enable_wake(pdev, PCI_D3hot, 0); | ||
| 2159 | pci_enable_wake(pdev, PCI_D3cold, 0); | ||
| 2160 | |||
| 2161 | iowrite32(0, adapter->hw.hw_addr + REG_WOL_CTRL); | ||
| 2162 | atl1_reset(adapter); | ||
| 2163 | |||
| 2164 | if (netif_running(netdev)) | ||
| 2165 | atl1_up(adapter); | ||
| 2166 | netif_device_attach(netdev); | ||
| 2167 | |||
| 2168 | atl1_via_workaround(adapter); | ||
| 2169 | |||
| 2170 | return 0; | ||
| 2060 | } | 2171 | } |
| 2172 | #else | ||
| 2173 | #define atl1_suspend NULL | ||
| 2174 | #define atl1_resume NULL | ||
| 2175 | #endif | ||
| 2061 | 2176 | ||
| 2062 | /* | 2177 | #ifdef CONFIG_NET_POLL_CONTROLLER |
| 2063 | * When ACPI resume on some VIA MotherBoard, the Interrupt Disable bit/0x400 | 2178 | static void atl1_poll_controller(struct net_device *netdev) |
| 2064 | * on PCI Command register is disable. | ||
| 2065 | * The function enable this bit. | ||
| 2066 | * Brackett, 2006/03/15 | ||
| 2067 | */ | ||
| 2068 | static void atl1_via_workaround(struct atl1_adapter *adapter) | ||
| 2069 | { | 2179 | { |
| 2070 | unsigned long value; | 2180 | disable_irq(netdev->irq); |
| 2071 | 2181 | atl1_intr(netdev->irq, netdev); | |
| 2072 | value = ioread16(adapter->hw.hw_addr + PCI_COMMAND); | 2182 | enable_irq(netdev->irq); |
| 2073 | if (value & PCI_COMMAND_INTX_DISABLE) | ||
| 2074 | value &= ~PCI_COMMAND_INTX_DISABLE; | ||
| 2075 | iowrite32(value, adapter->hw.hw_addr + PCI_COMMAND); | ||
| 2076 | } | 2183 | } |
| 2184 | #endif | ||
| 2077 | 2185 | ||
| 2078 | /* | 2186 | /* |
| 2079 | * atl1_probe - Device Initialization Routine | 2187 | * atl1_probe - Device Initialization Routine |
| @@ -2087,7 +2195,7 @@ static void atl1_via_workaround(struct atl1_adapter *adapter) | |||
| 2087 | * and a hardware reset occur. | 2195 | * and a hardware reset occur. |
| 2088 | */ | 2196 | */ |
| 2089 | static int __devinit atl1_probe(struct pci_dev *pdev, | 2197 | static int __devinit atl1_probe(struct pci_dev *pdev, |
| 2090 | const struct pci_device_id *ent) | 2198 | const struct pci_device_id *ent) |
| 2091 | { | 2199 | { |
| 2092 | struct net_device *netdev; | 2200 | struct net_device *netdev; |
| 2093 | struct atl1_adapter *adapter; | 2201 | struct atl1_adapter *adapter; |
| @@ -2141,7 +2249,7 @@ static int __devinit atl1_probe(struct pci_dev *pdev, | |||
| 2141 | } | 2249 | } |
| 2142 | /* get device revision number */ | 2250 | /* get device revision number */ |
| 2143 | adapter->hw.dev_rev = ioread16(adapter->hw.hw_addr + | 2251 | adapter->hw.dev_rev = ioread16(adapter->hw.hw_addr + |
| 2144 | (REG_MASTER_CTRL + 2)); | 2252 | (REG_MASTER_CTRL + 2)); |
| 2145 | dev_info(&pdev->dev, "version %s\n", DRIVER_VERSION); | 2253 | dev_info(&pdev->dev, "version %s\n", DRIVER_VERSION); |
| 2146 | 2254 | ||
| 2147 | /* set default ring resource counts */ | 2255 | /* set default ring resource counts */ |
| @@ -2294,7 +2402,8 @@ static void __devexit atl1_remove(struct pci_dev *pdev) | |||
| 2294 | * address, we need to save the permanent one. | 2402 | * address, we need to save the permanent one. |
| 2295 | */ | 2403 | */ |
| 2296 | if (memcmp(adapter->hw.mac_addr, adapter->hw.perm_mac_addr, ETH_ALEN)) { | 2404 | if (memcmp(adapter->hw.mac_addr, adapter->hw.perm_mac_addr, ETH_ALEN)) { |
| 2297 | memcpy(adapter->hw.mac_addr, adapter->hw.perm_mac_addr, ETH_ALEN); | 2405 | memcpy(adapter->hw.mac_addr, adapter->hw.perm_mac_addr, |
| 2406 | ETH_ALEN); | ||
| 2298 | atl1_set_mac_addr(&adapter->hw); | 2407 | atl1_set_mac_addr(&adapter->hw); |
| 2299 | } | 2408 | } |
| 2300 | 2409 | ||
| @@ -2306,112 +2415,11 @@ static void __devexit atl1_remove(struct pci_dev *pdev) | |||
| 2306 | pci_disable_device(pdev); | 2415 | pci_disable_device(pdev); |
| 2307 | } | 2416 | } |
| 2308 | 2417 | ||
| 2309 | #ifdef CONFIG_PM | ||
| 2310 | static int atl1_suspend(struct pci_dev *pdev, pm_message_t state) | ||
| 2311 | { | ||
| 2312 | struct net_device *netdev = pci_get_drvdata(pdev); | ||
| 2313 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 2314 | struct atl1_hw *hw = &adapter->hw; | ||
| 2315 | u32 ctrl = 0; | ||
| 2316 | u32 wufc = adapter->wol; | ||
| 2317 | |||
| 2318 | netif_device_detach(netdev); | ||
| 2319 | if (netif_running(netdev)) | ||
| 2320 | atl1_down(adapter); | ||
| 2321 | |||
| 2322 | atl1_read_phy_reg(hw, MII_BMSR, (u16 *) & ctrl); | ||
| 2323 | atl1_read_phy_reg(hw, MII_BMSR, (u16 *) & ctrl); | ||
| 2324 | if (ctrl & BMSR_LSTATUS) | ||
| 2325 | wufc &= ~ATL1_WUFC_LNKC; | ||
| 2326 | |||
| 2327 | /* reduce speed to 10/100M */ | ||
| 2328 | if (wufc) { | ||
| 2329 | atl1_phy_enter_power_saving(hw); | ||
| 2330 | /* if resume, let driver to re- setup link */ | ||
| 2331 | hw->phy_configured = false; | ||
| 2332 | atl1_set_mac_addr(hw); | ||
| 2333 | atl1_set_multi(netdev); | ||
| 2334 | |||
| 2335 | ctrl = 0; | ||
| 2336 | /* turn on magic packet wol */ | ||
| 2337 | if (wufc & ATL1_WUFC_MAG) | ||
| 2338 | ctrl = WOL_MAGIC_EN | WOL_MAGIC_PME_EN; | ||
| 2339 | |||
| 2340 | /* turn on Link change WOL */ | ||
| 2341 | if (wufc & ATL1_WUFC_LNKC) | ||
| 2342 | ctrl |= (WOL_LINK_CHG_EN | WOL_LINK_CHG_PME_EN); | ||
| 2343 | iowrite32(ctrl, hw->hw_addr + REG_WOL_CTRL); | ||
| 2344 | |||
| 2345 | /* turn on all-multi mode if wake on multicast is enabled */ | ||
| 2346 | ctrl = ioread32(hw->hw_addr + REG_MAC_CTRL); | ||
| 2347 | ctrl &= ~MAC_CTRL_DBG; | ||
| 2348 | ctrl &= ~MAC_CTRL_PROMIS_EN; | ||
| 2349 | if (wufc & ATL1_WUFC_MC) | ||
| 2350 | ctrl |= MAC_CTRL_MC_ALL_EN; | ||
| 2351 | else | ||
| 2352 | ctrl &= ~MAC_CTRL_MC_ALL_EN; | ||
| 2353 | |||
| 2354 | /* turn on broadcast mode if wake on-BC is enabled */ | ||
| 2355 | if (wufc & ATL1_WUFC_BC) | ||
| 2356 | ctrl |= MAC_CTRL_BC_EN; | ||
| 2357 | else | ||
| 2358 | ctrl &= ~MAC_CTRL_BC_EN; | ||
| 2359 | |||
| 2360 | /* enable RX */ | ||
| 2361 | ctrl |= MAC_CTRL_RX_EN; | ||
| 2362 | iowrite32(ctrl, hw->hw_addr + REG_MAC_CTRL); | ||
| 2363 | pci_enable_wake(pdev, PCI_D3hot, 1); | ||
| 2364 | pci_enable_wake(pdev, PCI_D3cold, 1); /* 4 == D3 cold */ | ||
| 2365 | } else { | ||
| 2366 | iowrite32(0, hw->hw_addr + REG_WOL_CTRL); | ||
| 2367 | pci_enable_wake(pdev, PCI_D3hot, 0); | ||
| 2368 | pci_enable_wake(pdev, PCI_D3cold, 0); /* 4 == D3 cold */ | ||
| 2369 | } | ||
| 2370 | |||
| 2371 | pci_save_state(pdev); | ||
| 2372 | pci_disable_device(pdev); | ||
| 2373 | |||
| 2374 | pci_set_power_state(pdev, PCI_D3hot); | ||
| 2375 | |||
| 2376 | return 0; | ||
| 2377 | } | ||
| 2378 | |||
| 2379 | static int atl1_resume(struct pci_dev *pdev) | ||
| 2380 | { | ||
| 2381 | struct net_device *netdev = pci_get_drvdata(pdev); | ||
| 2382 | struct atl1_adapter *adapter = netdev_priv(netdev); | ||
| 2383 | u32 ret_val; | ||
| 2384 | |||
| 2385 | pci_set_power_state(pdev, 0); | ||
| 2386 | pci_restore_state(pdev); | ||
| 2387 | |||
| 2388 | ret_val = pci_enable_device(pdev); | ||
| 2389 | pci_enable_wake(pdev, PCI_D3hot, 0); | ||
| 2390 | pci_enable_wake(pdev, PCI_D3cold, 0); | ||
| 2391 | |||
| 2392 | iowrite32(0, adapter->hw.hw_addr + REG_WOL_CTRL); | ||
| 2393 | atl1_reset(adapter); | ||
| 2394 | |||
| 2395 | if (netif_running(netdev)) | ||
| 2396 | atl1_up(adapter); | ||
| 2397 | netif_device_attach(netdev); | ||
| 2398 | |||
| 2399 | atl1_via_workaround(adapter); | ||
| 2400 | |||
| 2401 | return 0; | ||
| 2402 | } | ||
| 2403 | #else | ||
| 2404 | #define atl1_suspend NULL | ||
| 2405 | #define atl1_resume NULL | ||
| 2406 | #endif | ||
| 2407 | |||
| 2408 | static struct pci_driver atl1_driver = { | 2418 | static struct pci_driver atl1_driver = { |
| 2409 | .name = atl1_driver_name, | 2419 | .name = atl1_driver_name, |
| 2410 | .id_table = atl1_pci_tbl, | 2420 | .id_table = atl1_pci_tbl, |
| 2411 | .probe = atl1_probe, | 2421 | .probe = atl1_probe, |
| 2412 | .remove = __devexit_p(atl1_remove), | 2422 | .remove = __devexit_p(atl1_remove), |
| 2413 | /* Power Managment Hooks */ | ||
| 2414 | /* probably broken right now -- CHS */ | ||
| 2415 | .suspend = atl1_suspend, | 2423 | .suspend = atl1_suspend, |
| 2416 | .resume = atl1_resume | 2424 | .resume = atl1_resume |
| 2417 | }; | 2425 | }; |
diff --git a/drivers/net/ehea/ehea.h b/drivers/net/ehea/ehea.h index f03f070451de..6628fa622e2c 100644 --- a/drivers/net/ehea/ehea.h +++ b/drivers/net/ehea/ehea.h | |||
| @@ -39,13 +39,13 @@ | |||
| 39 | #include <asm/io.h> | 39 | #include <asm/io.h> |
| 40 | 40 | ||
| 41 | #define DRV_NAME "ehea" | 41 | #define DRV_NAME "ehea" |
| 42 | #define DRV_VERSION "EHEA_0067" | 42 | #define DRV_VERSION "EHEA_0070" |
| 43 | 43 | ||
| 44 | /* EHEA capability flags */ | 44 | /* eHEA capability flags */ |
| 45 | #define DLPAR_PORT_ADD_REM 1 | 45 | #define DLPAR_PORT_ADD_REM 1 |
| 46 | #define DLPAR_MEM_ADD 2 | 46 | #define DLPAR_MEM_ADD 2 |
| 47 | #define DLPAR_MEM_REM 4 | 47 | #define DLPAR_MEM_REM 4 |
| 48 | #define EHEA_CAPABILITIES (DLPAR_PORT_ADD_REM) | 48 | #define EHEA_CAPABILITIES (DLPAR_PORT_ADD_REM) |
| 49 | 49 | ||
| 50 | #define EHEA_MSG_DEFAULT (NETIF_MSG_LINK | NETIF_MSG_TIMER \ | 50 | #define EHEA_MSG_DEFAULT (NETIF_MSG_LINK | NETIF_MSG_TIMER \ |
| 51 | | NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR) | 51 | | NETIF_MSG_RX_ERR | NETIF_MSG_TX_ERR) |
| @@ -113,6 +113,8 @@ | |||
| 113 | /* Memory Regions */ | 113 | /* Memory Regions */ |
| 114 | #define EHEA_MR_ACC_CTRL 0x00800000 | 114 | #define EHEA_MR_ACC_CTRL 0x00800000 |
| 115 | 115 | ||
| 116 | #define EHEA_BUSMAP_START 0x8000000000000000ULL | ||
| 117 | |||
| 116 | #define EHEA_WATCH_DOG_TIMEOUT 10*HZ | 118 | #define EHEA_WATCH_DOG_TIMEOUT 10*HZ |
| 117 | 119 | ||
| 118 | /* utility functions */ | 120 | /* utility functions */ |
| @@ -186,6 +188,12 @@ struct h_epas { | |||
| 186 | set to 0 if unused */ | 188 | set to 0 if unused */ |
| 187 | }; | 189 | }; |
| 188 | 190 | ||
| 191 | struct ehea_busmap { | ||
| 192 | unsigned int entries; /* total number of entries */ | ||
| 193 | unsigned int valid_sections; /* number of valid sections */ | ||
| 194 | u64 *vaddr; | ||
| 195 | }; | ||
| 196 | |||
| 189 | struct ehea_qp; | 197 | struct ehea_qp; |
| 190 | struct ehea_cq; | 198 | struct ehea_cq; |
| 191 | struct ehea_eq; | 199 | struct ehea_eq; |
| @@ -382,6 +390,8 @@ struct ehea_adapter { | |||
| 382 | struct ehea_mr mr; | 390 | struct ehea_mr mr; |
| 383 | u32 pd; /* protection domain */ | 391 | u32 pd; /* protection domain */ |
| 384 | u64 max_mc_mac; /* max number of multicast mac addresses */ | 392 | u64 max_mc_mac; /* max number of multicast mac addresses */ |
| 393 | int active_ports; | ||
| 394 | struct list_head list; | ||
| 385 | }; | 395 | }; |
| 386 | 396 | ||
| 387 | 397 | ||
| @@ -431,6 +441,9 @@ struct port_res_cfg { | |||
| 431 | int max_entries_rq3; | 441 | int max_entries_rq3; |
| 432 | }; | 442 | }; |
| 433 | 443 | ||
| 444 | enum ehea_flag_bits { | ||
| 445 | __EHEA_STOP_XFER | ||
| 446 | }; | ||
| 434 | 447 | ||
| 435 | void ehea_set_ethtool_ops(struct net_device *netdev); | 448 | void ehea_set_ethtool_ops(struct net_device *netdev); |
| 436 | int ehea_sense_port_attr(struct ehea_port *port); | 449 | int ehea_sense_port_attr(struct ehea_port *port); |
diff --git a/drivers/net/ehea/ehea_main.c b/drivers/net/ehea/ehea_main.c index 383144db4d18..1d1571cf322e 100644 --- a/drivers/net/ehea/ehea_main.c +++ b/drivers/net/ehea/ehea_main.c | |||
| @@ -79,6 +79,11 @@ MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue " | |||
| 79 | MODULE_PARM_DESC(use_mcs, " 0:NAPI, 1:Multiple receive queues, Default = 1 "); | 79 | MODULE_PARM_DESC(use_mcs, " 0:NAPI, 1:Multiple receive queues, Default = 1 "); |
| 80 | 80 | ||
| 81 | static int port_name_cnt = 0; | 81 | static int port_name_cnt = 0; |
| 82 | static LIST_HEAD(adapter_list); | ||
| 83 | u64 ehea_driver_flags = 0; | ||
| 84 | struct workqueue_struct *ehea_driver_wq; | ||
| 85 | struct work_struct ehea_rereg_mr_task; | ||
| 86 | |||
| 82 | 87 | ||
| 83 | static int __devinit ehea_probe_adapter(struct ibmebus_dev *dev, | 88 | static int __devinit ehea_probe_adapter(struct ibmebus_dev *dev, |
| 84 | const struct of_device_id *id); | 89 | const struct of_device_id *id); |
| @@ -238,13 +243,17 @@ static int ehea_refill_rq_def(struct ehea_port_res *pr, | |||
| 238 | rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) | 243 | rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) |
| 239 | | EHEA_BMASK_SET(EHEA_WR_ID_INDEX, index); | 244 | | EHEA_BMASK_SET(EHEA_WR_ID_INDEX, index); |
| 240 | rwqe->sg_list[0].l_key = pr->recv_mr.lkey; | 245 | rwqe->sg_list[0].l_key = pr->recv_mr.lkey; |
| 241 | rwqe->sg_list[0].vaddr = (u64)skb->data; | 246 | rwqe->sg_list[0].vaddr = ehea_map_vaddr(skb->data); |
| 242 | rwqe->sg_list[0].len = packet_size; | 247 | rwqe->sg_list[0].len = packet_size; |
| 243 | rwqe->data_segments = 1; | 248 | rwqe->data_segments = 1; |
| 244 | 249 | ||
| 245 | index++; | 250 | index++; |
| 246 | index &= max_index_mask; | 251 | index &= max_index_mask; |
| 252 | |||
| 253 | if (unlikely(test_bit(__EHEA_STOP_XFER, &ehea_driver_flags))) | ||
| 254 | goto out; | ||
| 247 | } | 255 | } |
| 256 | |||
| 248 | q_skba->index = index; | 257 | q_skba->index = index; |
| 249 | 258 | ||
| 250 | /* Ring doorbell */ | 259 | /* Ring doorbell */ |
| @@ -253,7 +262,7 @@ static int ehea_refill_rq_def(struct ehea_port_res *pr, | |||
| 253 | ehea_update_rq2a(pr->qp, i); | 262 | ehea_update_rq2a(pr->qp, i); |
| 254 | else | 263 | else |
| 255 | ehea_update_rq3a(pr->qp, i); | 264 | ehea_update_rq3a(pr->qp, i); |
| 256 | 265 | out: | |
| 257 | return ret; | 266 | return ret; |
| 258 | } | 267 | } |
| 259 | 268 | ||
| @@ -1321,7 +1330,7 @@ static void write_swqe2_TSO(struct sk_buff *skb, | |||
| 1321 | sg1entry->len = skb_data_size - headersize; | 1330 | sg1entry->len = skb_data_size - headersize; |
| 1322 | 1331 | ||
| 1323 | tmp_addr = (u64)(skb->data + headersize); | 1332 | tmp_addr = (u64)(skb->data + headersize); |
| 1324 | sg1entry->vaddr = tmp_addr; | 1333 | sg1entry->vaddr = ehea_map_vaddr(tmp_addr); |
| 1325 | swqe->descriptors++; | 1334 | swqe->descriptors++; |
| 1326 | } | 1335 | } |
| 1327 | } else | 1336 | } else |
| @@ -1352,7 +1361,7 @@ static void write_swqe2_nonTSO(struct sk_buff *skb, | |||
| 1352 | sg1entry->l_key = lkey; | 1361 | sg1entry->l_key = lkey; |
| 1353 | sg1entry->len = skb_data_size - SWQE2_MAX_IMM; | 1362 | sg1entry->len = skb_data_size - SWQE2_MAX_IMM; |
| 1354 | tmp_addr = (u64)(skb->data + SWQE2_MAX_IMM); | 1363 | tmp_addr = (u64)(skb->data + SWQE2_MAX_IMM); |
| 1355 | sg1entry->vaddr = tmp_addr; | 1364 | sg1entry->vaddr = ehea_map_vaddr(tmp_addr); |
| 1356 | swqe->descriptors++; | 1365 | swqe->descriptors++; |
| 1357 | } | 1366 | } |
| 1358 | } else { | 1367 | } else { |
| @@ -1391,7 +1400,7 @@ static inline void write_swqe2_data(struct sk_buff *skb, struct net_device *dev, | |||
| 1391 | sg1entry->len = frag->size; | 1400 | sg1entry->len = frag->size; |
| 1392 | tmp_addr = (u64)(page_address(frag->page) | 1401 | tmp_addr = (u64)(page_address(frag->page) |
| 1393 | + frag->page_offset); | 1402 | + frag->page_offset); |
| 1394 | sg1entry->vaddr = tmp_addr; | 1403 | sg1entry->vaddr = ehea_map_vaddr(tmp_addr); |
| 1395 | swqe->descriptors++; | 1404 | swqe->descriptors++; |
| 1396 | sg1entry_contains_frag_data = 1; | 1405 | sg1entry_contains_frag_data = 1; |
| 1397 | } | 1406 | } |
| @@ -1406,7 +1415,7 @@ static inline void write_swqe2_data(struct sk_buff *skb, struct net_device *dev, | |||
| 1406 | 1415 | ||
| 1407 | tmp_addr = (u64)(page_address(frag->page) | 1416 | tmp_addr = (u64)(page_address(frag->page) |
| 1408 | + frag->page_offset); | 1417 | + frag->page_offset); |
| 1409 | sgentry->vaddr = tmp_addr; | 1418 | sgentry->vaddr = ehea_map_vaddr(tmp_addr); |
| 1410 | swqe->descriptors++; | 1419 | swqe->descriptors++; |
| 1411 | } | 1420 | } |
| 1412 | } | 1421 | } |
| @@ -1878,6 +1887,9 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
| 1878 | ehea_dump(swqe, 512, "swqe"); | 1887 | ehea_dump(swqe, 512, "swqe"); |
| 1879 | } | 1888 | } |
| 1880 | 1889 | ||
| 1890 | if (unlikely(test_bit(__EHEA_STOP_XFER, &ehea_driver_flags))) | ||
| 1891 | goto out; | ||
| 1892 | |||
| 1881 | ehea_post_swqe(pr->qp, swqe); | 1893 | ehea_post_swqe(pr->qp, swqe); |
| 1882 | pr->tx_packets++; | 1894 | pr->tx_packets++; |
| 1883 | 1895 | ||
| @@ -1892,7 +1904,7 @@ static int ehea_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
| 1892 | } | 1904 | } |
| 1893 | dev->trans_start = jiffies; | 1905 | dev->trans_start = jiffies; |
| 1894 | spin_unlock(&pr->xmit_lock); | 1906 | spin_unlock(&pr->xmit_lock); |
| 1895 | 1907 | out: | |
| 1896 | return NETDEV_TX_OK; | 1908 | return NETDEV_TX_OK; |
| 1897 | } | 1909 | } |
| 1898 | 1910 | ||
| @@ -2220,6 +2232,9 @@ out_dereg_bc: | |||
| 2220 | out_clean_pr: | 2232 | out_clean_pr: |
| 2221 | ehea_clean_all_portres(port); | 2233 | ehea_clean_all_portres(port); |
| 2222 | out: | 2234 | out: |
| 2235 | if (ret) | ||
| 2236 | ehea_info("Failed starting %s. ret=%i", dev->name, ret); | ||
| 2237 | |||
| 2223 | return ret; | 2238 | return ret; |
| 2224 | } | 2239 | } |
| 2225 | 2240 | ||
| @@ -2259,8 +2274,13 @@ static int ehea_down(struct net_device *dev) | |||
| 2259 | msleep(1); | 2274 | msleep(1); |
| 2260 | 2275 | ||
| 2261 | ehea_broadcast_reg_helper(port, H_DEREG_BCMC); | 2276 | ehea_broadcast_reg_helper(port, H_DEREG_BCMC); |
| 2262 | ret = ehea_clean_all_portres(port); | ||
| 2263 | port->state = EHEA_PORT_DOWN; | 2277 | port->state = EHEA_PORT_DOWN; |
| 2278 | |||
| 2279 | ret = ehea_clean_all_portres(port); | ||
| 2280 | if (ret) | ||
| 2281 | ehea_info("Failed freeing resources for %s. ret=%i", | ||
| 2282 | dev->name, ret); | ||
| 2283 | |||
| 2264 | return ret; | 2284 | return ret; |
| 2265 | } | 2285 | } |
| 2266 | 2286 | ||
| @@ -2292,15 +2312,11 @@ static void ehea_reset_port(struct work_struct *work) | |||
| 2292 | netif_stop_queue(dev); | 2312 | netif_stop_queue(dev); |
| 2293 | netif_poll_disable(dev); | 2313 | netif_poll_disable(dev); |
| 2294 | 2314 | ||
| 2295 | ret = ehea_down(dev); | 2315 | ehea_down(dev); |
| 2296 | if (ret) | ||
| 2297 | ehea_error("ehea_down failed. not all resources are freed"); | ||
| 2298 | 2316 | ||
| 2299 | ret = ehea_up(dev); | 2317 | ret = ehea_up(dev); |
| 2300 | if (ret) { | 2318 | if (ret) |
| 2301 | ehea_error("Reset device %s failed: ret=%d", dev->name, ret); | ||
| 2302 | goto out; | 2319 | goto out; |
| 2303 | } | ||
| 2304 | 2320 | ||
| 2305 | if (netif_msg_timer(port)) | 2321 | if (netif_msg_timer(port)) |
| 2306 | ehea_info("Device %s resetted successfully", dev->name); | 2322 | ehea_info("Device %s resetted successfully", dev->name); |
| @@ -2312,6 +2328,88 @@ out: | |||
| 2312 | return; | 2328 | return; |
| 2313 | } | 2329 | } |
| 2314 | 2330 | ||
| 2331 | static void ehea_rereg_mrs(struct work_struct *work) | ||
| 2332 | { | ||
| 2333 | int ret, i; | ||
| 2334 | struct ehea_adapter *adapter; | ||
| 2335 | |||
| 2336 | ehea_info("LPAR memory enlarged - re-initializing driver"); | ||
| 2337 | |||
| 2338 | list_for_each_entry(adapter, &adapter_list, list) | ||
| 2339 | if (adapter->active_ports) { | ||
| 2340 | /* Shutdown all ports */ | ||
| 2341 | for (i = 0; i < EHEA_MAX_PORTS; i++) { | ||
| 2342 | struct ehea_port *port = adapter->port[i]; | ||
| 2343 | |||
| 2344 | if (port) { | ||
| 2345 | struct net_device *dev = port->netdev; | ||
| 2346 | |||
| 2347 | if (dev->flags & IFF_UP) { | ||
| 2348 | ehea_info("stopping %s", | ||
| 2349 | dev->name); | ||
| 2350 | down(&port->port_lock); | ||
| 2351 | netif_stop_queue(dev); | ||
| 2352 | netif_poll_disable(dev); | ||
| 2353 | ehea_down(dev); | ||
| 2354 | up(&port->port_lock); | ||
| 2355 | } | ||
| 2356 | } | ||
| 2357 | } | ||
| 2358 | |||
| 2359 | /* Unregister old memory region */ | ||
| 2360 | ret = ehea_rem_mr(&adapter->mr); | ||
| 2361 | if (ret) { | ||
| 2362 | ehea_error("unregister MR failed - driver" | ||
| 2363 | " inoperable!"); | ||
| 2364 | goto out; | ||
| 2365 | } | ||
| 2366 | } | ||
| 2367 | |||
| 2368 | ehea_destroy_busmap(); | ||
| 2369 | |||
| 2370 | ret = ehea_create_busmap(); | ||
| 2371 | if (ret) | ||
| 2372 | goto out; | ||
| 2373 | |||
| 2374 | clear_bit(__EHEA_STOP_XFER, &ehea_driver_flags); | ||
| 2375 | |||
| 2376 | list_for_each_entry(adapter, &adapter_list, list) | ||
| 2377 | if (adapter->active_ports) { | ||
| 2378 | /* Register new memory region */ | ||
| 2379 | ret = ehea_reg_kernel_mr(adapter, &adapter->mr); | ||
| 2380 | if (ret) { | ||
| 2381 | ehea_error("register MR failed - driver" | ||
| 2382 | " inoperable!"); | ||
| 2383 | goto out; | ||
| 2384 | } | ||
| 2385 | |||
| 2386 | /* Restart all ports */ | ||
| 2387 | for (i = 0; i < EHEA_MAX_PORTS; i++) { | ||
| 2388 | struct ehea_port *port = adapter->port[i]; | ||
| 2389 | |||
| 2390 | if (port) { | ||
| 2391 | struct net_device *dev = port->netdev; | ||
| 2392 | |||
| 2393 | if (dev->flags & IFF_UP) { | ||
| 2394 | ehea_info("restarting %s", | ||
| 2395 | dev->name); | ||
| 2396 | down(&port->port_lock); | ||
| 2397 | |||
| 2398 | ret = ehea_up(dev); | ||
| 2399 | if (!ret) { | ||
| 2400 | netif_poll_enable(dev); | ||
| 2401 | netif_wake_queue(dev); | ||
| 2402 | } | ||
| 2403 | |||
| 2404 | up(&port->port_lock); | ||
| 2405 | } | ||
| 2406 | } | ||
| 2407 | } | ||
| 2408 | } | ||
| 2409 | out: | ||
| 2410 | return; | ||
| 2411 | } | ||
| 2412 | |||
| 2315 | static void ehea_tx_watchdog(struct net_device *dev) | 2413 | static void ehea_tx_watchdog(struct net_device *dev) |
| 2316 | { | 2414 | { |
| 2317 | struct ehea_port *port = netdev_priv(dev); | 2415 | struct ehea_port *port = netdev_priv(dev); |
| @@ -2573,6 +2671,8 @@ struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter, | |||
| 2573 | ehea_info("%s: Jumbo frames are %sabled", dev->name, | 2671 | ehea_info("%s: Jumbo frames are %sabled", dev->name, |
| 2574 | jumbo == 1 ? "en" : "dis"); | 2672 | jumbo == 1 ? "en" : "dis"); |
| 2575 | 2673 | ||
| 2674 | adapter->active_ports++; | ||
| 2675 | |||
| 2576 | return port; | 2676 | return port; |
| 2577 | 2677 | ||
| 2578 | out_unreg_port: | 2678 | out_unreg_port: |
| @@ -2596,6 +2696,7 @@ static void ehea_shutdown_single_port(struct ehea_port *port) | |||
| 2596 | ehea_unregister_port(port); | 2696 | ehea_unregister_port(port); |
| 2597 | kfree(port->mc_list); | 2697 | kfree(port->mc_list); |
| 2598 | free_netdev(port->netdev); | 2698 | free_netdev(port->netdev); |
| 2699 | port->adapter->active_ports--; | ||
| 2599 | } | 2700 | } |
| 2600 | 2701 | ||
| 2601 | static int ehea_setup_ports(struct ehea_adapter *adapter) | 2702 | static int ehea_setup_ports(struct ehea_adapter *adapter) |
| @@ -2788,6 +2889,8 @@ static int __devinit ehea_probe_adapter(struct ibmebus_dev *dev, | |||
| 2788 | goto out; | 2889 | goto out; |
| 2789 | } | 2890 | } |
| 2790 | 2891 | ||
| 2892 | list_add(&adapter->list, &adapter_list); | ||
| 2893 | |||
| 2791 | adapter->ebus_dev = dev; | 2894 | adapter->ebus_dev = dev; |
| 2792 | 2895 | ||
| 2793 | adapter_handle = of_get_property(dev->ofdev.node, "ibm,hea-handle", | 2896 | adapter_handle = of_get_property(dev->ofdev.node, "ibm,hea-handle", |
| @@ -2891,7 +2994,10 @@ static int __devexit ehea_remove(struct ibmebus_dev *dev) | |||
| 2891 | 2994 | ||
| 2892 | ehea_destroy_eq(adapter->neq); | 2995 | ehea_destroy_eq(adapter->neq); |
| 2893 | ehea_remove_adapter_mr(adapter); | 2996 | ehea_remove_adapter_mr(adapter); |
| 2997 | list_del(&adapter->list); | ||
| 2998 | |||
| 2894 | kfree(adapter); | 2999 | kfree(adapter); |
| 3000 | |||
| 2895 | return 0; | 3001 | return 0; |
| 2896 | } | 3002 | } |
| 2897 | 3003 | ||
| @@ -2939,9 +3045,18 @@ int __init ehea_module_init(void) | |||
| 2939 | printk(KERN_INFO "IBM eHEA ethernet device driver (Release %s)\n", | 3045 | printk(KERN_INFO "IBM eHEA ethernet device driver (Release %s)\n", |
| 2940 | DRV_VERSION); | 3046 | DRV_VERSION); |
| 2941 | 3047 | ||
| 3048 | ehea_driver_wq = create_workqueue("ehea_driver_wq"); | ||
| 3049 | |||
| 3050 | INIT_WORK(&ehea_rereg_mr_task, ehea_rereg_mrs); | ||
| 3051 | |||
| 2942 | ret = check_module_parm(); | 3052 | ret = check_module_parm(); |
| 2943 | if (ret) | 3053 | if (ret) |
| 2944 | goto out; | 3054 | goto out; |
| 3055 | |||
| 3056 | ret = ehea_create_busmap(); | ||
| 3057 | if (ret) | ||
| 3058 | goto out; | ||
| 3059 | |||
| 2945 | ret = ibmebus_register_driver(&ehea_driver); | 3060 | ret = ibmebus_register_driver(&ehea_driver); |
| 2946 | if (ret) { | 3061 | if (ret) { |
| 2947 | ehea_error("failed registering eHEA device driver on ebus"); | 3062 | ehea_error("failed registering eHEA device driver on ebus"); |
| @@ -2965,6 +3080,7 @@ static void __exit ehea_module_exit(void) | |||
| 2965 | { | 3080 | { |
| 2966 | driver_remove_file(&ehea_driver.driver, &driver_attr_capabilities); | 3081 | driver_remove_file(&ehea_driver.driver, &driver_attr_capabilities); |
| 2967 | ibmebus_unregister_driver(&ehea_driver); | 3082 | ibmebus_unregister_driver(&ehea_driver); |
| 3083 | ehea_destroy_busmap(); | ||
| 2968 | } | 3084 | } |
| 2969 | 3085 | ||
| 2970 | module_init(ehea_module_init); | 3086 | module_init(ehea_module_init); |
diff --git a/drivers/net/ehea/ehea_phyp.h b/drivers/net/ehea/ehea_phyp.h index d17a45a7e717..89b63531ff26 100644 --- a/drivers/net/ehea/ehea_phyp.h +++ b/drivers/net/ehea/ehea_phyp.h | |||
| @@ -60,6 +60,9 @@ static inline u32 get_longbusy_msecs(int long_busy_ret_code) | |||
| 60 | } | 60 | } |
| 61 | } | 61 | } |
| 62 | 62 | ||
| 63 | /* Number of pages which can be registered at once by H_REGISTER_HEA_RPAGES */ | ||
| 64 | #define EHEA_MAX_RPAGE 512 | ||
| 65 | |||
| 63 | /* Notification Event Queue (NEQ) Entry bit masks */ | 66 | /* Notification Event Queue (NEQ) Entry bit masks */ |
| 64 | #define NEQE_EVENT_CODE EHEA_BMASK_IBM(2, 7) | 67 | #define NEQE_EVENT_CODE EHEA_BMASK_IBM(2, 7) |
| 65 | #define NEQE_PORTNUM EHEA_BMASK_IBM(32, 47) | 68 | #define NEQE_PORTNUM EHEA_BMASK_IBM(32, 47) |
diff --git a/drivers/net/ehea/ehea_qmr.c b/drivers/net/ehea/ehea_qmr.c index 29eaa46948b0..a36fa6c23fdf 100644 --- a/drivers/net/ehea/ehea_qmr.c +++ b/drivers/net/ehea/ehea_qmr.c | |||
| @@ -31,6 +31,13 @@ | |||
| 31 | #include "ehea_phyp.h" | 31 | #include "ehea_phyp.h" |
| 32 | #include "ehea_qmr.h" | 32 | #include "ehea_qmr.h" |
| 33 | 33 | ||
| 34 | |||
| 35 | struct ehea_busmap ehea_bmap = { 0, 0, NULL }; | ||
| 36 | extern u64 ehea_driver_flags; | ||
| 37 | extern struct workqueue_struct *ehea_driver_wq; | ||
| 38 | extern struct work_struct ehea_rereg_mr_task; | ||
| 39 | |||
| 40 | |||
| 34 | static void *hw_qpageit_get_inc(struct hw_queue *queue) | 41 | static void *hw_qpageit_get_inc(struct hw_queue *queue) |
| 35 | { | 42 | { |
| 36 | void *retvalue = hw_qeit_get(queue); | 43 | void *retvalue = hw_qeit_get(queue); |
| @@ -547,18 +554,84 @@ int ehea_destroy_qp(struct ehea_qp *qp) | |||
| 547 | return 0; | 554 | return 0; |
| 548 | } | 555 | } |
| 549 | 556 | ||
| 557 | int ehea_create_busmap( void ) | ||
| 558 | { | ||
| 559 | u64 vaddr = EHEA_BUSMAP_START; | ||
| 560 | unsigned long abs_max_pfn = 0; | ||
| 561 | unsigned long sec_max_pfn; | ||
| 562 | int i; | ||
| 563 | |||
| 564 | /* | ||
| 565 | * Sections are not in ascending order -> Loop over all sections and | ||
| 566 | * find the highest PFN to compute the required map size. | ||
| 567 | */ | ||
| 568 | ehea_bmap.valid_sections = 0; | ||
| 569 | |||
| 570 | for (i = 0; i < NR_MEM_SECTIONS; i++) | ||
| 571 | if (valid_section_nr(i)) { | ||
| 572 | sec_max_pfn = section_nr_to_pfn(i); | ||
| 573 | if (sec_max_pfn > abs_max_pfn) | ||
| 574 | abs_max_pfn = sec_max_pfn; | ||
| 575 | ehea_bmap.valid_sections++; | ||
| 576 | } | ||
| 577 | |||
| 578 | ehea_bmap.entries = abs_max_pfn / EHEA_PAGES_PER_SECTION + 1; | ||
| 579 | ehea_bmap.vaddr = vmalloc(ehea_bmap.entries * sizeof(*ehea_bmap.vaddr)); | ||
| 580 | |||
| 581 | if (!ehea_bmap.vaddr) | ||
| 582 | return -ENOMEM; | ||
| 583 | |||
| 584 | for (i = 0 ; i < ehea_bmap.entries; i++) { | ||
| 585 | unsigned long pfn = section_nr_to_pfn(i); | ||
| 586 | |||
| 587 | if (pfn_valid(pfn)) { | ||
| 588 | ehea_bmap.vaddr[i] = vaddr; | ||
| 589 | vaddr += EHEA_SECTSIZE; | ||
| 590 | } else | ||
| 591 | ehea_bmap.vaddr[i] = 0; | ||
| 592 | } | ||
| 593 | |||
| 594 | return 0; | ||
| 595 | } | ||
| 596 | |||
| 597 | void ehea_destroy_busmap( void ) | ||
| 598 | { | ||
| 599 | vfree(ehea_bmap.vaddr); | ||
| 600 | } | ||
| 601 | |||
| 602 | u64 ehea_map_vaddr(void *caddr) | ||
| 603 | { | ||
| 604 | u64 mapped_addr; | ||
| 605 | unsigned long index = __pa(caddr) >> SECTION_SIZE_BITS; | ||
| 606 | |||
| 607 | if (likely(index < ehea_bmap.entries)) { | ||
| 608 | mapped_addr = ehea_bmap.vaddr[index]; | ||
| 609 | if (likely(mapped_addr)) | ||
| 610 | mapped_addr |= (((unsigned long)caddr) | ||
| 611 | & (EHEA_SECTSIZE - 1)); | ||
| 612 | else | ||
| 613 | mapped_addr = -1; | ||
| 614 | } else | ||
| 615 | mapped_addr = -1; | ||
| 616 | |||
| 617 | if (unlikely(mapped_addr == -1)) | ||
| 618 | if (!test_and_set_bit(__EHEA_STOP_XFER, &ehea_driver_flags)) | ||
| 619 | queue_work(ehea_driver_wq, &ehea_rereg_mr_task); | ||
| 620 | |||
| 621 | return mapped_addr; | ||
| 622 | } | ||
| 623 | |||
| 550 | int ehea_reg_kernel_mr(struct ehea_adapter *adapter, struct ehea_mr *mr) | 624 | int ehea_reg_kernel_mr(struct ehea_adapter *adapter, struct ehea_mr *mr) |
| 551 | { | 625 | { |
| 552 | int i, k, ret; | 626 | int ret; |
| 553 | u64 hret, pt_abs, start, end, nr_pages; | ||
| 554 | u32 acc_ctrl = EHEA_MR_ACC_CTRL; | ||
| 555 | u64 *pt; | 627 | u64 *pt; |
| 628 | void *pg; | ||
| 629 | u64 hret, pt_abs, i, j, m, mr_len; | ||
| 630 | u32 acc_ctrl = EHEA_MR_ACC_CTRL; | ||
| 556 | 631 | ||
| 557 | start = KERNELBASE; | 632 | mr_len = ehea_bmap.valid_sections * EHEA_SECTSIZE; |
| 558 | end = (u64)high_memory; | ||
| 559 | nr_pages = (end - start) / EHEA_PAGESIZE; | ||
| 560 | 633 | ||
| 561 | pt = kzalloc(PAGE_SIZE, GFP_KERNEL); | 634 | pt = kzalloc(EHEA_MAX_RPAGE * sizeof(u64), GFP_KERNEL); |
| 562 | if (!pt) { | 635 | if (!pt) { |
| 563 | ehea_error("no mem"); | 636 | ehea_error("no mem"); |
| 564 | ret = -ENOMEM; | 637 | ret = -ENOMEM; |
| @@ -566,7 +639,8 @@ int ehea_reg_kernel_mr(struct ehea_adapter *adapter, struct ehea_mr *mr) | |||
| 566 | } | 639 | } |
| 567 | pt_abs = virt_to_abs(pt); | 640 | pt_abs = virt_to_abs(pt); |
| 568 | 641 | ||
| 569 | hret = ehea_h_alloc_resource_mr(adapter->handle, start, end - start, | 642 | hret = ehea_h_alloc_resource_mr(adapter->handle, |
| 643 | EHEA_BUSMAP_START, mr_len, | ||
| 570 | acc_ctrl, adapter->pd, | 644 | acc_ctrl, adapter->pd, |
| 571 | &mr->handle, &mr->lkey); | 645 | &mr->handle, &mr->lkey); |
| 572 | if (hret != H_SUCCESS) { | 646 | if (hret != H_SUCCESS) { |
| @@ -575,49 +649,43 @@ int ehea_reg_kernel_mr(struct ehea_adapter *adapter, struct ehea_mr *mr) | |||
| 575 | goto out; | 649 | goto out; |
| 576 | } | 650 | } |
| 577 | 651 | ||
| 578 | mr->vaddr = KERNELBASE; | 652 | for (i = 0 ; i < ehea_bmap.entries; i++) |
| 579 | k = 0; | 653 | if (ehea_bmap.vaddr[i]) { |
| 580 | 654 | void *sectbase = __va(i << SECTION_SIZE_BITS); | |
| 581 | while (nr_pages > 0) { | 655 | unsigned long k = 0; |
| 582 | if (nr_pages > 1) { | 656 | |
| 583 | u64 num_pages = min(nr_pages, (u64)512); | 657 | for (j = 0; j < (PAGES_PER_SECTION / EHEA_MAX_RPAGE); |
| 584 | for (i = 0; i < num_pages; i++) | 658 | j++) { |
| 585 | pt[i] = virt_to_abs((void*)(((u64)start) + | 659 | |
| 586 | ((k++) * | 660 | for (m = 0; m < EHEA_MAX_RPAGE; m++) { |
| 587 | EHEA_PAGESIZE))); | 661 | pg = sectbase + ((k++) * EHEA_PAGESIZE); |
| 588 | 662 | pt[m] = virt_to_abs(pg); | |
| 589 | hret = ehea_h_register_rpage_mr(adapter->handle, | 663 | } |
| 590 | mr->handle, 0, | 664 | |
| 591 | 0, (u64)pt_abs, | 665 | hret = ehea_h_register_rpage_mr(adapter->handle, |
| 592 | num_pages); | 666 | mr->handle, |
| 593 | nr_pages -= num_pages; | 667 | 0, 0, pt_abs, |
| 594 | } else { | 668 | EHEA_MAX_RPAGE); |
| 595 | u64 abs_adr = virt_to_abs((void*)(((u64)start) + | 669 | if ((hret != H_SUCCESS) |
| 596 | (k * EHEA_PAGESIZE))); | 670 | && (hret != H_PAGE_REGISTERED)) { |
| 597 | 671 | ehea_h_free_resource(adapter->handle, | |
| 598 | hret = ehea_h_register_rpage_mr(adapter->handle, | 672 | mr->handle, |
| 599 | mr->handle, 0, | 673 | FORCE_FREE); |
| 600 | 0, abs_adr,1); | 674 | ehea_error("register_rpage_mr failed"); |
| 601 | nr_pages--; | 675 | ret = -EIO; |
| 602 | } | 676 | goto out; |
| 603 | 677 | } | |
| 604 | if ((hret != H_SUCCESS) && (hret != H_PAGE_REGISTERED)) { | 678 | } |
| 605 | ehea_h_free_resource(adapter->handle, | ||
| 606 | mr->handle, FORCE_FREE); | ||
| 607 | ehea_error("register_rpage_mr failed"); | ||
| 608 | ret = -EIO; | ||
| 609 | goto out; | ||
| 610 | } | 679 | } |
| 611 | } | ||
| 612 | 680 | ||
| 613 | if (hret != H_SUCCESS) { | 681 | if (hret != H_SUCCESS) { |
| 614 | ehea_h_free_resource(adapter->handle, mr->handle, | 682 | ehea_h_free_resource(adapter->handle, mr->handle, FORCE_FREE); |
| 615 | FORCE_FREE); | 683 | ehea_error("registering mr failed"); |
| 616 | ehea_error("register_rpage failed for last page"); | ||
| 617 | ret = -EIO; | 684 | ret = -EIO; |
| 618 | goto out; | 685 | goto out; |
| 619 | } | 686 | } |
| 620 | 687 | ||
| 688 | mr->vaddr = EHEA_BUSMAP_START; | ||
| 621 | mr->adapter = adapter; | 689 | mr->adapter = adapter; |
| 622 | ret = 0; | 690 | ret = 0; |
| 623 | out: | 691 | out: |
diff --git a/drivers/net/ehea/ehea_qmr.h b/drivers/net/ehea/ehea_qmr.h index c0eb3e03a102..b71f8452a5e3 100644 --- a/drivers/net/ehea/ehea_qmr.h +++ b/drivers/net/ehea/ehea_qmr.h | |||
| @@ -36,8 +36,14 @@ | |||
| 36 | * page size of ehea hardware queues | 36 | * page size of ehea hardware queues |
| 37 | */ | 37 | */ |
| 38 | 38 | ||
| 39 | #define EHEA_PAGESHIFT 12 | 39 | #define EHEA_PAGESHIFT 12 |
| 40 | #define EHEA_PAGESIZE 4096UL | 40 | #define EHEA_PAGESIZE (1UL << EHEA_PAGESHIFT) |
| 41 | #define EHEA_SECTSIZE (1UL << 24) | ||
| 42 | #define EHEA_PAGES_PER_SECTION (EHEA_SECTSIZE >> PAGE_SHIFT) | ||
| 43 | |||
| 44 | #if (1UL << SECTION_SIZE_BITS) < EHEA_SECTSIZE | ||
| 45 | #error eHEA module can't work if kernel sectionsize < ehea sectionsize | ||
| 46 | #endif | ||
| 41 | 47 | ||
| 42 | /* Some abbreviations used here: | 48 | /* Some abbreviations used here: |
| 43 | * | 49 | * |
| @@ -372,4 +378,8 @@ int ehea_rem_mr(struct ehea_mr *mr); | |||
| 372 | 378 | ||
| 373 | void ehea_error_data(struct ehea_adapter *adapter, u64 res_handle); | 379 | void ehea_error_data(struct ehea_adapter *adapter, u64 res_handle); |
| 374 | 380 | ||
| 381 | int ehea_create_busmap( void ); | ||
| 382 | void ehea_destroy_busmap( void ); | ||
| 383 | u64 ehea_map_vaddr(void *caddr); | ||
| 384 | |||
| 375 | #endif /* __EHEA_QMR_H__ */ | 385 | #endif /* __EHEA_QMR_H__ */ |
diff --git a/drivers/net/forcedeth.c b/drivers/net/forcedeth.c index 67046e8c21eb..136827f8dc2e 100644 --- a/drivers/net/forcedeth.c +++ b/drivers/net/forcedeth.c | |||
| @@ -550,6 +550,8 @@ union ring_type { | |||
| 550 | /* PHY defines */ | 550 | /* PHY defines */ |
| 551 | #define PHY_OUI_MARVELL 0x5043 | 551 | #define PHY_OUI_MARVELL 0x5043 |
| 552 | #define PHY_OUI_CICADA 0x03f1 | 552 | #define PHY_OUI_CICADA 0x03f1 |
| 553 | #define PHY_OUI_VITESSE 0x01c1 | ||
| 554 | #define PHY_OUI_REALTEK 0x01c1 | ||
| 553 | #define PHYID1_OUI_MASK 0x03ff | 555 | #define PHYID1_OUI_MASK 0x03ff |
| 554 | #define PHYID1_OUI_SHFT 6 | 556 | #define PHYID1_OUI_SHFT 6 |
| 555 | #define PHYID2_OUI_MASK 0xfc00 | 557 | #define PHYID2_OUI_MASK 0xfc00 |
| @@ -557,12 +559,36 @@ union ring_type { | |||
| 557 | #define PHYID2_MODEL_MASK 0x03f0 | 559 | #define PHYID2_MODEL_MASK 0x03f0 |
| 558 | #define PHY_MODEL_MARVELL_E3016 0x220 | 560 | #define PHY_MODEL_MARVELL_E3016 0x220 |
| 559 | #define PHY_MARVELL_E3016_INITMASK 0x0300 | 561 | #define PHY_MARVELL_E3016_INITMASK 0x0300 |
| 560 | #define PHY_INIT1 0x0f000 | 562 | #define PHY_CICADA_INIT1 0x0f000 |
| 561 | #define PHY_INIT2 0x0e00 | 563 | #define PHY_CICADA_INIT2 0x0e00 |
| 562 | #define PHY_INIT3 0x01000 | 564 | #define PHY_CICADA_INIT3 0x01000 |
| 563 | #define PHY_INIT4 0x0200 | 565 | #define PHY_CICADA_INIT4 0x0200 |
| 564 | #define PHY_INIT5 0x0004 | 566 | #define PHY_CICADA_INIT5 0x0004 |
| 565 | #define PHY_INIT6 0x02000 | 567 | #define PHY_CICADA_INIT6 0x02000 |
| 568 | #define PHY_VITESSE_INIT_REG1 0x1f | ||
| 569 | #define PHY_VITESSE_INIT_REG2 0x10 | ||
| 570 | #define PHY_VITESSE_INIT_REG3 0x11 | ||
| 571 | #define PHY_VITESSE_INIT_REG4 0x12 | ||
| 572 | #define PHY_VITESSE_INIT_MSK1 0xc | ||
| 573 | #define PHY_VITESSE_INIT_MSK2 0x0180 | ||
| 574 | #define PHY_VITESSE_INIT1 0x52b5 | ||
| 575 | #define PHY_VITESSE_INIT2 0xaf8a | ||
| 576 | #define PHY_VITESSE_INIT3 0x8 | ||
| 577 | #define PHY_VITESSE_INIT4 0x8f8a | ||
| 578 | #define PHY_VITESSE_INIT5 0xaf86 | ||
| 579 | #define PHY_VITESSE_INIT6 0x8f86 | ||
| 580 | #define PHY_VITESSE_INIT7 0xaf82 | ||
| 581 | #define PHY_VITESSE_INIT8 0x0100 | ||
| 582 | #define PHY_VITESSE_INIT9 0x8f82 | ||
| 583 | #define PHY_VITESSE_INIT10 0x0 | ||
| 584 | #define PHY_REALTEK_INIT_REG1 0x1f | ||
| 585 | #define PHY_REALTEK_INIT_REG2 0x19 | ||
| 586 | #define PHY_REALTEK_INIT_REG3 0x13 | ||
| 587 | #define PHY_REALTEK_INIT1 0x0000 | ||
| 588 | #define PHY_REALTEK_INIT2 0x8e00 | ||
| 589 | #define PHY_REALTEK_INIT3 0x0001 | ||
| 590 | #define PHY_REALTEK_INIT4 0xad17 | ||
| 591 | |||
| 566 | #define PHY_GIGABIT 0x0100 | 592 | #define PHY_GIGABIT 0x0100 |
| 567 | 593 | ||
| 568 | #define PHY_TIMEOUT 0x1 | 594 | #define PHY_TIMEOUT 0x1 |
| @@ -1096,6 +1122,28 @@ static int phy_init(struct net_device *dev) | |||
| 1096 | return PHY_ERROR; | 1122 | return PHY_ERROR; |
| 1097 | } | 1123 | } |
| 1098 | } | 1124 | } |
| 1125 | if (np->phy_oui == PHY_OUI_REALTEK) { | ||
| 1126 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT1)) { | ||
| 1127 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1128 | return PHY_ERROR; | ||
| 1129 | } | ||
| 1130 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG2, PHY_REALTEK_INIT2)) { | ||
| 1131 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1132 | return PHY_ERROR; | ||
| 1133 | } | ||
| 1134 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT3)) { | ||
| 1135 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1136 | return PHY_ERROR; | ||
| 1137 | } | ||
| 1138 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG3, PHY_REALTEK_INIT4)) { | ||
| 1139 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1140 | return PHY_ERROR; | ||
| 1141 | } | ||
| 1142 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT1)) { | ||
| 1143 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1144 | return PHY_ERROR; | ||
| 1145 | } | ||
| 1146 | } | ||
| 1099 | 1147 | ||
| 1100 | /* set advertise register */ | 1148 | /* set advertise register */ |
| 1101 | reg = mii_rw(dev, np->phyaddr, MII_ADVERTISE, MII_READ); | 1149 | reg = mii_rw(dev, np->phyaddr, MII_ADVERTISE, MII_READ); |
| @@ -1141,14 +1189,14 @@ static int phy_init(struct net_device *dev) | |||
| 1141 | /* phy vendor specific configuration */ | 1189 | /* phy vendor specific configuration */ |
| 1142 | if ((np->phy_oui == PHY_OUI_CICADA) && (phyinterface & PHY_RGMII) ) { | 1190 | if ((np->phy_oui == PHY_OUI_CICADA) && (phyinterface & PHY_RGMII) ) { |
| 1143 | phy_reserved = mii_rw(dev, np->phyaddr, MII_RESV1, MII_READ); | 1191 | phy_reserved = mii_rw(dev, np->phyaddr, MII_RESV1, MII_READ); |
| 1144 | phy_reserved &= ~(PHY_INIT1 | PHY_INIT2); | 1192 | phy_reserved &= ~(PHY_CICADA_INIT1 | PHY_CICADA_INIT2); |
| 1145 | phy_reserved |= (PHY_INIT3 | PHY_INIT4); | 1193 | phy_reserved |= (PHY_CICADA_INIT3 | PHY_CICADA_INIT4); |
| 1146 | if (mii_rw(dev, np->phyaddr, MII_RESV1, phy_reserved)) { | 1194 | if (mii_rw(dev, np->phyaddr, MII_RESV1, phy_reserved)) { |
| 1147 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | 1195 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); |
| 1148 | return PHY_ERROR; | 1196 | return PHY_ERROR; |
| 1149 | } | 1197 | } |
| 1150 | phy_reserved = mii_rw(dev, np->phyaddr, MII_NCONFIG, MII_READ); | 1198 | phy_reserved = mii_rw(dev, np->phyaddr, MII_NCONFIG, MII_READ); |
| 1151 | phy_reserved |= PHY_INIT5; | 1199 | phy_reserved |= PHY_CICADA_INIT5; |
| 1152 | if (mii_rw(dev, np->phyaddr, MII_NCONFIG, phy_reserved)) { | 1200 | if (mii_rw(dev, np->phyaddr, MII_NCONFIG, phy_reserved)) { |
| 1153 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | 1201 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); |
| 1154 | return PHY_ERROR; | 1202 | return PHY_ERROR; |
| @@ -1156,12 +1204,106 @@ static int phy_init(struct net_device *dev) | |||
| 1156 | } | 1204 | } |
| 1157 | if (np->phy_oui == PHY_OUI_CICADA) { | 1205 | if (np->phy_oui == PHY_OUI_CICADA) { |
| 1158 | phy_reserved = mii_rw(dev, np->phyaddr, MII_SREVISION, MII_READ); | 1206 | phy_reserved = mii_rw(dev, np->phyaddr, MII_SREVISION, MII_READ); |
| 1159 | phy_reserved |= PHY_INIT6; | 1207 | phy_reserved |= PHY_CICADA_INIT6; |
| 1160 | if (mii_rw(dev, np->phyaddr, MII_SREVISION, phy_reserved)) { | 1208 | if (mii_rw(dev, np->phyaddr, MII_SREVISION, phy_reserved)) { |
| 1161 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | 1209 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); |
| 1162 | return PHY_ERROR; | 1210 | return PHY_ERROR; |
| 1163 | } | 1211 | } |
| 1164 | } | 1212 | } |
| 1213 | if (np->phy_oui == PHY_OUI_VITESSE) { | ||
| 1214 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG1, PHY_VITESSE_INIT1)) { | ||
| 1215 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1216 | return PHY_ERROR; | ||
| 1217 | } | ||
| 1218 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT2)) { | ||
| 1219 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1220 | return PHY_ERROR; | ||
| 1221 | } | ||
| 1222 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, MII_READ); | ||
| 1223 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, phy_reserved)) { | ||
| 1224 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1225 | return PHY_ERROR; | ||
| 1226 | } | ||
| 1227 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, MII_READ); | ||
| 1228 | phy_reserved &= ~PHY_VITESSE_INIT_MSK1; | ||
| 1229 | phy_reserved |= PHY_VITESSE_INIT3; | ||
| 1230 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, phy_reserved)) { | ||
| 1231 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1232 | return PHY_ERROR; | ||
| 1233 | } | ||
| 1234 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT4)) { | ||
| 1235 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1236 | return PHY_ERROR; | ||
| 1237 | } | ||
| 1238 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT5)) { | ||
| 1239 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1240 | return PHY_ERROR; | ||
| 1241 | } | ||
| 1242 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, MII_READ); | ||
| 1243 | phy_reserved &= ~PHY_VITESSE_INIT_MSK1; | ||
| 1244 | phy_reserved |= PHY_VITESSE_INIT3; | ||
| 1245 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, phy_reserved)) { | ||
| 1246 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1247 | return PHY_ERROR; | ||
| 1248 | } | ||
| 1249 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, MII_READ); | ||
| 1250 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, phy_reserved)) { | ||
| 1251 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1252 | return PHY_ERROR; | ||
| 1253 | } | ||
| 1254 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT6)) { | ||
| 1255 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1256 | return PHY_ERROR; | ||
| 1257 | } | ||
| 1258 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT7)) { | ||
| 1259 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1260 | return PHY_ERROR; | ||
| 1261 | } | ||
| 1262 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, MII_READ); | ||
| 1263 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG4, phy_reserved)) { | ||
| 1264 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1265 | return PHY_ERROR; | ||
| 1266 | } | ||
| 1267 | phy_reserved = mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, MII_READ); | ||
| 1268 | phy_reserved &= ~PHY_VITESSE_INIT_MSK2; | ||
| 1269 | phy_reserved |= PHY_VITESSE_INIT8; | ||
| 1270 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG3, phy_reserved)) { | ||
| 1271 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1272 | return PHY_ERROR; | ||
| 1273 | } | ||
| 1274 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG2, PHY_VITESSE_INIT9)) { | ||
| 1275 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1276 | return PHY_ERROR; | ||
| 1277 | } | ||
| 1278 | if (mii_rw(dev, np->phyaddr, PHY_VITESSE_INIT_REG1, PHY_VITESSE_INIT10)) { | ||
| 1279 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1280 | return PHY_ERROR; | ||
| 1281 | } | ||
| 1282 | } | ||
| 1283 | if (np->phy_oui == PHY_OUI_REALTEK) { | ||
| 1284 | /* reset could have cleared these out, set them back */ | ||
| 1285 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT1)) { | ||
| 1286 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1287 | return PHY_ERROR; | ||
| 1288 | } | ||
| 1289 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG2, PHY_REALTEK_INIT2)) { | ||
| 1290 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1291 | return PHY_ERROR; | ||
| 1292 | } | ||
| 1293 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT3)) { | ||
| 1294 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1295 | return PHY_ERROR; | ||
| 1296 | } | ||
| 1297 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG3, PHY_REALTEK_INIT4)) { | ||
| 1298 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1299 | return PHY_ERROR; | ||
| 1300 | } | ||
| 1301 | if (mii_rw(dev, np->phyaddr, PHY_REALTEK_INIT_REG1, PHY_REALTEK_INIT1)) { | ||
| 1302 | printk(KERN_INFO "%s: phy init failed.\n", pci_name(np->pci_dev)); | ||
| 1303 | return PHY_ERROR; | ||
| 1304 | } | ||
| 1305 | } | ||
| 1306 | |||
| 1165 | /* some phys clear out pause advertisment on reset, set it back */ | 1307 | /* some phys clear out pause advertisment on reset, set it back */ |
| 1166 | mii_rw(dev, np->phyaddr, MII_ADVERTISE, reg); | 1308 | mii_rw(dev, np->phyaddr, MII_ADVERTISE, reg); |
| 1167 | 1309 | ||
diff --git a/drivers/net/gianfar_mii.c b/drivers/net/gianfar_mii.c index 5dd34a1a7b89..ac3596f45dd8 100644 --- a/drivers/net/gianfar_mii.c +++ b/drivers/net/gianfar_mii.c | |||
| @@ -31,7 +31,6 @@ | |||
| 31 | #include <linux/mm.h> | 31 | #include <linux/mm.h> |
| 32 | #include <linux/module.h> | 32 | #include <linux/module.h> |
| 33 | #include <linux/platform_device.h> | 33 | #include <linux/platform_device.h> |
| 34 | #include <asm/ocp.h> | ||
| 35 | #include <linux/crc32.h> | 34 | #include <linux/crc32.h> |
| 36 | #include <linux/mii.h> | 35 | #include <linux/mii.h> |
| 37 | #include <linux/phy.h> | 36 | #include <linux/phy.h> |
diff --git a/drivers/net/macb.c b/drivers/net/macb.c index 0e04f7ac3f2e..a4bb0264180a 100644 --- a/drivers/net/macb.c +++ b/drivers/net/macb.c | |||
| @@ -17,13 +17,12 @@ | |||
| 17 | #include <linux/init.h> | 17 | #include <linux/init.h> |
| 18 | #include <linux/netdevice.h> | 18 | #include <linux/netdevice.h> |
| 19 | #include <linux/etherdevice.h> | 19 | #include <linux/etherdevice.h> |
| 20 | #include <linux/mii.h> | ||
| 21 | #include <linux/mutex.h> | ||
| 22 | #include <linux/dma-mapping.h> | 20 | #include <linux/dma-mapping.h> |
| 23 | #include <linux/ethtool.h> | ||
| 24 | #include <linux/platform_device.h> | 21 | #include <linux/platform_device.h> |
| 22 | #include <linux/phy.h> | ||
| 25 | 23 | ||
| 26 | #include <asm/arch/board.h> | 24 | #include <asm/arch/board.h> |
| 25 | #include <asm/arch/cpu.h> | ||
| 27 | 26 | ||
| 28 | #include "macb.h" | 27 | #include "macb.h" |
| 29 | 28 | ||
| @@ -85,172 +84,202 @@ static void __init macb_get_hwaddr(struct macb *bp) | |||
| 85 | memcpy(bp->dev->dev_addr, addr, sizeof(addr)); | 84 | memcpy(bp->dev->dev_addr, addr, sizeof(addr)); |
| 86 | } | 85 | } |
| 87 | 86 | ||
| 88 | static void macb_enable_mdio(struct macb *bp) | 87 | static int macb_mdio_read(struct mii_bus *bus, int mii_id, int regnum) |
| 89 | { | 88 | { |
| 90 | unsigned long flags; | 89 | struct macb *bp = bus->priv; |
| 91 | u32 reg; | ||
| 92 | |||
| 93 | spin_lock_irqsave(&bp->lock, flags); | ||
| 94 | reg = macb_readl(bp, NCR); | ||
| 95 | reg |= MACB_BIT(MPE); | ||
| 96 | macb_writel(bp, NCR, reg); | ||
| 97 | macb_writel(bp, IER, MACB_BIT(MFD)); | ||
| 98 | spin_unlock_irqrestore(&bp->lock, flags); | ||
| 99 | } | ||
| 100 | |||
| 101 | static void macb_disable_mdio(struct macb *bp) | ||
| 102 | { | ||
| 103 | unsigned long flags; | ||
| 104 | u32 reg; | ||
| 105 | |||
| 106 | spin_lock_irqsave(&bp->lock, flags); | ||
| 107 | reg = macb_readl(bp, NCR); | ||
| 108 | reg &= ~MACB_BIT(MPE); | ||
| 109 | macb_writel(bp, NCR, reg); | ||
| 110 | macb_writel(bp, IDR, MACB_BIT(MFD)); | ||
| 111 | spin_unlock_irqrestore(&bp->lock, flags); | ||
| 112 | } | ||
| 113 | |||
| 114 | static int macb_mdio_read(struct net_device *dev, int phy_id, int location) | ||
| 115 | { | ||
| 116 | struct macb *bp = netdev_priv(dev); | ||
| 117 | int value; | 90 | int value; |
| 118 | 91 | ||
| 119 | mutex_lock(&bp->mdio_mutex); | ||
| 120 | |||
| 121 | macb_enable_mdio(bp); | ||
| 122 | macb_writel(bp, MAN, (MACB_BF(SOF, MACB_MAN_SOF) | 92 | macb_writel(bp, MAN, (MACB_BF(SOF, MACB_MAN_SOF) |
| 123 | | MACB_BF(RW, MACB_MAN_READ) | 93 | | MACB_BF(RW, MACB_MAN_READ) |
| 124 | | MACB_BF(PHYA, phy_id) | 94 | | MACB_BF(PHYA, mii_id) |
| 125 | | MACB_BF(REGA, location) | 95 | | MACB_BF(REGA, regnum) |
| 126 | | MACB_BF(CODE, MACB_MAN_CODE))); | 96 | | MACB_BF(CODE, MACB_MAN_CODE))); |
| 127 | 97 | ||
| 128 | wait_for_completion(&bp->mdio_complete); | 98 | /* wait for end of transfer */ |
| 99 | while (!MACB_BFEXT(IDLE, macb_readl(bp, NSR))) | ||
| 100 | cpu_relax(); | ||
| 129 | 101 | ||
| 130 | value = MACB_BFEXT(DATA, macb_readl(bp, MAN)); | 102 | value = MACB_BFEXT(DATA, macb_readl(bp, MAN)); |
| 131 | macb_disable_mdio(bp); | ||
| 132 | mutex_unlock(&bp->mdio_mutex); | ||
| 133 | 103 | ||
| 134 | return value; | 104 | return value; |
| 135 | } | 105 | } |
| 136 | 106 | ||
| 137 | static void macb_mdio_write(struct net_device *dev, int phy_id, | 107 | static int macb_mdio_write(struct mii_bus *bus, int mii_id, int regnum, |
| 138 | int location, int val) | 108 | u16 value) |
| 139 | { | 109 | { |
| 140 | struct macb *bp = netdev_priv(dev); | 110 | struct macb *bp = bus->priv; |
| 141 | |||
| 142 | dev_dbg(&bp->pdev->dev, "mdio_write %02x:%02x <- %04x\n", | ||
| 143 | phy_id, location, val); | ||
| 144 | |||
| 145 | mutex_lock(&bp->mdio_mutex); | ||
| 146 | macb_enable_mdio(bp); | ||
| 147 | 111 | ||
| 148 | macb_writel(bp, MAN, (MACB_BF(SOF, MACB_MAN_SOF) | 112 | macb_writel(bp, MAN, (MACB_BF(SOF, MACB_MAN_SOF) |
| 149 | | MACB_BF(RW, MACB_MAN_WRITE) | 113 | | MACB_BF(RW, MACB_MAN_WRITE) |
| 150 | | MACB_BF(PHYA, phy_id) | 114 | | MACB_BF(PHYA, mii_id) |
| 151 | | MACB_BF(REGA, location) | 115 | | MACB_BF(REGA, regnum) |
| 152 | | MACB_BF(CODE, MACB_MAN_CODE) | 116 | | MACB_BF(CODE, MACB_MAN_CODE) |
| 153 | | MACB_BF(DATA, val))); | 117 | | MACB_BF(DATA, value))); |
| 154 | 118 | ||
| 155 | wait_for_completion(&bp->mdio_complete); | 119 | /* wait for end of transfer */ |
| 120 | while (!MACB_BFEXT(IDLE, macb_readl(bp, NSR))) | ||
| 121 | cpu_relax(); | ||
| 156 | 122 | ||
| 157 | macb_disable_mdio(bp); | 123 | return 0; |
| 158 | mutex_unlock(&bp->mdio_mutex); | ||
| 159 | } | 124 | } |
| 160 | 125 | ||
| 161 | static int macb_phy_probe(struct macb *bp) | 126 | static int macb_mdio_reset(struct mii_bus *bus) |
| 162 | { | 127 | { |
| 163 | int phy_address; | 128 | return 0; |
| 164 | u16 phyid1, phyid2; | 129 | } |
| 165 | 130 | ||
| 166 | for (phy_address = 0; phy_address < 32; phy_address++) { | 131 | static void macb_handle_link_change(struct net_device *dev) |
| 167 | phyid1 = macb_mdio_read(bp->dev, phy_address, MII_PHYSID1); | 132 | { |
| 168 | phyid2 = macb_mdio_read(bp->dev, phy_address, MII_PHYSID2); | 133 | struct macb *bp = netdev_priv(dev); |
| 134 | struct phy_device *phydev = bp->phy_dev; | ||
| 135 | unsigned long flags; | ||
| 169 | 136 | ||
| 170 | if (phyid1 != 0xffff && phyid1 != 0x0000 | 137 | int status_change = 0; |
| 171 | && phyid2 != 0xffff && phyid2 != 0x0000) | 138 | |
| 172 | break; | 139 | spin_lock_irqsave(&bp->lock, flags); |
| 140 | |||
| 141 | if (phydev->link) { | ||
| 142 | if ((bp->speed != phydev->speed) || | ||
| 143 | (bp->duplex != phydev->duplex)) { | ||
| 144 | u32 reg; | ||
| 145 | |||
| 146 | reg = macb_readl(bp, NCFGR); | ||
| 147 | reg &= ~(MACB_BIT(SPD) | MACB_BIT(FD)); | ||
| 148 | |||
| 149 | if (phydev->duplex) | ||
| 150 | reg |= MACB_BIT(FD); | ||
| 151 | if (phydev->speed) | ||
| 152 | reg |= MACB_BIT(SPD); | ||
| 153 | |||
| 154 | macb_writel(bp, NCFGR, reg); | ||
| 155 | |||
| 156 | bp->speed = phydev->speed; | ||
| 157 | bp->duplex = phydev->duplex; | ||
| 158 | status_change = 1; | ||
| 159 | } | ||
| 173 | } | 160 | } |
| 174 | 161 | ||
| 175 | if (phy_address == 32) | 162 | if (phydev->link != bp->link) { |
| 176 | return -ENODEV; | 163 | if (phydev->link) |
| 164 | netif_schedule(dev); | ||
| 165 | else { | ||
| 166 | bp->speed = 0; | ||
| 167 | bp->duplex = -1; | ||
| 168 | } | ||
| 169 | bp->link = phydev->link; | ||
| 177 | 170 | ||
| 178 | dev_info(&bp->pdev->dev, | 171 | status_change = 1; |
| 179 | "detected PHY at address %d (ID %04x:%04x)\n", | 172 | } |
| 180 | phy_address, phyid1, phyid2); | ||
| 181 | 173 | ||
| 182 | bp->mii.phy_id = phy_address; | 174 | spin_unlock_irqrestore(&bp->lock, flags); |
| 183 | return 0; | 175 | |
| 176 | if (status_change) { | ||
| 177 | if (phydev->link) | ||
| 178 | printk(KERN_INFO "%s: link up (%d/%s)\n", | ||
| 179 | dev->name, phydev->speed, | ||
| 180 | DUPLEX_FULL == phydev->duplex ? "Full":"Half"); | ||
| 181 | else | ||
| 182 | printk(KERN_INFO "%s: link down\n", dev->name); | ||
| 183 | } | ||
| 184 | } | 184 | } |
| 185 | 185 | ||
| 186 | static void macb_set_media(struct macb *bp, int media) | 186 | /* based on au1000_eth. c*/ |
| 187 | static int macb_mii_probe(struct net_device *dev) | ||
| 187 | { | 188 | { |
| 188 | u32 reg; | 189 | struct macb *bp = netdev_priv(dev); |
| 190 | struct phy_device *phydev = NULL; | ||
| 191 | struct eth_platform_data *pdata; | ||
| 192 | int phy_addr; | ||
| 189 | 193 | ||
| 190 | spin_lock_irq(&bp->lock); | 194 | /* find the first phy */ |
| 191 | reg = macb_readl(bp, NCFGR); | 195 | for (phy_addr = 0; phy_addr < PHY_MAX_ADDR; phy_addr++) { |
| 192 | reg &= ~(MACB_BIT(SPD) | MACB_BIT(FD)); | 196 | if (bp->mii_bus.phy_map[phy_addr]) { |
| 193 | if (media & (ADVERTISE_100HALF | ADVERTISE_100FULL)) | 197 | phydev = bp->mii_bus.phy_map[phy_addr]; |
| 194 | reg |= MACB_BIT(SPD); | 198 | break; |
| 195 | if (media & ADVERTISE_FULL) | 199 | } |
| 196 | reg |= MACB_BIT(FD); | 200 | } |
| 197 | macb_writel(bp, NCFGR, reg); | 201 | |
| 198 | spin_unlock_irq(&bp->lock); | 202 | if (!phydev) { |
| 203 | printk (KERN_ERR "%s: no PHY found\n", dev->name); | ||
| 204 | return -1; | ||
| 205 | } | ||
| 206 | |||
| 207 | pdata = bp->pdev->dev.platform_data; | ||
| 208 | /* TODO : add pin_irq */ | ||
| 209 | |||
| 210 | /* attach the mac to the phy */ | ||
| 211 | if (pdata && pdata->is_rmii) { | ||
| 212 | phydev = phy_connect(dev, phydev->dev.bus_id, | ||
| 213 | &macb_handle_link_change, 0, PHY_INTERFACE_MODE_RMII); | ||
| 214 | } else { | ||
| 215 | phydev = phy_connect(dev, phydev->dev.bus_id, | ||
| 216 | &macb_handle_link_change, 0, PHY_INTERFACE_MODE_MII); | ||
| 217 | } | ||
| 218 | |||
| 219 | if (IS_ERR(phydev)) { | ||
| 220 | printk(KERN_ERR "%s: Could not attach to PHY\n", dev->name); | ||
| 221 | return PTR_ERR(phydev); | ||
| 222 | } | ||
| 223 | |||
| 224 | /* mask with MAC supported features */ | ||
| 225 | phydev->supported &= PHY_BASIC_FEATURES; | ||
| 226 | |||
| 227 | phydev->advertising = phydev->supported; | ||
| 228 | |||
| 229 | bp->link = 0; | ||
| 230 | bp->speed = 0; | ||
| 231 | bp->duplex = -1; | ||
| 232 | bp->phy_dev = phydev; | ||
| 233 | |||
| 234 | return 0; | ||
| 199 | } | 235 | } |
| 200 | 236 | ||
| 201 | static void macb_check_media(struct macb *bp, int ok_to_print, int init_media) | 237 | static int macb_mii_init(struct macb *bp) |
| 202 | { | 238 | { |
| 203 | struct mii_if_info *mii = &bp->mii; | 239 | struct eth_platform_data *pdata; |
| 204 | unsigned int old_carrier, new_carrier; | 240 | int err = -ENXIO, i; |
| 205 | int advertise, lpa, media, duplex; | ||
| 206 | 241 | ||
| 207 | /* if forced media, go no further */ | 242 | /* Enable managment port */ |
| 208 | if (mii->force_media) | 243 | macb_writel(bp, NCR, MACB_BIT(MPE)); |
| 209 | return; | ||
| 210 | 244 | ||
| 211 | /* check current and old link status */ | 245 | bp->mii_bus.name = "MACB_mii_bus", |
| 212 | old_carrier = netif_carrier_ok(mii->dev) ? 1 : 0; | 246 | bp->mii_bus.read = &macb_mdio_read, |
| 213 | new_carrier = (unsigned int) mii_link_ok(mii); | 247 | bp->mii_bus.write = &macb_mdio_write, |
| 248 | bp->mii_bus.reset = &macb_mdio_reset, | ||
| 249 | bp->mii_bus.id = bp->pdev->id, | ||
| 250 | bp->mii_bus.priv = bp, | ||
| 251 | bp->mii_bus.dev = &bp->dev->dev; | ||
| 252 | pdata = bp->pdev->dev.platform_data; | ||
| 214 | 253 | ||
| 215 | /* if carrier state did not change, assume nothing else did */ | 254 | if (pdata) |
| 216 | if (!init_media && old_carrier == new_carrier) | 255 | bp->mii_bus.phy_mask = pdata->phy_mask; |
| 217 | return; | ||
| 218 | 256 | ||
| 219 | /* no carrier, nothing much to do */ | 257 | bp->mii_bus.irq = kmalloc(sizeof(int)*PHY_MAX_ADDR, GFP_KERNEL); |
| 220 | if (!new_carrier) { | 258 | if (!bp->mii_bus.irq) { |
| 221 | netif_carrier_off(mii->dev); | 259 | err = -ENOMEM; |
| 222 | printk(KERN_INFO "%s: link down\n", mii->dev->name); | 260 | goto err_out; |
| 223 | return; | ||
| 224 | } | 261 | } |
| 225 | 262 | ||
| 226 | /* | 263 | for (i = 0; i < PHY_MAX_ADDR; i++) |
| 227 | * we have carrier, see who's on the other end | 264 | bp->mii_bus.irq[i] = PHY_POLL; |
| 228 | */ | ||
| 229 | netif_carrier_on(mii->dev); | ||
| 230 | 265 | ||
| 231 | /* get MII advertise and LPA values */ | 266 | platform_set_drvdata(bp->dev, &bp->mii_bus); |
| 232 | if (!init_media && mii->advertising) { | ||
| 233 | advertise = mii->advertising; | ||
| 234 | } else { | ||
| 235 | advertise = mii->mdio_read(mii->dev, mii->phy_id, MII_ADVERTISE); | ||
| 236 | mii->advertising = advertise; | ||
| 237 | } | ||
| 238 | lpa = mii->mdio_read(mii->dev, mii->phy_id, MII_LPA); | ||
| 239 | 267 | ||
| 240 | /* figure out media and duplex from advertise and LPA values */ | 268 | if (mdiobus_register(&bp->mii_bus)) |
| 241 | media = mii_nway_result(lpa & advertise); | 269 | goto err_out_free_mdio_irq; |
| 242 | duplex = (media & ADVERTISE_FULL) ? 1 : 0; | ||
| 243 | 270 | ||
| 244 | if (ok_to_print) | 271 | if (macb_mii_probe(bp->dev) != 0) { |
| 245 | printk(KERN_INFO "%s: link up, %sMbps, %s-duplex, lpa 0x%04X\n", | 272 | goto err_out_unregister_bus; |
| 246 | mii->dev->name, | 273 | } |
| 247 | media & (ADVERTISE_100FULL | ADVERTISE_100HALF) ? "100" : "10", | ||
| 248 | duplex ? "full" : "half", lpa); | ||
| 249 | 274 | ||
| 250 | mii->full_duplex = duplex; | 275 | return 0; |
| 251 | 276 | ||
| 252 | /* Let the MAC know about the new link state */ | 277 | err_out_unregister_bus: |
| 253 | macb_set_media(bp, media); | 278 | mdiobus_unregister(&bp->mii_bus); |
| 279 | err_out_free_mdio_irq: | ||
| 280 | kfree(bp->mii_bus.irq); | ||
| 281 | err_out: | ||
| 282 | return err; | ||
| 254 | } | 283 | } |
| 255 | 284 | ||
| 256 | static void macb_update_stats(struct macb *bp) | 285 | static void macb_update_stats(struct macb *bp) |
| @@ -265,16 +294,6 @@ static void macb_update_stats(struct macb *bp) | |||
| 265 | *p += __raw_readl(reg); | 294 | *p += __raw_readl(reg); |
| 266 | } | 295 | } |
| 267 | 296 | ||
| 268 | static void macb_periodic_task(struct work_struct *work) | ||
| 269 | { | ||
| 270 | struct macb *bp = container_of(work, struct macb, periodic_task.work); | ||
| 271 | |||
| 272 | macb_update_stats(bp); | ||
| 273 | macb_check_media(bp, 1, 0); | ||
| 274 | |||
| 275 | schedule_delayed_work(&bp->periodic_task, HZ); | ||
| 276 | } | ||
| 277 | |||
| 278 | static void macb_tx(struct macb *bp) | 297 | static void macb_tx(struct macb *bp) |
| 279 | { | 298 | { |
| 280 | unsigned int tail; | 299 | unsigned int tail; |
| @@ -519,9 +538,6 @@ static irqreturn_t macb_interrupt(int irq, void *dev_id) | |||
| 519 | spin_lock(&bp->lock); | 538 | spin_lock(&bp->lock); |
| 520 | 539 | ||
| 521 | while (status) { | 540 | while (status) { |
| 522 | if (status & MACB_BIT(MFD)) | ||
| 523 | complete(&bp->mdio_complete); | ||
| 524 | |||
| 525 | /* close possible race with dev_close */ | 541 | /* close possible race with dev_close */ |
| 526 | if (unlikely(!netif_running(dev))) { | 542 | if (unlikely(!netif_running(dev))) { |
| 527 | macb_writel(bp, IDR, ~0UL); | 543 | macb_writel(bp, IDR, ~0UL); |
| @@ -535,7 +551,8 @@ static irqreturn_t macb_interrupt(int irq, void *dev_id) | |||
| 535 | * until we have processed the buffers | 551 | * until we have processed the buffers |
| 536 | */ | 552 | */ |
| 537 | macb_writel(bp, IDR, MACB_RX_INT_FLAGS); | 553 | macb_writel(bp, IDR, MACB_RX_INT_FLAGS); |
| 538 | dev_dbg(&bp->pdev->dev, "scheduling RX softirq\n"); | 554 | dev_dbg(&bp->pdev->dev, |
| 555 | "scheduling RX softirq\n"); | ||
| 539 | __netif_rx_schedule(dev); | 556 | __netif_rx_schedule(dev); |
| 540 | } | 557 | } |
| 541 | } | 558 | } |
| @@ -765,7 +782,7 @@ static void macb_init_hw(struct macb *bp) | |||
| 765 | macb_writel(bp, TBQP, bp->tx_ring_dma); | 782 | macb_writel(bp, TBQP, bp->tx_ring_dma); |
| 766 | 783 | ||
| 767 | /* Enable TX and RX */ | 784 | /* Enable TX and RX */ |
| 768 | macb_writel(bp, NCR, MACB_BIT(RE) | MACB_BIT(TE)); | 785 | macb_writel(bp, NCR, MACB_BIT(RE) | MACB_BIT(TE) | MACB_BIT(MPE)); |
| 769 | 786 | ||
| 770 | /* Enable interrupts */ | 787 | /* Enable interrupts */ |
| 771 | macb_writel(bp, IER, (MACB_BIT(RCOMP) | 788 | macb_writel(bp, IER, (MACB_BIT(RCOMP) |
| @@ -776,18 +793,126 @@ static void macb_init_hw(struct macb *bp) | |||
| 776 | | MACB_BIT(TCOMP) | 793 | | MACB_BIT(TCOMP) |
| 777 | | MACB_BIT(ISR_ROVR) | 794 | | MACB_BIT(ISR_ROVR) |
| 778 | | MACB_BIT(HRESP))); | 795 | | MACB_BIT(HRESP))); |
| 796 | |||
| 779 | } | 797 | } |
| 780 | 798 | ||
| 781 | static void macb_init_phy(struct net_device *dev) | 799 | /* |
| 800 | * The hash address register is 64 bits long and takes up two | ||
| 801 | * locations in the memory map. The least significant bits are stored | ||
| 802 | * in EMAC_HSL and the most significant bits in EMAC_HSH. | ||
| 803 | * | ||
| 804 | * The unicast hash enable and the multicast hash enable bits in the | ||
| 805 | * network configuration register enable the reception of hash matched | ||
| 806 | * frames. The destination address is reduced to a 6 bit index into | ||
| 807 | * the 64 bit hash register using the following hash function. The | ||
| 808 | * hash function is an exclusive or of every sixth bit of the | ||
| 809 | * destination address. | ||
| 810 | * | ||
| 811 | * hi[5] = da[5] ^ da[11] ^ da[17] ^ da[23] ^ da[29] ^ da[35] ^ da[41] ^ da[47] | ||
| 812 | * hi[4] = da[4] ^ da[10] ^ da[16] ^ da[22] ^ da[28] ^ da[34] ^ da[40] ^ da[46] | ||
| 813 | * hi[3] = da[3] ^ da[09] ^ da[15] ^ da[21] ^ da[27] ^ da[33] ^ da[39] ^ da[45] | ||
| 814 | * hi[2] = da[2] ^ da[08] ^ da[14] ^ da[20] ^ da[26] ^ da[32] ^ da[38] ^ da[44] | ||
| 815 | * hi[1] = da[1] ^ da[07] ^ da[13] ^ da[19] ^ da[25] ^ da[31] ^ da[37] ^ da[43] | ||
| 816 | * hi[0] = da[0] ^ da[06] ^ da[12] ^ da[18] ^ da[24] ^ da[30] ^ da[36] ^ da[42] | ||
| 817 | * | ||
| 818 | * da[0] represents the least significant bit of the first byte | ||
| 819 | * received, that is, the multicast/unicast indicator, and da[47] | ||
| 820 | * represents the most significant bit of the last byte received. If | ||
| 821 | * the hash index, hi[n], points to a bit that is set in the hash | ||
| 822 | * register then the frame will be matched according to whether the | ||
| 823 | * frame is multicast or unicast. A multicast match will be signalled | ||
| 824 | * if the multicast hash enable bit is set, da[0] is 1 and the hash | ||
| 825 | * index points to a bit set in the hash register. A unicast match | ||
| 826 | * will be signalled if the unicast hash enable bit is set, da[0] is 0 | ||
| 827 | * and the hash index points to a bit set in the hash register. To | ||
| 828 | * receive all multicast frames, the hash register should be set with | ||
| 829 | * all ones and the multicast hash enable bit should be set in the | ||
| 830 | * network configuration register. | ||
| 831 | */ | ||
| 832 | |||
| 833 | static inline int hash_bit_value(int bitnr, __u8 *addr) | ||
| 782 | { | 834 | { |
| 835 | if (addr[bitnr / 8] & (1 << (bitnr % 8))) | ||
| 836 | return 1; | ||
| 837 | return 0; | ||
| 838 | } | ||
| 839 | |||
| 840 | /* | ||
| 841 | * Return the hash index value for the specified address. | ||
| 842 | */ | ||
| 843 | static int hash_get_index(__u8 *addr) | ||
| 844 | { | ||
| 845 | int i, j, bitval; | ||
| 846 | int hash_index = 0; | ||
| 847 | |||
| 848 | for (j = 0; j < 6; j++) { | ||
| 849 | for (i = 0, bitval = 0; i < 8; i++) | ||
| 850 | bitval ^= hash_bit_value(i*6 + j, addr); | ||
| 851 | |||
| 852 | hash_index |= (bitval << j); | ||
| 853 | } | ||
| 854 | |||
| 855 | return hash_index; | ||
| 856 | } | ||
| 857 | |||
| 858 | /* | ||
| 859 | * Add multicast addresses to the internal multicast-hash table. | ||
| 860 | */ | ||
| 861 | static void macb_sethashtable(struct net_device *dev) | ||
| 862 | { | ||
| 863 | struct dev_mc_list *curr; | ||
| 864 | unsigned long mc_filter[2]; | ||
| 865 | unsigned int i, bitnr; | ||
| 866 | struct macb *bp = netdev_priv(dev); | ||
| 867 | |||
| 868 | mc_filter[0] = mc_filter[1] = 0; | ||
| 869 | |||
| 870 | curr = dev->mc_list; | ||
| 871 | for (i = 0; i < dev->mc_count; i++, curr = curr->next) { | ||
| 872 | if (!curr) break; /* unexpected end of list */ | ||
| 873 | |||
| 874 | bitnr = hash_get_index(curr->dmi_addr); | ||
| 875 | mc_filter[bitnr >> 5] |= 1 << (bitnr & 31); | ||
| 876 | } | ||
| 877 | |||
| 878 | macb_writel(bp, HRB, mc_filter[0]); | ||
| 879 | macb_writel(bp, HRT, mc_filter[1]); | ||
| 880 | } | ||
| 881 | |||
| 882 | /* | ||
| 883 | * Enable/Disable promiscuous and multicast modes. | ||
| 884 | */ | ||
| 885 | static void macb_set_rx_mode(struct net_device *dev) | ||
| 886 | { | ||
| 887 | unsigned long cfg; | ||
| 783 | struct macb *bp = netdev_priv(dev); | 888 | struct macb *bp = netdev_priv(dev); |
| 784 | 889 | ||
| 785 | /* Set some reasonable default settings */ | 890 | cfg = macb_readl(bp, NCFGR); |
| 786 | macb_mdio_write(dev, bp->mii.phy_id, MII_ADVERTISE, | 891 | |
| 787 | ADVERTISE_CSMA | ADVERTISE_ALL); | 892 | if (dev->flags & IFF_PROMISC) |
| 788 | macb_mdio_write(dev, bp->mii.phy_id, MII_BMCR, | 893 | /* Enable promiscuous mode */ |
| 789 | (BMCR_SPEED100 | BMCR_ANENABLE | 894 | cfg |= MACB_BIT(CAF); |
| 790 | | BMCR_ANRESTART | BMCR_FULLDPLX)); | 895 | else if (dev->flags & (~IFF_PROMISC)) |
| 896 | /* Disable promiscuous mode */ | ||
| 897 | cfg &= ~MACB_BIT(CAF); | ||
| 898 | |||
| 899 | if (dev->flags & IFF_ALLMULTI) { | ||
| 900 | /* Enable all multicast mode */ | ||
| 901 | macb_writel(bp, HRB, -1); | ||
| 902 | macb_writel(bp, HRT, -1); | ||
| 903 | cfg |= MACB_BIT(NCFGR_MTI); | ||
| 904 | } else if (dev->mc_count > 0) { | ||
| 905 | /* Enable specific multicasts */ | ||
| 906 | macb_sethashtable(dev); | ||
| 907 | cfg |= MACB_BIT(NCFGR_MTI); | ||
| 908 | } else if (dev->flags & (~IFF_ALLMULTI)) { | ||
| 909 | /* Disable all multicast mode */ | ||
| 910 | macb_writel(bp, HRB, 0); | ||
| 911 | macb_writel(bp, HRT, 0); | ||
| 912 | cfg &= ~MACB_BIT(NCFGR_MTI); | ||
| 913 | } | ||
| 914 | |||
| 915 | macb_writel(bp, NCFGR, cfg); | ||
| 791 | } | 916 | } |
| 792 | 917 | ||
| 793 | static int macb_open(struct net_device *dev) | 918 | static int macb_open(struct net_device *dev) |
| @@ -797,6 +922,10 @@ static int macb_open(struct net_device *dev) | |||
| 797 | 922 | ||
| 798 | dev_dbg(&bp->pdev->dev, "open\n"); | 923 | dev_dbg(&bp->pdev->dev, "open\n"); |
| 799 | 924 | ||
| 925 | /* if the phy is not yet register, retry later*/ | ||
| 926 | if (!bp->phy_dev) | ||
| 927 | return -EAGAIN; | ||
| 928 | |||
| 800 | if (!is_valid_ether_addr(dev->dev_addr)) | 929 | if (!is_valid_ether_addr(dev->dev_addr)) |
| 801 | return -EADDRNOTAVAIL; | 930 | return -EADDRNOTAVAIL; |
| 802 | 931 | ||
| @@ -810,12 +939,11 @@ static int macb_open(struct net_device *dev) | |||
| 810 | 939 | ||
| 811 | macb_init_rings(bp); | 940 | macb_init_rings(bp); |
| 812 | macb_init_hw(bp); | 941 | macb_init_hw(bp); |
| 813 | macb_init_phy(dev); | ||
| 814 | 942 | ||
| 815 | macb_check_media(bp, 1, 1); | 943 | /* schedule a link state check */ |
| 816 | netif_start_queue(dev); | 944 | phy_start(bp->phy_dev); |
| 817 | 945 | ||
| 818 | schedule_delayed_work(&bp->periodic_task, HZ); | 946 | netif_start_queue(dev); |
| 819 | 947 | ||
| 820 | return 0; | 948 | return 0; |
| 821 | } | 949 | } |
| @@ -825,10 +953,11 @@ static int macb_close(struct net_device *dev) | |||
| 825 | struct macb *bp = netdev_priv(dev); | 953 | struct macb *bp = netdev_priv(dev); |
| 826 | unsigned long flags; | 954 | unsigned long flags; |
| 827 | 955 | ||
| 828 | cancel_rearming_delayed_work(&bp->periodic_task); | ||
| 829 | |||
| 830 | netif_stop_queue(dev); | 956 | netif_stop_queue(dev); |
| 831 | 957 | ||
| 958 | if (bp->phy_dev) | ||
| 959 | phy_stop(bp->phy_dev); | ||
| 960 | |||
| 832 | spin_lock_irqsave(&bp->lock, flags); | 961 | spin_lock_irqsave(&bp->lock, flags); |
| 833 | macb_reset_hw(bp); | 962 | macb_reset_hw(bp); |
| 834 | netif_carrier_off(dev); | 963 | netif_carrier_off(dev); |
| @@ -845,6 +974,9 @@ static struct net_device_stats *macb_get_stats(struct net_device *dev) | |||
| 845 | struct net_device_stats *nstat = &bp->stats; | 974 | struct net_device_stats *nstat = &bp->stats; |
| 846 | struct macb_stats *hwstat = &bp->hw_stats; | 975 | struct macb_stats *hwstat = &bp->hw_stats; |
| 847 | 976 | ||
| 977 | /* read stats from hardware */ | ||
| 978 | macb_update_stats(bp); | ||
| 979 | |||
| 848 | /* Convert HW stats into netdevice stats */ | 980 | /* Convert HW stats into netdevice stats */ |
| 849 | nstat->rx_errors = (hwstat->rx_fcs_errors + | 981 | nstat->rx_errors = (hwstat->rx_fcs_errors + |
| 850 | hwstat->rx_align_errors + | 982 | hwstat->rx_align_errors + |
| @@ -882,18 +1014,27 @@ static struct net_device_stats *macb_get_stats(struct net_device *dev) | |||
| 882 | static int macb_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) | 1014 | static int macb_get_settings(struct net_device *dev, struct ethtool_cmd *cmd) |
| 883 | { | 1015 | { |
| 884 | struct macb *bp = netdev_priv(dev); | 1016 | struct macb *bp = netdev_priv(dev); |
| 1017 | struct phy_device *phydev = bp->phy_dev; | ||
| 885 | 1018 | ||
| 886 | return mii_ethtool_gset(&bp->mii, cmd); | 1019 | if (!phydev) |
| 1020 | return -ENODEV; | ||
| 1021 | |||
| 1022 | return phy_ethtool_gset(phydev, cmd); | ||
| 887 | } | 1023 | } |
| 888 | 1024 | ||
| 889 | static int macb_set_settings(struct net_device *dev, struct ethtool_cmd *cmd) | 1025 | static int macb_set_settings(struct net_device *dev, struct ethtool_cmd *cmd) |
| 890 | { | 1026 | { |
| 891 | struct macb *bp = netdev_priv(dev); | 1027 | struct macb *bp = netdev_priv(dev); |
| 1028 | struct phy_device *phydev = bp->phy_dev; | ||
| 1029 | |||
| 1030 | if (!phydev) | ||
| 1031 | return -ENODEV; | ||
| 892 | 1032 | ||
| 893 | return mii_ethtool_sset(&bp->mii, cmd); | 1033 | return phy_ethtool_sset(phydev, cmd); |
| 894 | } | 1034 | } |
| 895 | 1035 | ||
| 896 | static void macb_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *info) | 1036 | static void macb_get_drvinfo(struct net_device *dev, |
| 1037 | struct ethtool_drvinfo *info) | ||
| 897 | { | 1038 | { |
| 898 | struct macb *bp = netdev_priv(dev); | 1039 | struct macb *bp = netdev_priv(dev); |
| 899 | 1040 | ||
| @@ -902,104 +1043,34 @@ static void macb_get_drvinfo(struct net_device *dev, struct ethtool_drvinfo *inf | |||
| 902 | strcpy(info->bus_info, bp->pdev->dev.bus_id); | 1043 | strcpy(info->bus_info, bp->pdev->dev.bus_id); |
| 903 | } | 1044 | } |
| 904 | 1045 | ||
| 905 | static int macb_nway_reset(struct net_device *dev) | ||
| 906 | { | ||
| 907 | struct macb *bp = netdev_priv(dev); | ||
| 908 | return mii_nway_restart(&bp->mii); | ||
| 909 | } | ||
| 910 | |||
| 911 | static struct ethtool_ops macb_ethtool_ops = { | 1046 | static struct ethtool_ops macb_ethtool_ops = { |
| 912 | .get_settings = macb_get_settings, | 1047 | .get_settings = macb_get_settings, |
| 913 | .set_settings = macb_set_settings, | 1048 | .set_settings = macb_set_settings, |
| 914 | .get_drvinfo = macb_get_drvinfo, | 1049 | .get_drvinfo = macb_get_drvinfo, |
| 915 | .nway_reset = macb_nway_reset, | ||
| 916 | .get_link = ethtool_op_get_link, | 1050 | .get_link = ethtool_op_get_link, |
| 917 | }; | 1051 | }; |
| 918 | 1052 | ||
| 919 | static int macb_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) | 1053 | static int macb_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) |
| 920 | { | 1054 | { |
| 921 | struct macb *bp = netdev_priv(dev); | 1055 | struct macb *bp = netdev_priv(dev); |
| 1056 | struct phy_device *phydev = bp->phy_dev; | ||
| 922 | 1057 | ||
| 923 | if (!netif_running(dev)) | 1058 | if (!netif_running(dev)) |
| 924 | return -EINVAL; | 1059 | return -EINVAL; |
| 925 | 1060 | ||
| 926 | return generic_mii_ioctl(&bp->mii, if_mii(rq), cmd, NULL); | 1061 | if (!phydev) |
| 927 | } | 1062 | return -ENODEV; |
| 928 | |||
| 929 | static ssize_t macb_mii_show(const struct device *_dev, char *buf, | ||
| 930 | unsigned long addr) | ||
| 931 | { | ||
| 932 | struct net_device *dev = to_net_dev(_dev); | ||
| 933 | struct macb *bp = netdev_priv(dev); | ||
| 934 | ssize_t ret = -EINVAL; | ||
| 935 | |||
| 936 | if (netif_running(dev)) { | ||
| 937 | int value; | ||
| 938 | value = macb_mdio_read(dev, bp->mii.phy_id, addr); | ||
| 939 | ret = sprintf(buf, "0x%04x\n", (uint16_t)value); | ||
| 940 | } | ||
| 941 | |||
| 942 | return ret; | ||
| 943 | } | ||
| 944 | |||
| 945 | #define MII_ENTRY(name, addr) \ | ||
| 946 | static ssize_t show_##name(struct device *_dev, \ | ||
| 947 | struct device_attribute *attr, \ | ||
| 948 | char *buf) \ | ||
| 949 | { \ | ||
| 950 | return macb_mii_show(_dev, buf, addr); \ | ||
| 951 | } \ | ||
| 952 | static DEVICE_ATTR(name, S_IRUGO, show_##name, NULL) | ||
| 953 | |||
| 954 | MII_ENTRY(bmcr, MII_BMCR); | ||
| 955 | MII_ENTRY(bmsr, MII_BMSR); | ||
| 956 | MII_ENTRY(physid1, MII_PHYSID1); | ||
| 957 | MII_ENTRY(physid2, MII_PHYSID2); | ||
| 958 | MII_ENTRY(advertise, MII_ADVERTISE); | ||
| 959 | MII_ENTRY(lpa, MII_LPA); | ||
| 960 | MII_ENTRY(expansion, MII_EXPANSION); | ||
| 961 | |||
| 962 | static struct attribute *macb_mii_attrs[] = { | ||
| 963 | &dev_attr_bmcr.attr, | ||
| 964 | &dev_attr_bmsr.attr, | ||
| 965 | &dev_attr_physid1.attr, | ||
| 966 | &dev_attr_physid2.attr, | ||
| 967 | &dev_attr_advertise.attr, | ||
| 968 | &dev_attr_lpa.attr, | ||
| 969 | &dev_attr_expansion.attr, | ||
| 970 | NULL, | ||
| 971 | }; | ||
| 972 | |||
| 973 | static struct attribute_group macb_mii_group = { | ||
| 974 | .name = "mii", | ||
| 975 | .attrs = macb_mii_attrs, | ||
| 976 | }; | ||
| 977 | |||
| 978 | static void macb_unregister_sysfs(struct net_device *net) | ||
| 979 | { | ||
| 980 | struct device *_dev = &net->dev; | ||
| 981 | 1063 | ||
| 982 | sysfs_remove_group(&_dev->kobj, &macb_mii_group); | 1064 | return phy_mii_ioctl(phydev, if_mii(rq), cmd); |
| 983 | } | 1065 | } |
| 984 | 1066 | ||
| 985 | static int macb_register_sysfs(struct net_device *net) | ||
| 986 | { | ||
| 987 | struct device *_dev = &net->dev; | ||
| 988 | int ret; | ||
| 989 | |||
| 990 | ret = sysfs_create_group(&_dev->kobj, &macb_mii_group); | ||
| 991 | if (ret) | ||
| 992 | printk(KERN_WARNING | ||
| 993 | "%s: sysfs mii attribute registration failed: %d\n", | ||
| 994 | net->name, ret); | ||
| 995 | return ret; | ||
| 996 | } | ||
| 997 | static int __devinit macb_probe(struct platform_device *pdev) | 1067 | static int __devinit macb_probe(struct platform_device *pdev) |
| 998 | { | 1068 | { |
| 999 | struct eth_platform_data *pdata; | 1069 | struct eth_platform_data *pdata; |
| 1000 | struct resource *regs; | 1070 | struct resource *regs; |
| 1001 | struct net_device *dev; | 1071 | struct net_device *dev; |
| 1002 | struct macb *bp; | 1072 | struct macb *bp; |
| 1073 | struct phy_device *phydev; | ||
| 1003 | unsigned long pclk_hz; | 1074 | unsigned long pclk_hz; |
| 1004 | u32 config; | 1075 | u32 config; |
| 1005 | int err = -ENXIO; | 1076 | int err = -ENXIO; |
| @@ -1073,6 +1144,7 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
| 1073 | dev->stop = macb_close; | 1144 | dev->stop = macb_close; |
| 1074 | dev->hard_start_xmit = macb_start_xmit; | 1145 | dev->hard_start_xmit = macb_start_xmit; |
| 1075 | dev->get_stats = macb_get_stats; | 1146 | dev->get_stats = macb_get_stats; |
| 1147 | dev->set_multicast_list = macb_set_rx_mode; | ||
| 1076 | dev->do_ioctl = macb_ioctl; | 1148 | dev->do_ioctl = macb_ioctl; |
| 1077 | dev->poll = macb_poll; | 1149 | dev->poll = macb_poll; |
| 1078 | dev->weight = 64; | 1150 | dev->weight = 64; |
| @@ -1080,10 +1152,6 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
| 1080 | 1152 | ||
| 1081 | dev->base_addr = regs->start; | 1153 | dev->base_addr = regs->start; |
| 1082 | 1154 | ||
| 1083 | INIT_DELAYED_WORK(&bp->periodic_task, macb_periodic_task); | ||
| 1084 | mutex_init(&bp->mdio_mutex); | ||
| 1085 | init_completion(&bp->mdio_complete); | ||
| 1086 | |||
| 1087 | /* Set MII management clock divider */ | 1155 | /* Set MII management clock divider */ |
| 1088 | pclk_hz = clk_get_rate(bp->pclk); | 1156 | pclk_hz = clk_get_rate(bp->pclk); |
| 1089 | if (pclk_hz <= 20000000) | 1157 | if (pclk_hz <= 20000000) |
| @@ -1096,20 +1164,9 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
| 1096 | config = MACB_BF(CLK, MACB_CLK_DIV64); | 1164 | config = MACB_BF(CLK, MACB_CLK_DIV64); |
| 1097 | macb_writel(bp, NCFGR, config); | 1165 | macb_writel(bp, NCFGR, config); |
| 1098 | 1166 | ||
| 1099 | bp->mii.dev = dev; | ||
| 1100 | bp->mii.mdio_read = macb_mdio_read; | ||
| 1101 | bp->mii.mdio_write = macb_mdio_write; | ||
| 1102 | bp->mii.phy_id_mask = 0x1f; | ||
| 1103 | bp->mii.reg_num_mask = 0x1f; | ||
| 1104 | |||
| 1105 | macb_get_hwaddr(bp); | 1167 | macb_get_hwaddr(bp); |
| 1106 | err = macb_phy_probe(bp); | ||
| 1107 | if (err) { | ||
| 1108 | dev_err(&pdev->dev, "Failed to detect PHY, aborting.\n"); | ||
| 1109 | goto err_out_free_irq; | ||
| 1110 | } | ||
| 1111 | |||
| 1112 | pdata = pdev->dev.platform_data; | 1168 | pdata = pdev->dev.platform_data; |
| 1169 | |||
| 1113 | if (pdata && pdata->is_rmii) | 1170 | if (pdata && pdata->is_rmii) |
| 1114 | #if defined(CONFIG_ARCH_AT91) | 1171 | #if defined(CONFIG_ARCH_AT91) |
| 1115 | macb_writel(bp, USRIO, (MACB_BIT(RMII) | MACB_BIT(CLKEN)) ); | 1172 | macb_writel(bp, USRIO, (MACB_BIT(RMII) | MACB_BIT(CLKEN)) ); |
| @@ -1131,9 +1188,11 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
| 1131 | goto err_out_free_irq; | 1188 | goto err_out_free_irq; |
| 1132 | } | 1189 | } |
| 1133 | 1190 | ||
| 1134 | platform_set_drvdata(pdev, dev); | 1191 | if (macb_mii_init(bp) != 0) { |
| 1192 | goto err_out_unregister_netdev; | ||
| 1193 | } | ||
| 1135 | 1194 | ||
| 1136 | macb_register_sysfs(dev); | 1195 | platform_set_drvdata(pdev, dev); |
| 1137 | 1196 | ||
| 1138 | printk(KERN_INFO "%s: Atmel MACB at 0x%08lx irq %d " | 1197 | printk(KERN_INFO "%s: Atmel MACB at 0x%08lx irq %d " |
| 1139 | "(%02x:%02x:%02x:%02x:%02x:%02x)\n", | 1198 | "(%02x:%02x:%02x:%02x:%02x:%02x)\n", |
| @@ -1141,8 +1200,15 @@ static int __devinit macb_probe(struct platform_device *pdev) | |||
| 1141 | dev->dev_addr[0], dev->dev_addr[1], dev->dev_addr[2], | 1200 | dev->dev_addr[0], dev->dev_addr[1], dev->dev_addr[2], |
| 1142 | dev->dev_addr[3], dev->dev_addr[4], dev->dev_addr[5]); | 1201 | dev->dev_addr[3], dev->dev_addr[4], dev->dev_addr[5]); |
| 1143 | 1202 | ||
| 1203 | phydev = bp->phy_dev; | ||
| 1204 | printk(KERN_INFO "%s: attached PHY driver [%s] " | ||
| 1205 | "(mii_bus:phy_addr=%s, irq=%d)\n", | ||
| 1206 | dev->name, phydev->drv->name, phydev->dev.bus_id, phydev->irq); | ||
| 1207 | |||
| 1144 | return 0; | 1208 | return 0; |
| 1145 | 1209 | ||
| 1210 | err_out_unregister_netdev: | ||
| 1211 | unregister_netdev(dev); | ||
| 1146 | err_out_free_irq: | 1212 | err_out_free_irq: |
| 1147 | free_irq(dev->irq, dev); | 1213 | free_irq(dev->irq, dev); |
| 1148 | err_out_iounmap: | 1214 | err_out_iounmap: |
| @@ -1153,7 +1219,9 @@ err_out_disable_clocks: | |||
| 1153 | clk_put(bp->hclk); | 1219 | clk_put(bp->hclk); |
| 1154 | #endif | 1220 | #endif |
| 1155 | clk_disable(bp->pclk); | 1221 | clk_disable(bp->pclk); |
| 1222 | #ifndef CONFIG_ARCH_AT91 | ||
| 1156 | err_out_put_pclk: | 1223 | err_out_put_pclk: |
| 1224 | #endif | ||
| 1157 | clk_put(bp->pclk); | 1225 | clk_put(bp->pclk); |
| 1158 | err_out_free_dev: | 1226 | err_out_free_dev: |
| 1159 | free_netdev(dev); | 1227 | free_netdev(dev); |
| @@ -1171,7 +1239,8 @@ static int __devexit macb_remove(struct platform_device *pdev) | |||
| 1171 | 1239 | ||
| 1172 | if (dev) { | 1240 | if (dev) { |
| 1173 | bp = netdev_priv(dev); | 1241 | bp = netdev_priv(dev); |
| 1174 | macb_unregister_sysfs(dev); | 1242 | mdiobus_unregister(&bp->mii_bus); |
| 1243 | kfree(bp->mii_bus.irq); | ||
| 1175 | unregister_netdev(dev); | 1244 | unregister_netdev(dev); |
| 1176 | free_irq(dev->irq, dev); | 1245 | free_irq(dev->irq, dev); |
| 1177 | iounmap(bp->regs); | 1246 | iounmap(bp->regs); |
diff --git a/drivers/net/macb.h b/drivers/net/macb.h index b3bb2182edd1..4e3283ebd97c 100644 --- a/drivers/net/macb.h +++ b/drivers/net/macb.h | |||
| @@ -383,11 +383,11 @@ struct macb { | |||
| 383 | 383 | ||
| 384 | unsigned int rx_pending, tx_pending; | 384 | unsigned int rx_pending, tx_pending; |
| 385 | 385 | ||
| 386 | struct delayed_work periodic_task; | 386 | struct mii_bus mii_bus; |
| 387 | 387 | struct phy_device *phy_dev; | |
| 388 | struct mutex mdio_mutex; | 388 | unsigned int link; |
| 389 | struct completion mdio_complete; | 389 | unsigned int speed; |
| 390 | struct mii_if_info mii; | 390 | unsigned int duplex; |
| 391 | }; | 391 | }; |
| 392 | 392 | ||
| 393 | #endif /* _MACB_H */ | 393 | #endif /* _MACB_H */ |
diff --git a/drivers/net/myri10ge/myri10ge.c b/drivers/net/myri10ge/myri10ge.c index e1732c164a40..deca65330b0f 100644 --- a/drivers/net/myri10ge/myri10ge.c +++ b/drivers/net/myri10ge/myri10ge.c | |||
| @@ -1060,7 +1060,6 @@ static inline void myri10ge_tx_done(struct myri10ge_priv *mgp, int mcp_index) | |||
| 1060 | struct myri10ge_tx_buf *tx = &mgp->tx; | 1060 | struct myri10ge_tx_buf *tx = &mgp->tx; |
| 1061 | struct sk_buff *skb; | 1061 | struct sk_buff *skb; |
| 1062 | int idx, len; | 1062 | int idx, len; |
| 1063 | int limit = 0; | ||
| 1064 | 1063 | ||
| 1065 | while (tx->pkt_done != mcp_index) { | 1064 | while (tx->pkt_done != mcp_index) { |
| 1066 | idx = tx->done & tx->mask; | 1065 | idx = tx->done & tx->mask; |
| @@ -1091,11 +1090,6 @@ static inline void myri10ge_tx_done(struct myri10ge_priv *mgp, int mcp_index) | |||
| 1091 | bus), len, | 1090 | bus), len, |
| 1092 | PCI_DMA_TODEVICE); | 1091 | PCI_DMA_TODEVICE); |
| 1093 | } | 1092 | } |
| 1094 | |||
| 1095 | /* limit potential for livelock by only handling | ||
| 1096 | * 2 full tx rings per call */ | ||
| 1097 | if (unlikely(++limit > 2 * tx->mask)) | ||
| 1098 | break; | ||
| 1099 | } | 1093 | } |
| 1100 | /* start the queue if we've stopped it */ | 1094 | /* start the queue if we've stopped it */ |
| 1101 | if (netif_queue_stopped(mgp->dev) | 1095 | if (netif_queue_stopped(mgp->dev) |
diff --git a/drivers/net/s2io.c b/drivers/net/s2io.c index 58bbfdd4f901..afef6c0c59fe 100644 --- a/drivers/net/s2io.c +++ b/drivers/net/s2io.c | |||
| @@ -796,12 +796,14 @@ static void free_shared_mem(struct s2io_nic *nic) | |||
| 796 | struct mac_info *mac_control; | 796 | struct mac_info *mac_control; |
| 797 | struct config_param *config; | 797 | struct config_param *config; |
| 798 | int lst_size, lst_per_page; | 798 | int lst_size, lst_per_page; |
| 799 | struct net_device *dev = nic->dev; | 799 | struct net_device *dev; |
| 800 | int page_num = 0; | 800 | int page_num = 0; |
| 801 | 801 | ||
| 802 | if (!nic) | 802 | if (!nic) |
| 803 | return; | 803 | return; |
| 804 | 804 | ||
| 805 | dev = nic->dev; | ||
| 806 | |||
| 805 | mac_control = &nic->mac_control; | 807 | mac_control = &nic->mac_control; |
| 806 | config = &nic->config; | 808 | config = &nic->config; |
| 807 | 809 | ||
diff --git a/drivers/net/usb/cdc_subset.c b/drivers/net/usb/cdc_subset.c index bc62b012602b..943988ed01d8 100644 --- a/drivers/net/usb/cdc_subset.c +++ b/drivers/net/usb/cdc_subset.c | |||
| @@ -305,6 +305,9 @@ static const struct usb_device_id products [] = { | |||
| 305 | USB_DEVICE (0x8086, 0x07d3), // "blob" bootloader | 305 | USB_DEVICE (0x8086, 0x07d3), // "blob" bootloader |
| 306 | .driver_info = (unsigned long) &blob_info, | 306 | .driver_info = (unsigned long) &blob_info, |
| 307 | }, { | 307 | }, { |
| 308 | USB_DEVICE (0x1286, 0x8001), // "blob" bootloader | ||
| 309 | .driver_info = (unsigned long) &blob_info, | ||
| 310 | }, { | ||
| 308 | // Linux Ethernet/RNDIS gadget on pxa210/25x/26x, second config | 311 | // Linux Ethernet/RNDIS gadget on pxa210/25x/26x, second config |
| 309 | // e.g. Gumstix, current OpenZaurus, ... | 312 | // e.g. Gumstix, current OpenZaurus, ... |
| 310 | USB_DEVICE_VER (0x0525, 0xa4a2, 0x0203, 0x0203), | 313 | USB_DEVICE_VER (0x0525, 0xa4a2, 0x0203, 0x0203), |
diff --git a/drivers/net/wireless/airo.c b/drivers/net/wireless/airo.c index 2d3a180dada0..1c54908fdc4c 100644 --- a/drivers/net/wireless/airo.c +++ b/drivers/net/wireless/airo.c | |||
| @@ -52,6 +52,8 @@ | |||
| 52 | 52 | ||
| 53 | #include "airo.h" | 53 | #include "airo.h" |
| 54 | 54 | ||
| 55 | #define DRV_NAME "airo" | ||
| 56 | |||
| 55 | #ifdef CONFIG_PCI | 57 | #ifdef CONFIG_PCI |
| 56 | static struct pci_device_id card_ids[] = { | 58 | static struct pci_device_id card_ids[] = { |
| 57 | { 0x14b9, 1, PCI_ANY_ID, PCI_ANY_ID, }, | 59 | { 0x14b9, 1, PCI_ANY_ID, PCI_ANY_ID, }, |
| @@ -71,7 +73,7 @@ static int airo_pci_suspend(struct pci_dev *pdev, pm_message_t state); | |||
| 71 | static int airo_pci_resume(struct pci_dev *pdev); | 73 | static int airo_pci_resume(struct pci_dev *pdev); |
| 72 | 74 | ||
| 73 | static struct pci_driver airo_driver = { | 75 | static struct pci_driver airo_driver = { |
| 74 | .name = "airo", | 76 | .name = DRV_NAME, |
| 75 | .id_table = card_ids, | 77 | .id_table = card_ids, |
| 76 | .probe = airo_pci_probe, | 78 | .probe = airo_pci_probe, |
| 77 | .remove = __devexit_p(airo_pci_remove), | 79 | .remove = __devexit_p(airo_pci_remove), |
| @@ -1092,7 +1094,7 @@ static int get_dec_u16( char *buffer, int *start, int limit ); | |||
| 1092 | static void OUT4500( struct airo_info *, u16 register, u16 value ); | 1094 | static void OUT4500( struct airo_info *, u16 register, u16 value ); |
| 1093 | static unsigned short IN4500( struct airo_info *, u16 register ); | 1095 | static unsigned short IN4500( struct airo_info *, u16 register ); |
| 1094 | static u16 setup_card(struct airo_info*, u8 *mac, int lock); | 1096 | static u16 setup_card(struct airo_info*, u8 *mac, int lock); |
| 1095 | static int enable_MAC( struct airo_info *ai, Resp *rsp, int lock ); | 1097 | static int enable_MAC(struct airo_info *ai, int lock); |
| 1096 | static void disable_MAC(struct airo_info *ai, int lock); | 1098 | static void disable_MAC(struct airo_info *ai, int lock); |
| 1097 | static void enable_interrupts(struct airo_info*); | 1099 | static void enable_interrupts(struct airo_info*); |
| 1098 | static void disable_interrupts(struct airo_info*); | 1100 | static void disable_interrupts(struct airo_info*); |
| @@ -1250,7 +1252,7 @@ static int flashputbuf(struct airo_info *ai); | |||
| 1250 | static int flashrestart(struct airo_info *ai,struct net_device *dev); | 1252 | static int flashrestart(struct airo_info *ai,struct net_device *dev); |
| 1251 | 1253 | ||
| 1252 | #define airo_print(type, name, fmt, args...) \ | 1254 | #define airo_print(type, name, fmt, args...) \ |
| 1253 | { printk(type "airo(%s): " fmt "\n", name, ##args); } | 1255 | printk(type DRV_NAME "(%s): " fmt "\n", name, ##args) |
| 1254 | 1256 | ||
| 1255 | #define airo_print_info(name, fmt, args...) \ | 1257 | #define airo_print_info(name, fmt, args...) \ |
| 1256 | airo_print(KERN_INFO, name, fmt, ##args) | 1258 | airo_print(KERN_INFO, name, fmt, ##args) |
| @@ -1926,28 +1928,54 @@ static int readStatsRid(struct airo_info*ai, StatsRid *sr, int rid, int lock) { | |||
| 1926 | return rc; | 1928 | return rc; |
| 1927 | } | 1929 | } |
| 1928 | 1930 | ||
| 1931 | static void try_auto_wep(struct airo_info *ai) | ||
| 1932 | { | ||
| 1933 | if (auto_wep && !(ai->flags & FLAG_RADIO_DOWN)) { | ||
| 1934 | ai->expires = RUN_AT(3*HZ); | ||
| 1935 | wake_up_interruptible(&ai->thr_wait); | ||
| 1936 | } | ||
| 1937 | } | ||
| 1938 | |||
| 1929 | static int airo_open(struct net_device *dev) { | 1939 | static int airo_open(struct net_device *dev) { |
| 1930 | struct airo_info *info = dev->priv; | 1940 | struct airo_info *ai = dev->priv; |
| 1931 | Resp rsp; | 1941 | int rc = 0; |
| 1932 | 1942 | ||
| 1933 | if (test_bit(FLAG_FLASHING, &info->flags)) | 1943 | if (test_bit(FLAG_FLASHING, &ai->flags)) |
| 1934 | return -EIO; | 1944 | return -EIO; |
| 1935 | 1945 | ||
| 1936 | /* Make sure the card is configured. | 1946 | /* Make sure the card is configured. |
| 1937 | * Wireless Extensions may postpone config changes until the card | 1947 | * Wireless Extensions may postpone config changes until the card |
| 1938 | * is open (to pipeline changes and speed-up card setup). If | 1948 | * is open (to pipeline changes and speed-up card setup). If |
| 1939 | * those changes are not yet commited, do it now - Jean II */ | 1949 | * those changes are not yet commited, do it now - Jean II */ |
| 1940 | if (test_bit (FLAG_COMMIT, &info->flags)) { | 1950 | if (test_bit(FLAG_COMMIT, &ai->flags)) { |
| 1941 | disable_MAC(info, 1); | 1951 | disable_MAC(ai, 1); |
| 1942 | writeConfigRid(info, 1); | 1952 | writeConfigRid(ai, 1); |
| 1943 | } | 1953 | } |
| 1944 | 1954 | ||
| 1945 | if (info->wifidev != dev) { | 1955 | if (ai->wifidev != dev) { |
| 1956 | clear_bit(JOB_DIE, &ai->jobs); | ||
| 1957 | ai->airo_thread_task = kthread_run(airo_thread, dev, dev->name); | ||
| 1958 | if (IS_ERR(ai->airo_thread_task)) | ||
| 1959 | return (int)PTR_ERR(ai->airo_thread_task); | ||
| 1960 | |||
| 1961 | rc = request_irq(dev->irq, airo_interrupt, IRQF_SHARED, | ||
| 1962 | dev->name, dev); | ||
| 1963 | if (rc) { | ||
| 1964 | airo_print_err(dev->name, | ||
| 1965 | "register interrupt %d failed, rc %d", | ||
| 1966 | dev->irq, rc); | ||
| 1967 | set_bit(JOB_DIE, &ai->jobs); | ||
| 1968 | kthread_stop(ai->airo_thread_task); | ||
| 1969 | return rc; | ||
| 1970 | } | ||
| 1971 | |||
| 1946 | /* Power on the MAC controller (which may have been disabled) */ | 1972 | /* Power on the MAC controller (which may have been disabled) */ |
| 1947 | clear_bit(FLAG_RADIO_DOWN, &info->flags); | 1973 | clear_bit(FLAG_RADIO_DOWN, &ai->flags); |
| 1948 | enable_interrupts(info); | 1974 | enable_interrupts(ai); |
| 1975 | |||
| 1976 | try_auto_wep(ai); | ||
| 1949 | } | 1977 | } |
| 1950 | enable_MAC(info, &rsp, 1); | 1978 | enable_MAC(ai, 1); |
| 1951 | 1979 | ||
| 1952 | netif_start_queue(dev); | 1980 | netif_start_queue(dev); |
| 1953 | return 0; | 1981 | return 0; |
| @@ -2338,14 +2366,13 @@ static int airo_set_mac_address(struct net_device *dev, void *p) | |||
| 2338 | { | 2366 | { |
| 2339 | struct airo_info *ai = dev->priv; | 2367 | struct airo_info *ai = dev->priv; |
| 2340 | struct sockaddr *addr = p; | 2368 | struct sockaddr *addr = p; |
| 2341 | Resp rsp; | ||
| 2342 | 2369 | ||
| 2343 | readConfigRid(ai, 1); | 2370 | readConfigRid(ai, 1); |
| 2344 | memcpy (ai->config.macAddr, addr->sa_data, dev->addr_len); | 2371 | memcpy (ai->config.macAddr, addr->sa_data, dev->addr_len); |
| 2345 | set_bit (FLAG_COMMIT, &ai->flags); | 2372 | set_bit (FLAG_COMMIT, &ai->flags); |
| 2346 | disable_MAC(ai, 1); | 2373 | disable_MAC(ai, 1); |
| 2347 | writeConfigRid (ai, 1); | 2374 | writeConfigRid (ai, 1); |
| 2348 | enable_MAC(ai, &rsp, 1); | 2375 | enable_MAC(ai, 1); |
| 2349 | memcpy (ai->dev->dev_addr, addr->sa_data, dev->addr_len); | 2376 | memcpy (ai->dev->dev_addr, addr->sa_data, dev->addr_len); |
| 2350 | if (ai->wifidev) | 2377 | if (ai->wifidev) |
| 2351 | memcpy (ai->wifidev->dev_addr, addr->sa_data, dev->addr_len); | 2378 | memcpy (ai->wifidev->dev_addr, addr->sa_data, dev->addr_len); |
| @@ -2392,6 +2419,11 @@ static int airo_close(struct net_device *dev) { | |||
| 2392 | disable_MAC(ai, 1); | 2419 | disable_MAC(ai, 1); |
| 2393 | #endif | 2420 | #endif |
| 2394 | disable_interrupts( ai ); | 2421 | disable_interrupts( ai ); |
| 2422 | |||
| 2423 | free_irq(dev->irq, dev); | ||
| 2424 | |||
| 2425 | set_bit(JOB_DIE, &ai->jobs); | ||
| 2426 | kthread_stop(ai->airo_thread_task); | ||
| 2395 | } | 2427 | } |
| 2396 | return 0; | 2428 | return 0; |
| 2397 | } | 2429 | } |
| @@ -2403,7 +2435,6 @@ void stop_airo_card( struct net_device *dev, int freeres ) | |||
| 2403 | set_bit(FLAG_RADIO_DOWN, &ai->flags); | 2435 | set_bit(FLAG_RADIO_DOWN, &ai->flags); |
| 2404 | disable_MAC(ai, 1); | 2436 | disable_MAC(ai, 1); |
| 2405 | disable_interrupts(ai); | 2437 | disable_interrupts(ai); |
| 2406 | free_irq( dev->irq, dev ); | ||
| 2407 | takedown_proc_entry( dev, ai ); | 2438 | takedown_proc_entry( dev, ai ); |
| 2408 | if (test_bit(FLAG_REGISTERED, &ai->flags)) { | 2439 | if (test_bit(FLAG_REGISTERED, &ai->flags)) { |
| 2409 | unregister_netdev( dev ); | 2440 | unregister_netdev( dev ); |
| @@ -2414,9 +2445,6 @@ void stop_airo_card( struct net_device *dev, int freeres ) | |||
| 2414 | } | 2445 | } |
| 2415 | clear_bit(FLAG_REGISTERED, &ai->flags); | 2446 | clear_bit(FLAG_REGISTERED, &ai->flags); |
| 2416 | } | 2447 | } |
| 2417 | set_bit(JOB_DIE, &ai->jobs); | ||
| 2418 | kthread_stop(ai->airo_thread_task); | ||
| 2419 | |||
| 2420 | /* | 2448 | /* |
| 2421 | * Clean out tx queue | 2449 | * Clean out tx queue |
| 2422 | */ | 2450 | */ |
| @@ -2554,8 +2582,7 @@ static int mpi_init_descriptors (struct airo_info *ai) | |||
| 2554 | * 2) Map PCI memory for issueing commands. | 2582 | * 2) Map PCI memory for issueing commands. |
| 2555 | * 3) Allocate memory (shared) to send and receive ethernet frames. | 2583 | * 3) Allocate memory (shared) to send and receive ethernet frames. |
| 2556 | */ | 2584 | */ |
| 2557 | static int mpi_map_card(struct airo_info *ai, struct pci_dev *pci, | 2585 | static int mpi_map_card(struct airo_info *ai, struct pci_dev *pci) |
| 2558 | const char *name) | ||
| 2559 | { | 2586 | { |
| 2560 | unsigned long mem_start, mem_len, aux_start, aux_len; | 2587 | unsigned long mem_start, mem_len, aux_start, aux_len; |
| 2561 | int rc = -1; | 2588 | int rc = -1; |
| @@ -2569,35 +2596,35 @@ static int mpi_map_card(struct airo_info *ai, struct pci_dev *pci, | |||
| 2569 | aux_start = pci_resource_start(pci, 2); | 2596 | aux_start = pci_resource_start(pci, 2); |
| 2570 | aux_len = AUXMEMSIZE; | 2597 | aux_len = AUXMEMSIZE; |
| 2571 | 2598 | ||
| 2572 | if (!request_mem_region(mem_start, mem_len, name)) { | 2599 | if (!request_mem_region(mem_start, mem_len, DRV_NAME)) { |
| 2573 | airo_print_err(ai->dev->name, "Couldn't get region %x[%x] for %s", | 2600 | airo_print_err("", "Couldn't get region %x[%x]", |
| 2574 | (int)mem_start, (int)mem_len, name); | 2601 | (int)mem_start, (int)mem_len); |
| 2575 | goto out; | 2602 | goto out; |
| 2576 | } | 2603 | } |
| 2577 | if (!request_mem_region(aux_start, aux_len, name)) { | 2604 | if (!request_mem_region(aux_start, aux_len, DRV_NAME)) { |
| 2578 | airo_print_err(ai->dev->name, "Couldn't get region %x[%x] for %s", | 2605 | airo_print_err("", "Couldn't get region %x[%x]", |
| 2579 | (int)aux_start, (int)aux_len, name); | 2606 | (int)aux_start, (int)aux_len); |
| 2580 | goto free_region1; | 2607 | goto free_region1; |
| 2581 | } | 2608 | } |
| 2582 | 2609 | ||
| 2583 | ai->pcimem = ioremap(mem_start, mem_len); | 2610 | ai->pcimem = ioremap(mem_start, mem_len); |
| 2584 | if (!ai->pcimem) { | 2611 | if (!ai->pcimem) { |
| 2585 | airo_print_err(ai->dev->name, "Couldn't map region %x[%x] for %s", | 2612 | airo_print_err("", "Couldn't map region %x[%x]", |
| 2586 | (int)mem_start, (int)mem_len, name); | 2613 | (int)mem_start, (int)mem_len); |
| 2587 | goto free_region2; | 2614 | goto free_region2; |
| 2588 | } | 2615 | } |
| 2589 | ai->pciaux = ioremap(aux_start, aux_len); | 2616 | ai->pciaux = ioremap(aux_start, aux_len); |
| 2590 | if (!ai->pciaux) { | 2617 | if (!ai->pciaux) { |
| 2591 | airo_print_err(ai->dev->name, "Couldn't map region %x[%x] for %s", | 2618 | airo_print_err("", "Couldn't map region %x[%x]", |
| 2592 | (int)aux_start, (int)aux_len, name); | 2619 | (int)aux_start, (int)aux_len); |
| 2593 | goto free_memmap; | 2620 | goto free_memmap; |
| 2594 | } | 2621 | } |
| 2595 | 2622 | ||
| 2596 | /* Reserve PKTSIZE for each fid and 2K for the Rids */ | 2623 | /* Reserve PKTSIZE for each fid and 2K for the Rids */ |
| 2597 | ai->shared = pci_alloc_consistent(pci, PCI_SHARED_LEN, &ai->shared_dma); | 2624 | ai->shared = pci_alloc_consistent(pci, PCI_SHARED_LEN, &ai->shared_dma); |
| 2598 | if (!ai->shared) { | 2625 | if (!ai->shared) { |
| 2599 | airo_print_err(ai->dev->name, "Couldn't alloc_consistent %d", | 2626 | airo_print_err("", "Couldn't alloc_consistent %d", |
| 2600 | PCI_SHARED_LEN); | 2627 | PCI_SHARED_LEN); |
| 2601 | goto free_auxmap; | 2628 | goto free_auxmap; |
| 2602 | } | 2629 | } |
| 2603 | 2630 | ||
| @@ -2742,7 +2769,7 @@ static int airo_networks_allocate(struct airo_info *ai) | |||
| 2742 | kzalloc(AIRO_MAX_NETWORK_COUNT * sizeof(BSSListElement), | 2769 | kzalloc(AIRO_MAX_NETWORK_COUNT * sizeof(BSSListElement), |
| 2743 | GFP_KERNEL); | 2770 | GFP_KERNEL); |
| 2744 | if (!ai->networks) { | 2771 | if (!ai->networks) { |
| 2745 | airo_print_warn(ai->dev->name, "Out of memory allocating beacons"); | 2772 | airo_print_warn("", "Out of memory allocating beacons"); |
| 2746 | return -ENOMEM; | 2773 | return -ENOMEM; |
| 2747 | } | 2774 | } |
| 2748 | 2775 | ||
| @@ -2770,7 +2797,6 @@ static int airo_test_wpa_capable(struct airo_info *ai) | |||
| 2770 | { | 2797 | { |
| 2771 | int status; | 2798 | int status; |
| 2772 | CapabilityRid cap_rid; | 2799 | CapabilityRid cap_rid; |
| 2773 | const char *name = ai->dev->name; | ||
| 2774 | 2800 | ||
| 2775 | status = readCapabilityRid(ai, &cap_rid, 1); | 2801 | status = readCapabilityRid(ai, &cap_rid, 1); |
| 2776 | if (status != SUCCESS) return 0; | 2802 | if (status != SUCCESS) return 0; |
| @@ -2778,12 +2804,12 @@ static int airo_test_wpa_capable(struct airo_info *ai) | |||
| 2778 | /* Only firmware versions 5.30.17 or better can do WPA */ | 2804 | /* Only firmware versions 5.30.17 or better can do WPA */ |
| 2779 | if ((cap_rid.softVer > 0x530) | 2805 | if ((cap_rid.softVer > 0x530) |
| 2780 | || ((cap_rid.softVer == 0x530) && (cap_rid.softSubVer >= 17))) { | 2806 | || ((cap_rid.softVer == 0x530) && (cap_rid.softSubVer >= 17))) { |
| 2781 | airo_print_info(name, "WPA is supported."); | 2807 | airo_print_info("", "WPA is supported."); |
| 2782 | return 1; | 2808 | return 1; |
| 2783 | } | 2809 | } |
| 2784 | 2810 | ||
| 2785 | /* No WPA support */ | 2811 | /* No WPA support */ |
| 2786 | airo_print_info(name, "WPA unsupported (only firmware versions 5.30.17" | 2812 | airo_print_info("", "WPA unsupported (only firmware versions 5.30.17" |
| 2787 | " and greater support WPA. Detected %s)", cap_rid.prodVer); | 2813 | " and greater support WPA. Detected %s)", cap_rid.prodVer); |
| 2788 | return 0; | 2814 | return 0; |
| 2789 | } | 2815 | } |
| @@ -2797,23 +2823,19 @@ static struct net_device *_init_airo_card( unsigned short irq, int port, | |||
| 2797 | int i, rc; | 2823 | int i, rc; |
| 2798 | 2824 | ||
| 2799 | /* Create the network device object. */ | 2825 | /* Create the network device object. */ |
| 2800 | dev = alloc_etherdev(sizeof(*ai)); | 2826 | dev = alloc_netdev(sizeof(*ai), "", ether_setup); |
| 2801 | if (!dev) { | 2827 | if (!dev) { |
| 2802 | airo_print_err("", "Couldn't alloc_etherdev"); | 2828 | airo_print_err("", "Couldn't alloc_etherdev"); |
| 2803 | return NULL; | 2829 | return NULL; |
| 2804 | } | ||
| 2805 | if (dev_alloc_name(dev, dev->name) < 0) { | ||
| 2806 | airo_print_err("", "Couldn't get name!"); | ||
| 2807 | goto err_out_free; | ||
| 2808 | } | 2830 | } |
| 2809 | 2831 | ||
| 2810 | ai = dev->priv; | 2832 | ai = dev->priv; |
| 2811 | ai->wifidev = NULL; | 2833 | ai->wifidev = NULL; |
| 2812 | ai->flags = 0; | 2834 | ai->flags = 1 << FLAG_RADIO_DOWN; |
| 2813 | ai->jobs = 0; | 2835 | ai->jobs = 0; |
| 2814 | ai->dev = dev; | 2836 | ai->dev = dev; |
| 2815 | if (pci && (pci->device == 0x5000 || pci->device == 0xa504)) { | 2837 | if (pci && (pci->device == 0x5000 || pci->device == 0xa504)) { |
| 2816 | airo_print_dbg(dev->name, "Found an MPI350 card"); | 2838 | airo_print_dbg("", "Found an MPI350 card"); |
| 2817 | set_bit(FLAG_MPI, &ai->flags); | 2839 | set_bit(FLAG_MPI, &ai->flags); |
| 2818 | } | 2840 | } |
| 2819 | spin_lock_init(&ai->aux_lock); | 2841 | spin_lock_init(&ai->aux_lock); |
| @@ -2821,14 +2843,11 @@ static struct net_device *_init_airo_card( unsigned short irq, int port, | |||
| 2821 | ai->config.len = 0; | 2843 | ai->config.len = 0; |
| 2822 | ai->pci = pci; | 2844 | ai->pci = pci; |
| 2823 | init_waitqueue_head (&ai->thr_wait); | 2845 | init_waitqueue_head (&ai->thr_wait); |
| 2824 | ai->airo_thread_task = kthread_run(airo_thread, dev, dev->name); | ||
| 2825 | if (IS_ERR(ai->airo_thread_task)) | ||
| 2826 | goto err_out_free; | ||
| 2827 | ai->tfm = NULL; | 2846 | ai->tfm = NULL; |
| 2828 | add_airo_dev(ai); | 2847 | add_airo_dev(ai); |
| 2829 | 2848 | ||
| 2830 | if (airo_networks_allocate (ai)) | 2849 | if (airo_networks_allocate (ai)) |
| 2831 | goto err_out_thr; | 2850 | goto err_out_free; |
| 2832 | airo_networks_initialize (ai); | 2851 | airo_networks_initialize (ai); |
| 2833 | 2852 | ||
| 2834 | /* The Airo-specific entries in the device structure. */ | 2853 | /* The Airo-specific entries in the device structure. */ |
| @@ -2851,27 +2870,22 @@ static struct net_device *_init_airo_card( unsigned short irq, int port, | |||
| 2851 | dev->base_addr = port; | 2870 | dev->base_addr = port; |
| 2852 | 2871 | ||
| 2853 | SET_NETDEV_DEV(dev, dmdev); | 2872 | SET_NETDEV_DEV(dev, dmdev); |
| 2873 | SET_MODULE_OWNER(dev); | ||
| 2854 | 2874 | ||
| 2855 | reset_card (dev, 1); | 2875 | reset_card (dev, 1); |
| 2856 | msleep(400); | 2876 | msleep(400); |
| 2857 | 2877 | ||
| 2858 | rc = request_irq( dev->irq, airo_interrupt, IRQF_SHARED, dev->name, dev ); | ||
| 2859 | if (rc) { | ||
| 2860 | airo_print_err(dev->name, "register interrupt %d failed, rc %d", | ||
| 2861 | irq, rc); | ||
| 2862 | goto err_out_nets; | ||
| 2863 | } | ||
| 2864 | if (!is_pcmcia) { | 2878 | if (!is_pcmcia) { |
| 2865 | if (!request_region( dev->base_addr, 64, dev->name )) { | 2879 | if (!request_region(dev->base_addr, 64, DRV_NAME)) { |
| 2866 | rc = -EBUSY; | 2880 | rc = -EBUSY; |
| 2867 | airo_print_err(dev->name, "Couldn't request region"); | 2881 | airo_print_err(dev->name, "Couldn't request region"); |
| 2868 | goto err_out_irq; | 2882 | goto err_out_nets; |
| 2869 | } | 2883 | } |
| 2870 | } | 2884 | } |
| 2871 | 2885 | ||
| 2872 | if (test_bit(FLAG_MPI,&ai->flags)) { | 2886 | if (test_bit(FLAG_MPI,&ai->flags)) { |
| 2873 | if (mpi_map_card(ai, pci, dev->name)) { | 2887 | if (mpi_map_card(ai, pci)) { |
| 2874 | airo_print_err(dev->name, "Could not map memory"); | 2888 | airo_print_err("", "Could not map memory"); |
| 2875 | goto err_out_res; | 2889 | goto err_out_res; |
| 2876 | } | 2890 | } |
| 2877 | } | 2891 | } |
| @@ -2899,6 +2913,7 @@ static struct net_device *_init_airo_card( unsigned short irq, int port, | |||
| 2899 | ai->bssListRidLen = sizeof(BSSListRid) - sizeof(BSSListRidExtra); | 2913 | ai->bssListRidLen = sizeof(BSSListRid) - sizeof(BSSListRidExtra); |
| 2900 | } | 2914 | } |
| 2901 | 2915 | ||
| 2916 | strcpy(dev->name, "eth%d"); | ||
| 2902 | rc = register_netdev(dev); | 2917 | rc = register_netdev(dev); |
| 2903 | if (rc) { | 2918 | if (rc) { |
| 2904 | airo_print_err(dev->name, "Couldn't register_netdev"); | 2919 | airo_print_err(dev->name, "Couldn't register_netdev"); |
| @@ -2921,8 +2936,6 @@ static struct net_device *_init_airo_card( unsigned short irq, int port, | |||
| 2921 | if (setup_proc_entry(dev, dev->priv) < 0) | 2936 | if (setup_proc_entry(dev, dev->priv) < 0) |
| 2922 | goto err_out_wifi; | 2937 | goto err_out_wifi; |
| 2923 | 2938 | ||
| 2924 | netif_start_queue(dev); | ||
| 2925 | SET_MODULE_OWNER(dev); | ||
| 2926 | return dev; | 2939 | return dev; |
| 2927 | 2940 | ||
| 2928 | err_out_wifi: | 2941 | err_out_wifi: |
| @@ -2940,14 +2953,9 @@ err_out_map: | |||
| 2940 | err_out_res: | 2953 | err_out_res: |
| 2941 | if (!is_pcmcia) | 2954 | if (!is_pcmcia) |
| 2942 | release_region( dev->base_addr, 64 ); | 2955 | release_region( dev->base_addr, 64 ); |
| 2943 | err_out_irq: | ||
| 2944 | free_irq(dev->irq, dev); | ||
| 2945 | err_out_nets: | 2956 | err_out_nets: |
| 2946 | airo_networks_free(ai); | 2957 | airo_networks_free(ai); |
| 2947 | err_out_thr: | ||
| 2948 | del_airo_dev(ai); | 2958 | del_airo_dev(ai); |
| 2949 | set_bit(JOB_DIE, &ai->jobs); | ||
| 2950 | kthread_stop(ai->airo_thread_task); | ||
| 2951 | err_out_free: | 2959 | err_out_free: |
| 2952 | free_netdev(dev); | 2960 | free_netdev(dev); |
| 2953 | return NULL; | 2961 | return NULL; |
| @@ -3529,9 +3537,11 @@ static u16 IN4500( struct airo_info *ai, u16 reg ) { | |||
| 3529 | return rc; | 3537 | return rc; |
| 3530 | } | 3538 | } |
| 3531 | 3539 | ||
| 3532 | static int enable_MAC( struct airo_info *ai, Resp *rsp, int lock ) { | 3540 | static int enable_MAC(struct airo_info *ai, int lock) |
| 3541 | { | ||
| 3533 | int rc; | 3542 | int rc; |
| 3534 | Cmd cmd; | 3543 | Cmd cmd; |
| 3544 | Resp rsp; | ||
| 3535 | 3545 | ||
| 3536 | /* FLAG_RADIO_OFF : Radio disabled via /proc or Wireless Extensions | 3546 | /* FLAG_RADIO_OFF : Radio disabled via /proc or Wireless Extensions |
| 3537 | * FLAG_RADIO_DOWN : Radio disabled via "ifconfig ethX down" | 3547 | * FLAG_RADIO_DOWN : Radio disabled via "ifconfig ethX down" |
| @@ -3547,7 +3557,7 @@ static int enable_MAC( struct airo_info *ai, Resp *rsp, int lock ) { | |||
| 3547 | if (!test_bit(FLAG_ENABLED, &ai->flags)) { | 3557 | if (!test_bit(FLAG_ENABLED, &ai->flags)) { |
| 3548 | memset(&cmd, 0, sizeof(cmd)); | 3558 | memset(&cmd, 0, sizeof(cmd)); |
| 3549 | cmd.cmd = MAC_ENABLE; | 3559 | cmd.cmd = MAC_ENABLE; |
| 3550 | rc = issuecommand(ai, &cmd, rsp); | 3560 | rc = issuecommand(ai, &cmd, &rsp); |
| 3551 | if (rc == SUCCESS) | 3561 | if (rc == SUCCESS) |
| 3552 | set_bit(FLAG_ENABLED, &ai->flags); | 3562 | set_bit(FLAG_ENABLED, &ai->flags); |
| 3553 | } else | 3563 | } else |
| @@ -3557,8 +3567,12 @@ static int enable_MAC( struct airo_info *ai, Resp *rsp, int lock ) { | |||
| 3557 | up(&ai->sem); | 3567 | up(&ai->sem); |
| 3558 | 3568 | ||
| 3559 | if (rc) | 3569 | if (rc) |
| 3560 | airo_print_err(ai->dev->name, "%s: Cannot enable MAC, err=%d", | 3570 | airo_print_err(ai->dev->name, "Cannot enable MAC"); |
| 3561 | __FUNCTION__, rc); | 3571 | else if ((rsp.status & 0xFF00) != 0) { |
| 3572 | airo_print_err(ai->dev->name, "Bad MAC enable reason=%x, " | ||
| 3573 | "rid=%x, offset=%d", rsp.rsp0, rsp.rsp1, rsp.rsp2); | ||
| 3574 | rc = ERROR; | ||
| 3575 | } | ||
| 3562 | return rc; | 3576 | return rc; |
| 3563 | } | 3577 | } |
| 3564 | 3578 | ||
| @@ -3902,12 +3916,9 @@ static u16 setup_card(struct airo_info *ai, u8 *mac, int lock) | |||
| 3902 | if ( status != SUCCESS ) return ERROR; | 3916 | if ( status != SUCCESS ) return ERROR; |
| 3903 | } | 3917 | } |
| 3904 | 3918 | ||
| 3905 | status = enable_MAC(ai, &rsp, lock); | 3919 | status = enable_MAC(ai, lock); |
| 3906 | if ( status != SUCCESS || (rsp.status & 0xFF00) != 0) { | 3920 | if (status != SUCCESS) |
| 3907 | airo_print_err(ai->dev->name, "Bad MAC enable reason = %x, rid = %x," | ||
| 3908 | " offset = %d", rsp.rsp0, rsp.rsp1, rsp.rsp2 ); | ||
| 3909 | return ERROR; | 3921 | return ERROR; |
| 3910 | } | ||
| 3911 | 3922 | ||
| 3912 | /* Grab the initial wep key, we gotta save it for auto_wep */ | 3923 | /* Grab the initial wep key, we gotta save it for auto_wep */ |
| 3913 | rc = readWepKeyRid(ai, &wkr, 1, lock); | 3924 | rc = readWepKeyRid(ai, &wkr, 1, lock); |
| @@ -3919,10 +3930,7 @@ static u16 setup_card(struct airo_info *ai, u8 *mac, int lock) | |||
| 3919 | rc = readWepKeyRid(ai, &wkr, 0, lock); | 3930 | rc = readWepKeyRid(ai, &wkr, 0, lock); |
| 3920 | } while(lastindex != wkr.kindex); | 3931 | } while(lastindex != wkr.kindex); |
| 3921 | 3932 | ||
| 3922 | if (auto_wep) { | 3933 | try_auto_wep(ai); |
| 3923 | ai->expires = RUN_AT(3*HZ); | ||
| 3924 | wake_up_interruptible(&ai->thr_wait); | ||
| 3925 | } | ||
| 3926 | 3934 | ||
| 3927 | return SUCCESS; | 3935 | return SUCCESS; |
| 3928 | } | 3936 | } |
| @@ -4004,7 +4012,7 @@ static int bap_setup(struct airo_info *ai, u16 rid, u16 offset, int whichbap ) | |||
| 4004 | } | 4012 | } |
| 4005 | if ( !(max_tries--) ) { | 4013 | if ( !(max_tries--) ) { |
| 4006 | airo_print_err(ai->dev->name, | 4014 | airo_print_err(ai->dev->name, |
| 4007 | "airo: BAP setup error too many retries\n"); | 4015 | "BAP setup error too many retries\n"); |
| 4008 | return ERROR; | 4016 | return ERROR; |
| 4009 | } | 4017 | } |
| 4010 | // -- PC4500 missed it, try again | 4018 | // -- PC4500 missed it, try again |
| @@ -5152,7 +5160,6 @@ static void proc_SSID_on_close( struct inode *inode, struct file *file ) { | |||
| 5152 | struct net_device *dev = dp->data; | 5160 | struct net_device *dev = dp->data; |
| 5153 | struct airo_info *ai = dev->priv; | 5161 | struct airo_info *ai = dev->priv; |
| 5154 | SsidRid SSID_rid; | 5162 | SsidRid SSID_rid; |
| 5155 | Resp rsp; | ||
| 5156 | int i; | 5163 | int i; |
| 5157 | int offset = 0; | 5164 | int offset = 0; |
| 5158 | 5165 | ||
| @@ -5177,7 +5184,7 @@ static void proc_SSID_on_close( struct inode *inode, struct file *file ) { | |||
| 5177 | SSID_rid.len = sizeof(SSID_rid); | 5184 | SSID_rid.len = sizeof(SSID_rid); |
| 5178 | disable_MAC(ai, 1); | 5185 | disable_MAC(ai, 1); |
| 5179 | writeSsidRid(ai, &SSID_rid, 1); | 5186 | writeSsidRid(ai, &SSID_rid, 1); |
| 5180 | enable_MAC(ai, &rsp, 1); | 5187 | enable_MAC(ai, 1); |
| 5181 | } | 5188 | } |
| 5182 | 5189 | ||
| 5183 | static inline u8 hexVal(char c) { | 5190 | static inline u8 hexVal(char c) { |
| @@ -5193,7 +5200,6 @@ static void proc_APList_on_close( struct inode *inode, struct file *file ) { | |||
| 5193 | struct net_device *dev = dp->data; | 5200 | struct net_device *dev = dp->data; |
| 5194 | struct airo_info *ai = dev->priv; | 5201 | struct airo_info *ai = dev->priv; |
| 5195 | APListRid APList_rid; | 5202 | APListRid APList_rid; |
| 5196 | Resp rsp; | ||
| 5197 | int i; | 5203 | int i; |
| 5198 | 5204 | ||
| 5199 | if ( !data->writelen ) return; | 5205 | if ( !data->writelen ) return; |
| @@ -5218,18 +5224,17 @@ static void proc_APList_on_close( struct inode *inode, struct file *file ) { | |||
| 5218 | } | 5224 | } |
| 5219 | disable_MAC(ai, 1); | 5225 | disable_MAC(ai, 1); |
| 5220 | writeAPListRid(ai, &APList_rid, 1); | 5226 | writeAPListRid(ai, &APList_rid, 1); |
| 5221 | enable_MAC(ai, &rsp, 1); | 5227 | enable_MAC(ai, 1); |
| 5222 | } | 5228 | } |
| 5223 | 5229 | ||
| 5224 | /* This function wraps PC4500_writerid with a MAC disable */ | 5230 | /* This function wraps PC4500_writerid with a MAC disable */ |
| 5225 | static int do_writerid( struct airo_info *ai, u16 rid, const void *rid_data, | 5231 | static int do_writerid( struct airo_info *ai, u16 rid, const void *rid_data, |
| 5226 | int len, int dummy ) { | 5232 | int len, int dummy ) { |
| 5227 | int rc; | 5233 | int rc; |
| 5228 | Resp rsp; | ||
| 5229 | 5234 | ||
| 5230 | disable_MAC(ai, 1); | 5235 | disable_MAC(ai, 1); |
| 5231 | rc = PC4500_writerid(ai, rid, rid_data, len, 1); | 5236 | rc = PC4500_writerid(ai, rid, rid_data, len, 1); |
| 5232 | enable_MAC(ai, &rsp, 1); | 5237 | enable_MAC(ai, 1); |
| 5233 | return rc; | 5238 | return rc; |
| 5234 | } | 5239 | } |
| 5235 | 5240 | ||
| @@ -5260,7 +5265,6 @@ static int set_wep_key(struct airo_info *ai, u16 index, | |||
| 5260 | const char *key, u16 keylen, int perm, int lock ) { | 5265 | const char *key, u16 keylen, int perm, int lock ) { |
| 5261 | static const unsigned char macaddr[ETH_ALEN] = { 0x01, 0, 0, 0, 0, 0 }; | 5266 | static const unsigned char macaddr[ETH_ALEN] = { 0x01, 0, 0, 0, 0, 0 }; |
| 5262 | WepKeyRid wkr; | 5267 | WepKeyRid wkr; |
| 5263 | Resp rsp; | ||
| 5264 | 5268 | ||
| 5265 | memset(&wkr, 0, sizeof(wkr)); | 5269 | memset(&wkr, 0, sizeof(wkr)); |
| 5266 | if (keylen == 0) { | 5270 | if (keylen == 0) { |
| @@ -5280,7 +5284,7 @@ static int set_wep_key(struct airo_info *ai, u16 index, | |||
| 5280 | 5284 | ||
| 5281 | if (perm) disable_MAC(ai, lock); | 5285 | if (perm) disable_MAC(ai, lock); |
| 5282 | writeWepKeyRid(ai, &wkr, perm, lock); | 5286 | writeWepKeyRid(ai, &wkr, perm, lock); |
| 5283 | if (perm) enable_MAC(ai, &rsp, lock); | 5287 | if (perm) enable_MAC(ai, lock); |
| 5284 | return 0; | 5288 | return 0; |
| 5285 | } | 5289 | } |
| 5286 | 5290 | ||
| @@ -5548,7 +5552,6 @@ static int proc_close( struct inode *inode, struct file *file ) | |||
| 5548 | changed. */ | 5552 | changed. */ |
| 5549 | static void timer_func( struct net_device *dev ) { | 5553 | static void timer_func( struct net_device *dev ) { |
| 5550 | struct airo_info *apriv = dev->priv; | 5554 | struct airo_info *apriv = dev->priv; |
| 5551 | Resp rsp; | ||
| 5552 | 5555 | ||
| 5553 | /* We don't have a link so try changing the authtype */ | 5556 | /* We don't have a link so try changing the authtype */ |
| 5554 | readConfigRid(apriv, 0); | 5557 | readConfigRid(apriv, 0); |
| @@ -5575,7 +5578,7 @@ static void timer_func( struct net_device *dev ) { | |||
| 5575 | } | 5578 | } |
| 5576 | set_bit (FLAG_COMMIT, &apriv->flags); | 5579 | set_bit (FLAG_COMMIT, &apriv->flags); |
| 5577 | writeConfigRid(apriv, 0); | 5580 | writeConfigRid(apriv, 0); |
| 5578 | enable_MAC(apriv, &rsp, 0); | 5581 | enable_MAC(apriv, 0); |
| 5579 | up(&apriv->sem); | 5582 | up(&apriv->sem); |
| 5580 | 5583 | ||
| 5581 | /* Schedule check to see if the change worked */ | 5584 | /* Schedule check to see if the change worked */ |
| @@ -5597,8 +5600,10 @@ static int __devinit airo_pci_probe(struct pci_dev *pdev, | |||
| 5597 | dev = _init_airo_card(pdev->irq, pdev->resource[0].start, 0, pdev, &pdev->dev); | 5600 | dev = _init_airo_card(pdev->irq, pdev->resource[0].start, 0, pdev, &pdev->dev); |
| 5598 | else | 5601 | else |
| 5599 | dev = _init_airo_card(pdev->irq, pdev->resource[2].start, 0, pdev, &pdev->dev); | 5602 | dev = _init_airo_card(pdev->irq, pdev->resource[2].start, 0, pdev, &pdev->dev); |
| 5600 | if (!dev) | 5603 | if (!dev) { |
| 5604 | pci_disable_device(pdev); | ||
| 5601 | return -ENODEV; | 5605 | return -ENODEV; |
| 5606 | } | ||
| 5602 | 5607 | ||
| 5603 | pci_set_drvdata(pdev, dev); | 5608 | pci_set_drvdata(pdev, dev); |
| 5604 | return 0; | 5609 | return 0; |
| @@ -5610,6 +5615,8 @@ static void __devexit airo_pci_remove(struct pci_dev *pdev) | |||
| 5610 | 5615 | ||
| 5611 | airo_print_info(dev->name, "Unregistering..."); | 5616 | airo_print_info(dev->name, "Unregistering..."); |
| 5612 | stop_airo_card(dev, 1); | 5617 | stop_airo_card(dev, 1); |
| 5618 | pci_disable_device(pdev); | ||
| 5619 | pci_set_drvdata(pdev, NULL); | ||
| 5613 | } | 5620 | } |
| 5614 | 5621 | ||
| 5615 | static int airo_pci_suspend(struct pci_dev *pdev, pm_message_t state) | 5622 | static int airo_pci_suspend(struct pci_dev *pdev, pm_message_t state) |
| @@ -5646,7 +5653,6 @@ static int airo_pci_resume(struct pci_dev *pdev) | |||
| 5646 | { | 5653 | { |
| 5647 | struct net_device *dev = pci_get_drvdata(pdev); | 5654 | struct net_device *dev = pci_get_drvdata(pdev); |
| 5648 | struct airo_info *ai = dev->priv; | 5655 | struct airo_info *ai = dev->priv; |
| 5649 | Resp rsp; | ||
| 5650 | pci_power_t prev_state = pdev->current_state; | 5656 | pci_power_t prev_state = pdev->current_state; |
| 5651 | 5657 | ||
| 5652 | pci_set_power_state(pdev, PCI_D0); | 5658 | pci_set_power_state(pdev, PCI_D0); |
| @@ -5679,7 +5685,7 @@ static int airo_pci_resume(struct pci_dev *pdev) | |||
| 5679 | ai->APList = NULL; | 5685 | ai->APList = NULL; |
| 5680 | } | 5686 | } |
| 5681 | writeConfigRid(ai, 0); | 5687 | writeConfigRid(ai, 0); |
| 5682 | enable_MAC(ai, &rsp, 0); | 5688 | enable_MAC(ai, 0); |
| 5683 | ai->power = PMSG_ON; | 5689 | ai->power = PMSG_ON; |
| 5684 | netif_device_attach(dev); | 5690 | netif_device_attach(dev); |
| 5685 | netif_wake_queue(dev); | 5691 | netif_wake_queue(dev); |
| @@ -5903,7 +5909,6 @@ static int airo_set_essid(struct net_device *dev, | |||
| 5903 | char *extra) | 5909 | char *extra) |
| 5904 | { | 5910 | { |
| 5905 | struct airo_info *local = dev->priv; | 5911 | struct airo_info *local = dev->priv; |
| 5906 | Resp rsp; | ||
| 5907 | SsidRid SSID_rid; /* SSIDs */ | 5912 | SsidRid SSID_rid; /* SSIDs */ |
| 5908 | 5913 | ||
| 5909 | /* Reload the list of current SSID */ | 5914 | /* Reload the list of current SSID */ |
| @@ -5935,7 +5940,7 @@ static int airo_set_essid(struct net_device *dev, | |||
| 5935 | /* Write it to the card */ | 5940 | /* Write it to the card */ |
| 5936 | disable_MAC(local, 1); | 5941 | disable_MAC(local, 1); |
| 5937 | writeSsidRid(local, &SSID_rid, 1); | 5942 | writeSsidRid(local, &SSID_rid, 1); |
| 5938 | enable_MAC(local, &rsp, 1); | 5943 | enable_MAC(local, 1); |
| 5939 | 5944 | ||
| 5940 | return 0; | 5945 | return 0; |
| 5941 | } | 5946 | } |
| @@ -6000,7 +6005,7 @@ static int airo_set_wap(struct net_device *dev, | |||
| 6000 | memcpy(APList_rid.ap[0], awrq->sa_data, ETH_ALEN); | 6005 | memcpy(APList_rid.ap[0], awrq->sa_data, ETH_ALEN); |
| 6001 | disable_MAC(local, 1); | 6006 | disable_MAC(local, 1); |
| 6002 | writeAPListRid(local, &APList_rid, 1); | 6007 | writeAPListRid(local, &APList_rid, 1); |
| 6003 | enable_MAC(local, &rsp, 1); | 6008 | enable_MAC(local, 1); |
| 6004 | } | 6009 | } |
| 6005 | return 0; | 6010 | return 0; |
| 6006 | } | 6011 | } |
| @@ -7454,7 +7459,6 @@ static int airo_config_commit(struct net_device *dev, | |||
| 7454 | char *extra) /* NULL */ | 7459 | char *extra) /* NULL */ |
| 7455 | { | 7460 | { |
| 7456 | struct airo_info *local = dev->priv; | 7461 | struct airo_info *local = dev->priv; |
| 7457 | Resp rsp; | ||
| 7458 | 7462 | ||
| 7459 | if (!test_bit (FLAG_COMMIT, &local->flags)) | 7463 | if (!test_bit (FLAG_COMMIT, &local->flags)) |
| 7460 | return 0; | 7464 | return 0; |
| @@ -7479,7 +7483,7 @@ static int airo_config_commit(struct net_device *dev, | |||
| 7479 | if (down_interruptible(&local->sem)) | 7483 | if (down_interruptible(&local->sem)) |
| 7480 | return -ERESTARTSYS; | 7484 | return -ERESTARTSYS; |
| 7481 | writeConfigRid(local, 0); | 7485 | writeConfigRid(local, 0); |
| 7482 | enable_MAC(local, &rsp, 0); | 7486 | enable_MAC(local, 0); |
| 7483 | if (test_bit (FLAG_RESET, &local->flags)) | 7487 | if (test_bit (FLAG_RESET, &local->flags)) |
| 7484 | airo_set_promisc(local); | 7488 | airo_set_promisc(local); |
| 7485 | else | 7489 | else |
| @@ -7746,7 +7750,6 @@ static int readrids(struct net_device *dev, aironet_ioctl *comp) { | |||
| 7746 | unsigned char *iobuf; | 7750 | unsigned char *iobuf; |
| 7747 | int len; | 7751 | int len; |
| 7748 | struct airo_info *ai = dev->priv; | 7752 | struct airo_info *ai = dev->priv; |
| 7749 | Resp rsp; | ||
| 7750 | 7753 | ||
| 7751 | if (test_bit(FLAG_FLASHING, &ai->flags)) | 7754 | if (test_bit(FLAG_FLASHING, &ai->flags)) |
| 7752 | return -EIO; | 7755 | return -EIO; |
| @@ -7758,7 +7761,7 @@ static int readrids(struct net_device *dev, aironet_ioctl *comp) { | |||
| 7758 | if (test_bit(FLAG_COMMIT, &ai->flags)) { | 7761 | if (test_bit(FLAG_COMMIT, &ai->flags)) { |
| 7759 | disable_MAC (ai, 1); | 7762 | disable_MAC (ai, 1); |
| 7760 | writeConfigRid (ai, 1); | 7763 | writeConfigRid (ai, 1); |
| 7761 | enable_MAC (ai, &rsp, 1); | 7764 | enable_MAC(ai, 1); |
| 7762 | } | 7765 | } |
| 7763 | break; | 7766 | break; |
| 7764 | case AIROGSLIST: ridcode = RID_SSID; break; | 7767 | case AIROGSLIST: ridcode = RID_SSID; break; |
| @@ -7815,7 +7818,6 @@ static int writerids(struct net_device *dev, aironet_ioctl *comp) { | |||
| 7815 | struct airo_info *ai = dev->priv; | 7818 | struct airo_info *ai = dev->priv; |
| 7816 | int ridcode; | 7819 | int ridcode; |
| 7817 | int enabled; | 7820 | int enabled; |
| 7818 | Resp rsp; | ||
| 7819 | static int (* writer)(struct airo_info *, u16 rid, const void *, int, int); | 7821 | static int (* writer)(struct airo_info *, u16 rid, const void *, int, int); |
| 7820 | unsigned char *iobuf; | 7822 | unsigned char *iobuf; |
| 7821 | 7823 | ||
| @@ -7849,7 +7851,7 @@ static int writerids(struct net_device *dev, aironet_ioctl *comp) { | |||
| 7849 | * same with MAC off | 7851 | * same with MAC off |
| 7850 | */ | 7852 | */ |
| 7851 | case AIROPMACON: | 7853 | case AIROPMACON: |
| 7852 | if (enable_MAC(ai, &rsp, 1) != 0) | 7854 | if (enable_MAC(ai, 1) != 0) |
| 7853 | return -EIO; | 7855 | return -EIO; |
| 7854 | return 0; | 7856 | return 0; |
| 7855 | 7857 | ||
diff --git a/drivers/net/wireless/ipw2100.c b/drivers/net/wireless/ipw2100.c index d51daf87450f..072ede71e575 100644 --- a/drivers/net/wireless/ipw2100.c +++ b/drivers/net/wireless/ipw2100.c | |||
| @@ -1768,7 +1768,8 @@ static int ipw2100_up(struct ipw2100_priv *priv, int deferred) | |||
| 1768 | 1768 | ||
| 1769 | if (priv->stop_rf_kill) { | 1769 | if (priv->stop_rf_kill) { |
| 1770 | priv->stop_rf_kill = 0; | 1770 | priv->stop_rf_kill = 0; |
| 1771 | queue_delayed_work(priv->workqueue, &priv->rf_kill, HZ); | 1771 | queue_delayed_work(priv->workqueue, &priv->rf_kill, |
| 1772 | round_jiffies(HZ)); | ||
| 1772 | } | 1773 | } |
| 1773 | 1774 | ||
| 1774 | deferred = 1; | 1775 | deferred = 1; |
| @@ -2098,7 +2099,7 @@ static void isr_indicate_rf_kill(struct ipw2100_priv *priv, u32 status) | |||
| 2098 | /* Make sure the RF Kill check timer is running */ | 2099 | /* Make sure the RF Kill check timer is running */ |
| 2099 | priv->stop_rf_kill = 0; | 2100 | priv->stop_rf_kill = 0; |
| 2100 | cancel_delayed_work(&priv->rf_kill); | 2101 | cancel_delayed_work(&priv->rf_kill); |
| 2101 | queue_delayed_work(priv->workqueue, &priv->rf_kill, HZ); | 2102 | queue_delayed_work(priv->workqueue, &priv->rf_kill, round_jiffies(HZ)); |
| 2102 | } | 2103 | } |
| 2103 | 2104 | ||
| 2104 | static void isr_scan_complete(struct ipw2100_priv *priv, u32 status) | 2105 | static void isr_scan_complete(struct ipw2100_priv *priv, u32 status) |
| @@ -4233,7 +4234,8 @@ static int ipw_radio_kill_sw(struct ipw2100_priv *priv, int disable_radio) | |||
| 4233 | /* Make sure the RF_KILL check timer is running */ | 4234 | /* Make sure the RF_KILL check timer is running */ |
| 4234 | priv->stop_rf_kill = 0; | 4235 | priv->stop_rf_kill = 0; |
| 4235 | cancel_delayed_work(&priv->rf_kill); | 4236 | cancel_delayed_work(&priv->rf_kill); |
| 4236 | queue_delayed_work(priv->workqueue, &priv->rf_kill, HZ); | 4237 | queue_delayed_work(priv->workqueue, &priv->rf_kill, |
| 4238 | round_jiffies(HZ)); | ||
| 4237 | } else | 4239 | } else |
| 4238 | schedule_reset(priv); | 4240 | schedule_reset(priv); |
| 4239 | } | 4241 | } |
| @@ -5969,7 +5971,8 @@ static void ipw2100_rf_kill(struct work_struct *work) | |||
| 5969 | if (rf_kill_active(priv)) { | 5971 | if (rf_kill_active(priv)) { |
| 5970 | IPW_DEBUG_RF_KILL("RF Kill active, rescheduling GPIO check\n"); | 5972 | IPW_DEBUG_RF_KILL("RF Kill active, rescheduling GPIO check\n"); |
| 5971 | if (!priv->stop_rf_kill) | 5973 | if (!priv->stop_rf_kill) |
| 5972 | queue_delayed_work(priv->workqueue, &priv->rf_kill, HZ); | 5974 | queue_delayed_work(priv->workqueue, &priv->rf_kill, |
| 5975 | round_jiffies(HZ)); | ||
| 5973 | goto exit_unlock; | 5976 | goto exit_unlock; |
| 5974 | } | 5977 | } |
| 5975 | 5978 | ||
diff --git a/drivers/net/wireless/ipw2200.c b/drivers/net/wireless/ipw2200.c index 7cb2052a55a5..aa32a97380ec 100644 --- a/drivers/net/wireless/ipw2200.c +++ b/drivers/net/wireless/ipw2200.c | |||
| @@ -1751,7 +1751,7 @@ static int ipw_radio_kill_sw(struct ipw_priv *priv, int disable_radio) | |||
| 1751 | /* Make sure the RF_KILL check timer is running */ | 1751 | /* Make sure the RF_KILL check timer is running */ |
| 1752 | cancel_delayed_work(&priv->rf_kill); | 1752 | cancel_delayed_work(&priv->rf_kill); |
| 1753 | queue_delayed_work(priv->workqueue, &priv->rf_kill, | 1753 | queue_delayed_work(priv->workqueue, &priv->rf_kill, |
| 1754 | 2 * HZ); | 1754 | round_jiffies(2 * HZ)); |
| 1755 | } else | 1755 | } else |
| 1756 | queue_work(priv->workqueue, &priv->up); | 1756 | queue_work(priv->workqueue, &priv->up); |
| 1757 | } | 1757 | } |
| @@ -4690,7 +4690,8 @@ static void ipw_rx_notification(struct ipw_priv *priv, | |||
| 4690 | else if (priv->config & CFG_BACKGROUND_SCAN | 4690 | else if (priv->config & CFG_BACKGROUND_SCAN |
| 4691 | && priv->status & STATUS_ASSOCIATED) | 4691 | && priv->status & STATUS_ASSOCIATED) |
| 4692 | queue_delayed_work(priv->workqueue, | 4692 | queue_delayed_work(priv->workqueue, |
| 4693 | &priv->request_scan, HZ); | 4693 | &priv->request_scan, |
| 4694 | round_jiffies(HZ)); | ||
| 4694 | 4695 | ||
| 4695 | /* Send an empty event to user space. | 4696 | /* Send an empty event to user space. |
| 4696 | * We don't send the received data on the event because | 4697 | * We don't send the received data on the event because |
diff --git a/drivers/net/wireless/libertas/cmd.c b/drivers/net/wireless/libertas/cmd.c index 13f6528abb00..4a8f5dc70239 100644 --- a/drivers/net/wireless/libertas/cmd.c +++ b/drivers/net/wireless/libertas/cmd.c | |||
| @@ -240,7 +240,7 @@ static int wlan_cmd_802_11_enable_rsn(wlan_private * priv, | |||
| 240 | if (*enable) | 240 | if (*enable) |
| 241 | penableRSN->enable = cpu_to_le16(cmd_enable_rsn); | 241 | penableRSN->enable = cpu_to_le16(cmd_enable_rsn); |
| 242 | else | 242 | else |
| 243 | penableRSN->enable = cpu_to_le16(cmd_enable_rsn); | 243 | penableRSN->enable = cpu_to_le16(cmd_disable_rsn); |
| 244 | } | 244 | } |
| 245 | 245 | ||
| 246 | lbs_deb_leave(LBS_DEB_CMD); | 246 | lbs_deb_leave(LBS_DEB_CMD); |
diff --git a/drivers/net/wireless/libertas/rx.c b/drivers/net/wireless/libertas/rx.c index 88d9d2d787d5..769c86fb9509 100644 --- a/drivers/net/wireless/libertas/rx.c +++ b/drivers/net/wireless/libertas/rx.c | |||
| @@ -439,7 +439,6 @@ static int process_rxed_802_11_packet(wlan_private * priv, struct sk_buff *skb) | |||
| 439 | ret = 0; | 439 | ret = 0; |
| 440 | 440 | ||
| 441 | done: | 441 | done: |
| 442 | skb->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ | ||
| 443 | lbs_deb_leave_args(LBS_DEB_RX, "ret %d", ret); | 442 | lbs_deb_leave_args(LBS_DEB_RX, "ret %d", ret); |
| 444 | return ret; | 443 | return ret; |
| 445 | } | 444 | } |
diff --git a/drivers/net/wireless/libertas/version.h b/drivers/net/wireless/libertas/version.h deleted file mode 100644 index 8b137891791f..000000000000 --- a/drivers/net/wireless/libertas/version.h +++ /dev/null | |||
| @@ -1 +0,0 @@ | |||
| 1 | |||
diff --git a/drivers/net/wireless/libertas/wext.c b/drivers/net/wireless/libertas/wext.c index f42b796b5e47..2fcc3bf21081 100644 --- a/drivers/net/wireless/libertas/wext.c +++ b/drivers/net/wireless/libertas/wext.c | |||
| @@ -1719,9 +1719,6 @@ static int wlan_set_encodeext(struct net_device *dev, | |||
| 1719 | pkey->type = KEY_TYPE_ID_TKIP; | 1719 | pkey->type = KEY_TYPE_ID_TKIP; |
| 1720 | } else if (alg == IW_ENCODE_ALG_CCMP) { | 1720 | } else if (alg == IW_ENCODE_ALG_CCMP) { |
| 1721 | pkey->type = KEY_TYPE_ID_AES; | 1721 | pkey->type = KEY_TYPE_ID_AES; |
| 1722 | } else { | ||
| 1723 | ret = -EINVAL; | ||
| 1724 | goto out; | ||
| 1725 | } | 1722 | } |
| 1726 | 1723 | ||
| 1727 | /* If WPA isn't enabled yet, do that now */ | 1724 | /* If WPA isn't enabled yet, do that now */ |
diff --git a/drivers/net/wireless/prism54/isl_ioctl.c b/drivers/net/wireless/prism54/isl_ioctl.c index 283be4a70524..585f5996d292 100644 --- a/drivers/net/wireless/prism54/isl_ioctl.c +++ b/drivers/net/wireless/prism54/isl_ioctl.c | |||
| @@ -1853,7 +1853,6 @@ prism54_del_mac(struct net_device *ndev, struct iw_request_info *info, | |||
| 1853 | islpci_private *priv = netdev_priv(ndev); | 1853 | islpci_private *priv = netdev_priv(ndev); |
| 1854 | struct islpci_acl *acl = &priv->acl; | 1854 | struct islpci_acl *acl = &priv->acl; |
| 1855 | struct mac_entry *entry; | 1855 | struct mac_entry *entry; |
| 1856 | struct list_head *ptr; | ||
| 1857 | struct sockaddr *addr = (struct sockaddr *) extra; | 1856 | struct sockaddr *addr = (struct sockaddr *) extra; |
| 1858 | 1857 | ||
| 1859 | if (addr->sa_family != ARPHRD_ETHER) | 1858 | if (addr->sa_family != ARPHRD_ETHER) |
| @@ -1861,11 +1860,9 @@ prism54_del_mac(struct net_device *ndev, struct iw_request_info *info, | |||
| 1861 | 1860 | ||
| 1862 | if (down_interruptible(&acl->sem)) | 1861 | if (down_interruptible(&acl->sem)) |
| 1863 | return -ERESTARTSYS; | 1862 | return -ERESTARTSYS; |
| 1864 | for (ptr = acl->mac_list.next; ptr != &acl->mac_list; ptr = ptr->next) { | 1863 | list_for_each_entry(entry, &acl->mac_list, _list) { |
| 1865 | entry = list_entry(ptr, struct mac_entry, _list); | ||
| 1866 | |||
| 1867 | if (memcmp(entry->addr, addr->sa_data, ETH_ALEN) == 0) { | 1864 | if (memcmp(entry->addr, addr->sa_data, ETH_ALEN) == 0) { |
| 1868 | list_del(ptr); | 1865 | list_del(&entry->_list); |
| 1869 | acl->size--; | 1866 | acl->size--; |
| 1870 | kfree(entry); | 1867 | kfree(entry); |
| 1871 | up(&acl->sem); | 1868 | up(&acl->sem); |
| @@ -1883,7 +1880,6 @@ prism54_get_mac(struct net_device *ndev, struct iw_request_info *info, | |||
| 1883 | islpci_private *priv = netdev_priv(ndev); | 1880 | islpci_private *priv = netdev_priv(ndev); |
| 1884 | struct islpci_acl *acl = &priv->acl; | 1881 | struct islpci_acl *acl = &priv->acl; |
| 1885 | struct mac_entry *entry; | 1882 | struct mac_entry *entry; |
| 1886 | struct list_head *ptr; | ||
| 1887 | struct sockaddr *dst = (struct sockaddr *) extra; | 1883 | struct sockaddr *dst = (struct sockaddr *) extra; |
| 1888 | 1884 | ||
| 1889 | dwrq->length = 0; | 1885 | dwrq->length = 0; |
| @@ -1891,9 +1887,7 @@ prism54_get_mac(struct net_device *ndev, struct iw_request_info *info, | |||
| 1891 | if (down_interruptible(&acl->sem)) | 1887 | if (down_interruptible(&acl->sem)) |
| 1892 | return -ERESTARTSYS; | 1888 | return -ERESTARTSYS; |
| 1893 | 1889 | ||
| 1894 | for (ptr = acl->mac_list.next; ptr != &acl->mac_list; ptr = ptr->next) { | 1890 | list_for_each_entry(entry, &acl->mac_list, _list) { |
| 1895 | entry = list_entry(ptr, struct mac_entry, _list); | ||
| 1896 | |||
| 1897 | memcpy(dst->sa_data, entry->addr, ETH_ALEN); | 1891 | memcpy(dst->sa_data, entry->addr, ETH_ALEN); |
| 1898 | dst->sa_family = ARPHRD_ETHER; | 1892 | dst->sa_family = ARPHRD_ETHER; |
| 1899 | dwrq->length++; | 1893 | dwrq->length++; |
| @@ -1960,7 +1954,6 @@ prism54_get_policy(struct net_device *ndev, struct iw_request_info *info, | |||
| 1960 | static int | 1954 | static int |
| 1961 | prism54_mac_accept(struct islpci_acl *acl, char *mac) | 1955 | prism54_mac_accept(struct islpci_acl *acl, char *mac) |
| 1962 | { | 1956 | { |
| 1963 | struct list_head *ptr; | ||
| 1964 | struct mac_entry *entry; | 1957 | struct mac_entry *entry; |
| 1965 | int res = 0; | 1958 | int res = 0; |
| 1966 | 1959 | ||
| @@ -1972,8 +1965,7 @@ prism54_mac_accept(struct islpci_acl *acl, char *mac) | |||
| 1972 | return 1; | 1965 | return 1; |
| 1973 | } | 1966 | } |
| 1974 | 1967 | ||
| 1975 | for (ptr = acl->mac_list.next; ptr != &acl->mac_list; ptr = ptr->next) { | 1968 | list_for_each_entry(entry, &acl->mac_list, _list) { |
| 1976 | entry = list_entry(ptr, struct mac_entry, _list); | ||
| 1977 | if (memcmp(entry->addr, mac, ETH_ALEN) == 0) { | 1969 | if (memcmp(entry->addr, mac, ETH_ALEN) == 0) { |
| 1978 | res = 1; | 1970 | res = 1; |
| 1979 | break; | 1971 | break; |
| @@ -2216,11 +2208,9 @@ prism54_wpa_bss_ie_init(islpci_private *priv) | |||
| 2216 | void | 2208 | void |
| 2217 | prism54_wpa_bss_ie_clean(islpci_private *priv) | 2209 | prism54_wpa_bss_ie_clean(islpci_private *priv) |
| 2218 | { | 2210 | { |
| 2219 | struct list_head *ptr, *n; | 2211 | struct islpci_bss_wpa_ie *bss, *n; |
| 2220 | 2212 | ||
| 2221 | list_for_each_safe(ptr, n, &priv->bss_wpa_list) { | 2213 | list_for_each_entry_safe(bss, n, &priv->bss_wpa_list, list) { |
| 2222 | struct islpci_bss_wpa_ie *bss; | ||
| 2223 | bss = list_entry(ptr, struct islpci_bss_wpa_ie, list); | ||
| 2224 | kfree(bss); | 2214 | kfree(bss); |
| 2225 | } | 2215 | } |
| 2226 | } | 2216 | } |
diff --git a/drivers/net/wireless/rtl8187_rtl8225.c b/drivers/net/wireless/rtl8187_rtl8225.c index e25a09f1b068..efc41207780e 100644 --- a/drivers/net/wireless/rtl8187_rtl8225.c +++ b/drivers/net/wireless/rtl8187_rtl8225.c | |||
| @@ -67,7 +67,7 @@ static void rtl8225_write_bitbang(struct ieee80211_hw *dev, u8 addr, u16 data) | |||
| 67 | msleep(2); | 67 | msleep(2); |
| 68 | } | 68 | } |
| 69 | 69 | ||
| 70 | static void rtl8225_write_8051(struct ieee80211_hw *dev, u8 addr, u16 data) | 70 | static void rtl8225_write_8051(struct ieee80211_hw *dev, u8 addr, __le16 data) |
| 71 | { | 71 | { |
| 72 | struct rtl8187_priv *priv = dev->priv; | 72 | struct rtl8187_priv *priv = dev->priv; |
| 73 | u16 reg80, reg82, reg84; | 73 | u16 reg80, reg82, reg84; |
| @@ -106,7 +106,7 @@ void rtl8225_write(struct ieee80211_hw *dev, u8 addr, u16 data) | |||
| 106 | struct rtl8187_priv *priv = dev->priv; | 106 | struct rtl8187_priv *priv = dev->priv; |
| 107 | 107 | ||
| 108 | if (priv->asic_rev) | 108 | if (priv->asic_rev) |
| 109 | rtl8225_write_8051(dev, addr, data); | 109 | rtl8225_write_8051(dev, addr, cpu_to_le16(data)); |
| 110 | else | 110 | else |
| 111 | rtl8225_write_bitbang(dev, addr, data); | 111 | rtl8225_write_bitbang(dev, addr, data); |
| 112 | } | 112 | } |
diff --git a/drivers/net/wireless/zd1211rw/zd_chip.c b/drivers/net/wireless/zd1211rw/zd_chip.c index 5b624bfc01a6..c39f1984b84d 100644 --- a/drivers/net/wireless/zd1211rw/zd_chip.c +++ b/drivers/net/wireless/zd1211rw/zd_chip.c | |||
| @@ -49,8 +49,9 @@ void zd_chip_clear(struct zd_chip *chip) | |||
| 49 | ZD_MEMCLEAR(chip, sizeof(*chip)); | 49 | ZD_MEMCLEAR(chip, sizeof(*chip)); |
| 50 | } | 50 | } |
| 51 | 51 | ||
| 52 | static int scnprint_mac_oui(const u8 *addr, char *buffer, size_t size) | 52 | static int scnprint_mac_oui(struct zd_chip *chip, char *buffer, size_t size) |
| 53 | { | 53 | { |
| 54 | u8 *addr = zd_usb_to_netdev(&chip->usb)->dev_addr; | ||
| 54 | return scnprintf(buffer, size, "%02x-%02x-%02x", | 55 | return scnprintf(buffer, size, "%02x-%02x-%02x", |
| 55 | addr[0], addr[1], addr[2]); | 56 | addr[0], addr[1], addr[2]); |
| 56 | } | 57 | } |
| @@ -61,10 +62,10 @@ static int scnprint_id(struct zd_chip *chip, char *buffer, size_t size) | |||
| 61 | int i = 0; | 62 | int i = 0; |
| 62 | 63 | ||
| 63 | i = scnprintf(buffer, size, "zd1211%s chip ", | 64 | i = scnprintf(buffer, size, "zd1211%s chip ", |
| 64 | chip->is_zd1211b ? "b" : ""); | 65 | zd_chip_is_zd1211b(chip) ? "b" : ""); |
| 65 | i += zd_usb_scnprint_id(&chip->usb, buffer+i, size-i); | 66 | i += zd_usb_scnprint_id(&chip->usb, buffer+i, size-i); |
| 66 | i += scnprintf(buffer+i, size-i, " "); | 67 | i += scnprintf(buffer+i, size-i, " "); |
| 67 | i += scnprint_mac_oui(chip->e2p_mac, buffer+i, size-i); | 68 | i += scnprint_mac_oui(chip, buffer+i, size-i); |
| 68 | i += scnprintf(buffer+i, size-i, " "); | 69 | i += scnprintf(buffer+i, size-i, " "); |
| 69 | i += zd_rf_scnprint_id(&chip->rf, buffer+i, size-i); | 70 | i += zd_rf_scnprint_id(&chip->rf, buffer+i, size-i); |
| 70 | i += scnprintf(buffer+i, size-i, " pa%1x %c%c%c%c%c", chip->pa_type, | 71 | i += scnprintf(buffer+i, size-i, " pa%1x %c%c%c%c%c", chip->pa_type, |
| @@ -366,64 +367,9 @@ error: | |||
| 366 | return r; | 367 | return r; |
| 367 | } | 368 | } |
| 368 | 369 | ||
| 369 | static int _read_mac_addr(struct zd_chip *chip, u8 *mac_addr, | ||
| 370 | const zd_addr_t *addr) | ||
| 371 | { | ||
| 372 | int r; | ||
| 373 | u32 parts[2]; | ||
| 374 | |||
| 375 | r = zd_ioread32v_locked(chip, parts, (const zd_addr_t *)addr, 2); | ||
| 376 | if (r) { | ||
| 377 | dev_dbg_f(zd_chip_dev(chip), | ||
| 378 | "error: couldn't read e2p macs. Error number %d\n", r); | ||
| 379 | return r; | ||
| 380 | } | ||
| 381 | |||
| 382 | mac_addr[0] = parts[0]; | ||
| 383 | mac_addr[1] = parts[0] >> 8; | ||
| 384 | mac_addr[2] = parts[0] >> 16; | ||
| 385 | mac_addr[3] = parts[0] >> 24; | ||
| 386 | mac_addr[4] = parts[1]; | ||
| 387 | mac_addr[5] = parts[1] >> 8; | ||
| 388 | |||
| 389 | return 0; | ||
| 390 | } | ||
| 391 | |||
| 392 | static int read_e2p_mac_addr(struct zd_chip *chip) | ||
| 393 | { | ||
| 394 | static const zd_addr_t addr[2] = { E2P_MAC_ADDR_P1, E2P_MAC_ADDR_P2 }; | ||
| 395 | |||
| 396 | ZD_ASSERT(mutex_is_locked(&chip->mutex)); | ||
| 397 | return _read_mac_addr(chip, chip->e2p_mac, (const zd_addr_t *)addr); | ||
| 398 | } | ||
| 399 | |||
| 400 | /* MAC address: if custom mac addresses are to to be used CR_MAC_ADDR_P1 and | 370 | /* MAC address: if custom mac addresses are to to be used CR_MAC_ADDR_P1 and |
| 401 | * CR_MAC_ADDR_P2 must be overwritten | 371 | * CR_MAC_ADDR_P2 must be overwritten |
| 402 | */ | 372 | */ |
| 403 | void zd_get_e2p_mac_addr(struct zd_chip *chip, u8 *mac_addr) | ||
| 404 | { | ||
| 405 | mutex_lock(&chip->mutex); | ||
| 406 | memcpy(mac_addr, chip->e2p_mac, ETH_ALEN); | ||
| 407 | mutex_unlock(&chip->mutex); | ||
| 408 | } | ||
| 409 | |||
| 410 | static int read_mac_addr(struct zd_chip *chip, u8 *mac_addr) | ||
| 411 | { | ||
| 412 | static const zd_addr_t addr[2] = { CR_MAC_ADDR_P1, CR_MAC_ADDR_P2 }; | ||
| 413 | return _read_mac_addr(chip, mac_addr, (const zd_addr_t *)addr); | ||
| 414 | } | ||
| 415 | |||
| 416 | int zd_read_mac_addr(struct zd_chip *chip, u8 *mac_addr) | ||
| 417 | { | ||
| 418 | int r; | ||
| 419 | |||
| 420 | dev_dbg_f(zd_chip_dev(chip), "\n"); | ||
| 421 | mutex_lock(&chip->mutex); | ||
| 422 | r = read_mac_addr(chip, mac_addr); | ||
| 423 | mutex_unlock(&chip->mutex); | ||
| 424 | return r; | ||
| 425 | } | ||
| 426 | |||
| 427 | int zd_write_mac_addr(struct zd_chip *chip, const u8 *mac_addr) | 373 | int zd_write_mac_addr(struct zd_chip *chip, const u8 *mac_addr) |
| 428 | { | 374 | { |
| 429 | int r; | 375 | int r; |
| @@ -444,12 +390,6 @@ int zd_write_mac_addr(struct zd_chip *chip, const u8 *mac_addr) | |||
| 444 | 390 | ||
| 445 | mutex_lock(&chip->mutex); | 391 | mutex_lock(&chip->mutex); |
| 446 | r = zd_iowrite32a_locked(chip, reqs, ARRAY_SIZE(reqs)); | 392 | r = zd_iowrite32a_locked(chip, reqs, ARRAY_SIZE(reqs)); |
| 447 | #ifdef DEBUG | ||
| 448 | { | ||
| 449 | u8 tmp[ETH_ALEN]; | ||
| 450 | read_mac_addr(chip, tmp); | ||
| 451 | } | ||
| 452 | #endif /* DEBUG */ | ||
| 453 | mutex_unlock(&chip->mutex); | 393 | mutex_unlock(&chip->mutex); |
| 454 | return r; | 394 | return r; |
| 455 | } | 395 | } |
| @@ -809,7 +749,7 @@ out: | |||
| 809 | 749 | ||
| 810 | static int hw_reset_phy(struct zd_chip *chip) | 750 | static int hw_reset_phy(struct zd_chip *chip) |
| 811 | { | 751 | { |
| 812 | return chip->is_zd1211b ? zd1211b_hw_reset_phy(chip) : | 752 | return zd_chip_is_zd1211b(chip) ? zd1211b_hw_reset_phy(chip) : |
| 813 | zd1211_hw_reset_phy(chip); | 753 | zd1211_hw_reset_phy(chip); |
| 814 | } | 754 | } |
| 815 | 755 | ||
| @@ -874,7 +814,7 @@ static int hw_init_hmac(struct zd_chip *chip) | |||
| 874 | if (r) | 814 | if (r) |
| 875 | return r; | 815 | return r; |
| 876 | 816 | ||
| 877 | return chip->is_zd1211b ? | 817 | return zd_chip_is_zd1211b(chip) ? |
| 878 | zd1211b_hw_init_hmac(chip) : zd1211_hw_init_hmac(chip); | 818 | zd1211b_hw_init_hmac(chip) : zd1211_hw_init_hmac(chip); |
| 879 | } | 819 | } |
| 880 | 820 | ||
| @@ -1136,8 +1076,15 @@ static int read_fw_regs_offset(struct zd_chip *chip) | |||
| 1136 | return 0; | 1076 | return 0; |
| 1137 | } | 1077 | } |
| 1138 | 1078 | ||
| 1079 | /* Read mac address using pre-firmware interface */ | ||
| 1080 | int zd_chip_read_mac_addr_fw(struct zd_chip *chip, u8 *addr) | ||
| 1081 | { | ||
| 1082 | dev_dbg_f(zd_chip_dev(chip), "\n"); | ||
| 1083 | return zd_usb_read_fw(&chip->usb, E2P_MAC_ADDR_P1, addr, | ||
| 1084 | ETH_ALEN); | ||
| 1085 | } | ||
| 1139 | 1086 | ||
| 1140 | int zd_chip_init_hw(struct zd_chip *chip, u8 device_type) | 1087 | int zd_chip_init_hw(struct zd_chip *chip) |
| 1141 | { | 1088 | { |
| 1142 | int r; | 1089 | int r; |
| 1143 | u8 rf_type; | 1090 | u8 rf_type; |
| @@ -1145,7 +1092,6 @@ int zd_chip_init_hw(struct zd_chip *chip, u8 device_type) | |||
| 1145 | dev_dbg_f(zd_chip_dev(chip), "\n"); | 1092 | dev_dbg_f(zd_chip_dev(chip), "\n"); |
| 1146 | 1093 | ||
| 1147 | mutex_lock(&chip->mutex); | 1094 | mutex_lock(&chip->mutex); |
| 1148 | chip->is_zd1211b = (device_type == DEVICE_ZD1211B) != 0; | ||
| 1149 | 1095 | ||
| 1150 | #ifdef DEBUG | 1096 | #ifdef DEBUG |
| 1151 | r = test_init(chip); | 1097 | r = test_init(chip); |
| @@ -1201,10 +1147,6 @@ int zd_chip_init_hw(struct zd_chip *chip, u8 device_type) | |||
| 1201 | goto out; | 1147 | goto out; |
| 1202 | #endif /* DEBUG */ | 1148 | #endif /* DEBUG */ |
| 1203 | 1149 | ||
| 1204 | r = read_e2p_mac_addr(chip); | ||
| 1205 | if (r) | ||
| 1206 | goto out; | ||
| 1207 | |||
| 1208 | r = read_cal_int_tables(chip); | 1150 | r = read_cal_int_tables(chip); |
| 1209 | if (r) | 1151 | if (r) |
| 1210 | goto out; | 1152 | goto out; |
| @@ -1259,7 +1201,7 @@ static int update_channel_integration_and_calibration(struct zd_chip *chip, | |||
| 1259 | r = update_pwr_int(chip, channel); | 1201 | r = update_pwr_int(chip, channel); |
| 1260 | if (r) | 1202 | if (r) |
| 1261 | return r; | 1203 | return r; |
| 1262 | if (chip->is_zd1211b) { | 1204 | if (zd_chip_is_zd1211b(chip)) { |
| 1263 | static const struct zd_ioreq16 ioreqs[] = { | 1205 | static const struct zd_ioreq16 ioreqs[] = { |
| 1264 | { CR69, 0x28 }, | 1206 | { CR69, 0x28 }, |
| 1265 | {}, | 1207 | {}, |
diff --git a/drivers/net/wireless/zd1211rw/zd_chip.h b/drivers/net/wireless/zd1211rw/zd_chip.h index 79d0288c193a..f4698576ab71 100644 --- a/drivers/net/wireless/zd1211rw/zd_chip.h +++ b/drivers/net/wireless/zd1211rw/zd_chip.h | |||
| @@ -704,7 +704,6 @@ struct zd_chip { | |||
| 704 | struct mutex mutex; | 704 | struct mutex mutex; |
| 705 | /* Base address of FW_REG_ registers */ | 705 | /* Base address of FW_REG_ registers */ |
| 706 | zd_addr_t fw_regs_base; | 706 | zd_addr_t fw_regs_base; |
| 707 | u8 e2p_mac[ETH_ALEN]; | ||
| 708 | /* EepSetPoint in the vendor driver */ | 707 | /* EepSetPoint in the vendor driver */ |
| 709 | u8 pwr_cal_values[E2P_CHANNEL_COUNT]; | 708 | u8 pwr_cal_values[E2P_CHANNEL_COUNT]; |
| 710 | /* integration values in the vendor driver */ | 709 | /* integration values in the vendor driver */ |
| @@ -715,7 +714,7 @@ struct zd_chip { | |||
| 715 | unsigned int pa_type:4, | 714 | unsigned int pa_type:4, |
| 716 | patch_cck_gain:1, patch_cr157:1, patch_6m_band_edge:1, | 715 | patch_cck_gain:1, patch_cr157:1, patch_6m_band_edge:1, |
| 717 | new_phy_layout:1, al2230s_bit:1, | 716 | new_phy_layout:1, al2230s_bit:1, |
| 718 | is_zd1211b:1, supports_tx_led:1; | 717 | supports_tx_led:1; |
| 719 | }; | 718 | }; |
| 720 | 719 | ||
| 721 | static inline struct zd_chip *zd_usb_to_chip(struct zd_usb *usb) | 720 | static inline struct zd_chip *zd_usb_to_chip(struct zd_usb *usb) |
| @@ -734,9 +733,15 @@ void zd_chip_init(struct zd_chip *chip, | |||
| 734 | struct net_device *netdev, | 733 | struct net_device *netdev, |
| 735 | struct usb_interface *intf); | 734 | struct usb_interface *intf); |
| 736 | void zd_chip_clear(struct zd_chip *chip); | 735 | void zd_chip_clear(struct zd_chip *chip); |
| 737 | int zd_chip_init_hw(struct zd_chip *chip, u8 device_type); | 736 | int zd_chip_read_mac_addr_fw(struct zd_chip *chip, u8 *addr); |
| 737 | int zd_chip_init_hw(struct zd_chip *chip); | ||
| 738 | int zd_chip_reset(struct zd_chip *chip); | 738 | int zd_chip_reset(struct zd_chip *chip); |
| 739 | 739 | ||
| 740 | static inline int zd_chip_is_zd1211b(struct zd_chip *chip) | ||
| 741 | { | ||
| 742 | return chip->usb.is_zd1211b; | ||
| 743 | } | ||
| 744 | |||
| 740 | static inline int zd_ioread16v_locked(struct zd_chip *chip, u16 *values, | 745 | static inline int zd_ioread16v_locked(struct zd_chip *chip, u16 *values, |
| 741 | const zd_addr_t *addresses, | 746 | const zd_addr_t *addresses, |
| 742 | unsigned int count) | 747 | unsigned int count) |
| @@ -825,8 +830,6 @@ static inline u8 _zd_chip_get_channel(struct zd_chip *chip) | |||
| 825 | } | 830 | } |
| 826 | u8 zd_chip_get_channel(struct zd_chip *chip); | 831 | u8 zd_chip_get_channel(struct zd_chip *chip); |
| 827 | int zd_read_regdomain(struct zd_chip *chip, u8 *regdomain); | 832 | int zd_read_regdomain(struct zd_chip *chip, u8 *regdomain); |
| 828 | void zd_get_e2p_mac_addr(struct zd_chip *chip, u8 *mac_addr); | ||
| 829 | int zd_read_mac_addr(struct zd_chip *chip, u8 *mac_addr); | ||
| 830 | int zd_write_mac_addr(struct zd_chip *chip, const u8 *mac_addr); | 833 | int zd_write_mac_addr(struct zd_chip *chip, const u8 *mac_addr); |
| 831 | int zd_chip_switch_radio_on(struct zd_chip *chip); | 834 | int zd_chip_switch_radio_on(struct zd_chip *chip); |
| 832 | int zd_chip_switch_radio_off(struct zd_chip *chip); | 835 | int zd_chip_switch_radio_off(struct zd_chip *chip); |
diff --git a/drivers/net/wireless/zd1211rw/zd_mac.c b/drivers/net/wireless/zd1211rw/zd_mac.c index 6753d240c168..f6c487aa8246 100644 --- a/drivers/net/wireless/zd1211rw/zd_mac.c +++ b/drivers/net/wireless/zd1211rw/zd_mac.c | |||
| @@ -86,38 +86,46 @@ out: | |||
| 86 | return r; | 86 | return r; |
| 87 | } | 87 | } |
| 88 | 88 | ||
| 89 | int zd_mac_init_hw(struct zd_mac *mac, u8 device_type) | 89 | int zd_mac_preinit_hw(struct zd_mac *mac) |
| 90 | { | 90 | { |
| 91 | int r; | 91 | int r; |
| 92 | struct zd_chip *chip = &mac->chip; | ||
| 93 | u8 addr[ETH_ALEN]; | 92 | u8 addr[ETH_ALEN]; |
| 93 | |||
| 94 | r = zd_chip_read_mac_addr_fw(&mac->chip, addr); | ||
| 95 | if (r) | ||
| 96 | return r; | ||
| 97 | |||
| 98 | memcpy(mac->netdev->dev_addr, addr, ETH_ALEN); | ||
| 99 | return 0; | ||
| 100 | } | ||
| 101 | |||
| 102 | int zd_mac_init_hw(struct zd_mac *mac) | ||
| 103 | { | ||
| 104 | int r; | ||
| 105 | struct zd_chip *chip = &mac->chip; | ||
| 94 | u8 default_regdomain; | 106 | u8 default_regdomain; |
| 95 | 107 | ||
| 96 | r = zd_chip_enable_int(chip); | 108 | r = zd_chip_enable_int(chip); |
| 97 | if (r) | 109 | if (r) |
| 98 | goto out; | 110 | goto out; |
| 99 | r = zd_chip_init_hw(chip, device_type); | 111 | r = zd_chip_init_hw(chip); |
| 100 | if (r) | 112 | if (r) |
| 101 | goto disable_int; | 113 | goto disable_int; |
| 102 | 114 | ||
| 103 | zd_get_e2p_mac_addr(chip, addr); | ||
| 104 | r = zd_write_mac_addr(chip, addr); | ||
| 105 | if (r) | ||
| 106 | goto disable_int; | ||
| 107 | ZD_ASSERT(!irqs_disabled()); | 115 | ZD_ASSERT(!irqs_disabled()); |
| 108 | spin_lock_irq(&mac->lock); | ||
| 109 | memcpy(mac->netdev->dev_addr, addr, ETH_ALEN); | ||
| 110 | spin_unlock_irq(&mac->lock); | ||
| 111 | 116 | ||
| 112 | r = zd_read_regdomain(chip, &default_regdomain); | 117 | r = zd_read_regdomain(chip, &default_regdomain); |
| 113 | if (r) | 118 | if (r) |
| 114 | goto disable_int; | 119 | goto disable_int; |
| 115 | if (!zd_regdomain_supported(default_regdomain)) { | 120 | if (!zd_regdomain_supported(default_regdomain)) { |
| 116 | dev_dbg_f(zd_mac_dev(mac), | 121 | /* The vendor driver overrides the regulatory domain and |
| 117 | "Regulatory Domain %#04x is not supported.\n", | 122 | * allowed channel registers and unconditionally restricts |
| 118 | default_regdomain); | 123 | * available channels to 1-11 everywhere. Match their |
| 119 | r = -EINVAL; | 124 | * questionable behaviour only for regdomains which we don't |
| 120 | goto disable_int; | 125 | * recognise. */ |
| 126 | dev_warn(zd_mac_dev(mac), "Unrecognised regulatory domain: " | ||
| 127 | "%#04x. Defaulting to FCC.\n", default_regdomain); | ||
| 128 | default_regdomain = ZD_REGDOMAIN_FCC; | ||
| 121 | } | 129 | } |
| 122 | spin_lock_irq(&mac->lock); | 130 | spin_lock_irq(&mac->lock); |
| 123 | mac->regdomain = mac->default_regdomain = default_regdomain; | 131 | mac->regdomain = mac->default_regdomain = default_regdomain; |
| @@ -164,14 +172,25 @@ int zd_mac_open(struct net_device *netdev) | |||
| 164 | { | 172 | { |
| 165 | struct zd_mac *mac = zd_netdev_mac(netdev); | 173 | struct zd_mac *mac = zd_netdev_mac(netdev); |
| 166 | struct zd_chip *chip = &mac->chip; | 174 | struct zd_chip *chip = &mac->chip; |
| 175 | struct zd_usb *usb = &chip->usb; | ||
| 167 | int r; | 176 | int r; |
| 168 | 177 | ||
| 178 | if (!usb->initialized) { | ||
| 179 | r = zd_usb_init_hw(usb); | ||
| 180 | if (r) | ||
| 181 | goto out; | ||
| 182 | } | ||
| 183 | |||
| 169 | tasklet_enable(&mac->rx_tasklet); | 184 | tasklet_enable(&mac->rx_tasklet); |
| 170 | 185 | ||
| 171 | r = zd_chip_enable_int(chip); | 186 | r = zd_chip_enable_int(chip); |
| 172 | if (r < 0) | 187 | if (r < 0) |
| 173 | goto out; | 188 | goto out; |
| 174 | 189 | ||
| 190 | r = zd_write_mac_addr(chip, netdev->dev_addr); | ||
| 191 | if (r) | ||
| 192 | goto disable_int; | ||
| 193 | |||
| 175 | r = zd_chip_set_basic_rates(chip, CR_RATES_80211B | CR_RATES_80211G); | 194 | r = zd_chip_set_basic_rates(chip, CR_RATES_80211B | CR_RATES_80211G); |
| 176 | if (r < 0) | 195 | if (r < 0) |
| 177 | goto disable_int; | 196 | goto disable_int; |
| @@ -251,9 +270,11 @@ int zd_mac_set_mac_address(struct net_device *netdev, void *p) | |||
| 251 | dev_dbg_f(zd_mac_dev(mac), | 270 | dev_dbg_f(zd_mac_dev(mac), |
| 252 | "Setting MAC to " MAC_FMT "\n", MAC_ARG(addr->sa_data)); | 271 | "Setting MAC to " MAC_FMT "\n", MAC_ARG(addr->sa_data)); |
| 253 | 272 | ||
| 254 | r = zd_write_mac_addr(chip, addr->sa_data); | 273 | if (netdev->flags & IFF_UP) { |
| 255 | if (r) | 274 | r = zd_write_mac_addr(chip, addr->sa_data); |
| 256 | return r; | 275 | if (r) |
| 276 | return r; | ||
| 277 | } | ||
| 257 | 278 | ||
| 258 | spin_lock_irqsave(&mac->lock, flags); | 279 | spin_lock_irqsave(&mac->lock, flags); |
| 259 | memcpy(netdev->dev_addr, addr->sa_data, ETH_ALEN); | 280 | memcpy(netdev->dev_addr, addr->sa_data, ETH_ALEN); |
| @@ -855,7 +876,7 @@ static int fill_ctrlset(struct zd_mac *mac, | |||
| 855 | /* ZD1211B: Computing the length difference this way, gives us | 876 | /* ZD1211B: Computing the length difference this way, gives us |
| 856 | * flexibility to compute the packet length. | 877 | * flexibility to compute the packet length. |
| 857 | */ | 878 | */ |
| 858 | cs->packet_length = cpu_to_le16(mac->chip.is_zd1211b ? | 879 | cs->packet_length = cpu_to_le16(zd_chip_is_zd1211b(&mac->chip) ? |
| 859 | packet_length - frag_len : packet_length); | 880 | packet_length - frag_len : packet_length); |
| 860 | 881 | ||
| 861 | /* | 882 | /* |
diff --git a/drivers/net/wireless/zd1211rw/zd_mac.h b/drivers/net/wireless/zd1211rw/zd_mac.h index faf4c7828d4e..9f9344eb50f9 100644 --- a/drivers/net/wireless/zd1211rw/zd_mac.h +++ b/drivers/net/wireless/zd1211rw/zd_mac.h | |||
| @@ -189,7 +189,8 @@ int zd_mac_init(struct zd_mac *mac, | |||
| 189 | struct usb_interface *intf); | 189 | struct usb_interface *intf); |
| 190 | void zd_mac_clear(struct zd_mac *mac); | 190 | void zd_mac_clear(struct zd_mac *mac); |
| 191 | 191 | ||
| 192 | int zd_mac_init_hw(struct zd_mac *mac, u8 device_type); | 192 | int zd_mac_preinit_hw(struct zd_mac *mac); |
| 193 | int zd_mac_init_hw(struct zd_mac *mac); | ||
| 193 | 194 | ||
| 194 | int zd_mac_open(struct net_device *netdev); | 195 | int zd_mac_open(struct net_device *netdev); |
| 195 | int zd_mac_stop(struct net_device *netdev); | 196 | int zd_mac_stop(struct net_device *netdev); |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf.c b/drivers/net/wireless/zd1211rw/zd_rf.c index 7407409b60b1..abe5d38f7f4d 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf.c +++ b/drivers/net/wireless/zd1211rw/zd_rf.c | |||
| @@ -34,7 +34,7 @@ static const char * const rfs[] = { | |||
| 34 | [AL2210_RF] = "AL2210_RF", | 34 | [AL2210_RF] = "AL2210_RF", |
| 35 | [MAXIM_NEW_RF] = "MAXIM_NEW_RF", | 35 | [MAXIM_NEW_RF] = "MAXIM_NEW_RF", |
| 36 | [UW2453_RF] = "UW2453_RF", | 36 | [UW2453_RF] = "UW2453_RF", |
| 37 | [UNKNOWN_A_RF] = "UNKNOWN_A_RF", | 37 | [AL2230S_RF] = "AL2230S_RF", |
| 38 | [RALINK_RF] = "RALINK_RF", | 38 | [RALINK_RF] = "RALINK_RF", |
| 39 | [INTERSIL_RF] = "INTERSIL_RF", | 39 | [INTERSIL_RF] = "INTERSIL_RF", |
| 40 | [RF2959_RF] = "RF2959_RF", | 40 | [RF2959_RF] = "RF2959_RF", |
| @@ -77,6 +77,7 @@ int zd_rf_init_hw(struct zd_rf *rf, u8 type) | |||
| 77 | r = zd_rf_init_rf2959(rf); | 77 | r = zd_rf_init_rf2959(rf); |
| 78 | break; | 78 | break; |
| 79 | case AL2230_RF: | 79 | case AL2230_RF: |
| 80 | case AL2230S_RF: | ||
| 80 | r = zd_rf_init_al2230(rf); | 81 | r = zd_rf_init_al2230(rf); |
| 81 | break; | 82 | break; |
| 82 | case AL7230B_RF: | 83 | case AL7230B_RF: |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf.h b/drivers/net/wireless/zd1211rw/zd_rf.h index c6dfd8227f6e..30502f26b71c 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf.h +++ b/drivers/net/wireless/zd1211rw/zd_rf.h | |||
| @@ -26,7 +26,7 @@ | |||
| 26 | #define AL2210_RF 0x7 | 26 | #define AL2210_RF 0x7 |
| 27 | #define MAXIM_NEW_RF 0x8 | 27 | #define MAXIM_NEW_RF 0x8 |
| 28 | #define UW2453_RF 0x9 | 28 | #define UW2453_RF 0x9 |
| 29 | #define UNKNOWN_A_RF 0xa | 29 | #define AL2230S_RF 0xa |
| 30 | #define RALINK_RF 0xb | 30 | #define RALINK_RF 0xb |
| 31 | #define INTERSIL_RF 0xc | 31 | #define INTERSIL_RF 0xc |
| 32 | #define RF2959_RF 0xd | 32 | #define RF2959_RF 0xd |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf_al2230.c b/drivers/net/wireless/zd1211rw/zd_rf_al2230.c index e7a4ecf7b6e2..006774de3202 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf_al2230.c +++ b/drivers/net/wireless/zd1211rw/zd_rf_al2230.c | |||
| @@ -21,6 +21,8 @@ | |||
| 21 | #include "zd_usb.h" | 21 | #include "zd_usb.h" |
| 22 | #include "zd_chip.h" | 22 | #include "zd_chip.h" |
| 23 | 23 | ||
| 24 | #define IS_AL2230S(chip) ((chip)->al2230s_bit || (chip)->rf.type == AL2230S_RF) | ||
| 25 | |||
| 24 | static const u32 zd1211_al2230_table[][3] = { | 26 | static const u32 zd1211_al2230_table[][3] = { |
| 25 | RF_CHANNEL( 1) = { 0x03f790, 0x033331, 0x00000d, }, | 27 | RF_CHANNEL( 1) = { 0x03f790, 0x033331, 0x00000d, }, |
| 26 | RF_CHANNEL( 2) = { 0x03f790, 0x0b3331, 0x00000d, }, | 28 | RF_CHANNEL( 2) = { 0x03f790, 0x0b3331, 0x00000d, }, |
| @@ -176,7 +178,7 @@ static int zd1211_al2230_init_hw(struct zd_rf *rf) | |||
| 176 | if (r) | 178 | if (r) |
| 177 | return r; | 179 | return r; |
| 178 | 180 | ||
| 179 | if (chip->al2230s_bit) { | 181 | if (IS_AL2230S(chip)) { |
| 180 | r = zd_iowrite16a_locked(chip, ioreqs_init_al2230s, | 182 | r = zd_iowrite16a_locked(chip, ioreqs_init_al2230s, |
| 181 | ARRAY_SIZE(ioreqs_init_al2230s)); | 183 | ARRAY_SIZE(ioreqs_init_al2230s)); |
| 182 | if (r) | 184 | if (r) |
| @@ -188,7 +190,7 @@ static int zd1211_al2230_init_hw(struct zd_rf *rf) | |||
| 188 | return r; | 190 | return r; |
| 189 | 191 | ||
| 190 | /* improve band edge for AL2230S */ | 192 | /* improve band edge for AL2230S */ |
| 191 | if (chip->al2230s_bit) | 193 | if (IS_AL2230S(chip)) |
| 192 | r = zd_rfwrite_locked(chip, 0x000824, RF_RV_BITS); | 194 | r = zd_rfwrite_locked(chip, 0x000824, RF_RV_BITS); |
| 193 | else | 195 | else |
| 194 | r = zd_rfwrite_locked(chip, 0x0005a4, RF_RV_BITS); | 196 | r = zd_rfwrite_locked(chip, 0x0005a4, RF_RV_BITS); |
| @@ -314,7 +316,7 @@ static int zd1211b_al2230_init_hw(struct zd_rf *rf) | |||
| 314 | if (r) | 316 | if (r) |
| 315 | return r; | 317 | return r; |
| 316 | 318 | ||
| 317 | if (chip->al2230s_bit) { | 319 | if (IS_AL2230S(chip)) { |
| 318 | r = zd_iowrite16a_locked(chip, ioreqs_init_al2230s, | 320 | r = zd_iowrite16a_locked(chip, ioreqs_init_al2230s, |
| 319 | ARRAY_SIZE(ioreqs_init_al2230s)); | 321 | ARRAY_SIZE(ioreqs_init_al2230s)); |
| 320 | if (r) | 322 | if (r) |
| @@ -328,7 +330,7 @@ static int zd1211b_al2230_init_hw(struct zd_rf *rf) | |||
| 328 | if (r) | 330 | if (r) |
| 329 | return r; | 331 | return r; |
| 330 | 332 | ||
| 331 | if (chip->al2230s_bit) | 333 | if (IS_AL2230S(chip)) |
| 332 | r = zd_rfwrite_locked(chip, 0x241000, RF_RV_BITS); | 334 | r = zd_rfwrite_locked(chip, 0x241000, RF_RV_BITS); |
| 333 | else | 335 | else |
| 334 | r = zd_rfwrite_locked(chip, 0x25a000, RF_RV_BITS); | 336 | r = zd_rfwrite_locked(chip, 0x25a000, RF_RV_BITS); |
| @@ -422,7 +424,7 @@ int zd_rf_init_al2230(struct zd_rf *rf) | |||
| 422 | struct zd_chip *chip = zd_rf_to_chip(rf); | 424 | struct zd_chip *chip = zd_rf_to_chip(rf); |
| 423 | 425 | ||
| 424 | rf->switch_radio_off = al2230_switch_radio_off; | 426 | rf->switch_radio_off = al2230_switch_radio_off; |
| 425 | if (chip->is_zd1211b) { | 427 | if (zd_chip_is_zd1211b(chip)) { |
| 426 | rf->init_hw = zd1211b_al2230_init_hw; | 428 | rf->init_hw = zd1211b_al2230_init_hw; |
| 427 | rf->set_channel = zd1211b_al2230_set_channel; | 429 | rf->set_channel = zd1211b_al2230_set_channel; |
| 428 | rf->switch_radio_on = zd1211b_al2230_switch_radio_on; | 430 | rf->switch_radio_on = zd1211b_al2230_switch_radio_on; |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf_al7230b.c b/drivers/net/wireless/zd1211rw/zd_rf_al7230b.c index f4e8b6ada854..73d0bb26f810 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf_al7230b.c +++ b/drivers/net/wireless/zd1211rw/zd_rf_al7230b.c | |||
| @@ -473,7 +473,7 @@ int zd_rf_init_al7230b(struct zd_rf *rf) | |||
| 473 | { | 473 | { |
| 474 | struct zd_chip *chip = zd_rf_to_chip(rf); | 474 | struct zd_chip *chip = zd_rf_to_chip(rf); |
| 475 | 475 | ||
| 476 | if (chip->is_zd1211b) { | 476 | if (zd_chip_is_zd1211b(chip)) { |
| 477 | rf->init_hw = zd1211b_al7230b_init_hw; | 477 | rf->init_hw = zd1211b_al7230b_init_hw; |
| 478 | rf->switch_radio_on = zd1211b_al7230b_switch_radio_on; | 478 | rf->switch_radio_on = zd1211b_al7230b_switch_radio_on; |
| 479 | rf->set_channel = zd1211b_al7230b_set_channel; | 479 | rf->set_channel = zd1211b_al7230b_set_channel; |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf_rf2959.c b/drivers/net/wireless/zd1211rw/zd_rf_rf2959.c index 2d736bdf707c..cc70d40684ea 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf_rf2959.c +++ b/drivers/net/wireless/zd1211rw/zd_rf_rf2959.c | |||
| @@ -265,7 +265,7 @@ int zd_rf_init_rf2959(struct zd_rf *rf) | |||
| 265 | { | 265 | { |
| 266 | struct zd_chip *chip = zd_rf_to_chip(rf); | 266 | struct zd_chip *chip = zd_rf_to_chip(rf); |
| 267 | 267 | ||
| 268 | if (chip->is_zd1211b) { | 268 | if (zd_chip_is_zd1211b(chip)) { |
| 269 | dev_err(zd_chip_dev(chip), | 269 | dev_err(zd_chip_dev(chip), |
| 270 | "RF2959 is currently not supported for ZD1211B" | 270 | "RF2959 is currently not supported for ZD1211B" |
| 271 | " devices\n"); | 271 | " devices\n"); |
diff --git a/drivers/net/wireless/zd1211rw/zd_rf_uw2453.c b/drivers/net/wireless/zd1211rw/zd_rf_uw2453.c index 414e40d571ab..857dcf3eae61 100644 --- a/drivers/net/wireless/zd1211rw/zd_rf_uw2453.c +++ b/drivers/net/wireless/zd1211rw/zd_rf_uw2453.c | |||
| @@ -486,7 +486,7 @@ static int uw2453_switch_radio_on(struct zd_rf *rf) | |||
| 486 | if (r) | 486 | if (r) |
| 487 | return r; | 487 | return r; |
| 488 | 488 | ||
| 489 | if (chip->is_zd1211b) | 489 | if (zd_chip_is_zd1211b(chip)) |
| 490 | ioreqs[1].value = 0x7f; | 490 | ioreqs[1].value = 0x7f; |
| 491 | 491 | ||
| 492 | return zd_iowrite16a_locked(chip, ioreqs, ARRAY_SIZE(ioreqs)); | 492 | return zd_iowrite16a_locked(chip, ioreqs, ARRAY_SIZE(ioreqs)); |
diff --git a/drivers/net/wireless/zd1211rw/zd_usb.c b/drivers/net/wireless/zd1211rw/zd_usb.c index ca24299a26c6..28d41a29d7b1 100644 --- a/drivers/net/wireless/zd1211rw/zd_usb.c +++ b/drivers/net/wireless/zd1211rw/zd_usb.c | |||
| @@ -71,6 +71,7 @@ static struct usb_device_id usb_ids[] = { | |||
| 71 | { USB_DEVICE(0x0586, 0x3412), .driver_info = DEVICE_ZD1211B }, | 71 | { USB_DEVICE(0x0586, 0x3412), .driver_info = DEVICE_ZD1211B }, |
| 72 | { USB_DEVICE(0x0586, 0x3413), .driver_info = DEVICE_ZD1211B }, | 72 | { USB_DEVICE(0x0586, 0x3413), .driver_info = DEVICE_ZD1211B }, |
| 73 | { USB_DEVICE(0x0053, 0x5301), .driver_info = DEVICE_ZD1211B }, | 73 | { USB_DEVICE(0x0053, 0x5301), .driver_info = DEVICE_ZD1211B }, |
| 74 | { USB_DEVICE(0x0411, 0x00da), .driver_info = DEVICE_ZD1211B }, | ||
| 74 | /* "Driverless" devices that need ejecting */ | 75 | /* "Driverless" devices that need ejecting */ |
| 75 | { USB_DEVICE(0x0ace, 0x2011), .driver_info = DEVICE_INSTALLER }, | 76 | { USB_DEVICE(0x0ace, 0x2011), .driver_info = DEVICE_INSTALLER }, |
| 76 | { USB_DEVICE(0x0ace, 0x20ff), .driver_info = DEVICE_INSTALLER }, | 77 | { USB_DEVICE(0x0ace, 0x20ff), .driver_info = DEVICE_INSTALLER }, |
| @@ -195,26 +196,27 @@ static u16 get_word(const void *data, u16 offset) | |||
| 195 | return le16_to_cpu(p[offset]); | 196 | return le16_to_cpu(p[offset]); |
| 196 | } | 197 | } |
| 197 | 198 | ||
| 198 | static char *get_fw_name(char *buffer, size_t size, u8 device_type, | 199 | static char *get_fw_name(struct zd_usb *usb, char *buffer, size_t size, |
| 199 | const char* postfix) | 200 | const char* postfix) |
| 200 | { | 201 | { |
| 201 | scnprintf(buffer, size, "%s%s", | 202 | scnprintf(buffer, size, "%s%s", |
| 202 | device_type == DEVICE_ZD1211B ? | 203 | usb->is_zd1211b ? |
| 203 | FW_ZD1211B_PREFIX : FW_ZD1211_PREFIX, | 204 | FW_ZD1211B_PREFIX : FW_ZD1211_PREFIX, |
| 204 | postfix); | 205 | postfix); |
| 205 | return buffer; | 206 | return buffer; |
| 206 | } | 207 | } |
| 207 | 208 | ||
| 208 | static int handle_version_mismatch(struct usb_device *udev, u8 device_type, | 209 | static int handle_version_mismatch(struct zd_usb *usb, |
| 209 | const struct firmware *ub_fw) | 210 | const struct firmware *ub_fw) |
| 210 | { | 211 | { |
| 212 | struct usb_device *udev = zd_usb_to_usbdev(usb); | ||
| 211 | const struct firmware *ur_fw = NULL; | 213 | const struct firmware *ur_fw = NULL; |
| 212 | int offset; | 214 | int offset; |
| 213 | int r = 0; | 215 | int r = 0; |
| 214 | char fw_name[128]; | 216 | char fw_name[128]; |
| 215 | 217 | ||
| 216 | r = request_fw_file(&ur_fw, | 218 | r = request_fw_file(&ur_fw, |
| 217 | get_fw_name(fw_name, sizeof(fw_name), device_type, "ur"), | 219 | get_fw_name(usb, fw_name, sizeof(fw_name), "ur"), |
| 218 | &udev->dev); | 220 | &udev->dev); |
| 219 | if (r) | 221 | if (r) |
| 220 | goto error; | 222 | goto error; |
| @@ -237,11 +239,12 @@ error: | |||
| 237 | return r; | 239 | return r; |
| 238 | } | 240 | } |
| 239 | 241 | ||
| 240 | static int upload_firmware(struct usb_device *udev, u8 device_type) | 242 | static int upload_firmware(struct zd_usb *usb) |
| 241 | { | 243 | { |
| 242 | int r; | 244 | int r; |
| 243 | u16 fw_bcdDevice; | 245 | u16 fw_bcdDevice; |
| 244 | u16 bcdDevice; | 246 | u16 bcdDevice; |
| 247 | struct usb_device *udev = zd_usb_to_usbdev(usb); | ||
| 245 | const struct firmware *ub_fw = NULL; | 248 | const struct firmware *ub_fw = NULL; |
| 246 | const struct firmware *uph_fw = NULL; | 249 | const struct firmware *uph_fw = NULL; |
| 247 | char fw_name[128]; | 250 | char fw_name[128]; |
| @@ -249,7 +252,7 @@ static int upload_firmware(struct usb_device *udev, u8 device_type) | |||
| 249 | bcdDevice = get_bcdDevice(udev); | 252 | bcdDevice = get_bcdDevice(udev); |
| 250 | 253 | ||
| 251 | r = request_fw_file(&ub_fw, | 254 | r = request_fw_file(&ub_fw, |
| 252 | get_fw_name(fw_name, sizeof(fw_name), device_type, "ub"), | 255 | get_fw_name(usb, fw_name, sizeof(fw_name), "ub"), |
| 253 | &udev->dev); | 256 | &udev->dev); |
| 254 | if (r) | 257 | if (r) |
| 255 | goto error; | 258 | goto error; |
| @@ -264,7 +267,7 @@ static int upload_firmware(struct usb_device *udev, u8 device_type) | |||
| 264 | dev_warn(&udev->dev, "device has old bootcode, please " | 267 | dev_warn(&udev->dev, "device has old bootcode, please " |
| 265 | "report success or failure\n"); | 268 | "report success or failure\n"); |
| 266 | 269 | ||
| 267 | r = handle_version_mismatch(udev, device_type, ub_fw); | 270 | r = handle_version_mismatch(usb, ub_fw); |
| 268 | if (r) | 271 | if (r) |
| 269 | goto error; | 272 | goto error; |
| 270 | } else { | 273 | } else { |
| @@ -275,7 +278,7 @@ static int upload_firmware(struct usb_device *udev, u8 device_type) | |||
| 275 | 278 | ||
| 276 | 279 | ||
| 277 | r = request_fw_file(&uph_fw, | 280 | r = request_fw_file(&uph_fw, |
| 278 | get_fw_name(fw_name, sizeof(fw_name), device_type, "uphr"), | 281 | get_fw_name(usb, fw_name, sizeof(fw_name), "uphr"), |
| 279 | &udev->dev); | 282 | &udev->dev); |
| 280 | if (r) | 283 | if (r) |
| 281 | goto error; | 284 | goto error; |
| @@ -294,6 +297,30 @@ error: | |||
| 294 | return r; | 297 | return r; |
| 295 | } | 298 | } |
| 296 | 299 | ||
| 300 | /* Read data from device address space using "firmware interface" which does | ||
| 301 | * not require firmware to be loaded. */ | ||
| 302 | int zd_usb_read_fw(struct zd_usb *usb, zd_addr_t addr, u8 *data, u16 len) | ||
| 303 | { | ||
| 304 | int r; | ||
| 305 | struct usb_device *udev = zd_usb_to_usbdev(usb); | ||
| 306 | |||
| 307 | r = usb_control_msg(udev, usb_rcvctrlpipe(udev, 0), | ||
| 308 | USB_REQ_FIRMWARE_READ_DATA, USB_DIR_IN | 0x40, addr, 0, | ||
| 309 | data, len, 5000); | ||
| 310 | if (r < 0) { | ||
| 311 | dev_err(&udev->dev, | ||
| 312 | "read over firmware interface failed: %d\n", r); | ||
| 313 | return r; | ||
| 314 | } else if (r != len) { | ||
| 315 | dev_err(&udev->dev, | ||
| 316 | "incomplete read over firmware interface: %d/%d\n", | ||
| 317 | r, len); | ||
| 318 | return -EIO; | ||
| 319 | } | ||
| 320 | |||
| 321 | return 0; | ||
| 322 | } | ||
| 323 | |||
| 297 | #define urb_dev(urb) (&(urb)->dev->dev) | 324 | #define urb_dev(urb) (&(urb)->dev->dev) |
| 298 | 325 | ||
| 299 | static inline void handle_regs_int(struct urb *urb) | 326 | static inline void handle_regs_int(struct urb *urb) |
| @@ -920,9 +947,42 @@ static int eject_installer(struct usb_interface *intf) | |||
| 920 | return 0; | 947 | return 0; |
| 921 | } | 948 | } |
| 922 | 949 | ||
| 950 | int zd_usb_init_hw(struct zd_usb *usb) | ||
| 951 | { | ||
| 952 | int r; | ||
| 953 | struct zd_mac *mac = zd_usb_to_mac(usb); | ||
| 954 | |||
| 955 | dev_dbg_f(zd_usb_dev(usb), "\n"); | ||
| 956 | |||
| 957 | r = upload_firmware(usb); | ||
| 958 | if (r) { | ||
| 959 | dev_err(zd_usb_dev(usb), | ||
| 960 | "couldn't load firmware. Error number %d\n", r); | ||
| 961 | return r; | ||
| 962 | } | ||
| 963 | |||
| 964 | r = usb_reset_configuration(zd_usb_to_usbdev(usb)); | ||
| 965 | if (r) { | ||
| 966 | dev_dbg_f(zd_usb_dev(usb), | ||
| 967 | "couldn't reset configuration. Error number %d\n", r); | ||
| 968 | return r; | ||
| 969 | } | ||
| 970 | |||
| 971 | r = zd_mac_init_hw(mac); | ||
| 972 | if (r) { | ||
| 973 | dev_dbg_f(zd_usb_dev(usb), | ||
| 974 | "couldn't initialize mac. Error number %d\n", r); | ||
| 975 | return r; | ||
| 976 | } | ||
| 977 | |||
| 978 | usb->initialized = 1; | ||
| 979 | return 0; | ||
| 980 | } | ||
| 981 | |||
| 923 | static int probe(struct usb_interface *intf, const struct usb_device_id *id) | 982 | static int probe(struct usb_interface *intf, const struct usb_device_id *id) |
| 924 | { | 983 | { |
| 925 | int r; | 984 | int r; |
| 985 | struct zd_usb *usb; | ||
| 926 | struct usb_device *udev = interface_to_usbdev(intf); | 986 | struct usb_device *udev = interface_to_usbdev(intf); |
| 927 | struct net_device *netdev = NULL; | 987 | struct net_device *netdev = NULL; |
| 928 | 988 | ||
| @@ -950,26 +1010,10 @@ static int probe(struct usb_interface *intf, const struct usb_device_id *id) | |||
| 950 | goto error; | 1010 | goto error; |
| 951 | } | 1011 | } |
| 952 | 1012 | ||
| 953 | r = upload_firmware(udev, id->driver_info); | 1013 | usb = &zd_netdev_mac(netdev)->chip.usb; |
| 954 | if (r) { | 1014 | usb->is_zd1211b = (id->driver_info == DEVICE_ZD1211B) != 0; |
| 955 | dev_err(&intf->dev, | ||
| 956 | "couldn't load firmware. Error number %d\n", r); | ||
| 957 | goto error; | ||
| 958 | } | ||
| 959 | 1015 | ||
| 960 | r = usb_reset_configuration(udev); | 1016 | r = zd_mac_preinit_hw(zd_netdev_mac(netdev)); |
| 961 | if (r) { | ||
| 962 | dev_dbg_f(&intf->dev, | ||
| 963 | "couldn't reset configuration. Error number %d\n", r); | ||
| 964 | goto error; | ||
| 965 | } | ||
| 966 | |||
| 967 | /* At this point the interrupt endpoint is not generally enabled. We | ||
| 968 | * save the USB bandwidth until the network device is opened. But | ||
| 969 | * notify that the initialization of the MAC will require the | ||
| 970 | * interrupts to be temporary enabled. | ||
| 971 | */ | ||
| 972 | r = zd_mac_init_hw(zd_netdev_mac(netdev), id->driver_info); | ||
| 973 | if (r) { | 1017 | if (r) { |
| 974 | dev_dbg_f(&intf->dev, | 1018 | dev_dbg_f(&intf->dev, |
| 975 | "couldn't initialize mac. Error number %d\n", r); | 1019 | "couldn't initialize mac. Error number %d\n", r); |
diff --git a/drivers/net/wireless/zd1211rw/zd_usb.h b/drivers/net/wireless/zd1211rw/zd_usb.h index 506ea6a74393..961a7a12ad68 100644 --- a/drivers/net/wireless/zd1211rw/zd_usb.h +++ b/drivers/net/wireless/zd1211rw/zd_usb.h | |||
| @@ -188,6 +188,7 @@ struct zd_usb { | |||
| 188 | struct zd_usb_rx rx; | 188 | struct zd_usb_rx rx; |
| 189 | struct zd_usb_tx tx; | 189 | struct zd_usb_tx tx; |
| 190 | struct usb_interface *intf; | 190 | struct usb_interface *intf; |
| 191 | u8 is_zd1211b:1, initialized:1; | ||
| 191 | }; | 192 | }; |
| 192 | 193 | ||
| 193 | #define zd_usb_dev(usb) (&usb->intf->dev) | 194 | #define zd_usb_dev(usb) (&usb->intf->dev) |
| @@ -236,6 +237,8 @@ int zd_usb_iowrite16v(struct zd_usb *usb, const struct zd_ioreq16 *ioreqs, | |||
| 236 | 237 | ||
| 237 | int zd_usb_rfwrite(struct zd_usb *usb, u32 value, u8 bits); | 238 | int zd_usb_rfwrite(struct zd_usb *usb, u32 value, u8 bits); |
| 238 | 239 | ||
| 240 | int zd_usb_read_fw(struct zd_usb *usb, zd_addr_t addr, u8 *data, u16 len); | ||
| 241 | |||
| 239 | extern struct workqueue_struct *zd_workqueue; | 242 | extern struct workqueue_struct *zd_workqueue; |
| 240 | 243 | ||
| 241 | #endif /* _ZD_USB_H */ | 244 | #endif /* _ZD_USB_H */ |
diff --git a/drivers/s390/net/qeth.h b/drivers/s390/net/qeth.h index b34eb82edd98..ec18bae05df0 100644 --- a/drivers/s390/net/qeth.h +++ b/drivers/s390/net/qeth.h | |||
| @@ -211,6 +211,10 @@ struct qeth_perf_stats { | |||
| 211 | /* initial values when measuring starts */ | 211 | /* initial values when measuring starts */ |
| 212 | unsigned long initial_rx_packets; | 212 | unsigned long initial_rx_packets; |
| 213 | unsigned long initial_tx_packets; | 213 | unsigned long initial_tx_packets; |
| 214 | /* inbound scatter gather data */ | ||
| 215 | unsigned int sg_skbs_rx; | ||
| 216 | unsigned int sg_frags_rx; | ||
| 217 | unsigned int sg_alloc_page_rx; | ||
| 214 | }; | 218 | }; |
| 215 | 219 | ||
| 216 | /* Routing stuff */ | 220 | /* Routing stuff */ |
| @@ -341,6 +345,9 @@ qeth_is_ipa_enabled(struct qeth_ipa_info *ipa, enum qeth_ipa_funcs func) | |||
| 341 | 345 | ||
| 342 | #define QETH_IP_HEADER_SIZE 40 | 346 | #define QETH_IP_HEADER_SIZE 40 |
| 343 | 347 | ||
| 348 | /* large receive scatter gather copy break */ | ||
| 349 | #define QETH_RX_SG_CB (PAGE_SIZE >> 1) | ||
| 350 | |||
| 344 | struct qeth_hdr_layer3 { | 351 | struct qeth_hdr_layer3 { |
| 345 | __u8 id; | 352 | __u8 id; |
| 346 | __u8 flags; | 353 | __u8 flags; |
| @@ -771,6 +778,7 @@ struct qeth_card_options { | |||
| 771 | int layer2; | 778 | int layer2; |
| 772 | enum qeth_large_send_types large_send; | 779 | enum qeth_large_send_types large_send; |
| 773 | int performance_stats; | 780 | int performance_stats; |
| 781 | int rx_sg_cb; | ||
| 774 | }; | 782 | }; |
| 775 | 783 | ||
| 776 | /* | 784 | /* |
| @@ -828,6 +836,7 @@ struct qeth_card { | |||
| 828 | int (*orig_hard_header)(struct sk_buff *,struct net_device *, | 836 | int (*orig_hard_header)(struct sk_buff *,struct net_device *, |
| 829 | unsigned short,void *,void *,unsigned); | 837 | unsigned short,void *,void *,unsigned); |
| 830 | struct qeth_osn_info osn_info; | 838 | struct qeth_osn_info osn_info; |
| 839 | atomic_t force_alloc_skb; | ||
| 831 | }; | 840 | }; |
| 832 | 841 | ||
| 833 | struct qeth_card_list_struct { | 842 | struct qeth_card_list_struct { |
diff --git a/drivers/s390/net/qeth_main.c b/drivers/s390/net/qeth_main.c index 86b0c44165c1..57f69434fbf9 100644 --- a/drivers/s390/net/qeth_main.c +++ b/drivers/s390/net/qeth_main.c | |||
| @@ -1054,6 +1054,7 @@ qeth_set_intial_options(struct qeth_card *card) | |||
| 1054 | else | 1054 | else |
| 1055 | card->options.layer2 = 0; | 1055 | card->options.layer2 = 0; |
| 1056 | card->options.performance_stats = 0; | 1056 | card->options.performance_stats = 0; |
| 1057 | card->options.rx_sg_cb = QETH_RX_SG_CB; | ||
| 1057 | } | 1058 | } |
| 1058 | 1059 | ||
| 1059 | /** | 1060 | /** |
| @@ -1934,6 +1935,7 @@ qeth_send_control_data(struct qeth_card *card, int len, | |||
| 1934 | atomic_inc(&reply->received); | 1935 | atomic_inc(&reply->received); |
| 1935 | wake_up(&reply->wait_q); | 1936 | wake_up(&reply->wait_q); |
| 1936 | } | 1937 | } |
| 1938 | cpu_relax(); | ||
| 1937 | }; | 1939 | }; |
| 1938 | rc = reply->rc; | 1940 | rc = reply->rc; |
| 1939 | qeth_put_reply(reply); | 1941 | qeth_put_reply(reply); |
| @@ -2258,6 +2260,89 @@ qeth_get_skb(unsigned int length, struct qeth_hdr *hdr) | |||
| 2258 | return skb; | 2260 | return skb; |
| 2259 | } | 2261 | } |
| 2260 | 2262 | ||
| 2263 | static inline int | ||
| 2264 | qeth_create_skb_frag(struct qdio_buffer_element *element, | ||
| 2265 | struct sk_buff **pskb, | ||
| 2266 | int offset, int *pfrag, int data_len) | ||
| 2267 | { | ||
| 2268 | struct page *page = virt_to_page(element->addr); | ||
| 2269 | if (*pfrag == 0) { | ||
| 2270 | /* the upper protocol layers assume that there is data in the | ||
| 2271 | * skb itself. Copy a small amount (64 bytes) to make them | ||
| 2272 | * happy. */ | ||
| 2273 | *pskb = dev_alloc_skb(64 + QETH_FAKE_LL_LEN_ETH); | ||
| 2274 | if (!(*pskb)) | ||
| 2275 | return -ENOMEM; | ||
| 2276 | skb_reserve(*pskb, QETH_FAKE_LL_LEN_ETH); | ||
| 2277 | if (data_len <= 64) { | ||
| 2278 | memcpy(skb_put(*pskb, data_len), element->addr + offset, | ||
| 2279 | data_len); | ||
| 2280 | } else { | ||
| 2281 | get_page(page); | ||
| 2282 | memcpy(skb_put(*pskb, 64), element->addr + offset, 64); | ||
| 2283 | skb_fill_page_desc(*pskb, *pfrag, page, offset + 64, | ||
| 2284 | data_len - 64); | ||
| 2285 | (*pskb)->data_len += data_len - 64; | ||
| 2286 | (*pskb)->len += data_len - 64; | ||
| 2287 | (*pskb)->truesize += data_len - 64; | ||
| 2288 | } | ||
| 2289 | } else { | ||
| 2290 | get_page(page); | ||
| 2291 | skb_fill_page_desc(*pskb, *pfrag, page, offset, data_len); | ||
| 2292 | (*pskb)->data_len += data_len; | ||
| 2293 | (*pskb)->len += data_len; | ||
| 2294 | (*pskb)->truesize += data_len; | ||
| 2295 | } | ||
| 2296 | (*pfrag)++; | ||
| 2297 | return 0; | ||
| 2298 | } | ||
| 2299 | |||
| 2300 | static inline struct qeth_buffer_pool_entry * | ||
| 2301 | qeth_find_free_buffer_pool_entry(struct qeth_card *card) | ||
| 2302 | { | ||
| 2303 | struct list_head *plh; | ||
| 2304 | struct qeth_buffer_pool_entry *entry; | ||
| 2305 | int i, free; | ||
| 2306 | struct page *page; | ||
| 2307 | |||
| 2308 | if (list_empty(&card->qdio.in_buf_pool.entry_list)) | ||
| 2309 | return NULL; | ||
| 2310 | |||
| 2311 | list_for_each(plh, &card->qdio.in_buf_pool.entry_list) { | ||
| 2312 | entry = list_entry(plh, struct qeth_buffer_pool_entry, list); | ||
| 2313 | free = 1; | ||
| 2314 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
| 2315 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
| 2316 | free = 0; | ||
| 2317 | break; | ||
| 2318 | } | ||
| 2319 | } | ||
| 2320 | if (free) { | ||
| 2321 | list_del_init(&entry->list); | ||
| 2322 | return entry; | ||
| 2323 | } | ||
| 2324 | } | ||
| 2325 | |||
| 2326 | /* no free buffer in pool so take first one and swap pages */ | ||
| 2327 | entry = list_entry(card->qdio.in_buf_pool.entry_list.next, | ||
| 2328 | struct qeth_buffer_pool_entry, list); | ||
| 2329 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
| 2330 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
| 2331 | page = alloc_page(GFP_ATOMIC|GFP_DMA); | ||
| 2332 | if (!page) { | ||
| 2333 | return NULL; | ||
| 2334 | } else { | ||
| 2335 | free_page((unsigned long)entry->elements[i]); | ||
| 2336 | entry->elements[i] = page_address(page); | ||
| 2337 | if (card->options.performance_stats) | ||
| 2338 | card->perf_stats.sg_alloc_page_rx++; | ||
| 2339 | } | ||
| 2340 | } | ||
| 2341 | } | ||
| 2342 | list_del_init(&entry->list); | ||
| 2343 | return entry; | ||
| 2344 | } | ||
| 2345 | |||
| 2261 | static struct sk_buff * | 2346 | static struct sk_buff * |
| 2262 | qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, | 2347 | qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, |
| 2263 | struct qdio_buffer_element **__element, int *__offset, | 2348 | struct qdio_buffer_element **__element, int *__offset, |
| @@ -2269,6 +2354,8 @@ qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, | |||
| 2269 | int skb_len; | 2354 | int skb_len; |
| 2270 | void *data_ptr; | 2355 | void *data_ptr; |
| 2271 | int data_len; | 2356 | int data_len; |
| 2357 | int use_rx_sg = 0; | ||
| 2358 | int frag = 0; | ||
| 2272 | 2359 | ||
| 2273 | QETH_DBF_TEXT(trace,6,"nextskb"); | 2360 | QETH_DBF_TEXT(trace,6,"nextskb"); |
| 2274 | /* qeth_hdr must not cross element boundaries */ | 2361 | /* qeth_hdr must not cross element boundaries */ |
| @@ -2293,23 +2380,43 @@ qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, | |||
| 2293 | 2380 | ||
| 2294 | if (!skb_len) | 2381 | if (!skb_len) |
| 2295 | return NULL; | 2382 | return NULL; |
| 2296 | if (card->options.fake_ll){ | 2383 | if ((skb_len >= card->options.rx_sg_cb) && |
| 2297 | if(card->dev->type == ARPHRD_IEEE802_TR){ | 2384 | (!(card->info.type == QETH_CARD_TYPE_OSN)) && |
| 2298 | if (!(skb = qeth_get_skb(skb_len+QETH_FAKE_LL_LEN_TR, *hdr))) | 2385 | (!atomic_read(&card->force_alloc_skb))) { |
| 2299 | goto no_mem; | 2386 | use_rx_sg = 1; |
| 2300 | skb_reserve(skb,QETH_FAKE_LL_LEN_TR); | 2387 | } else { |
| 2388 | if (card->options.fake_ll) { | ||
| 2389 | if (card->dev->type == ARPHRD_IEEE802_TR) { | ||
| 2390 | if (!(skb = qeth_get_skb(skb_len + | ||
| 2391 | QETH_FAKE_LL_LEN_TR, *hdr))) | ||
| 2392 | goto no_mem; | ||
| 2393 | skb_reserve(skb, QETH_FAKE_LL_LEN_TR); | ||
| 2394 | } else { | ||
| 2395 | if (!(skb = qeth_get_skb(skb_len + | ||
| 2396 | QETH_FAKE_LL_LEN_ETH, *hdr))) | ||
| 2397 | goto no_mem; | ||
| 2398 | skb_reserve(skb, QETH_FAKE_LL_LEN_ETH); | ||
| 2399 | } | ||
| 2301 | } else { | 2400 | } else { |
| 2302 | if (!(skb = qeth_get_skb(skb_len+QETH_FAKE_LL_LEN_ETH, *hdr))) | 2401 | skb = qeth_get_skb(skb_len, *hdr); |
| 2402 | if (!skb) | ||
| 2303 | goto no_mem; | 2403 | goto no_mem; |
| 2304 | skb_reserve(skb,QETH_FAKE_LL_LEN_ETH); | ||
| 2305 | } | 2404 | } |
| 2306 | } else if (!(skb = qeth_get_skb(skb_len, *hdr))) | 2405 | } |
| 2307 | goto no_mem; | 2406 | |
| 2308 | data_ptr = element->addr + offset; | 2407 | data_ptr = element->addr + offset; |
| 2309 | while (skb_len) { | 2408 | while (skb_len) { |
| 2310 | data_len = min(skb_len, (int)(element->length - offset)); | 2409 | data_len = min(skb_len, (int)(element->length - offset)); |
| 2311 | if (data_len) | 2410 | if (data_len) { |
| 2312 | memcpy(skb_put(skb, data_len), data_ptr, data_len); | 2411 | if (use_rx_sg) { |
| 2412 | if (qeth_create_skb_frag(element, &skb, offset, | ||
| 2413 | &frag, data_len)) | ||
| 2414 | goto no_mem; | ||
| 2415 | } else { | ||
| 2416 | memcpy(skb_put(skb, data_len), data_ptr, | ||
| 2417 | data_len); | ||
| 2418 | } | ||
| 2419 | } | ||
| 2313 | skb_len -= data_len; | 2420 | skb_len -= data_len; |
| 2314 | if (skb_len){ | 2421 | if (skb_len){ |
| 2315 | if (qeth_is_last_sbale(element)){ | 2422 | if (qeth_is_last_sbale(element)){ |
| @@ -2331,6 +2438,10 @@ qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, | |||
| 2331 | } | 2438 | } |
| 2332 | *__element = element; | 2439 | *__element = element; |
| 2333 | *__offset = offset; | 2440 | *__offset = offset; |
| 2441 | if (use_rx_sg && card->options.performance_stats) { | ||
| 2442 | card->perf_stats.sg_skbs_rx++; | ||
| 2443 | card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; | ||
| 2444 | } | ||
| 2334 | return skb; | 2445 | return skb; |
| 2335 | no_mem: | 2446 | no_mem: |
| 2336 | if (net_ratelimit()){ | 2447 | if (net_ratelimit()){ |
| @@ -2608,28 +2719,15 @@ qeth_process_inbound_buffer(struct qeth_card *card, | |||
| 2608 | } | 2719 | } |
| 2609 | } | 2720 | } |
| 2610 | 2721 | ||
| 2611 | static struct qeth_buffer_pool_entry * | 2722 | static int |
| 2612 | qeth_get_buffer_pool_entry(struct qeth_card *card) | ||
| 2613 | { | ||
| 2614 | struct qeth_buffer_pool_entry *entry; | ||
| 2615 | |||
| 2616 | QETH_DBF_TEXT(trace, 6, "gtbfplen"); | ||
| 2617 | if (!list_empty(&card->qdio.in_buf_pool.entry_list)) { | ||
| 2618 | entry = list_entry(card->qdio.in_buf_pool.entry_list.next, | ||
| 2619 | struct qeth_buffer_pool_entry, list); | ||
| 2620 | list_del_init(&entry->list); | ||
| 2621 | return entry; | ||
| 2622 | } | ||
| 2623 | return NULL; | ||
| 2624 | } | ||
| 2625 | |||
| 2626 | static void | ||
| 2627 | qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf) | 2723 | qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf) |
| 2628 | { | 2724 | { |
| 2629 | struct qeth_buffer_pool_entry *pool_entry; | 2725 | struct qeth_buffer_pool_entry *pool_entry; |
| 2630 | int i; | 2726 | int i; |
| 2631 | 2727 | ||
| 2632 | pool_entry = qeth_get_buffer_pool_entry(card); | 2728 | pool_entry = qeth_find_free_buffer_pool_entry(card); |
| 2729 | if (!pool_entry) | ||
| 2730 | return 1; | ||
| 2633 | /* | 2731 | /* |
| 2634 | * since the buffer is accessed only from the input_tasklet | 2732 | * since the buffer is accessed only from the input_tasklet |
| 2635 | * there shouldn't be a need to synchronize; also, since we use | 2733 | * there shouldn't be a need to synchronize; also, since we use |
| @@ -2648,6 +2746,7 @@ qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf) | |||
| 2648 | buf->buffer->element[i].flags = 0; | 2746 | buf->buffer->element[i].flags = 0; |
| 2649 | } | 2747 | } |
| 2650 | buf->state = QETH_QDIO_BUF_EMPTY; | 2748 | buf->state = QETH_QDIO_BUF_EMPTY; |
| 2749 | return 0; | ||
| 2651 | } | 2750 | } |
| 2652 | 2751 | ||
| 2653 | static void | 2752 | static void |
| @@ -2682,6 +2781,7 @@ qeth_queue_input_buffer(struct qeth_card *card, int index) | |||
| 2682 | int count; | 2781 | int count; |
| 2683 | int i; | 2782 | int i; |
| 2684 | int rc; | 2783 | int rc; |
| 2784 | int newcount = 0; | ||
| 2685 | 2785 | ||
| 2686 | QETH_DBF_TEXT(trace,6,"queinbuf"); | 2786 | QETH_DBF_TEXT(trace,6,"queinbuf"); |
| 2687 | count = (index < queue->next_buf_to_init)? | 2787 | count = (index < queue->next_buf_to_init)? |
| @@ -2692,9 +2792,27 @@ qeth_queue_input_buffer(struct qeth_card *card, int index) | |||
| 2692 | /* only requeue at a certain threshold to avoid SIGAs */ | 2792 | /* only requeue at a certain threshold to avoid SIGAs */ |
| 2693 | if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)){ | 2793 | if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)){ |
| 2694 | for (i = queue->next_buf_to_init; | 2794 | for (i = queue->next_buf_to_init; |
| 2695 | i < queue->next_buf_to_init + count; ++i) | 2795 | i < queue->next_buf_to_init + count; ++i) { |
| 2696 | qeth_init_input_buffer(card, | 2796 | if (qeth_init_input_buffer(card, |
| 2697 | &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]); | 2797 | &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q])) { |
| 2798 | break; | ||
| 2799 | } else { | ||
| 2800 | newcount++; | ||
| 2801 | } | ||
| 2802 | } | ||
| 2803 | |||
| 2804 | if (newcount < count) { | ||
| 2805 | /* we are in memory shortage so we switch back to | ||
| 2806 | traditional skb allocation and drop packages */ | ||
| 2807 | if (atomic_cmpxchg(&card->force_alloc_skb, 0, 1)) | ||
| 2808 | printk(KERN_WARNING | ||
| 2809 | "qeth: switch to alloc skb\n"); | ||
| 2810 | count = newcount; | ||
| 2811 | } else { | ||
| 2812 | if (atomic_cmpxchg(&card->force_alloc_skb, 1, 0)) | ||
| 2813 | printk(KERN_WARNING "qeth: switch to sg\n"); | ||
| 2814 | } | ||
| 2815 | |||
| 2698 | /* | 2816 | /* |
| 2699 | * according to old code it should be avoided to requeue all | 2817 | * according to old code it should be avoided to requeue all |
| 2700 | * 128 buffers in order to benefit from PCI avoidance. | 2818 | * 128 buffers in order to benefit from PCI avoidance. |
| @@ -6494,6 +6612,7 @@ qeth_hardsetup_card(struct qeth_card *card) | |||
| 6494 | 6612 | ||
| 6495 | QETH_DBF_TEXT(setup, 2, "hrdsetup"); | 6613 | QETH_DBF_TEXT(setup, 2, "hrdsetup"); |
| 6496 | 6614 | ||
| 6615 | atomic_set(&card->force_alloc_skb, 0); | ||
| 6497 | retry: | 6616 | retry: |
| 6498 | if (retries < 3){ | 6617 | if (retries < 3){ |
| 6499 | PRINT_WARN("Retrying to do IDX activates.\n"); | 6618 | PRINT_WARN("Retrying to do IDX activates.\n"); |
diff --git a/drivers/s390/net/qeth_proc.c b/drivers/s390/net/qeth_proc.c index 89d56c8ecdd2..f1ff165a5e05 100644 --- a/drivers/s390/net/qeth_proc.c +++ b/drivers/s390/net/qeth_proc.c | |||
| @@ -212,6 +212,12 @@ qeth_perf_procfile_seq_show(struct seq_file *s, void *it) | |||
| 212 | " Skb fragments sent in SG mode : %u\n\n", | 212 | " Skb fragments sent in SG mode : %u\n\n", |
| 213 | card->perf_stats.sg_skbs_sent, | 213 | card->perf_stats.sg_skbs_sent, |
| 214 | card->perf_stats.sg_frags_sent); | 214 | card->perf_stats.sg_frags_sent); |
| 215 | seq_printf(s, " Skbs received in SG mode : %u\n" | ||
| 216 | " Skb fragments received in SG mode : %u\n" | ||
| 217 | " Page allocations for rx SG mode : %u\n\n", | ||
| 218 | card->perf_stats.sg_skbs_rx, | ||
| 219 | card->perf_stats.sg_frags_rx, | ||
| 220 | card->perf_stats.sg_alloc_page_rx); | ||
| 215 | seq_printf(s, " large_send tx (in Kbytes) : %u\n" | 221 | seq_printf(s, " large_send tx (in Kbytes) : %u\n" |
| 216 | " large_send count : %u\n\n", | 222 | " large_send count : %u\n\n", |
| 217 | card->perf_stats.large_send_bytes >> 10, | 223 | card->perf_stats.large_send_bytes >> 10, |
diff --git a/include/asm-arm/arch-at91/board.h b/include/asm-arm/arch-at91/board.h index 0ce6ee98ed0b..d96b10fd449f 100644 --- a/include/asm-arm/arch-at91/board.h +++ b/include/asm-arm/arch-at91/board.h | |||
| @@ -64,6 +64,7 @@ extern void __init at91_add_device_mmc(short mmc_id, struct at91_mmc_data *data) | |||
| 64 | 64 | ||
| 65 | /* Ethernet (EMAC & MACB) */ | 65 | /* Ethernet (EMAC & MACB) */ |
| 66 | struct at91_eth_data { | 66 | struct at91_eth_data { |
| 67 | u32 phy_mask; | ||
| 67 | u8 phy_irq_pin; /* PHY IRQ */ | 68 | u8 phy_irq_pin; /* PHY IRQ */ |
| 68 | u8 is_rmii; /* using RMII interface? */ | 69 | u8 is_rmii; /* using RMII interface? */ |
| 69 | }; | 70 | }; |
diff --git a/include/asm-avr32/arch-at32ap/board.h b/include/asm-avr32/arch-at32ap/board.h index 9fd2e32f84b8..974480438849 100644 --- a/include/asm-avr32/arch-at32ap/board.h +++ b/include/asm-avr32/arch-at32ap/board.h | |||
| @@ -21,6 +21,7 @@ void at32_map_usart(unsigned int hw_id, unsigned int line); | |||
| 21 | struct platform_device *at32_add_device_usart(unsigned int id); | 21 | struct platform_device *at32_add_device_usart(unsigned int id); |
| 22 | 22 | ||
| 23 | struct eth_platform_data { | 23 | struct eth_platform_data { |
| 24 | u32 phy_mask; | ||
| 24 | u8 is_rmii; | 25 | u8 is_rmii; |
| 25 | }; | 26 | }; |
| 26 | struct platform_device * | 27 | struct platform_device * |
diff --git a/net/ieee80211/softmac/ieee80211softmac_assoc.c b/net/ieee80211/softmac/ieee80211softmac_assoc.c index cc8110bdd579..afb6c6698b27 100644 --- a/net/ieee80211/softmac/ieee80211softmac_assoc.c +++ b/net/ieee80211/softmac/ieee80211softmac_assoc.c | |||
| @@ -271,8 +271,11 @@ ieee80211softmac_assoc_work(struct work_struct *work) | |||
| 271 | */ | 271 | */ |
| 272 | dprintk(KERN_INFO PFX "Associate: Scanning for networks first.\n"); | 272 | dprintk(KERN_INFO PFX "Associate: Scanning for networks first.\n"); |
| 273 | ieee80211softmac_notify(mac->dev, IEEE80211SOFTMAC_EVENT_SCAN_FINISHED, ieee80211softmac_assoc_notify_scan, NULL); | 273 | ieee80211softmac_notify(mac->dev, IEEE80211SOFTMAC_EVENT_SCAN_FINISHED, ieee80211softmac_assoc_notify_scan, NULL); |
| 274 | if (ieee80211softmac_start_scan(mac)) | 274 | if (ieee80211softmac_start_scan(mac)) { |
| 275 | dprintk(KERN_INFO PFX "Associate: failed to initiate scan. Is device up?\n"); | 275 | dprintk(KERN_INFO PFX "Associate: failed to initiate scan. Is device up?\n"); |
| 276 | mac->associnfo.associating = 0; | ||
| 277 | mac->associnfo.associated = 0; | ||
| 278 | } | ||
| 276 | goto out; | 279 | goto out; |
| 277 | } else { | 280 | } else { |
| 278 | mac->associnfo.associating = 0; | 281 | mac->associnfo.associating = 0; |
